# Numbers

Swift contains three main number types: Int, Float, and Double.

### Integers (Int)

By default, numbers without fractional components are inferred to be Ints:

let myNumber = 5 // Inferred to be of type Int

Ints cannot have decimal values:

let myNumber: Int = 5.75 // ❌ error: cannot convert value of type 'Double' to specified type 'Int'

Arithmetic can be performed on integers, but decimal values will be ignored and rounded down.

print(20 / 3)   // 6 (not ~6.666666667)
print(9 / 4)    // 2 (not 2.25)
print(100 / 99) // 1 (not ~1.01010101)

### Floats (Float)

Floating-point numbers are represented by Floats and Doubles.

By default, numbers with fractional components are usually inferred to be Floats:

let myNumber1 = 5.0 // Inferred to be of type Float
let myNumber2 = 3.14 // Inferred to be of type Float

Arithmetic can be performed on Floats and is more accurate than arithmetic on Ints:

print(20.0 / 3.0)   // 6.666666666666667
print(9.0 / 4.0)    // 2.25
print(100.0 / 99.0) // 1.0101010101010102

### Doubles (Double)

Doubles are like Floats, but they are represented as 64-bits, rather than 32-bits. This means they are more precise but also take up more space.

A Double is usually defined with an explicit type:

let myNumber: Double = 5.0

### Converting between Int, Float, and Double

A Double can be converted to a Float, but precision may be lost:

let myDouble: Double = 834.4938247489
let myFloat = Float(myDouble)
print(myFloat) // 834.49384

A Float or Double can be converted to an Int but the fractional component is rounded down:

let myFloat = 9.999
let myInt = Int(myFloat)
print(myInt) // 9

An Int can be converted to a Float or Double:

let myInt = 5
print(Float(myInt))  // 5.0
print(Double(myInt)) // 5.0