In computer science, an integer is a
datum of integral data type, a
data type
In computer science and computer programming, a data type (or simply type) is a set of possible values and a set of allowed operations on it. A data type tells the compiler or interpreter how the programmer intends to use the data. Most progra ...
that represents some
range of mathematical
integers. Integral data types may be of different sizes and may or may not be allowed to contain negative values. Integers are commonly represented in a computer as a group of binary digits (bits). The size of the grouping varies so the set of integer sizes available varies between different types of computers. Computer hardware nearly always provides a way to represent a processor
register or memory address as an integer.
Value and representation
The ''value'' of an item with an integral type is the mathematical integer that it corresponds to. Integral types may be ''unsigned'' (capable of representing only non-negative integers) or ''signed'' (capable of representing negative integers as well).
An integer value is typically specified in the
source code of a program as a sequence of digits optionally prefixed with + or −. Some programming languages allow other notations, such as hexadecimal (base 16) or octal (base 8). Some programming languages also permit
digit group separator
Digit may refer to:
Mathematics and science
* Numerical digit, as used in mathematics or computer science
** Hindu-Arabic numerals, the most common modern representation of numerical digits
* Digit (anatomy), the most distal part of a limb, suc ...
s.
The ''internal representation'' of this datum is the way the value is stored in the computer's memory. Unlike mathematical integers, a typical datum in a computer has some minimal and maximum possible value.
The most common representation of a positive integer is a string of
bits, using the
binary numeral system
A binary number is a number expressed in the base-2 numeral system or binary numeral system, a method of mathematical expression which uses only two symbols: typically "0" (zero) and "1" ( one).
The base-2 numeral system is a positional notatio ...
. The order of the memory
bytes storing the bits varies; see
endianness. The ''width'' or ''precision'' of an integral type is the number of bits in its representation. An integral type with ''n'' bits can encode 2
''n'' numbers; for example an unsigned type typically represents the non-negative values 0 through 2
''n''−1. Other encodings of integer values to bit patterns are sometimes used, for example
binary-coded decimal
In computing and electronic systems, binary-coded decimal (BCD) is a class of binary encodings of decimal numbers where each digit is represented by a fixed number of bits, usually four or eight. Sometimes, special bit patterns are used for ...
or
Gray code
The reflected binary code (RBC), also known as reflected binary (RB) or Gray code after Frank Gray, is an ordering of the binary numeral system such that two successive values differ in only one bit (binary digit).
For example, the representati ...
, or as printed character codes such as
ASCII.
There are four well-known
ways to represent signed numbers in a binary computing system. The most common is
two's complement, which allows a signed integral type with ''n'' bits to represent numbers from −2
(''n''−1) through 2
(''n''−1)−1. Two's complement arithmetic is convenient because there is a perfect
one-to-one correspondence between representations and values (in particular, no separate +0 and −0), and because
addition
Addition (usually signified by the Plus and minus signs#Plus sign, plus symbol ) is one of the four basic Operation (mathematics), operations of arithmetic, the other three being subtraction, multiplication and Division (mathematics), division. ...
,
subtraction
Subtraction is an arithmetic operation that represents the operation of removing objects from a collection. Subtraction is signified by the minus sign, . For example, in the adjacent picture, there are peaches—meaning 5 peaches with 2 taken ...
and
multiplication
Multiplication (often denoted by the cross symbol , by the mid-line dot operator , by juxtaposition, or, on computers, by an asterisk ) is one of the four elementary mathematical operations of arithmetic, with the other ones being additi ...
do not need to distinguish between signed and unsigned types. Other possibilities include
offset binary,
sign-magnitude, and
ones' complement.
Some computer languages define integer sizes in a machine-independent way; others have varying definitions depending on the underlying processor word size. Not all language implementations define variables of all integer sizes, and defined sizes may not even be distinct in a particular implementation. An integer in one
programming language may be a different size in a different language or on a different processor.
Common integral data types
Different
CPUs support different integral data types. Typically, hardware will support both signed and unsigned types, but only a small, fixed set of widths.
The table above lists integral type widths that are supported in hardware by common processors. High level programming languages provide more possibilities. It is common to have a 'double width' integral type that has twice as many bits as the biggest hardware-supported type. Many languages also have ''bit-field'' types (a specified number of bits, usually constrained to be less than the maximum hardware-supported width) and ''range'' types (that can represent only the integers in a specified range).
Some languages, such as
Lisp
A lisp is a speech impairment in which a person misarticulates sibilants (, , , , , , , ). These misarticulations often result in unclear speech.
Types
* A frontal lisp occurs when the tongue is placed anterior to the target. Interdental lisping ...
,
Smalltalk
Smalltalk is an object-oriented, dynamically typed reflective programming language. It was designed and created in part for educational use, specifically for constructionist learning, at the Learning Research Group (LRG) of Xerox PARC by Alan Ka ...
,
REXX
Rexx (Restructured Extended Executor) is a programming language that can be interpreted or compiled. It was developed at IBM by Mike Cowlishaw. It is a structured, high-level programming language designed for ease of learning and reading. ...
,
Haskell,
Python, and
Raku, support ''arbitrary precision'' integers (also known as ''infinite precision integers'' or ''
bignum
In computer science, arbitrary-precision arithmetic, also called bignum arithmetic, multiple-precision arithmetic, or sometimes infinite-precision arithmetic, indicates that calculations are performed on numbers whose digits of precision are l ...
s''). Other languages that do not support this concept as a top-level construct may have libraries available to represent very large numbers using arrays of smaller variables, such as Java's class or
Perl's "" package. These use as much of the computer's memory as is necessary to store the numbers; however, a computer has only a finite amount of storage, so they, too, can only represent a finite subset of the mathematical integers. These schemes support very large numbers; for example one kilobyte of memory could be used to store numbers up to 2466 decimal digits long.
A
Boolean
Any kind of logic, function, expression, or theory based on the work of George Boole is considered Boolean.
Related to this, "Boolean" may refer to:
* Boolean data type, a form of data with only two possible values (usually "true" and "false" ...
or
Flag
A flag is a piece of fabric (most often rectangular or quadrilateral) with a distinctive design and colours. It is used as a symbol, a signalling device, or for decoration. The term ''flag'' is also used to refer to the graphic design empl ...
type is a type that can represent only two values: 0 and 1, usually identified with ''false'' and ''true'' respectively. This type can be stored in memory using a single bit, but is often given a full byte for convenience of addressing and speed of access.
A four-bit quantity is known as a ''
nibble
In computing, a nibble (occasionally nybble, nyble, or nybl to match the spelling of byte) is a four-bit aggregation, or half an octet. It is also known as half-byte or tetrade. In a networking or telecommunication context, the nibble is oft ...
'' (when eating, being smaller than a ''bite'') or ''nybble'' (being a pun on the form of the word ''byte''). One nibble corresponds to one digit in
hexadecimal
In mathematics and computing, the hexadecimal (also base-16 or simply hex) numeral system is a positional numeral system that represents numbers using a radix (base) of 16. Unlike the decimal system representing numbers using 10 symbols, hexa ...
and holds one digit or a sign code in binary-coded decimal.
Bytes and octets
The term ''byte'' initially meant 'the smallest addressable unit of memory'. In the past, 5-, 6-, 7-, 8-, and 9-bit bytes have all been used. There have also been computers that could address individual bits ('bit-addressed machine'), or that could only address 16- or 32-bit quantities ('word-addressed machine'). The term ''byte'' was usually not used at all in connection with bit- and word-addressed machines.
The term ''octet'' always refers to an 8-bit quantity. It is mostly used in the field of
computer networking, where computers with different byte widths might have to communicate.
In modern usage ''byte'' almost invariably means eight bits, since all other sizes have fallen into disuse; thus ''byte'' has come to be synonymous with ''octet''.
Words
The term 'word' is used for a small group of bits that are handled simultaneously by processors of a particular
architecture. The size of a word is thus CPU-specific. Many different word sizes have been used, including 6-, 8-, 12-, 16-, 18-, 24-, 32-, 36-, 39-, 40-, 48-, 60-, and 64-bit. Since it is architectural, the size of a ''word'' is usually set by the first CPU in a family, rather than the characteristics of a later compatible CPU. The meanings of terms derived from ''word'', such as ''longword'', ''doubleword'', ''quadword'', and ''halfword'', also vary with the CPU and OS.
Practically all new desktop processors are capable of using 64-bit words, though
embedded processors with 8- and 16-bit word size are still common. The
36-bit word length
36-bit computers were popular in the early mainframe computer era from the 1950s through the early 1970s.
Starting in the 1960s, but especially the 1970s, the introduction of 7-bit ASCII and 8-bit EBCDIC led to the move to machines using 8-bit ...
was common in the early days of computers.
One important cause of non-portability of software is the incorrect assumption that all computers have the same word size as the computer used by the programmer. For example, if a programmer using the C language incorrectly declares as a variable that will be used to store values greater than 2
15−1, the program will fail on computers with 16-bit integers. That variable should have been declared as , which has at least 32 bits on any computer. Programmers may also incorrectly assume that a pointer can be converted to an integer without loss of information, which may work on (some) 32-bit computers, but fail on 64-bit computers with 64-bit pointers and 32-bit integers. This issue is resolved by C99 in
stdint.h
In the C programming language, data types constitute the semantics and characteristics of storage of data elements. They are expressed in the language syntax in form of declarations for memory locations or variables. Data types also determin ...
in the form of .
Short integer
A ''short integer'' can represent a whole number that may take less storage, while having a smaller range, compared with a standard integer on the same machine.
In
C, it is denoted by . It is required to be at least 16 bits, and is often smaller than a standard integer, but this is not required.
A conforming program can assume that it can safely store values between −(2
15−1)
and 2
15−1,
but it may not assume that the range is not larger. In
Java, a is ''always'' a 16-bit integer. In the
Windows API, the datatype is defined as a 16-bit signed integer on all machines.
Long integer
A ''long integer'' can represent a whole
integer whose
range is greater than or equal to that of a standard integer on the same machine.
In
C, it is denoted by . It is required to be at least 32 bits, and may or may not be larger than a standard integer. A conforming program can assume that it can safely store values between −(2
31−1)
and 2
31−1,
but it may not assume that the range is not larger.
Long long
In the
C99 version of the
C programming language
''The C Programming Language'' (sometimes termed ''K&R'', after its authors' initials) is a computer programming book written by Brian Kernighan and Dennis Ritchie, the latter of whom originally designed and implemented the language, as well as ...
and the
C++11 version of
C++, a
long long
type is supported that has double the minimum capacity of the standard
long
. This type is not supported by compilers that require C code to be compliant with the previous C++ standard, C++03, because the type did not exist in C++03. For an ANSI/ISO compliant compiler, the minimum requirements for the specified ranges, that is, −(2
63−1)
to 2
63−1 for signed and 0 to 2
64−1 for unsigned,
must be fulfilled; however, extending this range is permitted. This can be an issue when exchanging code and data between platforms, or doing direct hardware access. Thus, there are several sets of headers providing platform independent exact width types. The C
standard library
In computer programming, a standard library is the library made available across implementations of a programming language. These libraries are conventionally described in programming language specifications; however, contents of a language's as ...
provides ''
stdint.h
In the C programming language, data types constitute the semantics and characteristics of storage of data elements. They are expressed in the language syntax in form of declarations for memory locations or variables. Data types also determin ...
''; this was introduced in C99 and C++11.
Syntax
Literals for integers can be written as regular
Arabic numerals
Arabic numerals are the ten numerical digits: , , , , , , , , and . They are the most commonly used symbols to write Decimal, decimal numbers. They are also used for writing numbers in other systems such as octal, and for writing identifiers ...
, consisting of a sequence of digits and with negation indicated by a
minus sign before the value. However, most programming languages disallow use of commas or spaces for
digit grouping. Examples of integer literals are:
*
42
*
10000
*
-233000
There are several alternate methods for writing integer literals in many programming languages:
* Most programming languages, especially those influenced by
C, prefix an integer literal with
0X
or
0x
to represent a
hexadecimal
In mathematics and computing, the hexadecimal (also base-16 or simply hex) numeral system is a positional numeral system that represents numbers using a radix (base) of 16. Unlike the decimal system representing numbers using 10 symbols, hexa ...
value, e.g.
0xDEADBEEF
. Other languages may use a different notation, e.g. some
assembly language
In computer programming, assembly language (or assembler language, or symbolic machine code), often referred to simply as Assembly and commonly abbreviated as ASM or asm, is any low-level programming language with a very strong correspondence be ...
s append an
H
or
h
to the end of a hexadecimal value.
*
Perl,
Ruby,
Java,
Julia,
D,
Go,
Rust and
Python (starting from version 3.6) allow embedded
underscores for clarity, e.g.
10_000_000
, and fixed-form
Fortran ignores embedded spaces in integer literals. C (starting from
C23) and C++ use single quotes for this purpose.
* In
C and
C++, a leading zero indicates an
octal value, e.g.
0755
. This was primarily intended to be used with
Unix modes; however, it has been criticized because normal integers may also lead with zero.
[ECMAScript 6th Edition draft: https://people.mozilla.org/~jorendorff/es6-draft.html#sec-literals-numeric-literals ] As such,
Python,
Ruby,
Haskell, and
OCaml
OCaml ( , formerly Objective Caml) is a general-purpose programming language, general-purpose, multi-paradigm programming language which extends the Caml dialect of ML (programming language), ML with object-oriented programming, object-oriented ...
prefix octal values with
0O
or
0o
, following the layout used by hexadecimal values.
* Several languages, including
Java,
C#,
Scala,
Python,
Ruby,
OCaml
OCaml ( , formerly Objective Caml) is a general-purpose programming language, general-purpose, multi-paradigm programming language which extends the Caml dialect of ML (programming language), ML with object-oriented programming, object-oriented ...
, C (starting from C23) and C++ can represent binary values by prefixing a number with
0B
or
0b
.
See also
*
Arbitrary-precision arithmetic
*
Binary-coded decimal
In computing and electronic systems, binary-coded decimal (BCD) is a class of binary encodings of decimal numbers where each digit is represented by a fixed number of bits, usually four or eight. Sometimes, special bit patterns are used for ...
(BCD)
*
C data types
*
Integer overflow
*
Signed number representations
Notes
References
{{DEFAULTSORT:Integer (Computer Science)
Data types
Computer arithmetic
Primitive types