In telecommunication, Hamming codes are a family of linear errorcorrecting codes that generalize the Hamming(7,4)code invented by Richard Hamming in 1950. Hamming codes can detect up to two and correct up to one bit errors. By contrast, the simple parity code cannot correct errors, and can detect only an odd number of errors. Hamming codes are special in that they are perfect codes, that is, they achieve the highest possible rate for codes with their block length and minimum distance 3.^{[1]}
In mathematical terms, Hamming codes are a class of binary linear codes. For each integer r \ge 2 there is a code with block length n=2^r1 and message length k=2^rr1. Hence the rate of Hamming codes is R=k/n=1r / (2^r1), which is highest possible for codes with distance 3 and block length 2^r1. The paritycheck matrix of a Hamming code is constructed by listing all columns of length r that are pairwise linearly independent.
Because of the simplicity of Hamming codes, they are widely used in computer memory (ECC memory). In this context, one often uses an extended Hamming code with one extra parity bit. Extended Hamming codes achieve a distance of 4, which allows the decoder to distinguish between the situation in which at most one bit error occurred and the situation in which two bit errors occurred. In this sense, extended Hamming codes are singleerror correcting and doubleerror detecting, and often referred to as SECDED.
History
Hamming worked at Bell Labs in the 1940s on the Bell Model V computer, an electromechanical relaybased machine with cycle times in seconds. Input was fed in on punched cards, which would invariably have read errors. During weekdays, special code would find errors and flash lights so the operators could correct the problem. During afterhours periods and on weekends, when there were no operators, the machine simply moved on to the next job.
Hamming worked on weekends, and grew increasingly frustrated with having to restart his programs from scratch due to the unreliability of the card reader. Over the next few years he worked on the problem of errorcorrection, developing an increasingly powerful array of algorithms. In 1950 he published what is now known as Hamming Code, which remains in use today in applications such as ECC memory.
Codes predating Hamming
A number of simple errordetecting codes were used before Hamming codes, but none were as effective as Hamming codes in the same overhead of space.
Parity
Parity adds a single bit that indicates whether the number of 1 bits in the preceding data was even or odd. If an odd number of bits is changed in transmission, the message will change parity and the error can be detected at this point. (Note that the bit that changed may have been the parity bit itself!) The most common convention is that a parity value of 1 indicates that there is an odd number of ones in the data, and a parity value of 0 indicates that there is an even number of ones in the data. In other words: the data and the parity bit together should contain an even number of 1s.
Parity checking is not very robust, since if the number of bits changed is even, the check bit will be valid and the error will not be detected. Moreover, parity does not indicate which bit contained the error, even when it can detect it. The data must be discarded entirely and retransmitted from scratch. On a noisy transmission medium, a successful transmission could take a long time or may never occur. However, while the quality of parity checking is poor, since it uses only a single bit, this method results in the least overhead. Furthermore, parity checking does allow for the restoration of an erroneous bit when its position is known.
Twooutoffive code
A twooutoffive code is an encoding scheme which uses five digits consisting of exactly three 0s and two 1s. This provides ten possible combinations, enough to represent the digits 0  9. This scheme can detect all single biterrors and all odd numbered biterrors. However it still cannot correct for these errors.
Repetition
Another code in use at the time repeated every data bit several times in order to ensure that it got through. For instance, if the data bit to be sent was a 1, an n=3 repetition code would send "111". If the three bits received were not identical, an error occurred. If the channel is clean enough, most of the time only one bit will change in each triple. Therefore, 001, 010, and 100 each correspond to a 0 bit, while 110, 101, and 011 correspond to a 1 bit, as though the bits counted as "votes" towards what the original bit was. A code with this ability to reconstruct the original message in the presence of errors is known as an errorcorrecting code. This triple repetition code is actually the simplest Hamming code with m = 2, since there are 2 parity bits, and 2^2  2  1 = 1 data bit.
Such codes cannot correctly repair all errors, however. In our example, if the channel flipped two bits and the receiver got "001", the system would detect the error, but conclude that the original bit was 0, which is incorrect. If we increase the number of times we duplicate each bit to four, we can detect all twobit errors but can't correct them (the votes "tie"); at five, we can correct all twobit errors, but not all threebit errors.
Moreover, the repetition code is extremely inefficient, reducing throughput by three times in our original case, and the efficiency drops drastically as we increase the number of times each bit is duplicated in order to detect and correct more errors.
Hamming codes
If more errorcorrecting bits are included with a message, and if those bits can be arranged such that different incorrect bits produce different error results, then bad bits could be identified. In a 7bit message, there are seven possible single bit errors, so three error control bits could potentially specify not only that an error occurred but also which bit caused the error.
Hamming studied the existing coding schemes, including twooffive, and generalized their concepts. To start with, he developed a nomenclature to describe the system, including the number of data bits and errorcorrection bits in a block. For instance, parity includes a single bit for any data word, so assuming ASCII words with 7bits, Hamming described this as an (8,7) code, with eight bits in total, of which 7 are data. The repetition example would be (3,1), following the same logic. The code rate is the second number divided by the first, for our repetition example, 1/3.
Hamming also noticed the problems with flipping two or more bits, and described this as the "distance" (it is now called the Hamming distance, after him). Parity has a distance of 2, as any two bit flips will be invisible. The (3,1) repetition has a distance of 3, as three bits need to be flipped in the same triple to obtain another code word with no visible errors. A (4,1) repetition (each bit is repeated four times) has a distance of 4, so flipping two bits can be detected, but not corrected. When three bits flip in the same group there can be situations where the code corrects towards the wrong code word.
Hamming was interested in two problems at once; increasing the distance as much as possible, while at the same time increasing the code rate as much as possible. During the 1940s he developed several encoding schemes that were dramatic improvements on existing codes. The key to all of his systems was to have the parity bits overlap, such that they managed to check each other as well as the data.
General algorithm
The following general algorithm generates a singleerror correcting (SEC) code for any number of bits.
 Number the bits starting from 1: bit 1, 2, 3, 4, 5, etc.
 Write the bit numbers in binary. 1, 10, 11, 100, 101, etc.
 All bit positions that are powers of two (have only one 1 bit in the binary form of their position) are parity bits.
 All other bit positions, with two or more 1 bits in the binary form of their position, are data bits.
 Each data bit is included in a unique set of 2 or more parity bits, as determined by the binary form of its bit position.
 Parity bit 1 covers all bit positions which have the least significant bit set: bit 1 (the parity bit itself), 3, 5, 7, 9, etc.
 Parity bit 2 covers all bit positions which have the second least significant bit set: bit 2 (the parity bit itself), 3, 6, 7, 10, 11, etc.
 Parity bit 4 covers all bit positions which have the third least significant bit set: bits 4 7, 12 15, 20 23, etc.
 Parity bit 8 covers all bit positions which have the fourth least significant bit set: bits 8 15, 24 31, 40 47, etc.
 In general each parity bit covers all bits where the binary AND of the parity position and the bit position is nonzero.
The form of the parity is irrelevant. Even parity is simpler from the perspective of theoretical mathematics, but there is no difference in practice.
This general rule can be shown visually:
 !colspan="2" Bit position ! 1 !! 2 !! 3 !! 4 !! 5 !! 6 !! 7 !! 8 !! 9 !! 10 !! 11 !! 12 !! 13 !! 14 !! 15 !! 16 !! 17 !! 18 !! 19 !! 20 rowspan="7" ...  !colspan="2" Encoded data bits !style="backgroundcolor: #90FF90;" p1 !style="backgroundcolor: #90FF90;" p2 !! d1 !style="backgroundcolor: #90FF90;" p4 !! d2 !! d3 !! d4 !style="backgroundcolor: #90FF90;" p8 !! d5 !! d6 !! d7 !! d8 !! d9 !! d10 !! d11 !style="backgroundcolor: #90FF90;" p16 !! d12 !! d13 !! d14 !! d15  !rowspan="5"Parity
bit
coverage !style="backgroundcolor: #90FF90;" p1  X   X   X   X   X   X   X   X   X   X   !style="backgroundcolor: #90FF90;" p2   X  X    X  X    X  X    X  X    X  X   !style="backgroundcolor: #90FF90;" p4     X  X  X  X      X  X  X  X      X  !style="backgroundcolor: #90FF90;" p8         X  X  X  X  X  X  X  X       !style="backgroundcolor: #90FF90;" p16                 X  X  X  X  X }
Shown are only 20 encoded bits (5 parity, 15 data) but the pattern continues indefinitely. The key thing about Hamming Codes that can be seen from visual inspection is that any given bit is included in a unique set of parity bits. To check for errors, check all of the parity bits. The pattern of errors, called the error syndrome, identifies the bit in error. If all parity bits are correct, there is no error. Otherwise, the sum of the positions of the erroneous parity bits identifies the erroneous bit. For example, if the parity bits in positions 1, 2 and 8 indicate an error, then bit 1+2+8=11 is in error. If only one parity bit indicates an error, the parity bit itself is in error.
As you can see, if you have m parity bits, it can cover bits from 1 up to 2^m1. If we subtract out the parity bits, we are left with 2^mm1 bits we can use for the data. As m varies, we get all the possible Hamming codes:
Parity bits 
Total bits 
Data bits 
Name 
Rate 
2 
3 
1 
Hamming(3,1) (Triple repetition code) 
1/3 0.333 
3 
7 
4 
Hamming(7,4) 
4/7 0.571 
4 
15 
11 
Hamming(15,11) 
11/15 0.733 
5 
31 
26 
Hamming(31,26) 
26/31 0.839 
... 
m 
2^m1 
2^mm1 
Hamming(2^m1,2^mm1)

1m/(2^m1) 
If, in addition, an overall parity bit (bit 0) is included, the code can detect (but not correct) any twobit error, making a SECDED code. The overall parity indicates whether the total number of errors is even or odd. If the basic Hamming code detects an error, but the overall parity says that there are an even number of errors, an uncorrectable 2bit error has occurred.
Hamming codes with additional parity (SECDED)
Hamming codes have a minimum distance of 3, which means that the decoder can detect and correct a single error, but it cannot distinguish a double bit error of some codeword from a single bit error of a different codeword. Thus, they can detect doublebit errors only if correction is not attempted.
To remedy this shortcoming, Hamming codes can be extended by an extra parity bit. This way, it is possible to increase the minimum distance of the Hamming code to 4, which allows the decoder to distinguish between single bit errors and twobit errors. Thus the decoder can detect and correct a single error and at the same time detect (but not correct) a double error. If the decoder does not attempt to correct errors, it can detect up to 3 errors.
This extended Hamming code is popular in computer memory systems, where it is known as SECDED ("single error correction, double error detection"). Particularly popular is the (72,64) code, a truncated (127,120) Hamming code plus an additional parity bit, which has the same space overhead as a (9,8) parity code.
Hamming(7,4) code
Graphical depiction of the 4 data bits and 3 parity bits and which parity bits apply to which data bits
In 1950, Hamming introduced the (7,4) code. It encodes 4 data bits into 7 bits by adding three parity bits. Hamming(7,4) can detect and correct singlebit errors. With the addition of an overall parity bit, it can also detect (but not correct) doublebit errors.
Construction of G and H
The matrix \mathbf{G} := \begin{pmatrix} I_k  A^T \\ \end{pmatrix} is called a (Canonical) generator matrix of a linear (n,k) code,
and \mathbf{H} := \begin{pmatrix} A  I_{nk} \\ \end{pmatrix} is called a paritycheck matrix.
This is the construction of G and H in standard (or systematic) form. Regardless of form, G and H for linear block codes must satisfy
\mathbf{H}\,\mathbf{G}^T = \mathbf{0}, an allzeros matrix [Moon, p. 89].
Since (7,4,3)=(n,k,d)=[2^{m} 1, 2^{m} 1m, m]. The paritycheck matrix H of a Hamming code is constructed by listing all columns of length m that are pairwise independent.
Thus H is a matrix whose left side is all of the nonzero ntuples where order of the ntuples in the columns of matrix does not matter. The right hand side is just the (nk)identity matrix.
So G can be obtained from H by taking the transpose of the left hand side of H with the identity kidentity matrix on the left hand side of G.
The code generator matrix \mathbf{G} and the paritycheck matrix \mathbf{H} are:
\mathbf{G} := \begin{pmatrix} 1 & 0 & 0 & 0 & 1 & 1 & 0 \\ 0 & 1 & 0 & 0 & 1 & 0 & 1 \\ 0 & 0 & 1 & 0 & 0 & 1 & 1 \\ 0 & 0 & 0 & 1 & 1 & 1 & 1 \\ \end{pmatrix}_{4,7} and $\backslash mathbf\{H\}\; :=\; \backslash begin\{pmatrix\}\; 1\; \&\; 1\; \&\; 0\; \&\; 1\; \&\; 1\; \&\; 0\; \&\; 0\; \backslash \backslash \; 1\; \&\; 0\; \&\; 1\; \&\; 1\; \&\; 0\; \&\; 1\; \&\; 0\; \backslash \backslash \; 0\; \&\; 1\; \&\; 1\; \&\; 1\; \&\; 0\; \&\; 0\; \&\; 1\; \backslash \backslash \; \backslash end\{pmatrix\}\_\{3,7\}.$
Finally, these matrices can be mutated into equivalent nonsystematic codes by the following operations [Moon, p. 85]:
 Column permutations (swapping columns)
 Elementary row operations (replacing a row with a linear combination of rows)
Encoding
Example
From the above matrix we have 2^{k}=2^{4}=16 codewords. The codewords \overrightarrow{x} of this binary code can be obtained from \overrightarrow{x}=\overrightarrow{a}G . With \overrightarrow{a}=a_1a_2a_3a_4 with a_i exist in F_2 ( A field with two elements namely 0 and 1).
Thus the codewords are all the 4tuples (ktuples).
Therefore,
(1,0,1,1) gets encoded as (1,0,1,1,0,1,0).
Hamming(7,4) code with an additional parity bit
The same (7,4) example from above with an extra parity bit
The Hamming(7,4) can easily be extended to an (8,4) code by adding an extra parity bit on top of the (7,4) encoded word (see Hamming(7,4)). This can be summed up with the revised matrices:
 \mathbf{G} := \begin{pmatrix} 1 & 1 & 1 & 0 & 0 & 0 & 0 & 1\\ 1 & 0 & 0 & 1 & 1 & 0 & 0 & 1\\ 0 & 1 & 0 & 1 & 0 & 1 & 0 & 1\\ 1 & 1 & 0 & 1 & 0 & 0 & 1 & 0 \end{pmatrix}_{4,8}
and
 \mathbf{H} := \begin{pmatrix} 1 & 0 & 1 & 0 & 1 & 0 & 1 & 0\\ 0 & 1 & 1 & 0 & 0 & 1 & 1 & 0\\ 0 & 0 & 0 & 1 & 1 & 1 & 1 & 0\\ 1 & 1 & 1 & 1 & 1 & 1 & 1 & 1 \end{pmatrix}_{4,8} .
Note that H is not in standard form. To obtain G, elementary row operations can be used to obtain an equivalent matrix to H in systematic form:
 \mathbf{H} = \left(\left.\begin{array}{cccc} 0 & 1 & 1 & 1\\ 1 & 0 & 1 & 1\\ 1 & 1 & 0 & 1\\ 1 & 1 & 1 & 0\end{array}\right\begin{array}{cccc} 1 & 0 & 0 & 0\\ 0 & 1 & 0 & 0\\ 0 & 0 & 1 & 0\\ 0 & 0 & 0 & 1\end{array}\right)_{4,8} .
For example, the first row in this matrix is the sum of the second and third rows of H in nonsystematic form. Using the systematic construction for Hamming codes from above, the matrix A is apparent and the systematic form of G is written as
 \mathbf{G} = \left(\left.\begin{array}{cccc} 1 & 0 & 0 & 0\\ 0 & 1 & 0 & 0\\ 0 & 0 & 1 & 0\\ 0 & 0 & 0 & 1\end{array}\right\begin{array}{cccc} 0 & 1 & 1 & 1\\ 1 & 0 & 1 & 1\\ 1 & 1 & 0 & 1\\ 1 & 1 & 1 & 0\end{array}\right)_{4,8} .
The nonsystematic form of G can be row reduced (using elementary row operations) to match this matrix.
The addition of the fourth row effectively computes the sum of all the codeword bits (data and parity) as the fourth parity bit.
For example, 1011 is encoded into 01100110 where blue digits are data; red digits are parity from the Hamming(7,4) code; and the green digit is the parity added by Hamming(8,4). The green digit makes the parity of the (7,4) code even.
Finally, it can be shown that the minimum distance has increased from 3, as with the (7,4) code, to 4 with the (8,4) code. Therefore, the code can be defined as Hamming(8,4,4).
See also
Notes
References
External links
ca:Codi de Hamming cs:Hamming v k d de:HammingCode es:C digo Hamming eu:Hamming kode fa: fr:Code de Hamming ko: it:Codice di Hamming he: ml: nl:Hammingcode ja: pl:Kod Hamminga pt:C digo de Hamming ru: simple:Hamming code tr:Hamming kodu vi:M Hamming zh:
