Definition - What does Hamming Code mean?
A hamming code is a linear code for error detection that can detect up to two simultaneous bit errors and is capable of correcting single-bit errors. Reliable communication is assured if the hamming distance between the transmitter and receiver is less than or equal to one.
Techopedia explains Hamming Code
Hamming code was invented by Richard Hamming in 1950. The method is useful for a single bit change, which is more probable than two or more bit changes.
The simplicity of hamming codes makes them suitable for use in computer memory and single-error correction. They use a double-error detection variant called SECDED. These codes have a minimum hamming distance of three, where the code detects and corrects single errors while double bit errors are detected only if a correction is not attempted. Adding an extra parity bit increases the minimum distance of the hamming code to four, which allows the code to detect and correct single errors while detecting double errors.
Hamming initially introduced code that enclosed four data bits into seven bits by adding three parity bits. It can easily be extended to eight and four bit code by adding an extra parity bit on top of the encoded word.
Experts Share the Top Cybersecurity Trends to Watch for in 2017
Join thousands of others with our weekly newsletter
The 4th Era of IT Infrastructure: Superconverged Systems:
Approaches and Benefits of Network Virtualization:
Free E-Book: Public Cloud Guide:
Free Tool: Virtual Health Monitor:
Free 30 Day Trial – Turbonomic: