The Hadoop Distributed File System (HDFS) is a distributed file system that runs on standard or low-end hardware. Developed by Apache Hadoop, HDFS works like a standard distributed file system but provides better data throughput and access through the MapReduce algorithm, high fault tolerance and native support of large data sets.
A hamming code is a linear code for error detection that can detect up to two simultaneous bit errors and is capable of correcting single-bit errors. Reliable communication is assured if the hamming distance between the transmitter and receiver is less than or equal to one.
Hamming code was invented by Richard Hamming in 1950. The method is useful for a single bit change, which is more probable than two or more bit changes. The simplicity of hamming codes makes them suitable for use in computer memory and single-error correction. They use a double-error detection variant called SECDED. These codes have a minimum hamming distance of three, where the code detects and corrects single errors while double bit errors are detected only if a correction is not attempted. Adding an extra parity bit increases the minimum distance of the hamming code to four, which allows the code to detect and correct single errors while detecting double errors.Hamming initially introduced code that enclosed four data bits into seven bits by adding three parity bits. It can easily be extended to eight and four bit code by adding an extra parity bit on top of the encoded word.
Read More »
Join 138,000+ IT pros on our weekly newsletter
Home | Advertising Info | Write for Us | About | Contact Us
2010 - 2014
Janalta Interactive Sites: