The central problem regarding constant-weight codes is the following: what is the maximum number of codewords in a binary constant-weight code with length , Hamming distance , and weight ? This number is called .
Apart from some trivial observations, it is generally impossible to compute these numbers in a straightforward way. Upper bounds are given by several important theorems such as the first and second Johnson bounds,[2] and better upper bounds can sometimes be found in other ways. Lower bounds are most often found by exhibiting specific codes, either with use of a variety of methods from discrete mathematics, or through heavy computer searching. A large table of such record-breaking codes was published in 1990,[3] and an extension to longer codes (but only for those values of and which are relevant for the GSM application) was published in 2006.[1]
A special case of constant weight codes are the one-of-N codes, that encode bits in a code-word of bits. The one-of-two code uses the code words 01 and 10 to encode the bits '0' and '1'. A one-of-four code can use the words 0001, 0010, 0100, 1000 in order to encode two bits 00, 01, 10, and 11. An example is dual rail encoding, and chain link [4] used in delay insensitive circuits. For these codes, and .
Some of the more notable uses of one-hot codes include
biphase mark code uses a 1-of-2 code;
pulse-position modulation uses a 1-of-n code;
address decoder,
etc.
In coding theory, a balanced code is a binary forward error correction code for which each codeword contains an equal number of zero and one bits. Balanced codes have been introduced by Donald Knuth;[5] they are a subset of so-called unordered codes, which are codes having the property that the positions of ones in a codeword are never a subset of the positions of the ones in another codeword. Like all unordered codes, balanced codes are suitable for the detection of all unidirectional errors in an encoded message. Balanced codes allow for particularly efficient decoding, which can be carried out in parallel.[5][6][7]
Some of the more notable uses of balanced-weight codes include
biphase mark code uses a 1 of 2 code;
6b/8b encoding uses a 4 of 8 code;
the Hadamard code is a of code (except for the zero codeword),
the three-of-six code;
etc.
The 3-wire lane encoding used in MIPI C-PHY can be considered a generalization of constant-weight code to ternary -- each wire transmits a ternary signal, and at any one instant one of the 3 wires is transmitting a low, one is transmitting a middle, and one is transmitting a high signal.[8]
An m-of-n code is a separable error detection code with a code word length of n bits, where each code word contains exactly m instances of a "one". A single bit error will cause the code word to have either m + 1 or m − 1 "ones". An example m-of-n code is the 2-of-5 code used by the United States Postal Service.
The simplest implementation is to append a string of ones to the original data until it contains m ones, then append zeros to create a code of length n.
Example:
More information Original 3 data bits, Appended bits ...
3-of-6 code
Original 3 data bits | Appended bits |
000 | 111 |
001 | 110 |
010 | 110 |
011 | 100 |
100 | 110 |
101 | 100 |
110 | 100 |
111 | 000 |
Close
Some of the more notable uses of constant-weight codes, other than the one-hot and balanced-weight codes already mentioned above, include
Code 39 uses a 3-of-9 code;
bi-quinary coded decimal code uses a 2-of-7 code,
the 2-of-5 code,
etc.
See pp. 526–527 of F. J. MacWilliams and N. J. A. Sloane (1979). The Theory of Error-Correcting Codes. Amsterdam: North-Holland.
A. E. Brouwer, James B. Shearer, N. J. A. Sloane and Warren D. Smith (1990). "A New Table of Constant Weight Codes". IEEE Transactions of Information Theory 36.
Sulaiman Al-Bassam; Bella Bose (March 1990). "On Balanced Codes". IEEE Transactions on Information Theory. 36 (2): 406–408. doi:10.1109/18.52490.