Concatenated error correction code: Difference between revisions

Content deleted Content added
m MOS:MATHSPECIAL / convert special characters found by Wikipedia:Typo Team/moss (via WP:JWB)
 
(87 intermediate revisions by 36 users not shown)
Line 1:
{{Use dmy dates|date=July 2014}}
In [[coding theory]], '''concatenated codes''' form a class of [[Error detection and correction|error-correcting codes]] that are derived by combining an '''inner code''' and an '''outer code'''. They were conceived in 1966 by [[Dave Forney]] as a solution for the problem of finding a code that has both exponentially decreasing error probability with increasing block length and [[polynomial-time]] decoding [[Computational_complexity_theory|complexity]].
In [[coding theory]], '''concatenated codes''' form a class of [[error-correcting code]]s that are derived by combining an '''inner code''' and an '''outer code'''. They were conceived in 1966 by [[Dave Forney]] as a solution to the problem of finding a code that has both exponentially decreasing error probability with increasing block length and [[polynomial-time]] decoding [[Computational complexity theory|complexity]].<ref name="Forney">
{{cite journal
|author=G. D. Forney
|author-link=Dave Forney
|title=Concatenated codes
|publisher=MIT Press
|___location=Cambridge, Massachusetts
|year=1967
}}
</ref>
Concatenated codes became widely used in space communications in the 1970s.
 
==Background==
The field of [[channel coding]] is concerned with sending a stream of data at the highest possible rate over a given [[communications channel]], and then decoding the original data reliably at the receiver, using encoding and decoding algorithms that are feasible to implement in a given technology.
 
[[Noisy-channel coding theorem|Shannon's channel coding theorem]] shows that over many common channels there exist channel coding schemes that are able to transmit data reliably at all rates <math>R</math> less than a certain threshold <math>C</math>, called the [[channel capacity]] of the given channel. In fact, the probability of decoding error can be made to decrease exponentially as the block length <math>N</math> of the coding scheme goes to infinity. However, the complexity of a naive optimum decoding scheme that simply computes the likelihood of every possible transmitted codeword increases exponentially with <math>N</math>, so such an optimum decoder rapidly becomes infeasible.
 
In his [https://web.archive.org/web/20121012080412/http://mitpress.mit.edu/catalog/item/default.asp?tid=5813&ttype=2 doctoral thesis], [[Dave Forney]] showed that concatenated codes could be used to achieve exponentially decreasing error probabilities at all data rates less than capacity, with decoding complexity that increases only polynomially with the code block length.
 
==Description==
[[File:Concatenated codes diagram.png|thumb|upright=2|Schematic depiction of a concatenated code built upon an inner code and an outer code.]]
Let ''C'' be a code with length ''N'' and rate ''R'' over an alphabet ''A'' with ''K''=''N*R'' symbols. Let ''I'' be another code with length ''n'' and rate ''r'' over an alphabet ''B'' with ''k''=''n*r'' symbols.
[[File:Concatenation of Reed–Solomon code with Hadamard code.svg|thumb|400px|This is a pictorial representation of a code concatenation, and, in particular, the [[Reed–Solomon code]] with n=q=4 and k=2 is used as the outer code and the [[Hadamard code]] with n=q and k=log q is used as the inner code. Overall, the concatenated code is a <math>[q^2,k \log q]</math>-code.]]
 
Let ''C''<sub>''in''</sub> be a [''n'', ''k'', ''d''] code, that is, a [[block code]] of length ''n'', [[dimension (vector space)|dimension]] ''k'', minimum [[Hamming distance]] ''d'', and [[code rate|rate]] ''r'' = ''k''/''n'', over an alphabet ''A'':
The inner code ''I'' takes one of ''k'' possible inputs, encodes onto an ''n''-tuple from ''B'', transmits, and decodes into one of ''k'' possible outputs. We regard this as a channel which can transmit one symbol from the alphabet ''A'', also of size ''k''. We use this channel ''N'' times to transmit each of the ''N'' symbols in a codeword of ''C''. The ''concatenation'' of ''C'' (as outer code) with ''I'' as (inner code) is thus a code of length ''Nn'' over the alphabet ''B''.
:<math>C_{in}: A^k \rightarrow A^n</math>
Let ''C''<sub>''out''</sub> be a [''N'', ''K'', ''D''] code over an alphabet ''B'' with |''B''| = |''A''|<sup>''k''</sup> symbols:
:<math>C_{out}: B^K \rightarrow B^N</math>
The inner code ''C''<sub>''in''</sub> takes one of |''A''|<sup>''k''</sup> = |''B''| possible inputs, encodes into an ''n''-tuple over ''A'', transmits, and decodes into one of |''B''| possible outputs. We regard this as a (super) channel which can transmit one symbol from the alphabet ''B''. We use this channel ''N'' times to transmit each of the ''N'' symbols in a codeword of ''C''<sub>''out''</sub>. The ''concatenation'' of ''C''<sub>''out''</sub> (as outer code) with ''C''<sub>''in''</sub> (as inner code), denoted ''C''<sub>''out''</sub><math>\circ</math>''C''<sub>''in''</sub>, is thus a code of length ''Nn'' over the alphabet ''A'':<ref name="Forney"/>
:<math>C_{out} \circ C_{in}: A^{kK} \rightarrow A^{nN}</math>
It maps each input message ''m'' = (''m''<sub>1</sub>, ''m''<sub>2</sub>, ..., ''m''<sub>K</sub>) to a codeword (''C''<sub>''in''</sub>(''m''<nowiki>'</nowiki><sub>1</sub>), ''C''<sub>''in''</sub>(''m''<nowiki>'</nowiki><sub>2</sub>), ..., ''C''<sub>''in''</sub>(''m''<nowiki>'</nowiki><sub>N</sub>)),
where (''m''<nowiki>'</nowiki><sub>1</sub>, ''m''<nowiki>'</nowiki><sub>2</sub>, ..., ''m''<nowiki>'</nowiki><sub>N</sub>) = ''C''<sub>''out''</sub>(''m''<sub>1</sub>, ''m''<sub>2</sub>, ..., ''m''<sub>K</sub>).
 
The ''key insight'' in this approach is that if ''IC''<sub>''in''</sub> is decoded using a [[maximum likelihood decoding|maximum-likelihood approach]] (thus showing an exponentially decreasing error probability with increasing length), and ''C''<sub>''out''</sub> is a code with length ''N'' = 2^<sup>''(n*r)nr''</sup> that can be decoded in polynomial time of ''N'', then the concatenated code can be decoded in polynomial time of its combined length ''n*''2^<sup>''nr''</sup> = [[O notation|''O'']](n*r)''N''⋅log(''N'')) and shows an exponentially decreasing error probability, even if ''IC''<sub>''in''</sub> has exponential decoding complexity.<ref name="Forney"/> This is discussed in more detail in section [[#Decoding concatenated codes|Decoding concatenated codes]].
 
In a generalisationgeneralization of above concatenation, there are ''N'' possible inner codes ''IC''<sub>''in'',''i''</sub> and the ''i''-th symbol in a codeword of ''C''<sub>''out''</sub> is transmitted across the inner channel using the ''i''-th inner code. The [[Justesen code]]s are examples of generalisedgeneralized concatenated codes, where the outer code is a [[Reed-SolomonReed–Solomon code]].
 
==Properties==
== Simplified mathematical overview ==
'''1.''' The distance of the concatenated code ''C''<sub>''out''</sub><math>\circ</math>''C''<sub>''in''</sub> is at least ''dD'', that is, it is a [''nN'', ''kK'', ''D''<nowiki>'</nowiki>] code with ''D''<nowiki>'</nowiki> ≥ ''dD''.
How does concatenated error correction work in real transmission situations? In essence you have a datastream, typically composed of packets containing and obeying the general process or function
 
''Proof:''
* Packet = Outer Code (Inner Code(data))
Consider two different messages ''m''<sup>1</sup> ≠ ''m''<sup>2</sup> ∈ ''B''<sup>''K''</sup>. Let Δ denote the distance between two codewords. Then
:<math>\Delta(C_{out}(m^1), C_{out}(m^2)) \ge D.</math>
 
Thus, there are at least ''D'' positions in which the sequence of ''N'' symbols of the codewords ''C''<sub>''out''</sub>(''m''<sup>1</sup>) and ''C''<sub>''out''</sub>(''m''<sup>2</sup>) differ. For these positions, denoted ''i'', we have
That may or may not at some further point be defined as
:<math>\Delta(C_{in}(C_{out}(m^1)_i), C_{in}(C_{out}(m^2)_i)) \ge d.</math>
 
Consequently, there are at least ''d''⋅''D'' positions in the sequence of ''n''⋅''N'' symbols taken from the alphabet ''A'' in which the two codewords differ, and hence
* datastream = randomizer(Packet)
:<math>\Delta(C_{in}(C_{out}(m^1)), C_{in}(C_{out}(m^2))) \ge dD.</math>
* datastream = interleaver(Packet)
 
'''2.''' If ''C''<sub>''out''</sub> and ''C''<sub>''in''</sub> are [[linear block code]]s, then ''C''<sub>''out''</sub><math>\circ</math>''C''<sub>''in''</sub> is also a linear block code.
The receiver and decoder at the other end merely reverse the process, weather this be on a [[Compact Disc]] player or a [[Deep Space Network]] telemetry system.
 
This property can be easily shown based on the idea of defining a [[generator matrix]] for the concatenated code in terms of the generator matrices of ''C''<sub>''out''</sub> and ''C''<sub>''in''</sub>.
 
==Decoding concatenated codes==
The randomizer or interleaver essentially perform the same function of removing any DC biases in datastream.
 
A natural concept for a decoding algorithm for concatenated codes is to first decode the inner code and then the outer code. For the algorithm to be practical it must be [[polynomial-time]] in the final block length. Consider that there is a polynomial-time unique decoding algorithm for the outer code. Now we have to find a polynomial-time decoding algorithm for the inner code. It is understood that polynomial running time here means that running time is polynomial in the final block length. The main idea is that if the inner block length is selected to be logarithmic in the size of the outer code then the decoding algorithm for the inner code may run in [[exponential time]] of the inner block length, and we can thus use an exponential-time but optimal [[Decoding methods#Maximum likelihood decoding|maximum likelihood decoder]] (MLD) for the inner code.
* Interleavers scramble data but do not alter it in any way.
* Randomizers alter data, but mostly do not touch its structure.
* Concatenating an interleaver and a randomizer can improve link margins.
 
In detail, let the input to the decoder be the vector ''y'' = (''y''<sub>1</sub>, ..., ''y''<sub>''N''</sub>) ∈ (''A''<sup>''n''</sup>)<sup>''N''</sup>. Then the decoding algorithm is a two-step process:
The [[CCSDS]] Standard and [[NICAM]] Standard interleavers are closely related to each other. A good interleaver works just as well for a TV broadcaster as it would for a deep space mission. CCSDS recommends always using the CCSDS interleaver -- and NICAM's interleaver cannot be turned off. Space missions that have not used interleavers have run into telecom problems, so their usage is essentially mandatory as far as the CCSDS is concerned.
# Use the MLD of the inner code ''C''<sub>in</sub> to reconstruct a set of inner code words ''y''<nowiki>'</nowiki> = (''y''<nowiki>'</nowiki><sub>1</sub>, ..., ''y''<nowiki>'</nowiki><sub>''N''</sub>), with ''y''<nowiki>'</nowiki><sub>''i''</sub> = MLD<sub>''C''<sub>in</sub></sub>(''y''<sub>i</sub>), 1 ≤ ''i'' ≤ ''N''.
# Run the unique decoding algorithm for ''C''<sub>out</sub> on ''y''<nowiki>'</nowiki>.
 
Now, the time complexity of the first step is [[O notation|''O'']](''N''⋅exp(''n'')), where ''n'' = ''O''(log(''N'')) is the inner block length. In other words, it is ''N''<sup>''O''(1)</sup> (i.e., polynomial-time) in terms of the outer block length ''N''. As the outer decoding algorithm in step two is assumed to run in polynomial time the complexity of the overall decoding algorithm is polynomial-time as well.
==Applications==
Concatenated codes were first implemented for [[deep space]] communication in the [[Voyager program]], which launched their first probe in 1977, but all images were transmitted with a Golay Code.<ref name="deep-space-codes">K. Andrews et al., ''The Development of Turbo and LDPC Codes for Deep-Space Applications'', Proceedings of the IEEE, Vol. 95, No. 11, Nov. 2007.</ref> Since then, concatenated codes became the workhorse for efficient error correction coding, and stayed so at least until the invention of [[turbo codes]] and [[LDPC codes]].
 
===Remarks===
Typically, the inner code is not a block code but a soft-decision [[convolutional code|convolutional]] [[Viterbi decoder|Viterbi-decoded]] code with a short constraint length. For the outer code, a longer hard-decision block code, frequently [[Reed Solomon]] with 8-bit symbols, is selected. The larger symbol size makes the outer code more robust to [[error burst|burst error]]s that may occur due to channel impairments, and because erroneous output of the convolutional code itself is bursty. Additionally, an [[interleaving]] layer (possibly coupled with a randomizer) may be used to spread burst errors across a wider range.
 
The decoding algorithm described above can be used to correct all errors up to less than ''dD''/4 in number. Using [[minimum distance decoding]], the outer decoder can correct all inputs ''y''<nowiki>'</nowiki> with less than ''D''/2 symbols ''y''<nowiki>'</nowiki><sub>''i''</sub> in error. Similarly, the inner code can reliably correct an input ''y''<sub>''i''</sub> if less than ''d''/2 inner symbols are erroneous. Thus, for an outer symbol ''y''<nowiki>'</nowiki><sub>''i''</sub> to be incorrect after inner decoding at least ''d''/2 inner symbols must have been in error, and for the outer code to fail this must have happened for at least ''D''/2 outer symbols. Consequently, the total number of inner symbols that must be received incorrectly for the concatenated code to fail must be at least ''d''/2⋅''D''/2 = ''dD''/4.
The combination of an inner Viterbi convolutional code with an outer Reed-Solomon code (known as an RSV code) was to became the most commonly used Error Correction Code construction outside of the Space sector. The RSV code was thoroughly proven on the Voyager Program spacecraft over a 20 year time span.
 
The algorithm also works if the inner codes are different, e.g., for [[Justesen code]]s. The [[generalized minimum distance decoding|generalized minimum distance algorithm]], developed by Forney, can be used to correct up to ''dD''/2 errors.<ref name="gmd">
The RSV code is still in use today for deep-space and [[satellite communication]], notably the [[DVB-S]] [[digital television]] broadcast standard.
{{cite journal
|first=G. David
|last=Forney
|title=Generalized Minimum Distance Decoding
|journal=IEEE Transactions on Information Theory
|volume=12
|issue=2
|pages=125–131
|date=April 1966
|doi=10.1109/TIT.1966.1053873
}}</ref>
It uses [[erasure code|erasure]] information from the inner code to improve performance of the outer code, and was the first example of an algorithm using [[soft-decision decoding]].<ref>{{cite journal
|first1=Christopher C.H.
|last1=Yu
|first2=Daniel J.
|last2=Costello
|title=Generalized Minimum Distance Decoding for ''Q''ary Output Channels
|journal=IEEE Transactions on Information Theory
|volume=26
|issue=2
|pages=238–243
|date=March 1980
|doi=10.1109/TIT.1980.1056148
}}</ref><ref>{{cite journal
|first1=Yingquan
|last1=Wu
|first2=Christoforos
|last2=Hadjicostis
|title=Soft-Decision Decoding of Linear Block Codes Using Preprocessing and Diversification
|journal=IEEE Transactions on Information Theory
|volume=53
|issue=1
|pages=387–393
|date=January 2007
|doi=10.1109/tit.2006.887478
|s2cid=8338433
}}</ref>
 
==Applications==
In a more loose sense, any (serial) combination of two or more codes may be referred to as a concatenated code. For example, within the [[DVB-S2]] standard, a highly efficient [[LDPC code]] is combined with an algebraic outer code in order to remove any resilient errors left over from the inner LDPC code due to its inherent [[error floor]].
Although a simple concatenation scheme was implemented already for the 1971 [[Mariner 8|Mariner]] Mars orbiter mission,<ref name="McEliece"/> concatenated codes were starting to be regularly used for [[Deep Space Network|deep space]] communication with the [[Voyager program]], which launched two [[space probe]]s in 1977.<ref name="deep-space-codes">K. Andrews et al., ''The Development of Turbo and LDPC Codes for Deep-Space Applications'', Proceedings of the IEEE, Vol. 95, No. 11, Nov. 2007.</ref> Since then, concatenated codes became the workhorse for efficient error correction coding, and stayed so at least until the invention of [[turbo codes]] and [[LDPC codes]].<ref name="McEliece"/><ref name="deep-space-codes"/>
 
Typically, the inner code is not a block code but a [[soft-decision decoder|soft-decision]] [[convolutional code|convolutional]] [[Viterbi decoder|Viterbi-decoded]] code with a short constraint length.<ref name="Odenwalder">
{{cite journal
|author=J. P. Odenwalder
|title=Optimal decoding of convolutional codes
|publisher=[[U.C.L.A.]], Systems Science Dept. (dissertation)
|year=1970
}}
</ref>
For the outer code, a longer hard-decision block code, frequently a [[Reed-Solomon code]] with eight-bit symbols, is used.<ref name="Forney"/><ref name="McEliece">
{{cite journal
|author1=Robert J. McEliece
|author-link=Robert McEliece
|author2=Laif Swanson
|title=Reed–Solomon Codes and the Exploration of the Solar System
|publisher=JPL
|date=20 August 1993
}}
</ref>
The larger symbol size makes the outer code more robust to [[error burst]]s that can occur due to channel impairments, and also because erroneous output of the convolutional code itself is bursty.<ref name="Forney"/><ref name="McEliece"/> An [[forward error correction#Interleaving|interleaving layer]] is usually added between the two codes to spread error bursts across a wider range.<ref name="McEliece"/>
 
The combination of an inner Viterbi convolutional code with an outer [[Reed–Solomon code]] (known as an RSV code) was first used in ''[[Voyager 2]]'',<ref name="McEliece"/><ref>R. Ludwig, J. Taylor, [http://descanso.jpl.nasa.gov/DPSummary/Descanso4--Voyager_new.pdf Voyager Telecommunications Manual], [[JPL]] DESCANSO ''(Design and Performance Summary Series)'', March 2002.</ref> and it became a popular construction both within and outside of the space sector. It is still notably used today for [[satellite communication]]s, such as the [[DVB-S]] [[digital television]] broadcast standard.<ref>[http://www.etsi.org/deliver/etsi_en/300400_300499/300421/01.01.02_60/en_300421v010102p.pdf Digital Video Broadcasting (DVB); Framing structure, channel coding and modulation for 11/12 GHz satellite services], [[ETSI]] EN 300 421, V1.1.2, August 1997.</ref>
 
In a looser sense, any (serial) combination of two or more codes may be referred to as a concatenated code. For example, within the [[DVB-S2]] standard, a highly efficient [[LDPC code]] is combined with an algebraic outer code in order to remove any resilient errors left over from the inner LDPC code due to its inherent [[error floor]].<ref>[http://www.etsi.org/deliver/etsi_en/302300_302399/302307/01.02.01_60/en_302307v010201p.pdf Digital Video Broadcasting (DVB); Second generation framing structure, channel coding and modulation systems for Broadcasting, Interactive Services, News Gathering and other broadband satellite applications (DVB-S2)], [[ETSI]] EN 302 307, V1.2.1, April 2009.</ref>
 
A simple concatenation scheme is also used on the compact disc (CD), where an interleaving layer between two Reed–Solomon codes of different sizes spreads errors across various blocks.
The Compact Disc uses concatenated Reed-Solomon codes (of different lengths), with no Viterbi code involvement at all.
 
== Turbo codes: A parallel concatenation approach ==
The description above is given for what is now called a serially concatenated code. [[Turbo code]]s, as described first in 1993, implemented a parallel concatenation of two convolutional codes, with an interleaver between the two codes and an iterative decoder that would passpasses information forth and back between the codes.<ref name="deep-space-codes"/> This constructiondesign hadhas mucha higherbetter performance than allany previously conveivedconceived concatenated codes. However, a key aspect of turbo codes is their iterated decoding approach.
 
However, a key aspect of turbo codes is their iterated decoding approach. Iterated decoding is now also applied to serial concatenations in order to achieve higher coding gains, such as withwithin theserially 2concatenated convolutional codes (SCCCs). An early form of iterated decoding was implemented with two to 5five iterationiterations ''Galileoin Code''the as"Galileo usedcode" byof the [[Galileo (spacecraft)|Galileo space probe]].<ref name="McEliece"/>
 
==See Notes also==
*[[Gilbert–Varshamov bound]]
<references/>
*[[Justesen code]]
*[[Singleton bound]]
*[[Zyablov bound]]
 
== References ==
{{reflist|33em}}
* {{cite book | author=F.J. MacWilliams | authorlink=Jessie MacWilliams | coauthors=N.J.A. Sloane | title=The Theory of Error-Correcting Codes | publisher=North-Holland | date=1977 | isbn=0-444-85193-3 | pages=307–316 }}
 
== ExternalFurther linksreading ==
* {{cite book |author1=Shu Lin |author2=Daniel J. Costello Jr. | title=Error Control Coding: Fundamentals and Applications |url=https://archive.org/details/errorcontrolcodi00lins_044 |url-access=limited | publisher=[[Prentice Hall]] | year=1983 | isbn=978-0-13-283796-5 | pages=[https://archive.org/details/errorcontrolcodi00lins_044/page/n296 278]–280}}
* {{cite book | author=F.J. MacWilliams | authorlink=Jessie MacWilliams |author2=N.J.A. Sloane | title=The Theory of Error-Correcting Codes | url=https://archive.org/details/theoryoferrorcor0000macw | url-access=registration | publisher=North-Holland | year=1977 | isbn=978-0-444-85193-2 | pages=[https://archive.org/details/theoryoferrorcor0000macw/page/307 307–316] }}
 
== External links ==
* {{scholarpedia|title=Concatenated codes|urlname=Concatenated_codes|curator=[[Dave Forney]]}}
* [https://web.archive.org/web/20110606191907/http://www.cse.buffalo.edu/~atri/courses/coding-theory/fall07.html University at Buffalo Lecture Notes on Coding Theory – Dr. Atri Rudra]
* [http://www.youtube.com/watch?v=Pa3-vrCEbW0 Why space missions fail (some detail on Galileo ECC)]
 
<br>
{{CCSDS|state=collapsed}}
 
Line 66 ⟶ 162:
[[Category:Error detection and correction]]
[[Category:Coding theory]]
[[Category:Finite fields]]
[[Category:Information theory]]