Opposed to block codes, convolutional codes don’t divide data bits into blocks. In convolutional codes, data bits are still in continuous sequence. In some occasions, convolution codes maybe more convenient than a block code, because it generates redundant bits and corrects errors continuously.
Convolutional code can be marked by (n, k, K), which means for every k bits, there are an output of n bits and K is called constraint length. Basically, convolutional code is generated by passing the information sequentially through a series of shift registers. K stands for the number of the shift registers. Because of the shift registers, convolutional code has memory, the current n-bit output depends not only on the value of the current block of k input bits but also on the previous K-1 blocks of k input bits. So, the current output of n bits is a function of the last K×k bits. As shown in Figure, 3 most recent bits have used and two possible outputs are generated. Typically, n and k are very small number. Here, the figure shows (2,1,3) codes.
For any (n,k,K) code, the shift register keeps most recent K×k input bits. Starting from initial states, the encoder produces n output bits, after that, the oldest k bits from the register are discarded and k new bits are shifted in. For any given input of k bits, there are different functions that map k bits into n output bits. What kind of functions is used depends on the history of the last (K-1) input blocks. Since the convolutional encoder can be identified as a K-memory, n output machine, it can be expressed as a 0-1 sequence, which is called generator sequence g(.). With 2 output in Figure 1, there are 2 sequences:. Then the generator matrix is defined as:
So, the output of the encoder is
Where, u is the input.
Viterbi decoder algorithm is a kind of maximum likelihood decoder and the most popular decoder for convolutional codes. The easily way to introduce Viterbi algorithm is using state diagram of the encoder, namely trellis diagram.
Figure 2 trellis diagram of sample rate 0.5, constraint length K = 3 convolutional encoder
Figure 2 is a simple example of a sample rate 0.5 and constraint length 3 convolutional encoder. At start time, the system is at state 00. With different input, the state will transient to different state. At each time step, there is only one bit input, either be 0 or 1. So the state transient paths will separate to two at each time step. The solid line means with input signal 1 and the dashed line is with input signal 0. At last it will return to the state 00 (because the registers are all zeros), and there will multiple paths from the start state to the end state.
The key idea of Viterbi algorithm is to find the most likelihood sample path with the received codes. The formal description of Viterbi algorithm can be found at . An error metric is setup for each path. When decoder receives a data bit, it will compute the partial metric for the single path entering each state, and store the path with its metric. At next time step, re-compute the partial metric for all paths entering a state by adding the accumulated error metric.
For each state, decoder only stores the path with the largest metric and eliminates all other paths. Figure 3 is a simple example, where the at time unit 5, the state 11 has the smallest error, which means the path to state 11 is most likely correct path.
Figure 3 Trellis diagram with accumulated error metric
At last the decoder will get a most likelihood path from initial state to ended state. Tracing back the sample path, decoder can decide the data from each state transient step.
Detailed algorithm and performance analysis are discussed at . Also  is a very nice and detailed tutorial for convolutional encoder and Viterbi decoder.
The program is developed with Java applet. Basically, the implementation involves three steps: Encoder, Error adding, Decoder.
Constraint length (K) and message length is available
for adjusting. Considering the limitation of displaying, K is set to range from
3 to 7. The codeword in convolution coding is usually pretty long. User will be
able to choose either to manually type data stream in Hex (user generate) or
allow program to randomly generate data stream (randomly generate). If
“randomly generator” is selected, upon K and data length being settled, the
program will randomly generate a data stream and encoder it into codeword.
With the “random gen” or “user gen” button being clicked, the original signal, convolutional and its generator sequence (details have been discussed above) will be displayed. Meanwhile, the data stream and codeword will be show the plot graph underneath.
· Error Adding
Error Adding will be implemented in term of two styles: error inferred by probability and fixed-length error.
· Error inferred by probability
In this category, user will be allowed to set “failure probability for each bit”, and the error bits will be applied according to the probability.
· Fixed-length error
In this category, user is allowed to decide number of errors that will happen to codeword. Besides, two error-assignment forms will be available to be selected:
Random: The faulty bits will be randomly selected from the codeword;
Burst: The faulty bits will be consecutive, and user is allowed to set start position for burst error. In random styles, the start position will be set to “-1”, meaning not available for adjustment. If the start position that user set exceeds the length of the codeword, the program will coerce the start position back to signal range. It is well know that burst errors are hard for normal other error correction codes to deal with. But user will find burst or separate errors make no difference for convolutional code, which is an advantage of such code. When “adding error” is clicked, the receiving code with error bits will be displayed. Meanwhile, corresponding receiving code will be shown in the plot graph. On the receiving curve, a red “E” will appear on the error bit.
The decoder will be triggered by user clicking the “decode”. The decoder will give the results of error corrections, including:
· decoded signal;
· The number of corrected errors;
· The text message, which indicates whether all the error bits have been corrected.
· The trellis diagram: it shows the maximum likely-hood path from Viterbi algorithm, which is identical to the one in Figure 3 and will be shown under the plot graph.
The decoded signal will also be shown in the plot graph. When the number of errors exceeds the error-correction capability decoder will fail to correct all the errors. In that case, some of “E”s will be left to indicate that some errors are still there.
 Lin, Shu, and Costello, Deniel, “Error Control Coding: Fundamentals and Applications”, Prentice-Hall Inc., 1983.
 Tutorial of Convolutional Codes: http://home.netcom.com/~chip.f/viterbi/algrthms.html#algorithms