Citation
Abstract
A model for predicting the computational performance of a maximum likelihood convolutional decoder (MCD) operating in a noisy carrier reference environment is described. This model is used to develop a subroutine that will be utilized by the Telemetry Analysis Program (TAP) to compute the MCD bit error rate. When this computational model is averaged over noisy reference phase errors using a high-rate interpolation scheme, the results are found to agree quite favorably with experimental measurements.
Details
- Volume
- 42-34
- Published
- August 15, 1976
- Pages
- 108–118
- File Size
- 730.1 KB