Browsing by Subject "Rateless Codes"
Now showing 1 - 2 of 2
Results Per Page
Sort Options
Item Exit charts based analysis and design of rateless codes for the erasure and Gaussian channels(2009-06-02) Mothi Venkatesan, SabaresanLuby Transform Codes were the first class of universal erasure codes introduced to fully realize the concept of scalable and fault?tolerant distribution of data over computer networks, also called Digital Fountain. Later Raptor codes, a generalization of the LT codes were introduced to trade off complexity with performance. In this work, we show that an even broader class of codes exists that are near optimal for the erasure channel and that the Raptor codes form a special case. More precisely, Raptorlike codes can be designed based on an iterative (joint) decoding schedule wherein information is transferred between the LT decoder and an outer decoder in an iterative manner. The design of these codes can be formulated as a LP problem using EXIT Charts and density evolution. In our work, we show the existence of codes, other than the Raptor codes, that perform as good as the existing ones. We extend this framework of joint decoding of the component codes to the additive white Gaussian noise channels and introduce the design of Rateless codes for these channels. Under this setting, for asymptotic lengths, it is possible to design codes that work for a class of channels defined by the signal?to?noise ratio. In our work, we show that good profiles can be designed using density evolution and Gaussian approximation. EXIT charts prove to be an intuitive tool and aid in formulating the code design problem as a LP problem. EXIT charts are not exact because of the inherent approximations. Therefore, we use density evolution to analyze the performance of these codes. In the Gaussian case, we show that for asymptotic lengths, a range of designs of Rateless codes exists to choose from based on the required complexity and the overhead. Moreover, under this framework, we can design incrementally redundant schemes for already existing outer codes to make the communication system more robust to channel noise variations.Item Outage Capacity and Code Design for Dying Channels(2012-10-19) Zeng, MengIn wireless networks, communication links may be subject to random fatal impacts: for example, sensor networks under sudden power losses or cognitive radio networks with unpredictable primary user spectrum occupancy. Under such circumstances, it is critical to quantify how fast and reliably the information can be collected over attacked links. For a single point-to-point channel subject to a random attack, named as a dying channel, we model it as a block-fading (BF) channel with a finite and random channel length. First, we study the outage probability when the coding length K is fixed and uniform power allocation is assumed. Furthermore, we discuss the optimization over K and the power allocation vector PK to minimize the outage probability. In addition, we extend the single point to-point dying channel case to the parallel multi-channel case where each sub-channel is a dying channel, and investigate the corresponding asymptotic behavior of the overall outage probability with two different attack models: the independent-attack case and the m-dependent-attack case. It can be shown that the overall outage probability diminishes to zero for both cases as the number of sub-channels increases if the rate per unit cost is less than a certain threshold. The outage exponents are also studied to reveal how fast the outage probability improves over the number of sub-channels. Besides the information-theoretical results, we also study a practical coding scheme for the dying binary erasure channel (DBEC), which is a binary erasure channel (BEC) subject to a random fatal failure. We consider the rateless codes and optimize the degree distribution to maximize the average recovery probability. In particular, we first study the upper bound of the average recovery probability, based on which we define the objective function as the gap between the upper bound and the average recovery probability achieved by a particular degree distribution. We then seek the optimal degree distribution by minimizing the objective function. A simple and heuristic approach is also proposed to provide a suboptimal but good degree distribution.