On packet lengths and overhead for random linear coding over the erasure channel
Abstract
We assess the practicality of random network coding by illuminating the issue of overhead and considering it in conjunction with increasingly long packets sent over the erasure channel. We show that the transmission of increasingly long packets, consisting of either of an increasing number of symbols per packet or an increasing symbol alphabet size, results in a data rate approaching zero over the erasure channel. This result is due to an erasure probability that increases with packet length. Numerical results for a particular modulation scheme demonstrate a data rate of approximately zero for a large, but finite-length packet. Our results suggest a reduction in the performance gains offered by random network coding.
I Introduction
It has been shown that network coding allows a source node to multicast information at a rate approaching the maximum flow of a network as the symbol alphabet size approaches infinity [1, 2]. An infinitely large symbol alphabet would correspond to transmission of infinitely long blocks of data, which is neither possible nor practical. Subsequent works have shown that a finite alphabet is sufficient to achieve the maximum flow and a number of works have provided bounds on the necessary alphabet size. For example, in [3] the alphabet size is upper bounded by the product of the number of sources with the number of receivers, while in [4] a lower bound on the alphabet size is given by the number of receivers.
Random linear coding was proposed to allow for distributed implementation [5]. In random network coding, a randomly generated encoding vector is typically communicated to the receiver by appending it to the header of the transmitted packet. The overhead inherent in communicating the encoding vector becomes negligible as the number of symbols in the packet grows large. Regarding the alphabet size for random network coding, a lower bound of 2 times the number of receivers is given in [4].
The “length” of a transmitted packet (i.e., the number of bits conveyed by the packet) depends on both the symbol alphabet and the number of symbols per packet. From the previous works listed above, it is clear that transmitting sufficiently long packets is crucial to ensuring the existence of network codes and to allowing random linear coding to operate with low overhead.
However, long packets are more susceptible to noise, interference, congestion, and other adverse channel effects. This point has been ignored in previous works on network coding. In particular, a number of previous works [6, 7, 8] consider transmission over an erasure channel, in which a packet is either dropped with a probability or received without error. In a wireline network, the erasure channel is used to model dropped packets due to buffer overflows. Clearly, longer packets take up more space in memory, so the erasure probability will increase as the packet length grows. In the case of a wireless network, there are a number of reasons that the erasure probability will increase with the length of the packet, as listed below.
-
•
If we try to fit more bits into the channel using modulation, then for fixed transmission power, points in the signal constellation will move closer together and errors are more likely.
-
•
If we try to fit more bits into the channel by decreasing symbol duration, then we are constrained by bandwidth, a carefully-controlled resource in wireless systems.
-
•
Longer packets are more susceptible to the effects of fading.
In this work, we model the erasure probability as a function of the packet length and investigate the implications on network coding performance. This is somewhat reminiscent of [9], in which the erasure probability is a function of the link distance in a wireless network, and the effect on capacity is investigated. Our emphasis will be on a wireless channel with a fixed bandwidth, for which we will associate erasure probability with the probability of symbol error for a given modulation scheme.
We note that a careful examination of packet lengths in data networks is not a new idea; most notably, many researchers participated in a dispute over the packet size for the Asynchronous Transfer Mode (ATM) standard in the 1980s. And still many tradeoffs are being studied today at the network level regarding the packet length. The interplay between packet lengths and coding arises because of the way that network coding unifies different layers of the protocol stack. Another work which examines a similar problem is [10], in which packet headers and packet lengths are analyzed for Reed-Solomon coding in video applications.
II Throughput of random linear coding
We consider the following setting. A source node has units of information that it wants to transmit. We will refer to each of these information units as packets and let each packet be given by a vector of -ary symbols, where is the symbol alphabet (the size of a finite field) and is the number of symbols per packet. We consider values of which are powers of 2, i.e., for some . The length of a packet is given by bits. In Section III we will consider the case where are packets of uncoded information, whereas in Section IV we will assume that a (deterministic) error-correcting code has been applied in order to form .
The source generates random linear combinations by forming the modulo- sum , where are chosen randomly and uniformly from the set . Note that the resulting random linear combination is a packet with length bits. With each random linear combination transmitted, the source appends a packet header identifying , which requires an additional bits of overhead with every bits transmitted. A receiver will collect these random linear combinations until it has received enough to decode. For transmission over an erasure channel, decoding can be performed by Gaussian elimination. Let denote the number of random linear combinations needed for decoding. Each of the random linear combinations represents a linear equation in and the distribution of is given by
Note that Pr is equal to zero for . For , the distribution can be found following the procedure in [11] and is given by
The expected value of can be found from the above distribution and is shown to be given by
(1) | |||||
Clearly, as , .
We let denote the erasure probability on the channel, which is an increasing function of the packet length . The expected number of transmissions needed for the receiver to decode the original packets is given by
Over the course of these transmissions, the average number of packets received for each transmission is
We account for the overhead by scaling the number of packets received per transmission by , which is the ratio of number of information symbols to the total number of symbols (information plus overhead) sent with each transmission. We define as the effective portion of each transmission which contains message information; is a measure of throughput in packets per transmission and takes values between 0 and 1.
(2) |
If the erasure probability is constant, , then for , , corresponding to infinitely long packets, the value of approaches , which is the Shannon capacity of the erasure channel.
From the expression in (2), we can identify a tradeoff between the packet length and the throughput. As the number of information symbols per packet grows large, the effect of overhead becomes negligible, but the erasure probability grows and the transmissions are more likely to fail. Alternatively, if approaches zero, transmissions are more likely to succeed, but the overwhelming amount of overhead means that no information can be transmitted. In a similar manner, as the alphabet size grows, the random linear coding becomes more efficient in the sense that , but again, the erasure probability increases and transmissions are likely to be unsuccessful. This tradeoff demonstrates that the alphabet size, packet length, and overhead must be carefully weighed in determining the performance of random linear coding over the erasure channel.
III Performance without pre-coding
In this section we consider the case where are uncoded information symbols. We define the data rate in bits per transmission as , where
(3) |
The data rate accounts for the fact that each received packet contains bits of information. Since the packets consist of uncoded information, for every random linear combination sent, all symbols must be received without error. Then
and will correspond to the symbol error probability for -ary modulation over the channel. We denote the probability of symbol error by and note that it is independent of but depends on as well as features of the channel such as pathloss, signal-to-noise ratio (SNR), and fading effects. We will consider a wireless channel of limited bandwidth, for which modulation techniques such as pulse amplitude modulation (PAM), phase shift keying (PSK), and quadrature amplitude modulation (QAM) are appropriate. For these modulation techniques, as [12].
In this setting, the erasure probability is given by
(4) |
From the above expression we note that, as suggested by our discussion on the tradeoff between throughput and packet length, as , both and approach zero exponentially fast. Additionally, for the modulation schemes mentioned above, and approach zero as , albeit at a slower rate.
As a numerical example, we have plotted the throughput and data rate for QAM modulation over an additive white Gaussian noise (AWGN) channel. In this case the symbol error probability for the optimum detector is approximated by [12]
(5) |
where is the SNR per bit and is the complementary cumulative distribution function for the Gaussian distribution. The above expression for holds with equality for even. The results are shown in Figure 1, where is fixed and is varied, and in Figure 2, where is fixed and is varied. In all cases, the throughput and data rate are concave functions, admitting optimum values for and . Furthermore, if grows large as is fixed or if grows large as is fixed, the throughput approaches zero.


To improve the performance described above, a number of techniques might be employed, including enhanced receivers, diversity techniques, and channel coding.
IV Performance with pre-coding
In this section we examine the effect of using a deterministic pre-code prior to performing random linear coding. We assume that an uncoded packet , consisting of -ary information symbols, is passed through an encoder to add redundancy, producing the packet consisting of -ary symbols. Thus a -ary code of rate is used to produce . Random linear coding is again performed on and we assume that the codebook for the pre-code is known to the decoder and does not need to be transmitted over the channel. At the receiver, decoding of the random linear code will be performed first using the coefficients , sent with each random linear combination. Subsequently the pre-code will be decoded. The use of the pre-code means that the system will be tolerant to some errors in the symbols sent in each random linear combination.
To account for the pre-code, we modify the throughput expressed in (2) by multiplying it be , which results in
(6) |
To obtain the data rate in bits per transmission, is now scaled by , the number of information symbols present in each packet.
(7) |
We can bound the erasure probability, and thus the throughput and data rate , using bounds on the error-correcting capabilities of the pre-code. Let denote the number of symbol errors that the pre-code can correct. For a given code, errors, where is the minimum distance of the code. The erasure probability can be bounded as follows.
In the above expression, holds with equality for the class of perfect codes [12], follows by assuming that errors are independent, identically distributed among the symbols, and holds for . We make use of the Gilbert-Varshamov bound to provide a lower bound on as follows.
(8) |
Lower bounds on the throughput and data rate are given as follows.
(9) |
(10) |
For while all other parameters are fixed, if the modulation scheme has , then . In the limit of increasing symbols per packet , we identify two different cases. First, if the pre-code rate is fixed (i.e, as with fixed ratio between the two) then (multiplied by and increases without bound. On the other hand, if is fixed (i.e., as , ) then .
We have computed numerical examples of and assuming QAM modulation with as given in (5). In Fig. 3 we display the throughput and data rate as a function of . Clearly, pre-coding allows the system to support higher with non-zero throughput, although the throughput does eventually go to zero for large . In comparing this result to Fig. 2, note that all other parameters are fixed and the pre-code rate results in a halving of the throughput and data rate.
In Fig. 4 we have displayed the two different cases for performance as a function of increasing symbols per packet . In Fig. 4(a), the rate is fixed at and while increases without bound. In this case, pre-coding ensures that the throughput and data rate increase with . However, there is a caveat: the result in Fig. 4(a) requires that , which means that the source must have an infinite supply of uncoded information to be transmitted. On the other hand, if the source has a finite amount of information to be transmitted, as shown in Fig. 4(b), then as , the pre-code rate and the data rate also approaches zero. In summary, the use of a pre-code can improve performance on the erasure channel; indeed, pre-coding is the mechanism which allows us to represent a noisy channel as an erasure channel. However, pre-coding can only ensure that the data rate increases with packet length if there is an infinite supply of (uncoded) information at the source. This is a strong assumption for a network, where data often arrives in bursts.



V Discussion
The primary contribution of this work is the elucidation of the fact that the widely-used assumption of the erasure probability being constant as the packet length increases leads to idealized performance that cannot be obtained in a practical scenario. We have shown that the assumption of increasingly long packets, either due to increasingly many symbols per packet or to an increasing alphabet size, can result in a data rate of zero for random linear coding if the erasure probability increases with packet length. While the use of a pre-code prior to performing random linear coding can improve performance, it can only ensure that the throughput increases with packet length if there is an infinite supply of data awaiting transmission at the source. We have also demonstrated that the overhead needed for random linear coding can have a devastating impact on the throughput. We have focused our attention on the erasure channel, but the tradeoff between packet length and throughput will arise in other channels, particularly in wireless channels where these effects can be exacerbated by fading. A number of techniques, such as pre-coding, can be employed to combat the adverse effects of the channel, but in making use of these techniques, the performance gains offered by network coding will be tempered.
Acknowledgment
This work is supported by the Office of Naval Research through grant N000140610065, by the Department of Defense under MURI grant S0176941, and by NSF grant CNS0626620. Prepared through collaborative participation in the Communications and Networks Consortium sponsored by the U.S. Army Research Laboratory under The Collaborative Technology Alliance Program, Cooperative Agreement DAAD19-01-2-0011. The U.S. Government is authorized to reproduce and distribute reprints for Government purposes notwithstanding any copyright notation thereon. The views and conclusions contained in this document are those of the authors and should not be interpreted as representing the official policies, either expressed or implied, of the Army Research Laboratory or the U. S. Government.
References
- [1] R. Ahlswede, N. Cai, S.-Y. R. Li and R. W. Yeung, “Network information flow,” IEEE Trans. Inform. Theory, vol. 46, no. 4, July 2000.
- [2] S.-Y. R. Li, R. W. Yeung, and N. Cai, “Linear network coding,” IEEE Trans. Inform. Theory, vol. 49, Feb 2003.
- [3] R. Koetter and M. Medard, “Beyond routing: an algebraic approach to network coding,” INFOCOM 2002.
- [4] S. Jaggi, P. Sanders, P. A. Chou, M. Effros, S. Egner, K. Jain and L. M. G. M. Tolhuizen, “Polynomial time algortihms for multicast network code construction,” IEEE Trans. Inform. Theory, vol. 51, no. 6, June 2005.
- [5] T. Ho, R. Koetter, M. Medard, D. R. Karger and M. Effros, “The benefits of coding over routing in a randomized setting,” ISIT 2003.
- [6] D. S. Lun, M. Medard, R. Koetter, and M. Effros, “On coding for reliable communication over packet networks,” submitted to IEEE Trans. Inform. Theory.
- [7] A. Eryilmaz, A. Ozdaglar and M. Medard, “On delay performance gains from network coding,” CISS, 2006.
- [8] A. F. Dana, R. Gowaikar, R. Palanki, B. Hassibi and M. Effros, “Capacity of wireless erasure networks,” IEEE Trans. Inform. Theory, vol. 52, no. 3, pp. 789-804, 2006.
- [9] B. Smith and S. Vishwanath, “Asymptotic transport capacity of wireless erasure networks,” Allerton conference, 2006.
- [10] B. Hong and A. Nosratinia, “Rate-constrained scalable video transmission over the Internet,” IEEE Packet Video Workshop, 2002.
- [11] D. J. C. MacKay, “Fountain codes,” IEE Workshop on Discrete Event Systems, 1998.
- [12] J. G. Proakis, Digital Communications, Third edition, McGraw-Hill, Boston, 1995.