Outer Code Designs for Augmented and Local-Global Polar Code Architectures
Abstract
In this paper, we introduce two novel methods to design outer polar codes for two previously proposed concatenated polar code architectures: augmented polar codes and local-global polar codes. These methods include a stopping set (SS) construction and a nonstationary density evolution (NDE) construction. Simulation results demonstrate the advantage of these methods over previously proposed constructions based on density evolution (DE) and LLR evolution.
I Introduction
Polar codes, introduced by E. Arıkan [1], occupy a unique place in the history of error correction codes as the first family of codes to achieve Shannon capacity of arbitrary binary symmetric memoryless channels (BSMs). Arıkan [2] also introduced the concept of systematic polar encoding, achieved through the solution of linear encoding equations that ensures the codewords contain the information bits at designated positions. Concatenated polar codes have been proposed that leverage the error floor performance of polar codes in conjunction with other powerful codes such as Low-Density Parity-Check codes [3] and Reed-Solomon codes [4]. Expanding upon [5], Elkelesh et al. [6] introduced an augmented polar code architecture that concatenates two polar codes, using an outer auxiliary polar code to further protect the semipolarized bit-channels within the inner polar code. In the same work, they also suggested connecting several inner polar codes through a single auxiliary polar code, offering the flexibility of codeword lengths other than a power of two. Motivated by practical applications in data storage and low-latency communication systems, Zhu et al. [7] proposed an architecture for polar codes offering local-global decoding. In this scheme, a codeword comprising several inner polar codes is concatenated with a systematic outer polar code, thus enabling both local decoding of the inner codes and global decoding of the codeword.
The belief propagation (BP) decoder for polar codes was introduced to increase throughput through a pipelined decoding process [8]. While the BP decoder surpasses the error rate performance of the original successive-cancellation (SC) decoder, it still falls short of the SC-List (SCL) decoder [9]. The BP-List (BPL) decoder [10], which incorporates different permutation patterns of BP decoding units, significantly enhances error rate performance, bridging the performance gap between BP-based and SC-based decoders.
Polar codes and Reed-Muller (RM) codes share the same basic encoding matrix before selecting the information set: RM codes select rows according to their Hamming weights, while polar codes select rows by comparing their associated Bhattacharyya parameters [1]. Another frozen set selection method, introduced by Mori et al. [11], uses density evolution (DE) to analyze BP results for each decoding tree corresponding to the SC decoding process. The high computational complexity of DE motivated the Gaussian approximation (GA) algorithm [12], which assumes that the log-likelihood ratio (LLR) distribution corresponding to each variable node is a Gaussian with mean and variance , thus reducing the convolution of densities to a one-dimensional computation of mean values. In [13], Dai et al. proposed a modification to GA to address the performance loss incurred when applying GA to long polar codes.
While design methods based on the Bhattacharyya parameter [1], DE, and GA were originally used in the context of SC decoding, they have also been applied to code design for BP decoding. Eslami et al. [3] introduced a construction method based on stopping sets in the sparse polar code factor graph, aimed at increasing the stopping distance of the polar code. They provided empirical evidence showing improved performance under BP decoding, compared with the conventional code design. Another design approach based on LLR evolution was proposed by Qin et al. [14]. In this method, weak bit-channels, identified using LLR distributions obtained from BP decoder simulations, are swapped with stronger ones. While these algorithms have been shown to improve polar code design for BP decoding, the quest for an optimal design method remains an ongoing exploration.
In this paper, we propose two construction methods for designing outer codes within concatenated polar code architectures under BP decoding. The first construction, stopping set (SS) design, uses an analysis of stopping sets in the concatenated factor graph to identify the information set of the outer code. The nonstationary density evolution (NDE), initializes the DE construction of the outer code with empirical LLR densities from BP decoding of the inner code(s) to better reflect bit-channel reliabilities. Error rate simulations demonstrate that both of these methods can improve the performance of augmented and local-global polar codes.
The paper is organized as follows. Section II briefly reviews background results and notations used in the rest of the paper. In Section III, the SS and NDE construction methods are presented and discussed. Section IV provides error rate simulation results, and Section V concludes the paper.
II Preliminaries
II-A Polar Codes and Systematic Polar Codes
In conventional polar code design, independent copies of a channel are combined in a recursive manner into a vector channel , which is then split into channels , referred to as bit-channels. The Bhattacharyya parameter is used to identify the quality of bit-channel . A polar code of rate selects the most reliable bit-channels (with the smallest ) to input information bits, and the remaining bit-channel inputs are frozen to zero. We use to denote the set of information indices, and to denote the frozen indices. Let be the matrix that is the -th Kronecker power of , where . The polar encoding process is specified by , where .
Arıkan showed that a systematic encoder can be realized that maps information bits to positions in the set in the codeword . To be specific, is set to 0, is set to the information vector, and and are found by solving a system of equations [2].
II-B Concatenated Polar Codes
Our focus in this paper is on concatenated code architectures in which all component codes are polar codes. The augmented and flexible length architectures were introduced in [6]. In an augmented polar code, a short, rate auxiliary outer polar code is connected to an inner polar code of length . The bits of the outer codeword are assigned to the semipolarized channels within the inner code (through an interleaver). An additional message bits are assigned to the good bit-channels within the inner code. The total code rate for the augmented structure is .
In the flexible length architecture, two inner codes of length are coupled through a rate auxiliary outer code . Information words of length are assigned to the good bit-channels of the two inner codes, respectively. The outer codeword is divided into two parts which are assigned to the semipolarized bit-channels of the inner codes. The total encoding rate for the flexible length structure is .
Inspired by the flexible length architecture, the local-global polar code architecture, introduced in [7], connects multiple inner codes through a systematic outer polar code. We asssume these codes have the same length . A word of information bits is divided into parts of bits that are assigned to good bit-channels within the inner codes. The outer information bits are divided into parts of bits that are mapped to semipolarized bit-channels in the inner codes, respectively. The parity bits of the outer codeword are similar partitioned into parts of bits and mapped to remaining semipolarized bit-channels within the inner codes. This architecture supports local decoding of information bits within each inner code , with the option of improved decoding of the inner codewords via global decoding using the outer code.
II-C Stopping Set Analysis
We briefly review the stopping set analysis of polar codes in [3]. Recall that a stopping set (SS) is a non-empty set of variable nodes such that each neighboring check node is connected to this set at least twice. A stopping tree (ST) is a SS that contains one and only one information bit, i.e., variable node in the leftmost stage of the sparse polar code factor graph. For each information bit , there is a unique stopping tree denoted by . The size of the leaf set (variable nodes on the rightmost stage) of is denoted by . Only variable nodes on the right are observed nodes, with all other variables nodes hidden. The set of observed variable nodes in a SS form a variable-node SS (VSS). Accordingly, we define a minimum VSS (MVSS) to be a VSS with a minimum number of observed variable nodes, among all the VSSs. The size of a MVSS is the stopping distance of the code. For any given set , we denote the set of SSs whose information nodes are precisely as . The set of observed variable nodes in each of these SSs is a VSS for , and the collection of these VSSs is denoted as . A minimum size VSS in is called a minimum VSS for , denoted .
III Construction methods
In this section we describe the stopping set and NDE methods for designing outer codes for augmented and local-global polar codes.
III-A Stopping Set Construction
Consider the augmented polar code structure in Fig. 1, where an inner polar code is concatenated with an outer polar code. Let denote the -th node on the leftmost stage within the original outer factor graph, and be the -th node on the leftmost stage within the original inner factor graph. Let denote the set of inner information nodes such that each element in is connected to one of the leaves in the stopping tree defined on the original outer factor graph. For example in Fig. 1, is the set of nodes . Let be the MVSS defined by set on the original inner factor graph. Let denote the submatrix of the inner encoding matrix consisting of the rows that correspond to . Again taking Fig. 1 as an example, is the set of nodes . The resulting consists of the second and sixth rows of inner encoding matrix . Note that is also a VSS for , but it is not the minimum VSS. The following theorem gives a bound on the size of .
Theorem III.1.
Given any outer code position on the leftmost stage of the factor graph and its corresponding , , we have , where
(1) |
(2) |
Proof.
See Appendix. ∎
In words, the function counts the number of columns in a matrix that have weight one. Thus, for any given information position we easily calculate the lower bound on the size of the MVSS that only contains on the leftmost stage by looking at the generator matrix of the inner code.
Theorem 2 suggests a practical way to design outer polar codes based on the size of stopping sets. We first initialize an unfrozen set for the outer code using the conventional DE algorithm, for example. Then we swap a specified number of unfrozen bits with the smallest “stopping distance” with some positions such that .
Before presenting the design algorithm, we introduce some notation. Let be a length vector that contains the indices of bit channels ordered according to channel reliability calculated by DE. The indices are ordered by descending channel reliability, i.e., stores the index for the strongest bit channel, stores the index for the second strongest, and so on. Let denote the number of bits we are going to swap. For convenience, we denote by . Let denote the size of the desired unfrozen set. Let be the function that returns the -th smallest value in a vector, while returns the smallest value along with its index. Note that should be chosen such that there are more than frozen bits that have value larger than . The detailed swapping algorithm is presented in Algorithm 1.
We can easily extend Theorem 2 to the case when inner codes are connected by a single outer code. For example, assume and , where and are connected nodes in the first and second inner codes, respectively. Then .
The design method of Algorithm 1 can be extended to the local-global polar code architecture, but some care is needed. The systematic outer polar code assigns information vectors , to the inner codes. Directly applying Algorithm 1 can potentially swap bits with (), causing to be assigned to the same inner code. For example, assume the unfrozen set represents the most reliable positions according to DE, and . Then, if the partition of is according to bit index, the first half of will correspond to and the second half to . If the parity bits are partitioned similarly, we have and . If, after calculating the value for each position, we swap positions 3 and 5, this would yield . This assignment is now inconsistent with the local-global architecture because part of (position 5) is connected with the second inner code. To avoid this problem, one needs to carefully choose the connections in the local-global encoder to ensure that positions in are only swapped with positions in .
Example III.2.
For the case , there is a mapping and partition rule that works for any , , with . Let be the information set of size according to DE. Assign to the first half according to natural index order, and to the second half. Assign to the first half of the parity positions in natural index order, and to the second half. The semipolarized bit-channels in the first inner code are connected in natural index order to and those in the second inner code connect to . See Fig. 2.
III-B Nonstationary density evolution (NDE)
III-B1 NDE for augmented polar codes
In prior formulations of augmented polar codes in [5],[6], the outer polar codes were designed independently from the code concatenation. Since the inputs to the outer code are the LLRs passed from the inner code and the LLR densities may vary within the semipolarized bit-channels, there is a potential mismatch with the assumed reliability ordering of the outer code bit-channels which could lead to inferior performance of the augmented code.
In order to improve the performance of the concatenated code under BP decoding, we introduce a small modification of the conventional DE design method introduced in [11] which we call the nonstationary DE (NDE) algorithm. The NDE algorithm does not assume that the inital LLR distributions are identical; rather, the LLR distributions coming from the inner code at the rightmost stage of the outer code factor graph correspond to separate binary symmetric memoryless channels . (A similar nonstationary scenario in code construction via polarization was considered in [15].)
Using the notation of [11], the densities on the rightmost stage are initialized as . We then apply DE over the SC decoding trees to compute the LLR densities for the bit-channels. As in [11], assuming symmetry and an all-zero codeword, we compute a probability of error, denoted , for bit-channel by integrating over the interval . The unfrozen set is chosen to minimize , subject to .
In practice, we replace each initial LLR density of the outer code with the empirical LLR density of the corresponding bit-channel of the inner code after iterations of BP decoding under assumption of an all-zero codeword. If the semipolarized bit channels of the inner code are denoted by index set , then the initial LLR density is replaced by the empirical density of the corresponding semipolarized bit-channel . Details of the NDE implementation are given in Algorithm 2.
III-B2 NDE for local-global polar codes
Let be the positions within the inner codes that are connected to the bits of the outer codeword. The connection mapping requires knowledge of the unfrozen set of the outer code. We assume here that the mapping, which we denote by , respects the natural ordering of bit-channels within the subsets in and in and their corresponding subsets within the semipolarized bit-channels of the associated inner codes. Once the connections are established and a set of empirical LLR densities are found for the bit-channels in , we can apply the NDE algorithm to determine a probability of error, denoted , for bit-channel of the outer code. (All-zero codewords are again assumed in computing the empirical LLRs for the inner code and in computing probabilities of error for the bit-channels of the outer code.) We can then determine a set of bit-channels that minimizes .
The unfrozen set that we seek for the outer code is the solution to the equation
(3) |
It is not clear if (3) has a solution, and, if it does, whether the solution is unique. However, we have used the iterative procedure shown in Algorithm 3 to find a solution in the code structures we considered. Different initial values of the unfrozen set can lead to different solutions, and we have also observed cases with periodic solutions, but they differ in only a small number of bit-channels and the corresponding values of are nearly the same. We arbitrarily choose one solution for the outer code design.
IV Empirical results
This section gives empirical results under BP decoding for augmented and local-global polar codes constructed using the proposed methods, along with comparisons to codes designed using conventional DE and LLR evolution methods. The BP decoding schedules are the same as those in [6] for augmented codes and in [7] for local-global codes. The maximum number of BP decoder iterations is set at 100. In the SS design, we set the number of bit-channel swaps to . In the NDE design for the augmented code, we use BP decoder iterations on the inner code to generate the required empirical LLRs, and for the local-global code, we use iterations.
Fig. 3 shows frame error rate (FER) results for augmented code constructions. The outer code length is with code rate . The inner code length is . Its bit-channel ordering is based on DE on the AWGN channel at dB, computed using the algorithm in [11]. The design rate of the augmented code is . The connections between the bit-channels of the inner code and the bits of the outer codeword are based on the natural index ordering within the set of semipolarized bit-channels. For the conventional DE and NDE designs, DE is simplified by using Gaussian approximation (GA), using the 4-segment approximation function in [13].
At FER = , the stopping set design and NDE design offer gains of 0.12 dB and 0.18 dB over the conventional DE design, respectively, and both outperform the LLR evolution design. At this FER they also perform comparably to SCL decoding, although SCL becomes superior at higher FERs. We remark that when the outer codes designed using the stopping set and NDE methods are disconnected from the concatenation architecture, their performance is inferior to that of a code designed using conventional DE. This confirms their inherent relationship with the concatenation structure.
Figs. 4 and 5 present the results for local and global decoding, respectively, for a local-global code with component code lengths . The connections between the inner codes and the outer code are as described in Example III.2. Note that Algorithm 3 for NDE typically converges in fewer than 10 iterations to a unique or oscillating solution. Local decoding results for the different outer code design methods are similar, as expected, since local decoding does not rely on the outer code. Under global decoding, at FER = , the stopping set and NDE designs provide gains of 0.22 dB and 0.09 dB over conventional DE, respectively. Both outperform the LLR evolution design. The result for a length-2048 conventional polar code is also shown for reference. In summary, the improved global decoding performance provided by the new outer code constructions does not reduce the local decoding performance.
V Conclusion
In this paper, we proposed two methods to design outer polar codes within concatenated structures. The stopping set and NDE designs both outperform existing construction methods on augmented and local-global architectures with connection patterns based on the natural index ordering of semipolarized bit-channels. Although the natural ordering is attractive from an implementation standpoint, it may not provide the best starting point for these design methods, and experiments with other connection patterns, both structured and randomly generated, show that gains achieved with the proposed methods vary. Determining the choice of connection pattern and outer code design method that yields the best performance remains a problem for further research.
Acknowledgment
This research was supported in part by NSF Grants CCF-1764104 and CCF-2212437.
References
- [1] E. Arıkan, “Channel polarization: A method for constructing capacity-achieving codes for symmetric binary-input memoryless channels,” IEEE Trans. Inf. Theory, vol. 55, no. 7, pp. 3051-3073, Jul. 2009.
- [2] E. Arıkan, “Systematic polar coding,” IEEE Commun. Lett., vol. 15, no. 8, pp. 860-862, Aug. 2011.
- [3] A. Eslami and H. Pishro-Nik, “On finite-length performance of polar codes: Stopping sets, error floor, and concatenated design,” IEEE Trans. Commun., vol. 61, no. 3, pp. 919-929, Mar. 2013.
- [4] M. Bakshi, S. Jaggi and M. Effros, “Concatenated polar codes,” Proc. IEEE Int. Symp. Inf. Theory (ISIT), pp. 918-922, Jun. 2010.
- [5] J. Guo, M. Qin, A. Guillén i Fàbregas and P. H. Siegel, “Enhanced belief propagation decoding of polar codes through concatenation,” Proc. IEEE Int. Symp. Inf. Theory (ISIT), pp. 2987-2991, Jun.-Jul. 2014.
- [6] A. Elkelesh, M. Ebada, S. Cammerer and S. t. Brink, “Flexible length polar codes through graph based augmentation,” Proc. 11th Int. ITG Conf. on Syst., Commun. and Coding (SCC 2017), pp. 1-6, Feb. 2017.
- [7] Z. Zhu, W. Wu and P. H. Siegel, “Polar codes with local-global decoding,” Proc. 56th Asilomar Conf. Signals Syst. Comput., pp. 392-396, Oct. 2022.
- [8] E. Arıkan, “Polar codes: A pipelined implementation,” Proc. Int. Symp. Broadcast Commun. (ISBC), pp. 11–14, Jul. 2010.
- [9] I. Tal and A. Vardy, “List decoding of polar codes,” Proc. IEEE Int. Symp. Inf. Theory (ISIT), pp. 1-5, Jul. 2011.
- [10] A. Elkelesh, M. Ebada, S. Cammerer and S. ten Brink, “Belief propagation list decoding of polar codes,” IEEE Commun. Lett., vol. 22, no. 8, pp. 1536-1539, Aug. 2018.
- [11] R. Mori and T. Tanaka, “Performance of polar codes with the construction using density evolution,” IEEE Commun. Lett., vol. 13, no. 7, pp. 519-521, Jul. 2009.
- [12] S.-Y. Chung, T. J. Richardson and R. L. Urbanke, “Analysis of sum-product decoding of low-density parity-check codes using a Gaussian approximation,” IEEE Trans. Inf. Theory, vol. 47, no. 2, pp. 657-670, Feb. 2001.
- [13] J. Dai, K. Niu, Z. Si, C. Dong and J. Lin, “Does gaussian approximation work well for the long-length polar code construction?,” IEEE Access, vol. 5, pp. 7950-7963, Apr. 2017.
- [14] M. Qin, J. Guo, A. Bhatia, A. Guillén i Fàbregas and P. H. Siegel, “Polar code constructions based on LLR evolution,” IEEE Commun. Lett., vol. 21, no. 6, pp. 1221-1224, Jun. 2017.
- [15] M. Zorgui, M. E. Fouda, Z. Wang, A. M. Eltawil and F. Kurdahi, “Non-stationary polar codes for resistive memories,” IEEE Global Commun. Conf. (GLOBECOM), pp. 1-6, Dec. 2019.
Proof of Theorem 2: The definition of stopping set implies that the union of stopping sets is a stopping set. For a given set of information bits , consider the set which is the union of all the stopping trees defined by the elements in , i.e., . To find a minimum size VSS for , we can try to find a stopping set that is properly contained in that preserves but has fewer leaf nodes (i.e., observed variable nodes). We claim that the only leaf nodes that can possibly be deleted from are shared leaf nodes, i.e., leaf nodes that belong to at least two stopping trees in .
To see this, suppose that unshared leaf node belongs only to the stopping tree , where . The variable nodes on the branch of that traces back to the root node must also be unshared.111We are using the indexing convention of [11] for stages in the polar code factor graph. This is the reverse of the indexing convention in [3]. This is because if one of the nodes on that branch is shared by two trees or more, then all the children nodes of , i.e., nodes to the right of along the tree , must be shared nodes, including the leaf node . This contradicts the assumption that is unshared. Since this branch belongs only to , the result of deleting or any subset of nodes from this branch other than the root node could not produce a stopping set, for this would mean that the remaining subset of nodes in would still constitute a stopping tree, call it , with root . However, this would violate Fact 2 in [3], which states that every information bit has a unique stopping tree.
A simple example is shown in Fig. 6, where and . There is no stopping set properly contained in that does not include .
To complete the proof, we need to characterize the indices of the leaf nodes in . This can be done by noting that the indices for the leaf nodes in the stopping tree are given by the indices for the ones in , where is the -th row of , the encoding matrix for length polar codes. This fact was stated in [3]. For completeness, we provide a detailed proof here. The proof proceeds by induction. For the case , the statement follows immediately from inspection of the matrix and inspection of the corresponding factor graph for .
Now, suppose the result is true for a given . For a length polar code, we denote the upper and lower halves of the factor graph by and , as shown in Fig. 7. The recursive construction of the factor graph implies that and are isomorphic, and is the encoding matrix for each of the subgraphs and . Let be a length binary vector in which the indices of ones are the positions of the leaf nodes contained in in . By the induction hypothesis, .
Note that is recursively represented as
(4) |
Referring to Fig. 7, we see that, for , , with the last equality following from (4). Similarly, for , since the subgraphs and are isomorphic, we have , where again the last equality follows from (4). This completes the induction.
In summary, the number of leaf nodes in is given by the number of columns in that have non-zero weight. The number of unshared leaf nodes in is given by the number of columns in that have weight exactly one. Thus is precisely the number of unshared leaf nodes, which by the previous discussion must belong to a minimum size VSS for . This implies , as desired. ∎
The following result shows that the bound in Theorem 2 is tight when .
Proposition .1.
When , .
Proof: Let . Then . Assume is a shared leaf node belonging to both and . Then there must exist a degree-3 check node in the graph of , with all three neighboring variable nodes in , that lies in the intersection of the graphs of and and has as a child node. Otherwise, one could trace back from to a single root node, contradicting the fact that it is a shared leaf node. All of the children nodes to the right of that degree-3 check node must be shared by the two trees. This implies that we can delete those children nodes, which include the shared node , and the remaining structure will still be a stopping set in . This procedure can be repeated for any remaining shared leaf nodes, until all of the original shared leaf nodes in have been deleted. The only remaining leaf nodes are the original unshared leaf nodes, completing the proof. ∎
Fig. 8 illustrates the proof procedure. Here , and . The shared leaf nodes are and , while the unshared leaf nodes are and . The green variable nodes in are the children of the orange degree-3 check node that lies in the intersection of the graphs of and . Both and are children variable nodes of this check node. The orange nodes represents a that remains after deleting the green variable nodes.
We note that Proposition .1 can not be extended to the case when . For example, if , then the leaf node is shared by and . However, any proper stopping set in that contains must also contain .