The Computational and Latency Advantage
of Quantum Communication Networks
Abstract
This article summarises the current status of classical communication networks and identifies some critical open research challenges that can only be solved by leveraging quantum technologies. By now, the main goal of quantum communication networks has been security. However, quantum networks can do more than just exchange secure keys or serve the needs of quantum computers. In fact, the scientific community is still investigating on the possible use cases/benefits that quantum communication networks can bring. Thus, this article aims at pointing out and clearly describing how quantum communication networks can enhance in-network distributed computing and reduce the overall end-to-end latency, beyond the intrinsic limits of classical technologies. Furthermore, we also explain how entanglement can reduce the communication complexity (overhead) that future classical virtualised networks will experience.
I Introduction
The history of telecommunications has already experienced a fundamental progress-driven paradigm shift from circuit switching to packet switching. However, the necessity to interconnect very heterogeneous networks and to target different verticals (mobile broadband, augmented reality, vehicular networks, Tactile Internet, Industry 4.0, etc.), with different concurrent and stringent requirements, have raised the need for a new generation of networks, called 5G and beyond (B5G). The scope of these networks is to provide an ecosystem of networks flexibly, efficiently and effectively interconnecting heterogeneous radio access networks (RANs) and wired networks (edge, core and the Internet). At the same time, requirements of very low latency, significantly greater throughput, increase of energy efficiency, and of ubiquitous connectivity have also pushed research and industrial community to investigate new paradigms for telecommunications.

In order to solve these problems, since 2010 the idea of network softwarization has become popular via software-defined networking (SDN) and network function virtualization (NFV), as a realization of the paradigm shift from store-and-forward to compute-and-forward (Fig. 1). The idea behind such technologies is the transformation of communication networks from employing dedicated hardware to using general purpose hardware, running software-based network components (instantiated inside virtual machines, containers, unikernels, etc.). Nevertheless, software-based future generation networks will not be able to satisfy the expectations, because of classical software and network-theoretic intrinsic limits, while they will also introduce some critical drawbacks. Firstly, the communication complexity (i.e. the number of bits exchanged among distributed network nodes to compute a function) of these networks, relying on massive distributed computing, will significantly increase. Secondly, the security of the software functions will be weaker than that of dedicated hardware, requiring a continuous growth of security level at all layers of the software-based network stack, which will add overhead, delay and resource usage.
A radical reform of the network nature is necessary to target existing and future expectations. This intrinsic reform can come from exploiting quantum-mechanical resources such as superposition of quantum states, quantum entanglement, and distributed quantum computing, beyond targeted quantum security use cases. Quantum networks will be built on top of classical ones, in a unique hybrid infrastructure, where quantum virtual machines will consist of a high number of entangled spatially-distributed qubits and scaling with the number of interconnected devices. This hybrid classical-quantum communication network is normally called by the research community, the Quantum Internet [1]. Generally, the goal of the Quantum Internet is enabling the transmission of qubits between distant quantum devices to achieve the tasks that are impossible using classical communication. However, the definition of the classical 5G and B5G networks is much broader. Since we do not want to develop quantum versions of the classical Internet components, the term Quantum Internet is rather misleading. In this work, we thus prefer to use the term quantum communication network (QCN) [2].
Many quantum network projects have the only goal of securely exchanging cryptographic keys. Entanglement-based protocols like Ekert ’91, as opposed to Bennet-Brassard ’84 like protocols, are often used because they work directly with quantum repeaters. Other popular quantum network projects focus on using the entanglement for teleportation. However, using the quantum network for just these applications is limiting its potential.

II In-Network Computing and Big Data
As previously mentioned, a critical aspect is low latency [3]. Latency can be seen as the sum of different parameters, referred to all the network levels. The propagation latency depends on spatial distance and packet size. While source data is constant, each layer of the network stack adds its own overhead, contributing to this delay. The transmission delay at each link is proportional to the inverse of the available capacity. Next, the queuing delay is due to the queues of packets at all the nodes of the network, which varies for different frames/datagrams/packets depending on their Quality-of-Service (QoS). The processing delay is the physical and link layer delay, which mainly depends on the hardware’s processing capacity of network nodes and on the signal-processing algorithms. Currently, reduction of latency is done per layer. For example, at the physical layer propagation delay is reduced via the transmission of short packets at the cost of increasing the overhead per frame and thus the transmission delay. However, when a single layer is overloaded, optimising the network per layer prevents the other layers to overtake some of the load efficiently, and may result in the whole network being overloaded. For example, the medium-access layer is responsible for synchronisation, initial access, interference management, scheduling, rate adaptation, delay-to-access and management of retransmission, while the transport layer is responsible for the flow and congestion controls. If congestion causes significant losses at the transport layer, it can increase queuing, processing and retransmission delays (losses increase both retransmission time and redundancy for error correction) at the medium-access layer.
Future softwarized NFV-SDN communication networks will consist of a data plane and a control plane (Fig. 2), which will be obtained by a resource mapping from general-purpose hardware owned by various infrastructure providers (also called tenants). Network nodes and traffic in the data plane will be managed and processed by SDN controllers and hypervisor in the control-plane network, with the help of machine learning (ML) towards self-organising (autonomic) networks (SONs) [4]. Moreover, in the protocol stack, the number of active layers will be optimised at each network node improving latency. Fig. 2 depicts the logical architecture of future softwarized classical networks, highlighting the internal structure of their main components. The complete softwarization of networks and the realisation of in-network intelligence will transform the network stack and structure from static to adaptive, through the constant processing of Big Data collected from the enormous amount of devices of the network.
Sometimes these virtual-network optimisation problems can be efficiently distributed, allowing to distribute the overhead load and delays within the network by having distributed SDN controllers into various network nodes or as virtual network functions (VNFs) in different servers or data centres [5]. However, the flexibility introduced by softwarization will inevitably come with an overhead trade-off, which will reduce the improvement in terms of latency and energy efficiency [6]. This overhead can be potentially resolved, or greatly reduced, in a classical-quantum hybrid network, by exploiting the fact that even some classical tasks can be solved much more efficiently in the quantum network, rather than the classical one. All the previously mentioned delays are not only in the network but also in data centres and distributed databases, which are networks themselves (Fig. 2).
III From Classical to Quantum
Classically, the smallest information element is the bit, an object that stores information by using two possible values: and . In quantum systems bits do not exist. The state of any quantum-mechanical system shows a linear behaviour similar to that of continuous waves and signals111Thus the name wavefunction in quantum mechanics.. The unit of quantum information is thus the qubit, an object that stores information on the unit vectors of a two-dimensional complex vector space. The connection with the classical bit is that and now index the two standard basis vectors as and 222 is the Dirac’s bra-ket notation for vectors.. Just like for continuous signals, the sum of and states is another state that is different from probabilistically generating either of the two. The standard representation of the state space of a qubit is displayed in Fig. 3.
The difference between a signal and a quantum state lies in their behaviour. The state of a qubit should be thought as the polarisation of a single photon, where the only collectable information is whether the photon is present or not. We can distinguish two orthogonal states of the photon such as the vertical and horizontal polarisation using a polarising beam splitter, and detecting on which side the photon appears. A qubit can thus carry one bit of information. However, if we try to collect information about a diagonally polarised photon (with the same vertical/horizontal measurement), we will still register the photon only in either the vertical or the horizontal measurement. Only repeating the experiment with many photons, we probabilistically observe either polarisation, but never both simultaneously. Still, while in the above measurement the photon behaves probabilistically, there exists the diagonal-polarisation measurement in which the photon behaves deterministically. This probabilistic/deterministic behaviour of the qubit depends on how we measure it, and thus we refer to those states that behave deterministically in some basis as pure state. The fact that measuring a photon only gives a single bit of information333This is not in contradiction with superdense coding, where one qubit carries two bits of information, but two qubits must be measured to obtain them. implies that we cannot distinguish arbitrary pure states and, after the measurement, the original pure state is destroyed: the quantum information of arbitrary pure states cannot be cloned/copied.
The different behaviour of qubits becomes even more drastic when we consider multiple qubits. There exist states that behave probabilistically under any measurement performed independently on each qubit, but deterministically under a joint measurement that can be performed only with global access to the involved qubits. These are the entangled pure states. Now, it is important to clarify that the randomness produced by a pure state is fundamentally different from the randomness produced by a classical system: the classical system knows the value of the randomness (which is simply hidden), while in a pure state not even the quantum system itself knows the value before the measurement is performed. Said otherwise, the randomness does not even exist before measuring the qubit. This is, for example, how Quantum Key Distribution (QKD) protocols share keys that are claimed provable-secure against any adversary: the protocols prove that the keys at the receivers were produced by entangled pure states.
The manipulation of quantum information gives access to exponentially large continuous systems (where before we had bits of classical values, the -bit-strings now index the standard basis of a dimensional vector space) and provably stronger correlations than classical systems (entanglement). However, we cannot just access all this information. Everything is hidden behind the no-cloning theorem [2] and we can only access the information partially via a measurement. In particular still bits of information can only be collected by measuring qubits.
IV The Quantum Control Plane
Quantum information processing does not give an automatic advantage over classical information processing, and exactly determining for which problems the advantage exists is still an open research. Nonetheless, we already have important quantum algorithms of independent use and wide applications, with quantum simulations in chemistry likely becoming the first real applications with an exponential speedup [7, 8]. A quantum network capable of connecting quantum computers will be essential to support these applications of quantum computing and it will rely on the existent classical network infrastructure. Most importantly (the aspect to which we want to bring our attention), quantum computation and communication have the potential to return some benefits to the classical network itself, by reducing overhead, latency, and congestion, while increasing energy efficiency.
Various classes of unsupervised ML and optimisation problems have been proven to have an exponentially lower computational cost, once the expensive step of encoding the large amount of data in the coefficients of a quantum state has been performed [9, 10, 11, 12]. Notice, that these are not exponential speedups in the strict sense, because the gain is obtained by changing the input to the algorithm, thus making the classical and the quantum algorithm incomparable. To gain this exponential advantage, the type of encoding is crucial. We have seen that we can encode each bit-string onto a standard basis vector, this encoding is lossless but also uses more expensive -qubit strings. Alternatively, we can take any dimensional vector, normalise it, and construct a quantum state of qubits corresponding to this vector. The encoding is extremely lossy, because by measuring the encoded state we can only obtain bits of information, but also extremely storage efficient. For example, a vector of one petabyte with one-byte coefficients can be encoded in just 15 qubits. Since in order to construct the quantum state, the classical data must be accessed, this encoding is still an expensive step. Once the expensive encoding is done, -means [9], principal component analysis (PCA) [10], support vector machines (SVM)[11], and any semidefinite program (SDP) – a general class of optimisation problems that include linear programs – can be run efficiently, namely polynomially in the number of qubits, with a quantum computer.
By relying the network optimisation on algorithms that fit this class, we can experience an exponential reduction on the traffic and the computational load introduced by the intelligent control-plane operations. Even any other algorithm that relies on the classical data, being encoded in the coefficients of the superposition, which does not have the exponential speed up, will incur an exponential reduction of the traffic load. In particular, a quantum convolutional neural network will at least gain access to exponentially deeper networks at the same training cost [13]. The required flexibility of the network and the amount of data analysed make neural network and unsupervised ML algorithms the main choice for ML in the hypervisor [4]. This makes the quantum algorithms mentioned above particularly relevant for this application.
The expensive quantum encoding of classical Big Data would be distributed and performed on-site at the nodes of the network where the data is collected. Namely, each data-source node would collect their data on a few qubits rather than streaming a large number of classical data via the classical network. At this point only a logarithmic number of qubits, compared to the number of classical bits that would otherwise be required, needs to be sent to a processing quantum data centre, hosting the hypervisor. A large load on the classical network thus becomes a small load on the quantum network.
To take full advantage of this encoding, the intermediate nodes will have to be capable of using the quantum analog of the random access memory (QRAM) [14] and all nodes need to be able to encode in superposition a vector of data into a state where each vector coefficient gives a coefficient in the superposition. A typical B5G data-plane node will be controlled by up to parameters [15] that can be encoded in 11 qubits. Via QRAM, the quantum states can be collected at the SDN nodes. If the SDN controller receives data from switches then with qubits the SDN node can use QRAM to join the qubits from all the nodes into qubits. Namely, the 11 qubits of all the switches are joined into the same 11 qubits via entanglement with the qubits of QRAM address. The nodes then can send the newly encoded quantum states with minimum overhead to the central hypervisor, which will in turn join the states received by the quantum SDN controllers, into a state of qubits. This advantage is displayed in Fig. 4. The communication at every link is reduced exponentially, compared to each SDN controller sending parameters and the central hypervisor receiving parameters. Indeed, if each SDN controller receives switches and the SDN controllerhypervisor receives from SDN controllers, then the final quantum states only occupies qubits at the hypervisor. The total amount of sent qubits is , however distributed over for example control-plane links, in contrast to the classical parameters.

The quantum algorithms, however, are probabilistic and usually work by collecting the expectation value over the repetition of the algorithms many times. Because of the no cloning theorem [2], the hypervisor will not be able to generate the copies of the encoded data needed for the repetition, thus the drawback is that for every repetition the data must be re-encoded in the quantum states at the source. If for example samples are needed, then the whole operation must be repeated times, thus multiplying by the amount of sent qubits. However, the number of repetition is generally the inverse polynomial on an approximation/error parameter and thus as the number of parameters grows exponentially with future generation networks, the number of repetitions will scale polynomially. Namely, the quantum algorithm is scalable to the increasing size and complexity of future generation networks. Already at the B5G nodes with parameters, this quantum encoding allows samples to be taken before the quantum communication exceeds the needed classical communication at the leaf nodes. The bandwidth reduction within the control-plane links is exponentially larger if we consider nodes that send all data to the hypervisor. If instead we consider SDN controllers that classically reduce the data via some ML preprocessing before sending it to the hypervisor, the bandwidth reduction might be comparable. However, in this comparison the quantum ML algorithm at the hypervisor has a full-global view of the network status, while the classical algorithm has a reduced and indirect view of the global status, depending on the reduced data.
This reduction of packets sent, stored and processed within the control plane will reduce congestion and reduce almost all sources of delay such as queuing and transmission delay in the network, and storage and processing delay at the processing centre. As we have seen, each additional load to one layer also contributes to load the other. In particular, each additional processing and softwarization itself introduces new delays and loads on other layers of the network. Thus the smaller queues due to the quantum processing further contributes to the reduction of processing and communication needed for queues’ management. The processing required by the ML algorithm at the hypervisor to handle Big Data of the network will consume significant energy. In general, the amount of energy saving directly depends on the number of bytes sent to the hypervisor and the number of instructions for computations, and inversely depends on the available bandwidth at the communication links [6]. Thus, the reduction due to quantum deployment described above will not only reduce latency but it will also imply higher energy efficiency (in line with 5G and B5G goals).
Furthermore, shared entanglement will allow the classical and the quantum networks to transfer each others’ load. Namely, stored entanglement allows to exchange quantum and classical communication through the teleportation and dense-coding protocols, in order to balance the load in the hybrid network. In quantum teleportation, one qubit can be communicated using previously-shared entanglement and sending two (classical) bits. In dense-coding two bits of communication can be achieved by sending one qubit of the previously-shared entanglement. This is one additional reason to consider obsolete the static-layer view of the network. The classical and quantum networks cannot be two distinct layers of the network that are optimised independently because in this way the flexibility we just described is lost.
Finally, shared entanglement has the potential of even reducing communication complexity, by allowing nodes in the network to send data that is more correlated than classically to the processing centres. This is not the case of extracting a small amount of information from large data, and is not in contradiction with a measurement on qubits only being able to provide bits of information. Sometimes the required amount of communication to perform a task is less than the communication needed to solve the same task classically, as demonstrated by non-local games such as the Clauser-Horne-Shimony-Holt (CHSH) and the Mermin game [2]. However, for this advantage to exist, strong non-local correlations between the environments, adversarial setting, or strong limitations on the type of channels must be present. Therefore further research is needed to determine whether such advantages are realistic.
V Conclusion
We have seen that quantum computation and communication will not necessarily be only a new resource but can also contribute to reduce overheads, latency, and increase energy efficiency in the classical softwarized networks. However, quantum information processing does not magically improve any computation or communication tasks and only a few classes of problems have been shown to take advantage from quantum technologies. Notably, Grover’s search algorithm [2] is an extremely general algorithm, capable of providing up to quadratic quantum speedup on almost any algorithm, that will surely find its applications in every field, including future softwarized networks. Our aim was to clearly point out where there is a huge potential for future networks to benefit from quantum technologies. In parallel, even the classical part of future network infrastructures should be designed with quantum technologies in mind, to efficiently and effectively merge the two into a unique hybrid architecture. The use of quantum technologies in computation and communication is a young field but is now quickly maturing so that possible applications will appear in the near future444See the European Quantum Flagship https://qt.eu/ (Feb. 2021).. Thus, the final goal is to see a paradigm shift in the network development, where classical and quantum technologies become two sides of the same hybrid communication network.
Acknowledgements
This work has been partially funded by the German Research Foundation (DFG, Deutsche Forschungsgemeinschaft) as part of Germany’s Excellence Strategy – EXC2050/1 – Project ID 390696704 – Cluster of Excellence “Centre for Tactile Internet with Human-in-the-Loop” (CeTI) of Technische Universität Dresden. Holger Boche, Christian Deppe and Roberto Ferrara were supported by the Bundesministerium für Bildung und Forschung (BMBF) through Grants 16KIS0856 (Deppe, Ferrara), 16KIS0858 (Boche), and 16KIS0948 (Boche). This work has also been partially funded by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) within Germany’s Excellence Strategy under Grant EXC-2111 390814868.
References
- [1] A. S. Cacciapuoti, M. Caleffi, F. Tafuri, F. S. Cataliotti, S. Gherardini, and G. Bianchi, “Quantum internet: Networking challenges in distributed quantum computing,” IEEE Network, vol. 34, no. 1, pp. 137–143, January 2020.
- [2] R. Bassoli, H. Boche, C. Deppe, R. Ferrara, F. H. P. Fitzek, G. Janßen, and S. Saeedinaeen, Quantum Communication Networks, 1st ed. Springer International Publishing, Jan. 2021. [Online]. Available: https://www.springer.com/gp/book/9783030629373
- [3] X. Jiang, H. Shokri-Ghadikolaei, G. Fodor, E. Modiano, Z. Pang, M. Zorzi, and C. Fischione, “Low-latency networking: Where latency lurks and how to tame it,” Proceedings of the IEEE, vol. 107, no. 2, pp. 280–306, 2019.
- [4] J. Xie, F. R. Yu, T. Huang, R. Xie, J. Liu, C. Wang, and Y. Liu, “A survey of machine learning techniques applied to software defined networking (SDN): Research issues and challenges,” IEEE Communications Surveys Tutorials, vol. 21, no. 1, pp. 393–430, 2019.
- [5] F. Bannour, S. Souihi, and A. Mellouk, “Distributed SDN control: Survey, taxonomy, and challenges,” IEEE Communications Surveys Tutorials, vol. 20, no. 1, pp. 333–354, 2018.
- [6] X. Cao, L. Liu, Y. Cheng, and X. Shen, “Towards Energy-Efficient Wireless Networking in the Big Data Era: A Survey,” IEEE Communications Surveys Tutorials, vol. 20, no. 1, pp. 303–332, 2018.
- [7] J. Preskill, “Quantum computing and the entanglement frontier,” California Institute of Technology, Rapporteur talk at the 25th Solvay Conference CaltechAUTHORS:20120516-084322874, 2012. [Online]. Available: https://resolver.caltech.edu/CaltechAUTHORS:20120516-084322874
- [8] ——, “Quantum Computing in the NISQ era and beyond,” Quantum, vol. 2, p. 79, Aug 2018. [Online]. Available: http://dx.doi.org/10.22331/q-2018-08-06-79
- [9] S. Lloyd, M. Mohseni, and P. Rebentrost, “Quantum algorithms for supervised and unsupervised machine learning,” 2013.
- [10] ——, “Quantum principal component analysis,” Nature Physics, vol. 10, no. 9, p. 631–633, Jul 2014. [Online]. Available: http://dx.doi.org/10.1038/nphys3029
- [11] P. Rebentrost, M. Mohseni, and S. Lloyd, “Quantum support vector machine for big data classification,” Physical Review Letters, vol. 113, no. 13, p. 130503, Sep 2014. [Online]. Available: http://dx.doi.org/10.1103/PhysRevLett.113.130503
- [12] F. G. Brandao and K. M. Svore, “Quantum speed-ups for solving semidefinite programs,” in 2017 IEEE 58th Annual Symposium on Foundations of Computer Science (FOCS). IEEE, 2017, pp. 415–426.
- [13] I. Kerenidis, J. Landman, and A. Prakash, “Quantum algorithms for deep convolutional neural networks,” 2019.
- [14] V. Giovannetti, S. Lloyd, and L. Maccone, “Quantum random access memory,” Physical Review Letters, vol. 100, no. 16, p. 160501, Apr 2008. [Online]. Available: http://dx.doi.org/10.1103/PhysRevLett.100.160501
- [15] A. Imran, A. Zoha, and A. Abu-Dayya, “Challenges in 5G: how to empower SON with big data for enabling 5G,” IEEE Network, vol. 28, no. 6, pp. 27–33, 2014.
Roberto Ferrara obtained his Ph.D. in science at the Department of Mathematical Sciences of the University of Copenhagen. Since 2019 he is at the Department of Communications Engineering at the Technische Universität München. |
Riccardo Bassoli is a senior researcher with the Deutsche Telekom Chair of Communication Networks at the Faculty of Electrical and Computer Engineering, Technische Universität Dresden (Germany). He received his Ph.D. degree from 5G Innovation Centre (5GIC) at University of Surrey (UK), in 2016. |
Christian Deppe received the Dr.-Math. degree in mathematics from the Universität Bielefeld, Bielefeld in 1998. He was a Research and Teaching Assistant with the Universität Bielefeld. Since 2018 he is at the Department of Communications Engineering, Technical University of Munich, Munich, Germany. |
Frank H.P. Fitzek received the Dr.-Ing. degree in Electrical Engineering from the Technical University Berlin, Germany. He is a Professor and head of the Deutsche Telekom Chair of Communication Networks at Technische Universität Dresden. |
Holger Boche received the Dr.-Ing. degree in electrical engineering from the Technische Universität Dresden in 1990 and 1994, respectively, and the Dr. rer. nat. degree in pure mathematics from the Technische Universität Berlin in 1998. Since 2010, he is Full Professor at the Technische Universität München, Munich, Germany. |