Quantum Key Distribution Protocols
Quantum computing poses a serious threat to the basis of contemporary encryption, which uses mathematical complexity to secure secret information. Existing key distribution techniques have significant drawbacks that quantum computing could potentially exploit and are dependent on classical computing.
Quantum Key Distribution (QKD), on the other hand, offers a secure communication technique based on the basic principles of quantum mechanics. This area of data and information security is one that is rapidly growing and showing promise. With QKD, two communicating parties, typically referred to as Alice (sender) and Bob (receiver), can generate a shared, secret, random key that is only known to them.
The Heisenberg uncertainty principle, quantum entanglement, superposition, and the no-cloning theorem are among the fundamental ideas of quantum mechanics that are utilized by QKD methods. The capacity to identify the existence of any third party (Eve, the eavesdropper) trying to obtain information about the key is a crucial and distinctive feature of QKD. This ability is based on the idea that measuring a quantum system would always cause it to be disturbed, which makes the eavesdropper visible through observable defects.
The key generation process is stopped and the communication is terminated if the error rate exceeds a predetermined threshold. It is crucial to remember that QKD simply creates and disseminates this secret key, which is subsequently utilized for message encryption by traditional encryption techniques like Advanced Encryption Standard (AES) and the one-time pad.
QKD Protocols Based on Quantum Uncertainty: BB84 and B92
The “prepare-and-measure” category of Quantum Key Distribution (QKD) procedures, which is based on the measurement-disturbance principle that results from the Heisenberg uncertainty principle, includes several of them.
The most well-known and fundamental QKD protocol is the BB84 protocol, which was first presented by Charles Bennett and Gilles Brassard in 1984. For encoding, BB84 uses single photons and two conjugate pairs of non-orthogonal states, such as the diagonal basis (45°/135° polarisation) and the rectilinear basis (vertical/horizontal polarisation). Since it is typically impossible to measure these non-orthogonal states without affecting the initial state, security is ensured.
Alice encodes her photon by choosing at random a bit value (0 or 1) and a matching basis, which she then transmits to Bob. Bob chooses one of the two bases at random to measure the received photon since he is unaware of Alice’s selection. Following transmission, they make their bases known to the public (sifting), eliminating any bits whose bases did not match. In an ideal channel, their remaining bit strings should be the same if there were no interference or eavesdropping.
Bennett presented the B92 protocol in 1992 as a revision to BB84. By using only two non-orthogonal quantum states, B92 streamlines the procedure. The objective of B92’s design was to enable variable parameter modification based on channel conditions in order to possibly increase the data exchange rate. In contrast to BB84, the B92 protocol is used more simply in real-world applications on systems such as the IBM Quantum Composer. A parameter pertaining to the angle between the non-orthogonal states is what distinguishes the B92 protocol; altering this angle permits trade-offs between resistance to interference and a faster data exchange rate.
QKD Protocols Based on Quantum Entanglement: E91
Entanglement-based protocols, as opposed to prepare-and-measure schemes, are based on the idea that two or more particles are inherently connected, regardless of their distance from one another, according to quantum entanglement.
In 1991, Artur Ekert developed the E91 protocol, which makes use of maximally entangled photon pairs. Alice, Bob, or an outside source may create these entangled pairs. The entangled pair sends one photon each to Alice and Bob, who use randomly selected bases to make measurements. Their measurement results will be perfectly synchronized if they select the same basis because of entanglement.
Bell’s inequality and the theorem violations provide the foundation of E91’s key security mechanism. Alice and Bob can identify these correlations by comparing a subset of their results, but Eve destroys them if she tries to measure the entangled state or intercept one of the photons. They come to the conclusion that Eve has violated the fundamental quantum correlations by introducing local realism into the system if the Bell test statistic is not maximized.
One of E91’s main advantages is that its security is “device-independent,” which means it depends on the basic characteristics of quantum entanglement rather than the assumed reliability or flawless calibration of the actual devices used to generate the keys.
The Key Generation Process: Post-Processing Steps
To finalize the secret key, QKD necessitates a number of post-processing procedures carried out over an authorized classical channel, regardless of the underlying quantum principle (entanglement or uncertainty).
- Sifting: To create a raw, correlated key known as the sifted key, Alice and Bob openly disclose which states or bases were used, eliminating the non-matching or inconclusive results.
- Error Estimation: The Quantum Bit Error Rate (QBER) is then calculated by Alice and Bob publicly comparing a little, random portion of their sifted key bits. Eavesdropping or channel flaws (such as quantum noise) might result in errors. Usually, it is thought that Eve is to blame for all mistakes. The process is stopped if the QBER rises above a predetermined level.
- Information Reconciliation (Error Correction): Alice and Bob utilize error correction methods, like the Cascade protocol, to make sure their key strings are exactly the same if the error rate is acceptable. However, they frequently provide Eve some incomplete information in the process.
- Privacy Amplification: In this last stage, Eve’s residual partial knowledge obtained from parity checks during reconciliation or the quantum channel is diminished. Eve has a very small chance of knowing the final secret key because a universal hash function is applied to the reconciled key to create a new, shorter final key.
Also Read About National Quantum Computing Centre Gets Ion Trap From NPL
Practical Difficulties and Security
Unlike traditional public key cryptography, QKD protocols use information theory based on quantum mechanical rules to provide provable security.
Despite this theoretical power, there are significant obstacles to real-world implementation. Channel loss, decoherence, and flaws in the preparation, transmission, and measurement of quantum states are challenges that result in low secret key rates and transmission distance restrictions. Furthermore, even though the underlying protocol is mathematically safe, real-world QKD systems have been shown to be susceptible to a variety of attacks that take advantage of hardware vulnerabilities, including Trojan-horse attacks and the Photon Number Splitting (PNS) attack.
For the majority of applications, the US National Security Agency (NSA) and the UK National Cyber Security Center (NCSC) have advised switching from QKD to Post-Quantum Cryptography (PQC) due to these security validation issues. High infrastructure costs, the requirement for specialized hardware, and the challenge of safely verifying the hardware of the deployed system, which frequently lacks the adaptability for simple upgrades or security patches, are some of their worries.
However, QKD is still developing quickly, with both commercial devices and in-depth research being conducted to increase security, distance, and efficiency against real-world threats.