Next Article in Journal
A New Hybrid Automated Security Framework to Cloud Storage System
Previous Article in Journal
Beyond Bitcoin: Recent Trends and Perspectives in Distributed Ledger Technology
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Post-Quantum and Code-Based Cryptography—Some Prospective Research Directions

by
Chithralekha Balamurugan
1,
Kalpana Singh
2,*,
Ganeshvani Ganesan
1 and
Muttukrishnan Rajarajan
3
1
Department of Computer Science, Pondicherry University, Chinna Kalapet, Kalapet, Puducherry 605014, India
2
IRT SystemX, 2 Boulevard Thomas Gobert, 91120 Palaiseau, France
3
Institute for Cyber Security, University of London, Northampton Square, Clerkenwell, London EC1V 0HB, UK
*
Author to whom correspondence should be addressed.
Cryptography 2021, 5(4), 38; https://doi.org/10.3390/cryptography5040038
Submission received: 2 August 2021 / Revised: 20 November 2021 / Accepted: 9 December 2021 / Published: 20 December 2021
(This article belongs to the Special Issue Cryptographic Systems Based on Coding Theory)

Abstract

:
Cryptography has been used from time immemorial for preserving the confidentiality of data/information in storage or transit. Thus, cryptography research has also been evolving from the classical Caesar cipher to the modern cryptosystems, based on modular arithmetic to the contemporary cryptosystems based on quantum computing. The emergence of quantum computing poses a major threat to the modern cryptosystems based on modular arithmetic, whereby even the computationally hard problems which constitute the strength of the modular arithmetic ciphers could be solved in polynomial time. This threat triggered post-quantum cryptography research to design and develop post-quantum algorithms that can withstand quantum computing attacks. This paper provides an overview of the various research directions that have been explored in post-quantum cryptography and, specifically, the various code-based cryptography research dimensions that have been explored. Some potential research directions that are yet to be explored in code-based cryptography research from the perspective of codes is a key contribution of this paper.

1. Introduction

Cryptographic systems are built on complex mathematical problems such as integer factorization and computing discrete logarithms [1,2], which can only be solved if knowledge of some secret data is available; typically a very large number. Without these numbers, it is impossible to reverse-engineer encrypted data or create a fraudulent digital signature. These numbers are what we know as cryptographic keys. For instance, the RSA algorithm [3] works by using pairs of very large prime numbers to generate public and private keys. The public key can be used to create a mathematical challenge which can only be solved by someone who holds the private key. Attempting to guess the answer, by way of a brute-force search, would take thousands of years using contemporary computers. Unlike their classical counterparts, quantum computers will be able to solve these mathematical problems incredibly quickly. The asymmetric algorithms we use today for digital signatures and key exchange will no longer be strong enough to keep data secret once a sufficiently powerful quantum computer can be built.
This means that core cryptographic technologies that we have to rely on, RSA and elliptic curve cryptography, will become insecure. Even symmetric cryptographic algorithms have now been found possible to be attacked by quantum computers [4]. This context alludes to the fact that both symmetric and asymmetric algorithms which are in widespread use today can succumb to quantum attacks and hence, quantum attack resistant, or in other words, post-quantum cryptographic algorithms, need to be evolved.
Certain candidate families of post-quantum schemes have been realized including code-based [5], hash-based [6], multivariate [7], lattice-based [8,9] and isogeny-based [10] solutions. The maturity in post-quantum research has led to the formulation of various post-quantum cryptographic algorithms, the standardization of post-quantum algorithms by various standardization bodies world-wide, industry adoption of post-quantum technology, and the development of open-source post-quantum libraries.
In this paper, an overview of these research dimensions that have been explored in post-quantum cryptography has been provided. Furthermore, with a specific focus on code-based cryptography (CBC), the key milestones in CBC research have been provided. Although code-based cryptography focuses on the use of codes, other potential research dimensions could be explored in CBC research leveraging the codes—which is yet to receive focus in CBC research. Delineating this aspect has been the important contribution of this paper, and has been achieved by the study of the linear codes, their operations and relationships, and other kinds of codes that could be deployed in CBC. This paper also highlights that privacy-preserving CBC is another prospective research dimension that is yet to receive enough focus in CBC research.
Section 2 briefs about quantum computing and alludes to the motivation for post-quantum cryptographic research. Section 3 reviews post-quantum research dimensions with due summaries and comparisons. Section 4 explains the CBC research dimensions that have been explored. Section 5 describes the white spaces that are yet to be addressed in CBC research from the perspective of codes and privacy aspects. Section 6 concludes the paper.

2. Quantum Computing

A quantum computer is a machine which employs quantum-physical phenomena to perform computations in a way that is fundamentally different from a “normal”, classical computer [11]. Whereas a classical computer is, at any point in time, in a fixed state—such as a bit string representing its memory contents—the state of a quantum computer can be a “mixture”, a so-called superposition, of several states. Classical computers carry out logical operations using the definite position of a physical state. These are usually binary, meaning its operations are based on one of two positions. A single state—such as on or off, up or down, 1 or 0—is called a bit. Note that the internal state is hidden: the only way to obtain information about the state is to perform a measurement, which will return a single non-superimposed classical output, such as a bit string, that is randomly distributed according to the internal state, and the internal state is replaced by the measurement outcome.
The work was initiated by several mathematicians and physicists such as Paul Benioff (1980) [12], Yuri Manin (1980) [13], Richard Feynman (1982) [14], and David Deutsch (1985) [15]. With decades of research, the development of quantum computing has been challenging yet groundbreaking.
Thus, quantum computing constitutes a new computing paradigm, which is expected to solve complex problems that require far more computational power than what is possible with the current generation of computer technologies. Advance research in materials science, molecular modelling, and deep learning are a few examples of complex problems that quantum computing can solve.

Shor’s 1994 and Grover’s 1996 Algorithms

On the quantum algorithmic development, there are two groundbreaking algorithms which have laid out a strong foundation towards breaking today’s number of theoretically based public-key cryptosystems. In 1994, Shor proposed a polynomial-time (efficient) algorithm [16] for solving integer factorization and discrete logarithm problems. The algorithm relies on the existence of quantum computers, and hence this type of algorithm is called quantum algorithms in this article. Shor’s quantum algorithm and its variants can be used for breaking most of the currently used public-key cryptosystems.
In 1996, Grover proposed an O ( N ) -query complexity of quantum algorithm for functions with N-bit domains [17]. This quantum algorithm once realized on quantum computers can be used for breaking symmetric-key cryptosystems, and to defend against attacks based on Grover’s algorithm, we need to double the key sizes in order to achieve a similar level of security against conventional computers.
For example, for 128-bit symmetric-key security, we need to use symmetric-key cryptosystems which are originally designed for achieving 256-bit security against attacks based on Grover’s quantum algorithm. It is also predicted that quantum computers will be able to break several of today’s cryptographic algorithms that are used to secure communications over the Internet, provide the root of trust for secure transactions in the digital economy and encrypt data. To protect against attacks from quantum computers, vendors of security products and service providers must constantly assess the risk associated with the choice of cryptographic algorithms and evolve entirely new quantum-resistant algorithms for the post-quantum world.

3. Post-Quantum Cryptography

Post-quantum cryptography (PQC) is about devising cryptographic algorithms that are secure in the quantum era with security against both classical/conventional and quantum computers. There are several candidate approaches for building the post-quantum cryptographic schemes, as described below in Section 3.1 [18,19,20].

3.1. Post-Quantum Cryptography Candidates

This subsection delineates the candidates of the PQC schemes. These are hash-based cryptography, code-based cryptography, multivariate cryptography, lattice-based cryptography, and isogeny-based cryptography schemes.

3.1.1. Hash-Based Cryptography

Hash-based cryptography focuses on designing digital signature schemes based on the security of cryptographic hash functions, e.g., SHA-3. These schemes are based on the security of hash functions (as a one-way function, collision-resistant property, and hardness of second pre-image attacks), and require fewer security assumptions than the number-theoretic signature schemes (e.g. RSA, DSA). Ralph Merkle in 1989 introduced Merkle Signature Scheme (MSS) [21], which is based on one-time signatures (e.g., the Lamport signature scheme) and uses a binary hash tree (Merkle tree). The MSS is resistant to quantum computer algorithms. More details can be found in this survey on hash-based schemes Butin (2017) [22]. Sphincs+ hash-based signature [23] is chosen as an alternate solution in the outcome of the third round of the NIST standardization process [24].

3.1.2. Code-Based Cryptography

Code-based cryptography [25,26] has its security relying on the hardness of problems from coding theory, for example, syndrome decoding (SD) and learning parity with noise (LPN). These cryptosystems are based on error-correcting codes to construct a one-way function. The security is based on the hardness of decoding a message which contains random errors and recovering the code structure. A classic McEliece code-based encryption scheme [5] is chosen as a finalist scheme in the outcome of the third round of the NIST standardization process [24].

3.1.3. Multivariate Cryptography

Multivariate cryptography has its security relying on the hardness of solving multivariate systems of equations. These schemes are based on systems of multivariate polynomial equations over a finite field F. There are several variants of multivariate cryptography schemes based on hidden field equations (HFE) trapdoor functions, such as the unbalanced oil and vinegar cryptosystems (UOV). UOV is used for signatures. Other examples of multivariate cryptography are Rainbow, TTS, or MPKC schemes. More about the current state of the multivariate cryptography schemes can be found in the paper by Ding and Petzoldt (2017) [27]. Two multivariate signature schemes are chosen in the outcome of the third round of the NIST. Rainbow [28] is one of the finalists. GeMSS [29] is one of the alternate finalist schemes.

3.1.4. Lattice-Based Cryptography

Lattice-based cryptography seems to be one of the most active directions in recent years, for several key reasons. First, it has strong security guarantees from some well-known lattice problems, for example, shortest vector problem (SVP) and the ring learning with errors (RLWE) problem [30]. Second, it enables powerful cryptographic primitives; for example, fully homomorphic encryption (FHE) and functional encryption [31]. Third, some new lattice-based cryptographic schemes have become quite practical recently, for example, the key exchange protocol NewHope [32], and a signature scheme BLISS [33].
Lattice-based cryptography is one of the successful schemes in the third round result of the NIST standardisation process. Kyber [34], NTRU [35], SABER [36] lattice-based encryption schemes are chosen as the finalists’ schemes. NTRUprime [37] is in the alternate finalist lattice-based encryption scheme. Dilithium [38] and Falcon [39] lattice-based signature schemes are also finalist schemes.

3.1.5. Isogeny-Based Cryptography

Isogeny-based cryptography is a specific type of post-quantum cryptography that uses certain well-behaved maps between abelian varieties over finite fields (typically elliptic curves) as its core building block. Its main advantages are relatively small keys and its rich mathematical structure, which pose some extremely interesting questions to cryptographers and computer allegorists. These schemes are based on supersingular elliptic curve isogenies [10] that are secure against quantum adversaries. These schemes are secured under the problem of constructing an isogeny between two supersingular curves with the same number of points. Isogeny-based schemes may serve as digital signatures or key exchange, such as the supersingular isogeny Diffie–Hellman (SIDH) scheme [40]. SIKE [41] is the only isogeny-based encryption scheme in the alternate list of NIST third round results. There is no isogeny-based signature scheme identified in the NIST third round outcome.

3.1.6. Comparison of Post-Quantum Cryptography Algorithms

The five post-quantum cryptographic algorithm categories have been compared considering only the key algorithm(s) in each category, and the comparison has been presented in Table 1. This is to provide a quick summary of the key algorithms in the five post-quantum algorithm categories, and is not intended for a complete review of the existing algorithms in each of the post-quantum algorithm categories. An earlier comparison of post-quantum cryptographic algorithms has also been attempted in [42].

3.2. Industry Adoption of Post-Quantum Cryptography

The industry adoption of post-quantum cryptography is happening very aggressively. On this front, the following lines of works are found to be available:
  • Industry survey of post-quantum cryptography,
  • Revenue Assessment of post-quantum cryptography,
  • Industry initiatives in PQC—PQC R&D, PQC-based products, PQC products, PQC consulting.

3.2.1. Industry Survey

The industry survey has been carried out by Digicert involving IT Directors, IT Generalists, IT security professionals, and other professionals belonging to the USA, Germany, and Japan. The survey focused on identifying the following:
  • The awareness or the understanding about PQC with industry professionals,
  • The industry professionals’ prediction of timelines by which quantum computers would break the existing modular arithmetic cryptographic algorithms,
  • The understanding among the industry professionals about the significance of threat imposed by quantum computing on existing cryptographic algorithms,
  • The study of industry readiness to adopt PQC.
The results obtained indicate that the awareness of PQC among industry professionals is reasonably good and they have a clear understanding of an appropriate timeline by which quantum computers would break the existing cryptographic algorithms. The impact of the threat imposed by PQC is also well perceived by the industry professionals. The survey also reveals the industry readiness in the adoption of PQC to be beyond 50%. These aspects indicate that the industry survey has helped to comprehend the line of thought and the industry awareness and preparedness for PQC among industry professionals; this is very important when it comes to starting to be precautious and working out plans for adoption of PQC given the discernment of the threat due to quantum computing technology.

3.2.2. Revenue Assessment of Post-Quantum Cryptography

As per the ten-year Market and Technology Forecast Report in [50], a comprehensive study about the prospective markets for PQC products and services has been carried out. The IT industry, cybersecurity industry, telecommunications industry, financial services industry, healthcare industry, manufacturing industry, PQC in IoT, and public sector applications of PQC have been identified as prospective markets for PQC in the report. An elaborate study of how PQC could augment or enhance the functioning of the above industries has been detailed in the report. A ten-year forecast of revenue assessment of PQC in each of the above industries is detailed in the said report, which is indicative of the prospective industry market and trend for PQC.

3.2.3. Industry Initiatives in PQC–PQC R& D, PQC-Based Products, PQC Products, PQC Consulting

The IT industry has been closely following up with post-quantum cryptographic research and the standardization process. The industry initiatives could be observed in terms of the following.
  • Research and development in post-quantum cryptography—organizations like IBM [51], Microsoft [52], etc. conduct research and development in post-quantum cryptography.
  • Development of post-quantum-based products—for example, Avaya has tied up with post-quantum (a leading organization developing post-quantum solutions), to incorporate post-quantum security into its products.
  • PQC products—Organizations like Infineon, Qualcomm (OnBoard Security), Thales, Envieta, etc. have developed post-quantum security hardware/software products [50].
  • Post-quantum consulting—Utimaco is one of the leading players which provides for post-quantum cryptography consulting.
Table 2 provides a comparison of the various industry initiatives of PQC concerning PQC-, R&D-, PQC-based security products development, PQC product development, and PQC consulting.

3.3. Standardization Efforts in PQC

The standardization of post-quantum algorithms has been taken up by different standardization bodies across the globe. The following section provides an overview of the standardization activities taken up by the following standardization bodies viz. NIST, ITU, ISO, ETSI, and CRYPTREC.

3.3.1. NIST

NIST is one of the primary bodies involved in the standardization of post-quantum algorithms [58]. The standardization process began in 2016, and it is currently in the third round after two previous rounds of post-quantum algorithm evaluations. The third round finalists [24] comprise 4 public-key encryption algorithms and 3 digital signature algorithms, along with 5 and 4 alternate candidates for public-key encryption and digital signature algorithms, respectively. These are listed in the Table 3. For standardization, the algorithm submissions were first ensured to fulfill certain minimum acceptability requirements before evaluation, and evicted otherwise. In each round, the evaluations were carried out using a set of criteria under security, cost, and efficiency concerning algorithm implementation aspects, respectively. From Table 3, it is obvious that lattice-based post-quantum technology has the majority contribution among the list of standardized post-quantum algorithms.

3.3.2. International Telecommunication Union (ITU)

International Telecommunication Union (ITU) has two study groups SG13 and SG17 for PQC and a focus group FG QIT4N [60], which works on pre-standardization activities. SG13 has provided a plethora of standards under the categories of (i) Architecture, Framework, Function of Quantum Key Distribution Network and (ii) Quality of Service of Quantum Key Distribution Network. Some of these standards are published and many are work-in-progress. SG17 has provided a set of standards related to the security aspects of the Quantum Key Distribution Network. FG QIT4N group focuses on the pre-standardization activities related to Quantum Information Technology for Networks. It has two working groups, WG1 and WG2; working on this and relevant technical reports has been published by these working groups. A detailed listing of the standards and pre-standardization reports of ITU could be found in [60].

3.3.3. European Telecommunications Standards Institute (ETSI)

European Telecommunications Standards Institute (ETSI) ETSI develops ETSI group specifications and group reports describing quantum cryptography for ICT networks. The ETSI has been involved in the standardization activities of QKD since 2008. The standards developed by ETSI pertain to Quantum-Safe Cryptography, CYBER, and Quantum Key Distribution. Various standards under the said three categories have been developed by ETSI. The listing and details of the standards under the said categories could be found in [61].

3.3.4. ISO

ISO/IEC JTC 1/SC 27 IT Security techniques include 5 working groups [60,62]. SC27 has developed many cryptography standards in the past 28 years. In ISO/IEC JTC 1/SC27, WG2 is for the standardization of cryptography and security mechanisms. The standards cover a large scope, from relatively advanced topics such as homomorphic encryption, group signatures to some essential functions such as block ciphers and hash functions. A six-month study period on Quantum Resistant Cryptography was initiated at a SC 27/WG 2 meeting held in Jaipur, India October 2015. After the first six months, the study period was extended three times and determined to close at SC 27/WG 2 meeting held in Berlin, Germany, in November 2017. As a result of the study period, it was determined to generate a WG2 standing document (SD). An outcome of this post-quantum cryptography study is SD8. SD8 provides a survey on different categories/families of post-quantum cryptography, and is intended to prepare WG2 experts for standardization. SD8 is created in multiple parts, where each part corresponds to each of the post-quantum cryptographic techniques vires. Hash, Lattice, Code, etc. ISO/IEC JTC1 SC27 WG3 (ISO/IEC 23837) focus on security requirements, test, and evaluation methods for quantum key distribution. This addresses QKD implementation security issues. A high-level framework for the security evaluation of the QKD module under the Common Criteria (CC) (ISO/IEC 15408) framework has evolved.

3.3.5. CRYPTREC

CRYPTREC is the Cryptography Research and Evaluation Committees set up by the Japanese Government to evaluate and recommend cryptographic techniques for government and industrial use. The following activities have been carried out as part of CRYPTREC. The CRYPTREC cipher list was published in 2013, followed by the CRYPTREC Report 2014 on lattice problems. The Cryptanalysis Evaluation Working Group was formed in 2015. This group has published a Report on PQC in 2018. CRYPTREC plans to revise the CRYPTREC Cipher List to include post-quantum algorithms by 2022–2024. The details of the above activities are detailed in [63].
Table 4 provides a summary of all the standardization efforts described.
In addition to the above efforts, the following works are being carried out by other organizations for PQC:
  • IETF has formulated a Framework to Integrate Post-quantum Key Exchanges into Internet Key Exchange Protocol Version 2 (IKEv2), refs. [64,65].
  • libpqcrypto [66] is a new cryptographic software library produced by the PQCRYPTO project. libpqcrypto collects this software into an integrated library, with (i) a unified compilation framework; (ii) an automatic test framework; (iii) automatic selection of the fastest implementation of each system; (iv) a unified C interface following the NaCl/TweetNaCl/SUPERCOP/libsodium API; (v) a unified Python interface (vi) command-line signature/verification/encryption/decryption tools, and (vii) command-line benchmarking tools.
  • The Cloud Security Alliance Quantum-Safe Security Working Group’s [67] goal is to address key generation and transmission methods that will aid the industry in understanding quantum-safe methods for protecting data through quantum key distribution (QKD) and post-quantum cryptography (PQC). The goal of the working group is to support the quantum-safe cryptography community in the development and deployment of a framework to protect data, whether in movement or at rest. Several reports and whitepapers on quantum safe cryptography have been published.
  • NSA is publicly sharing guidance on quantum key distribution (QKD) and quantum cryptography (QC) as it relates to secure National Security Systems (NSS). NSA is responsible for the cybersecurity of NSS, i.e., systems that transmit classified and/or otherwise sensitive data. Due to the nature of these systems, NSS owners require especially robust assurance in their cryptographic solutions; some amount of uncertainty may be acceptable for other system owners, but not for NSS. While it has great theoretical interest, and has been the subject of many widely publicized demonstrations, it suffers from limitations and implementation challenges that make it impractical for use in NSS operational networks.

3.4. Post-Quantum Cryptography Tools and Technology

Numerous open-source projects and libraries implement quantum-safe schemes with free public access to the source code and packages. This subsection provides a walkthrough of such existing tools and libraries which may benefit future research in this domain.

3.4.1. Codecrypt

Codecrypt is the first free-of-cost software package available for quantum-safe encryption and digital signatures. Codecrypt [68] was originally written by Miroslav Kratochvíl in 2013–2017, and is available in the Github repository (https://github.com/exaexa/codecrypt, accessed on 18 November 2020). The C++ language has been chosen for the development of this package. It uses McEliece Cryptosystem with IND CCA2 secure QD—MDPC variant for encryption and Hash-based Merkle Tree Algorithm (FMTSeq variant) for digital signatures. It has a command-line interface which provides multiple commands under the following categories:
  • Encrypt, decrypt, sign and verify data
  • Key management operations
  • Input and output from standard I/O, as well as files
  • View options including help and ASCII formatting
The cryptographic primitives of Codecrypt are similar to GnuPG software. The advantage of Codecrypt is that the complexity of all these quantum-safe cryptographic operations was scaled down to linear complexity from the respective exponential complexities in GnuPG. The speed of operations has increased significantly in Codecrypt due to this factor. The disadvantage of this software is that, although the signature schemes perform better than GnuPG, the encryption timing seems to be varying when the size of the message increases, except for the key generation part.

3.4.2. Open Quantum Safe

Open Quantum Safe (OQS) is a project for implementing and evaluating various quantum-resistant cryptographic primitives. It is created by a team headed by researchers Douglas Stebila and Michele Mosca [69]. The entire project is bifurcated into an open-source C library named ‘liboqs’ and a prototype integration suite of quantum-resistant algorithms in existing protocols and applications. The ‘liboqs’ library (https://github.com/open-quantum-safe/liboqs, accessed on 18 November 2020) provides the open-source implementation of various post-quantum key encapsulation mechanisms (KEM) and digital signature algorithms through a common API which can build on Windows, Linux and macOS. These can be executed overwork multiple platforms, including Intel x86, AMD, and ARM. The latest version of this library is ‘liboqs 0.6.0’ updated in June 2021, which is available in the Github repository, and is open to contributions. It includes all algorithms which made their entry to NIST Round 3 finalists or alternate candidates, except GeMSS. ‘liboqs’ can be integrated into different languages using language wrappers available with the package.
Prototype integration is achieved by modifying original protocols and applications to provide post-quantum primitives. The developers forked OpenSSL and BoringSSL to integrate liboqs supporting quantum-safe hybrid key exchange and authentication (KEA) and cipher suites in TLS protocol. It also supports post-quantum algorithms in X.509 certificate generation and S/MIME/CMS message handling. Test servers are provided to test the interoperability of these prototypes. OpenSSH is also modified into OQS-OpenSSH for implementing KEA. New algorithms can be added to this suite at any time through the options available in the interface.

3.4.3. jLBC

The Java Lattice-based Cryptography Library (jLBC) (http://gas.dia.unisa.it/projects/jlbc/buildHowto.html, accessed on 17 November 2020) is a free Java library created by Angelo De Caro (https://github.com/adecaro/jlbc, accessed on 17 November 2020), a researcher at IBM Research Zurich. It is released under GNU public license. It has several modules providing API for the implementation of some lattice-based cryptosystems. ‘jlbc-crypto’ is the main supporting library implementing such algorithms. It utilizes the Bouncy Castle Cryptographic API framework (https://www.bouncycastle.org/, accessed on 18 November 2020), which acts as an intermediary providing Java Cryptography Extension (JCE) and Java Cryptography Architecture (JCA) (https://docs.oracle.com/javase/7/docs/technotes/guides/security/crypto/CryptoSpec.html, accessed on 18 November 2020) services. JCA provides a platform-independent, interoperable, and extensible architecture, and a set of APIs for implementing various cryptographic primitives. JCE supports advanced cryptographic operations. It also contains generators/processors for integration into other protocols and applications as well. All classes belonging to JCA and JCE are called engines. JCA engines are located in the ‘java.security’ package, whereas the JCE classes are located in the ‘javax.crypto’ package. The implementations of cryptographic algorithms are done via provider classes for these libraries. jLBC also implements few Homomorphic Encryption Schemes with the help of the Bouncy Castle framework (https://www.bouncycastle.org/java.html, accessed on 18 November 2020).

3.4.4. Microsoft’s Lattice Cryptography Library

Microsoft product LatticeCrypto (https://www.microsoft.com/en-us/research/project/lattice-cryptography-library/, accessed on 20 November 2020) is a post-quantum cryptographic software that concentrates on lattice-based algorithms. It is created by a team of researchers including Michael Naehrig and Patrick Longa as principal contributors. The software is portable across a variety of platforms and is interoperable. It implements lattice-based cryptographic primitives for key exchange and authentication KEA built on R-LWE for its security. The ring-based Learning With Errors (R-LWE) Problem is unbreakable with quantum computers, and thus the cryptographic schemes in this software are supposed to be efficient against resisting quantum attacks. The implementation includes novel techniques for computing the number theoretic transform to achieve higher performance. The library successfully resists all timing and cache attacks reported so far. It supports Windows and Linux, and can be used on a wide range of platforms, including x86, x64, and ARM, and provides at least 128 bits of classical and quantum security.

3.4.5. libPQP

libPQP (https://github.com/grocid/libPQP/blob/869dfbf86b8fe3a56ba3dbddc86f1291ee0263d1/README.md, accessed on 20 November 2020) is a Python post-quantum library that stands for post-quantum PGP. The source code of this library has not yet been audited and can be used only for testing purposes, as declared by the developer. It concentrates on improving the efficiency of the QC- MDPC variant of the McEliece cryptosystem and has several vulnerabilities. The final product was supposed to use Salsa-20 and Poly 1305 cryptographic primitives. The current version of the library was deprecated due to a key recovery attack10 on QC- MDPC variant published in 2016.
Besides these, there are other packages like ISARA Radiate Quantum-Safe Library that give a free sample code license, Palisade Homomorphic Encryption Library (https://palisade-crypto.org/, accessed on 20 November 2020) which is an open-source lattice crypto-library, and Lattigo, which is a Go module that implements ring-learning-with-errors-based homomorphic-encryption primitives and multiparty-homomorphic-encryption-based secure protocols.
The implementation of post-quantum cryptographic schemes on the Android platform is highly necessary, as 90% of the mobile devices work on it. The liboqs, jLBC, and LatticeCrypto are all capable of adapting the algorithms to the ARM platform, and are efficient candidates for developing quantum-safe schemes for Android devices. ‘liboqs’ library is mostly used by developers for building such applications. The source code can be compiled and integrated into several applications via cross-platform compilers and APIs. A licensed library of quantum-safe cryptographic schemes has been created by O.S.Estrada and team14, which is a ready-to-use package for quantum-safe development in an Android environment. Moreover, numerous independent researchers are working to integrate quantum safety on android devices.

4. Code-Based Cryptography

Linear codes [19] are originally used for digital communication and are based on coding theory. Coding theory is an important study that attempts to minimize data loss due to errors introduced in transmission from noise, interference, or other forces. Data to be transmitted are encoded by the sender as linear codes which are decoded by the receiver. Data encoding is accomplished by adding additional information to each transmitted message, to enable the message to be decoded even if errors occur.
Different codes are being studied to provide solutions to various problems occurring in applications. The most prominent type of error-correcting code is called linear code. The linear codes can be represented by k x n matrices, where k is the length of the original messages, and n is the length of the encoded message. It is computationally difficult to decode messages without knowing the underlying linear code. This hardness underpins the security of the code-based cryptosystem, which includes all cryptosystems, symmetric or asymmetric, whose security relies, partially or totally, on the hardness of decoding in a linear error-correcting code, possibly chosen with some particular structure or in a specific family of linear codes.
Linear codes [70,71] are linear block codes over an alphabet A = F q , where F q denotes the finite field with q = p l elements l N x , p prime. The alphabet is often assumed to be binary, that is p = 2 , l = 1 , q = 2 , F 2 = { 0 , 1 } . The encoding of the source bits is done in blocks of predefined length k, giving rise to the name “block code”.
The following are the matrices used in code-based cryptography.
  • A generator matrix G of an [ n , k ] code C is a k × n matrix G such that C = { x G : x F 2 k } . Generator matrix is of the form ( I k Q ) , where I k is the ( k × k ) identity matrix and Q is a k × ( n - k ) matrix (redundant part).
  • A parity-check matrix H of an [ n , k ] code C is an ( n k ) × n matrix H, such that C = { c : F 2 n : H c T = 0 } .
  • Parity-check matrix H is generated from the generator matrix as H = ( Q T I n k ) .
Encoding process applies an injective F 2 –linear function f c : F 2 k F 2 n on an input block of length k, i.e., every codeword can be generated by multiplying a source vector x F 2 k with G C = x · G x F 2 k F 2 n . Hence, the matrix G corresponds to a map F 2 k F 2 n mapping a message of length k to an n-bit string. This encoding process corresponds to encryption in code-based cryptography.
The decoding process is about finding the closest codeword x C to a given y F 2 n , assuming that there is a unique closest codeword. Decoding a generic binary code of length n and without knowing anything about its structure requires about 2 ( 0.5 + o ( 1 ) ) × n l o g 2 ( n ) binary operations, assuming a rate 1 / 2 . The following are the common decoding techniques [71]:
  • List Decoding—Given C and x, outputs the list L c ( x , t ) : = { c C d ( x , c ) t } , of all codewords at distance at most t to the vector x with decoding radius t.
  • Minimum Distance Decoding—Minimum distance decoding (MDD) is also known as nearest neighbour decoding, and tries to minimize the Hamming distance d ( x ; y ) for all codewords y C given a received F 2 n .
  • Maximum Likelihood Decoding—Given a received codeword x F 2 n , maximum likelihood decoding (MLD) tries to find the codeword y C to maximize the probability that x was received, given that y was sent.
  • Syndrome Decoding—For an [ n ; k ; d ] code C, we can assume that the parity-check matrix H is given. Syndrome S = [ Y ] [ H T ] , Y is received code and x = e + Y where e is the error bit.
The decoding process corresponds to the decryption in code-based cryptography.

4.1. Different Types of Error-Correcting Codes

Error-correcting codes could be broadly classified into block codes and convolutional codes [71].
In block codes, the input is divided into blocks of k digits. The coder then produces a block of n digits for transmission, and the code is described as an ( n , k ) code, as depicted in Figure 1. Linear block codes and non-linear block codes are types of block codes.
In convolutions coding, the coder input and output are continuous streams of digits. The coder outputs n output digits for every k digits input, and the code is described as a rate k / n code.

4.2. Operations on Codes

In many applications, the allowed length of the error control code is determined by system constraints unrelated to error control. When the length of the code that one wishes to use is unsuitable, the code’s length can be modified by the following operations on codes viz. (i) puncturing, (ii) extending, (iii) shortening, (iv) lengthening, (v) expurgating, or (vi) augmenting, which can be carried out in the following ways:
  • An ( n , k ) code is punctured by deleting any of its parity bits to become a ( n 1 , k ) code.
  • An ( n , k ) code is extended by adding an additional parity bit to become a ( n + 1 , k ) code.
  • An ( n , k ) code is shortened by deleting any of its information bits to become a ( n 1 , k 1 ) code.
  • An ( n , k ) code is lengthened by adding an additional information bit to become a ( n + 1 , k + 1 ) code.
  • An ( n , k ) code is expurgated by deleting some of its codewords. If half of the codewords are deleted such that the remainder form a linear subcode, then the code becomes a ( n , k 1 ) code.
  • An ( n , k ) code is augmented by adding new codewords. If the number of codewords added is 2k such that the resulting code is linear, then the code becomes a ( n , k + 1 ) code.

4.3. Properties to Be Fulfilled by Linear Codes

The properties to be fulfilled by linear codes depend on the code metric based upon which they are designed. Hamming metric, Rank metric and Lee metric codes are the most widely used codes. The description below provides an overview of each of these metrics.

4.3.1. Hamming Metric

The Hamming metric is based on the Hamming Distance which is described below [72].
  • The Hamming Distance between two linear codes in F 2 n is the number of coordinates where they differ.
  • The Hamming Weight of a linear code is the number of non-zero coordinates.
  • The Minimum Distance d m i n of a linear code C is the smallest Hamming weight of a nonzero codeword in C.
  • A code is called maximum distance separable (MDS) code when its d m i n is equal to n k + 1 .
There exist many bounds for linear codes, which are mentioned below. Plotkin bound and the Hamming bound are upper bounds on d m i n for a given fixed value of n and k. The Hamming bound is a tighter bound for high rate codes, but the Plotkin bound is for low rate codes.
  • Singleton bound [70,72]: The minimum distance d m i n for an ( n , k ) Binary Linear Block Code is bounded by d m i n ( n k + 1 ) .
  • Plotkin bound [72]: For any ( n , k ) binary linear block code, d m i n n × 2 k 1 ( 2 k 1 ) , the minimum distance of a code cannot exceed the average weight of all nonzero codewords.
  • Gilbert Varshamov Bound [72]: For a fixed value of n and k, Gilbert–Varshamov Bound gives a lower bound on d m i n . According to this bound, if j = 0 ( d m i n 2 ) n 1 j 2 n k then there exists an ( n , k ) binary linear block code whose minimum distance is outlast d m i n .
  • A Hamming sphere of radius t contains all possible received vectors that are at a Hamming distance less than t from a code word. The size of a Hamming sphere for an ( n , k ) Binary Linear Block Code is, V ( n , t ) , where V ( n , t ) = j = 0 t n j .
  • The Hamming bound: A t-error correcting ( n , k ) Binary Linear Block Code must have redundancy n k such that ( n k ) log 2 V ( n , t ) . An ( n , k ) Binary Linear Block Code which satisfies the Hamming bound is called a perfect code.
The Hamming metric codes which were used in code-based cryptography include [19] Reed–Muller Codes, Generalized Reed–Solomon Codes, Binary Goppa Codes, BCH Codes, Generalized Shrivastava Codes. More recently, to reduce the key size, some new kinds of codes with additional structures viz. Quasi-Cyclic Medium Density Parity Check Codes, Quasi-Cyclic Low Density Parity Check Codes, Quasi-Dyadic Generalized Shrivastava codes, Quasi-Cyclic Goppa Codes have been used for code-based cryptography.

4.3.2. Rank Metric

In coding theory [73,74], the most common metric is the Hamming metric, where the distance between two codewords is given by the number of positions in which they differ. An alternative to this is the rank metric. Let X n be an n-dimensional vector space over finite field G F ( q N ) , where q is a power of prime and N is a positive integer. Let ( u 1 , u 2 , u 3 , u n ) with u i G F ( q N ) be a base of G F ( q N ) as a vector space over the field G F ( q ) . Every element x i G F ( q N ) can be represented as x i = a 1 i u 1 + a 2 i u 2 + a N i u N . Hence, every vector x = ( x 1 , x 2 , x 3 , x n ) over the field G F ( q N ) can be written as a matrix.
Every vector x over the field G F ( q N ) is a rank of the corresponding matrix A ( x ) over the field G F ( q ) denoted by r ( x ; q ) · ( x , y ) = r ( x y ; q ) . A set ( x 1 , x 2 , x 3 , x n ) of vectors from X n is called a code with distance d = m i n ( x i , x j ) . If the set also forms a k-dimensional subspace of X n , then it is called a linear ( n , k ) code with distance d. Such a linear rank metric code always satisfies the Singleton bound d n k + 1 . Such codes are called maximum rank distance (MRD) codes.
x = a 1 , 1 a 1 , 2 a 1 , n a 2 , 1 a 2 , 1 a 2 , n a N , 1 a N , 1 a N , n
Non-binary linear codes have been defined using rank-metric. Such codes are called rank codes or Gabidulin codes. The rank metric defines a systematic way of building codes that could detect and correct multiple random rank errors. By adding redundancy with coding k-symbol word to an-symbol word, a rank code can correct any errors of rank up to t = [ ( d 1 ) / 2 ] , where d is a code distance. Some rank metric codes using code-based cryptography are [19] Rank Gabidulin Codes, Low Rank Parity Check Codes, Rank Quasi-Cyclic Codes, Ideal LRPC Codes.

4.3.3. Lee Metric

Lee metric was introduced by Lee as an alternative to Hamming metric for certain noisy channels [75]. Let Z = q { 0 , 1 , q 1 } be the set of representatives of the integer equivalence classes modulo q. The Lee weight of any element a Z q is given by W L ( a ) = m i n { a , q a } . Given an element c = ( c 1 , c 2 , c 3 c n ) Z q , the Lee weight of c denoted as W L ( c ) is given by
W L c = i = 1 n min a , q a
For c , e Z q n , the Lee distance d L ( c , e ) between c and e is defined to be the Lee weight of their difference
d L c , e = W L c e = i = 1 n min c i e i mod q , e i c i mod q
The codes defined using Lee metric are called Lee codes. A Lee code will be specified by ( n , k , d L ) q . If q = 2 or 3, then the Hamming Metric and Lee Metric are identical.The Preparata Code and Kerdock Code are examples of Lee Codes. However, they have not been used in code-based cryptography.

4.4. Relationship between Codes

The relationship between codes has been abstractly specified in [19], as shown in Figure 2.
Based on our study of the various linear codes, the following relationship between codes has been identified as part of this research work, which is depicted in Figure 3. The relationship between the special codes is depicted in Figure 4.

4.5. Common Code-Based Cryptographic Algorithms

Research in code-based cryptography provides for public-key encryption and digital signature algorithms. An overview of the same is provided in subsequent sections.

4.5.1. Code-Based Encryption

McEliece and Niederreiter are two of the earliest cryptographic algorithms developed in code-based cryptography. The McEliece was initially built in 1978 using the binary goppa code with [n,k] = [1024, 524] [76]. Subsequently, several variants of McEliece were built using different linear codes [26]. However, those variants were proven to be susceptible to attacks [19,77], and only the McEliece built using the Binary Goppa Code is found to be quantum attack resistant to date. Thus, it has also been chosen for the third round of standardization by NIST [24]. The McEliece has quadratic complexity in block length, and no polynomial-time quantum algorithm is found to decode the general linear block code [78]. Moreover, the algorithm incorporates an element of randomness in every encryption by the use of e, which is a randomly generated error vector [79]. These are the advantages of McEliece [79]. The large key size is the limitation of McEliece [78,79].
The Niederreiter cryptographic algorithm [80] was developed in 1986, and is very similar to McEliece, and encrypts messages using parity check matrices, unlike generator matrices used in McEliece, and uses the Generalized Reed–Solomon Codes. Table 5 shows an indicative comparison of the two cryptosystems.

4.5.2. Code-Based Signature Schemes

Signature schemes based on linear codes have been developed based on the FDH-like (full domain hash) approach by Courtois–Finiasz–Sendrier (called CFS) [84], and uses Goppa Codes. The modified CFS signature—mCFS—was developed by Dallot [85]. Signature schemes based on Fiat–Shamir Transformation on zero-knowledge identification schemes have been developed by Stern et al. [86], Jain et al. [87], and Cayrel et al. [88]. However, none of the code-based signature schemes have been shortlisted by NIST for the third round of standardization. A comparison of the latter three signature schemes is provided in Table 6 [89].
In addition to these works, a signature scheme with the name RankSign in the rank metric setting was proposed [90]. The security of RankSign builds on the assumption that the special codes are indistinguishable from random linear rank metric codes. However, this scheme was attacked with a structural key-recovery attack in 2018 [91]. The Random Code-based Signature Scheme (RaCoSS) was submitted to NIST [92]. The RaCoSS scheme builds upon random linear codes. This scheme was attacked [93] two days after submission. Subsequently, the scheme was patched [94] and attacked again in the same year—2018 [95]. The main issue of the schemes [92,94] is that the weight of valid signatures is large.
In the same year of 2018, Persichetti [96] conformed the Lyubashevsky scheme to random quasi-cyclic Hamming metric codes. Persichetti’s proposal does not suffer from the weakness of RaCoSS, because the weight of a valid signature is below the GV bound of the code [97]. This scheme [96] was attacked in the two subsequent independent works [98,99].
In 2019, Anguil et al. [100] proposed a signature scheme with the name Durandal in the rank metric context. The security of Durandal builds on a complicated new problem—PSSI+. It is not proven that the signature distribution is independent of the secret key.
The Wave [101] signature scheme was proposed in 2019, which follows the hash-and-sign framework. The security of Wave builds on the new assumption that generalized ( U , U + V ) codes are independent of random linear codes. The hardness of distinguishing generalized ( U , U + V ) codes from random linear codes is still unclear [97]. The output signature is proven to be independent of the secret key. Different from traditional code-based cryptography, the Wave signature chooses the desired weight of the underlying syndrome decoding problem to lie at the large end [97].
In summary, the existing works on secure code-based signature schemes build the security on immature intractability assumptions.

4.6. Attacks in Code-Based Cryptography

The attacks on code-based cryptography can be classified into critical [102] and non-critical attacks (Chapter 4 in [103]).
  • Critical Attacks:
The following are the different types of critical attacks possible in code-based cryptography. Critical attacks aim to recover the key or cryptanalysis of the ciphertexts to decrypt the ciphertext.
1.
Broadcast attack: This attack [104] aims to recover a single message sent to several recipients. Here, the cryptanalyst knows only several ciphertexts of the same message. Since the same message is encrypted with several public keys, it was found that it is possible to recover the message. This attack has been used to break the Niederreiter and HyMES (Hybrid McEliece Encryption Scheme) cryptosystems.
2.
Known partial plaintext attack: A known partial plaintext attack [105] is an attack for which only a part of the plaintext is known.
3.
Sidelnikov–Shestakov attack: This attack [106] aims to recover an alternative private key from the public key.
4.
Generalised known partial plaintext attack: This attack [102] allows to recover the plaintext by knowing a bit’s positions of the original message.
5.
Message-resend attack: A message-resend condition [102] is given if the same message is encrypted and sent twice (or several times) with two different random error vectors to the same recipient.
6.
Related-message: In a related-message attack against a cryptosystem, the attacker obtains several ciphertexts such that there exists a known relation between the corresponding plaintexts.
7.
Chosen plaintext (CPA): A chosen-plaintext attack [102] is an attack model for cryptanalysis which presumes that the attacker can choose arbitrary plaintexts to be encrypted and obtain the corresponding ciphertexts. The goal of the attack is to gain some further information that reduces the security of the encryption scheme. In the worst case, a chosen-plaintext attack could reveal the scheme’s secret key. For some chosen-plaintext attacks, only a small part of the plaintext needs to be chosen by the attacker: such attacks are known as plaintext injection attacks. Two forms of chosen-plaintext attack can be distinguished: Batch chosen-plaintext attack, where the cryptanalyst chooses all plaintexts before any of them are encrypted.
Adaptive chosen plaintext attack, where the cryptanalyst makes a series of interactive queries, choosing subsequent plaintexts based on the information from the previous encryptions.
8.
Chosen-ciphertext attack (CCA) In a chosen-ciphertext attack [102], an attacker has access to a decryption oracle that allows decrypting any chosen ciphertext (except the one that the attacker attempts to reveal). In the general setting, the attacker has to choose all cipher texts in advance before querying the oracle.
In the adaptive chosen-ciphertext attack, formalized by Rackoff and Simon (1991), one can adapt this selection depending on the interaction with the oracle. An especially noted variant of the chosen-ciphertext attack is the lunchtime, midnight, or indifferent attack, in which an attacker may make adaptive chosen-ciphertext queries, but only up until a certain point, after which the attacker must demonstrate some improved ability to attack the system.
9.
Reaction attack: This attack [107] can be considered as a weaker version of the chosen-ciphertext attack. Here, instead of receiving the decrypted ciphertexts from the oracle, the attacker only observes the reaction of this one. Usually, this means whether the oracle was able to decrypt the ciphertext.
10.
Malleability attack: A cryptosystem is vulnerable to a malleability attack [107] of its ciphertexts if an attacker can create new valid ciphertexts from a given one, and if the new ciphertexts decrypt to a clear text which is related to the original message.
  • Non-Critical Attacks:
These kinds of attacks attempt to decode a ciphertext by solving either the general decoding problem or the syndrome decoding problem, and do not focus on recovering the secret key.
1.
Information Set Decoding (ISD) Attack: ISD algorithms [103]) are the most efficient attacks against the code-based cryptosystems. They attempt to solve the general decoding problem. That is, if m is a plaintext and c = m G + e is a ciphertext, where e is a vector of weight t and G a generator matrix, then ISD algorithms take c as input and recover m (or, equivalently, e).
2.
Generalized Birthday Algorithm (GBA): The GBA algorithm [108] is named after the famous birthday paradox which allows one to quickly find common entries in lists. This algorithm tries to solve the syndrome decoding problem. That is, for given parity check matrix H, syndrome s, and integer t, it tries to find a vector e of weight t such that HeT = s T. In other words, it tries to find a set of columns of H whose weighted sum equals the given syndrome.
3.
Support Splitting Algorithm (SSA): The SSA algorithm [109] decides the question of whether two given codes are permutation equivalent, i.e., one can be obtained from the other by permuting the coordinates. For this, SSA makes use of invariants and signatures (not digital signatures). An invariant is a property of a code that is invariant under permutation, while a signature is a local property of a code and one of its coordinates. The difficulty of using invariants and signatures to decide whether two codes are permutation equivalent is that most invariants are either too coarse (i.e., they take the same value for too many codes which are not permutation equivalent) or the complexity to compute them is very high. The SSA solves this issue by starting with a coarse signature and adaptively refines it in every iteration.

4.7. Related Work

This section provides an overview of the various surveys that have been conducted in CBC. This has been provided to substantiate the fact that the prospective research dimensions to be explored that have been delineated by us in the subsequent section have not been in the focus of any of the existing works in CBC. Towards this, we reviewed the surveys from 2008 till now. In 2008, Overbeck and Sendrier [110] published a comprehensive state-of-the-art of code-based cryptography. In this work, the authors illustrate the theory and the practice of code-based cryptographic systems. In 2011, Cayrel et al. [25] presented a survey paper which has included state-of-the-art of publications since 2008 in code-based cryptography, including encryption and identification schemes, digital signatures, secret-key cryptography, and cryptanalysis. This work provides a comprehensive study and an extension of the chapter “Code-based cryptography” of the book [111].
In the same year, 2014, Repka and Cayrel [112] proposed a survey paper on cryptography based on error-correcting codes. This paper surveys the code-based cryptography existing schemes, as well as implementations and side-channel attacks. This work also recalls briefly the basic ideas and provides a roadmap to readers.
In 2015, PQCRYPTO project Horizon 2020 ICT-645622 [113] published a report on post-quantum cryptography for long-term security which provides the PQCRYPTO project’s intermediate report on optimized software. This report also provides the preliminary software implementation results of selected post-quantum schemes and the corresponding parameters for embedded systems. This report surveys modern post-quantum schemes regarding their implementation on such small embedded microcontrollers. This report reviews the most popular schemes in post-quantum cryptography such as encryption and digital signatures schemes.
In 2017, Sendrier [26] published a survey paper that focuses on the McEliece public-key encryption scheme and its variants which are the candidates of post-quantum public-key encryption standard. This paper also focuses on other cryptographic primitives using codes such as zero-knowledge authentication and digital signature.
In parallel to this work in 2017, Bucerzan et al. [114] analyzed the evolution of the main encryption variants coming from code-based cryptography field. The authors focus on security issues and Rank-based cryptography. This paper provides the details and survey on the McEliece encryption scheme. In addition to this, these papers detail the main security threats for the scheme and for each of the mentioned variants.
The most recent survey in code-based cryptography was published in 2018 [115]. In this paper, the authors survey code-based cryptography, essentially for encryption and signature schemes. The authors also provide the main ideas for theoretical and physical cryptanalysis.
According to the best of our knowledge, these are the surveys available in the direction of code-based cryptography research. A review of these surveys helped to understand that the existing works in code-based cryptography have mostly identified the following as the prospective research dimensions that could be explored in CBC research:
  • reduction of key size—large key size is one of the important limitations of CBC and reducing the key size is an important research direction explored
  • use of new kinds of linear and non-linear codes in CBC, viz. QC-MDPC, QC-LDPC, etc.—recently CBC using these kinds of codes have been proposed to overcome various kinds of attacks
  • algorithms for resolving new kinds of security attacks—there are various security attacks possible in CBC and various techniques and algorithms to counteract the same have been proposed
  • evolving new signature schemes—signature schemes using CBC were a recent addition to CBC research
The above research directions proposed in the existing works help to emphasize the fact that the ones which have been provided in this paper are novel and yet to receive the interest of the CBC research community.

5. Research Directions Identified in Code-Based Cryptography

In this section, we lay out some of the research directions which have been least explored and remain as white spaces in the code-based cryptographic research. Though this paper elaborates on both PQC and code-based cryptography, the future research directions confine only to code-based cryptography for two reasons (i) future research direction in PQC ultimately boils down to any one of the PQC schemes viz. code-based, lattice-based, etc and (ii) our current research directions centers around code-based cryptography.

5.1. Dynamic Code-Based Cryptographic Algorithms

The linear codes are many in number and various code-based cryptographic algorithms using these code variants have been proposed. However, these cryptosystems except the McEleice cryptosystem which uses binary Goppa codes have been reported to be broken, discouraging the use of other linear codes. Even The variants of the McEliece algorithm using the other different types of linear codes apart from binary Goppa codes are susceptible to attacks. This is because the static code used in the algorithm is known earlier and also it results in a known structure of the linear code which could be cryptanalysed easily. However, the study of linear codes and the relationships between them, as described above explicate that it is possible to transform one code to another utilizing some operations on codes viz. augmenting, puncturing, extending, … etc as mentioned in Section 4.2. The existing variants of code-based cryptographic algorithms, for example, the McEliece uses a single code (binary goppa) as the basis for the encryption algorithm. Since, it is possible to transform one linear code to another using the possible code transformation operations, the same could be exploited in the encryption. Thereby, the cryptographic algorithm can dynamically choose to use any type of linear code to perform the encryption operation. This dynamic code transformation may produce any other existing linear codes or a new code that fulfills the properties for linear codes viz. Gilbert Varshmov bound, Singleton bound, etc. – for example, from the alternant code one can transform to the Generalized Shrivastava code or a new code fulfilling the linear code bounds. This dynamic approach provides two-fold advantages viz. i) the cryptographic algorithm can dynamically choose to use a particular type of linear code randomly with every session or even in between sessions so that it becomes very difficult to break the cipher since the structure of the linear code keeps varying ii)renders the otherwise unsafe linear codes to provide for quantum attack resistance thereby augmenting the utility of the various types of linear codes available in code-based cryptographic algorithms.
All aspects of this proposal including the study of various relationships between codes, (in this paper we have provided the relationship between only a subset of linear codes) the transformations that can help a linear code to transform into another code dynamically, the design and engineering of such cryptographic algorithms efficiently are the research directions intended to be explored here.

5.2. Use of Other Types of Codes in Code-Based Cryptography

Codes have existed for a long in the computing domain, and there are a variety of codes available, which have been used for a variety of encoding purposes. The following is a representative list of encoding purposes commonly encountered in the computing domain [116]. (i) To encode data for digital data communication (ii) To encode data for digital data communication with error correction capabilities (iii) To encode data in a compressed format for faster message communication (iv) To represent data in a digital system (v) To store and manipulate data in a digital system (vi) Programmatic representation of Character set (vii) To communicate digital data confidentially (viii) To represent data or data set features to be used in machine learning (ix)To represent data in a format comprehensible for visually challenged persons.
In each of the above said seven purposes, the encoding achieves either one of the following encoding (i) Alphabets/character set encoded to another alphabet (ii) Alphabets/character set encoded to a sequence of alphabets (iii) character set encoded to a number (iv) Alphabets/character set encoded to binary/BCD/Hexadecimal/Octal through ordinal encoding, (v) Alphabets/character set encoded to image(s)/symbol(s)/pattern(s) (vi) Alphabets/character set encoded to a compressed code (vii) Binary data encoded to linear codes (viii) Data encoded to a vector (ix) Data encoded to a non-numerical label.
As is evident from above, the said encoding types work at multiple levels of abstractions viz. The encoding may work to encode alphabets or a complete character set, or it may encode binary data, or it may encode a data unit.
Since we consider encoding from a cryptographic perspective, the following requirements are to be fulfilled by the code/encoding technique to constitute a complete and secure code/encoding technique. These requirements have already been identified in our earlier work in [117].
Support for encoding of complete character set—Some coding techniques provide code for encoding only the alphabets like the Caesar cipher, A1Z26, Atbash codes, etc. However, since a plain text message may be alphanumerically comprising of alphabets, numbers, and special characters, the coding technique must help to encode the complete character set.
Possible for representation, manipulation, and storage in digital systems—Some encoding techniques result in the production of codes which are in the form of an image(s) or symbol(s) or pattern(s) for the alphabet or character set that is encoded. However, these image(s) or symbol(s) or pattern(s) cannot be directly represented, manipulated, and stored in digital systems. Dorabella cipher, Morse codes, Dice Cipher, Rosicrucian Cipher are some examples of such coding techniques which produce symbols or patterns as codes that are not suitable for direct representation, manipulation, and storage in digital systems.
Enables Data Hiding—Since the codes are approached from the perspective of cryptography, enabling data hiding as a result of encoding is one of the important properties expected to be fulfilled by the code and hence the encoding technique.
Support for multiple data type—Since, the plain text may not only be plain text and also be of any data type viz., audio, video, etc. it is essential that the encoding technique can encode any data type.
Dynamicity of encoding—Usually, the encoding techniques provide for static encoding whereby a particular element of a character set or data is always encoded to the same code any time encoding happens. This may enable ease of encoding – but approached from a cryptographic perspective, a dynamic encoding scheme which produces a different code for an element of the character set or data for different data communications sessions should help augment the strength and effectiveness of the encoding.
Variable Encoding—Even in a single session of communication, the plain text to be encoded should be subject to the production of varied codes unlike the Caesar cipher codes, Atbash Cipher codes which produce the same code for the same character in all of the plain text which leaves a lot of ques for easy decoding.
Hard Decoding—The encoding technique should be able to produce codes which are hard to decode. In other words, the decoding should be a computationally hard problem to solve.
Randomness of Encoding—Usually, an encoding technique comprises steps constituting for the generation of code in some particular static order. If the steps of encoding itself are randomized, then it constitutes a higher strength of the generated code making the decoding a computationally hard problem to solve.
Possibility for Decoding—Though the decoding is required to be a computationally hard problem, it should be possible to decode the code given the key – unlike the hash codes which cannot be decoded since the process of hashing is a one-way function.
Random Character Set and Collating Sequence—In encoding, character set is considered and its collating sequence is usually adopted. To provide for a high degree of randomness and hence to render strength to the code, it should be possible to use alphabets across character sets and specify a user-defined collating sequence of those alphabets chosen.
Table 7 shows the various requirements for data encoding listed above and the types of codes available in each of them and the properties fulfilled by each of them.
From the comparison in Table 7, it is observed that more than the linear codes which are presently used in code-based cryptography, the DNA codes provide promising scope to be used for cryptography. This has been described in detail in [117]. Though research in DNA cryptography is active and the domain has been explored in interesting dimensions, DNA cryptography has not been proven to be quantum attack resistant. If DNA cryptography is proved to be quantum attack resistant, then it provides for a bio-inspired, best value addition to the field of code-based cryptography. This dimension needs to be explored in further detail.
Thus, DNA codes and identifying other kinds of codes amenable for code-based cryptography and designing and engineering code-based cryptographic algorithms using these other kinds of codes efficiently are prospective research directions to be explored.

5.3. Privacy-Preserving Code-Based Cryptography

Privacy-preserving encryption algorithms are the need of the hour owing to the growing privacy issues and concerns globally. Privacy-preserving encryption can be achieved in the following ways:
  • Attribute-based Encryption—Key Policy-based encryption and Cipher Policy-based encryption [118]
  • Homomorphic encryption [119]
Whereas attribute-based encryption provides for selective decryption of ciphertext based on the fulfillment of attributes by the receiver, homomorphic encryption enables one to perform computations on the encrypted data itself, eliminating the requirement for decryption.
In code-based cryptography, only the McEliece cryptosystem has been proven to be somewhat homomorphic [120]. However, attribute-based encryption in code-based cryptography is yet to be explored. Hence, there is a need for a lot of research to enable the maturity of this dimension of code-based cryptography and its practical use in privacy demanding domains such as healthcare, the financial sector, etc. This constitutes another prospective line of research in code-based cryptography.

5.4. Prospective Applicability of Codes with Lattice-Based Cryptography

Codes and Lattices are having similar mathematical properties. A linear code C of length n and dimension k is a k-dimensional subspace of finite field typically endowed with hamming metric.
Given n-linearly independent vectors b 1 , , b n in R n , the lattice generated by them is the set of vectors
L b 1 , , b n = i = 1 n x i b i : x i Z   f o r   1 i n
The vectors b 1 , , b n are known as a basis of the lattice.
Both of them are vector spaces over some finite fields.
1.
Typical lattice-based cryptographic schemes have used q-ary lattices to solve SIS and LWE problems [133]. Linear code of length n and dimension k is a linear subspace which is called a q-ary code. The possibility of using q-ary lattices [134] to implement ternary codes i.e., q-ary codes in code-based cryptographic schemes is an unexplored area. It may be noted here that DNA cryptography is a Quaternary code which has received due exploration from the authors but only needs to be ascertained for its quantum attack resistance.
2.
There is a major lattice algorithmic technique that has no clear counterpart for codes, namely, basis reduction. There seems to be no analogue notions of reduction for codes, or at least they are not explicit nor associated with reduction algorithms. We are also unaware of any study of how such reduced bases would help with decoding tasks. This observation leads to two questions.
  • Is there an algorithmic reduction theory for codes, analogue to one of the lattices?
  • If so, can it be useful for decoding tasks?
These questions are potential leads towards prospective research directions in code-based cryptography.

6. Conclusions

Post-quantum cryptography research has branched out in many dimensions and a considerable research outcome has been emerging in each of these dimensions. While this evinces the maturity of post-quantum cryptography research, each of these outcomes is available in discrete sources hindering the broad spectrum view and comprehension of these outcomes. This paper addresses this limitation, whereby, it provides a one-stop reference of the entire spectrum of post-quantum cryptography research and briefs the research happening in those directions.
Furthermore, from the NIST standardization, it has been observed that though code-based cryptography provides scope to be recognized as a complete cryptosystem with the availability of encryption, key exchange, and digital signature schemes, unlike its post-quantum counterparts which provide for a subset of these. Hence, an overview of the research directions that have been explored in code-based cryptography has been provided and the promising research directions that can augment the prospects of code-based cryptography from the codes perspective have been identified and described. Thereby, this paper provides two solid contributions in the roadmap of post-quantum computing research.

Author Contributions

Idea Conceptualization and Writing, C.B. and K.S.; methodology, C.B. and K.S.; formal analysis, C.B.; investigation, C.B. and K.S.; resources, C.B. and K.S.; writing—original draft preparation, C.B. and K.S.; writing—review and editing, C.B., K.S. and G.G.; supervision, C.B., K.S. and M.R.; project administration, K.S.; funding acquisition, K.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received funding from IRT SystemX.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available in artcile.

Acknowledgments

This research work has been carried out under the leadership of the Institute for Technological Research SystemX, and therefore granted within the scope of the program “Recherche Exploratoire”.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Shor, P.W. Polynomial-Time Algorithms for Prime Factorization and Discrete Logarithms on a Quantum Computer. SIAM J. Comput. 1997, 26, 1484–1509. [Google Scholar] [CrossRef] [Green Version]
  2. Yan, S.Y. Integer factorization and discrete logarithms. Primality Testing and Integer Factorization in Public-Key Cryptography; Springer: Berlin/Heidelberg, Germany, 2004; pp. 139–191. [Google Scholar]
  3. Rivest, R.; Shamir, A.; Adleman, L. A method for obtaining digital for signatures and public-Key cryptosystems. Commun. ACM 1978, 21, 120–126. [Google Scholar] [CrossRef]
  4. Kuwakado, H.; Morii, M. Quantum distinguisher between the 3-round Feistel cipher and the random permutation. In Proceedings of the IEEE International Symposium on Information Theory, Austin, TX, USA, 12–18 June 2010; pp. 2682–2685. [Google Scholar]
  5. McEliece, R.J. A public-key cryptosystem based on algebraic. Coding Thv 1978, 4244, 114–116. [Google Scholar]
  6. Merkle, R. Secrecy, Authentication, and Public Key Systems; Computer Science Series; UMI Research Press: Ann Arbor, MI, USA, 1982. [Google Scholar]
  7. Patarin, J. Hidden fields equations (HFE) and isomorphisms of polynomials (IP): Two new families of asymmetric algorithms. In Proceedings of the International Conference on the Theory and Applications of Cryptographic Techniques, Saragossa, Spain, 12–16 May 1996; pp. 33–48. [Google Scholar]
  8. Hoffstein, J.; Pipher, J.; Silverman, J.H. NTRU: A ring-based public key cryptosystem. In International Algorithmic Number Theory Symposium; Springer: Berlin/Heidelberg, Germany, 1998; pp. 267–288. [Google Scholar]
  9. Regev, O. On lattices, learning with errors, random linear codes, and cryptography. JACM 2009, 56, 34. [Google Scholar] [CrossRef]
  10. Jao, D.; Feo, L.D. Towards Quantum-Resistant Cryptosystems from Supersingular Elliptic Curve Isogenies. PQCrypto 2011, 7071, 19–34. [Google Scholar]
  11. Nielsen, M.A.; Chuang, I. Quantum Computation and Quantum Information; Cambridge University Press: Cambridge, UK, 2002. [Google Scholar]
  12. Benioff, P. The computer as a physical system: A microscopic quantum mechanical Hamiltonian model of computers as represented by Turing machines. J. Stat. Phys. 1980, 22, 563–591. [Google Scholar] [CrossRef]
  13. Manin, Y. Mathematics and Physics; American Mathematical Society: Providence, RI, USA, 1981. [Google Scholar]
  14. Feynman, R.P.U. Simulating physics with computers. Int. J. Theor. Phys. 1982, 21, 467–488. [Google Scholar] [CrossRef]
  15. Deutsch, D. Quantum Theory, the Church-Turing Principle and the Universal Quantum Computer. Proc. R. Soc. Lond. 1985, A400, 97–117. [Google Scholar]
  16. Shor, P.W. Algorithms for quantum computation: Discrete Logarithms and Factoring. In Proceedings of the 35th Annual Symposium on Foundations of Computer Science, Santa Fe, NM, USA, 20–22 November 1994; pp. 124–134. [Google Scholar]
  17. Grover, L.K. A fast quantum mechanical algorithm for database search. In Proceedings of the 28th Annual ACM Symposium on Theory of Computing, Philadephia, PA, USA, 22–24 May 1996; pp. 212–219. [Google Scholar]
  18. Buchmann, J.; Lauter, K.; Mosca, M. Postquantum Cryptography-State of the Art. IEEE Secur. Priv. 2017, 15, 12–13. [Google Scholar] [CrossRef] [Green Version]
  19. Umana, V.G. Post Quantum Cryptography. Ph.D. Thesis, Technical University of Denmark, Lyngby, Denmark, 2011. [Google Scholar]
  20. Beullens, W.; D’Anvers, J.; Hülsing, A.; Lange, T.; Panny, L.; Guilhem, C.d.S.; Smart, N.P. Post-Quantum Cryptography: Current State and Quantum Mitigation; Technical Report; European Union Agency for Cybersecurity: Athens, Greece, 2021. [Google Scholar]
  21. Merkle, R. A certified digital signature. In Advances in Cryptology – CRYPTO’89; Springer: Berlin/Heidelberg, Germany, 1989; pp. 218–238. [Google Scholar]
  22. Butin, D. Hash-based signatures: State of play. IEEE Secur. Priv. 2007, 15, 37–43. [Google Scholar] [CrossRef]
  23. Bernstein, D.J.; Hülsing, A.; Kölbl, S.; Niederhagen, R.; Rijneveld, J.; Schwabe, P. The SPHINCS+ Signature Framework. Available online: http://www.informationweek.com/news/201202317 (accessed on 20 November 2020).
  24. NIST. PQC Standardization Process: Third Round Candidate Announcement. 2020. Available online: https://csrc.nist.gov/News/2020/pqc-third-round-candidate-announcement (accessed on 20 November 2020).
  25. Cayrel, P.L.; ElYousfi, M.; Hoffmann, G.; Meziani, M.; Niebuhr, R. Recent Progress in Code-Based Cryptography. In International Conference on Information Security and Assurance; Springer: Berlin/Heidelberg, Germany, 2011; pp. 21–32. [Google Scholar]
  26. Sendrier, N. Code-Based Cryptography: State of the Art and Perspectives. IEEE Secur. Priv. 2017, 15, 44–50. [Google Scholar] [CrossRef]
  27. Ding, J.; Petzoldt, A. Current state of multivariate cryptography. IEEE Secur. Priv. 2017, 15, 28–36. [Google Scholar] [CrossRef]
  28. Chen, M.; Ding, J.; Kannwischer, M.; Patarin, J.; Petzoldt, A.; Schmidt, D.; Yang, B. Rainbow Signature. Available online: https://www.pqcrainbow.org/ (accessed on 27 August 2020).
  29. Casanova, A.; Faueère, J.C.; Macario-Rat, G.; Patarin, J.; Perret, L.; Ryckeghem, J. GeMSS: A great multivariate short signature. Available online: https://www-polsys.lip6.fr/Links/NIST/GeMSS.html (accessed on 8 December 2020).
  30. Chi, D.P.; Choi, J.W.; Kim, J.S.; Kim, T. Lattice Based Cryptography for Beginners. Available online: https://eprint.iacr.org/2015/938 (accessed on 20 November 2020).
  31. Lepoint, T. Design and Implementation of Lattice-Based Cryptography. Ph.D. Thesis, Ecole Normale Euérieure de Paris—ENS, Paris, France, 2014. [Google Scholar]
  32. Alkim, D.; Ducas, L.; Pöppelmann, T.; Schwabe, P. Post-Quantum Key Exchange—A New Hope. Available online: https://eprint.iacr.org/2015/1092 (accessed on 20 November 2020).
  33. Ducas, L.; Durmus, A.; Lepoint, T.; Lyubashevsky, V. Lattice Signatures and Bimodal Gaussians. Available online: https://eprint.iacr.org/2013/383 (accessed on 20 November 2020).
  34. Bos, J.; Ducas, L.; Kiltz, E.; Lepoint, T.; Lyubashevsky, V.; Schanck, J.M.; Schwabe, P.; Seiler, G.; Stehlé, D. Cryptology ePrint Archive: Report 2017/634. Available online: https://eprint.iacr.org/2017/634/20170627:201157 (accessed on 20 November 2020).
  35. Chen, C.; Danba, O.; Hoffstein, J.; Hülsing, A.; Rijneveld, J.; Saito, T.; Schanck, J.M.; Schwabe, P.; Whyte, W.; Xagawa, K.; et al. NTRU: A Submission to the NIST Post-Quantum Standardization Effort. Available online: https://ntru.org/ (accessed on 12 July 2020).
  36. D’Anvers, J.P.; Karmakar, A.; Roy, S.S.; Vercauteren, F. Saber: Module-LWR Based Key Exchange, CPA-Secure Encryption and CCA-Secure KEM. Available online: https://eprint.iacr.org/2018/230/20181026:121404 (accessed on 20 November 2020).
  37. Bernstein, D.J.; Chuengsatiansup, C.; Lange, T.; Vredendaal, C.V. NTRU Prime: Reducing Attack Surface at Low Cost. Available online: https://eprint.iacr.org/2016/461 (accessed on 20 November 2020).
  38. Ducas, L.; Lepoint, T.; Lyubashevsky, V.; Schwabe, P.; Seiler, G.; Stehle, D. CRYSTALS—Dilithium: Digital Signatures from Module Lattices. Available online: https://eprint.iacr.org/2017/633 (accessed on 20 November 2020).
  39. Fouque, P.A.; Hoffstein, J.; Kirchner, P.; Lyubashevsky, V.; Pornin, T.; Prest, T.; Ricosset, T.; Seiler, G.; Whyte, W.; Zhang, Z. Falcon: Fast-Fourier Lattice-Based Compact Signatures over NTRU. Available online: https://www.di.ens.fr/~prest/Publications/falcon.pdf (accessed on 3 January 2021).
  40. Supersingular Isogeny Diffie–Hellman Key Exchange (SIDH). Available online: https://en.wikipedia.org/wiki/Supersingular_isogeny_key_exchange (accessed on 4 February 2021).
  41. Costello, C.; Longa, P.; Naehrig, M. Efficient algorithms for supersingular isogeny Diffie-Hellman. In Proceedings of the Annual International Cryptology Conference, Santa Barbara, CA, USA, 14–18 August 2016. [Google Scholar]
  42. Valyukh, V. Performance and Compari-Son of Post-Quantum Crypto-Graphic Algorithms. 2017. Available online: http://www.liu.se (accessed on 18 August 2020).
  43. Gaithuru, J.N.; Bakhtiari, M. Insight into the operation of NTRU and a comparative study of NTRU, RSA and ECC public key cryptosystems. In Proceedings of the 2014 8th Malaysian Software Engineering Conference (MySEC), Langkawi, Malaysia, 23–24 September 2014; pp. 273–278. [Google Scholar]
  44. Espitau, T.; Fouque, P.A.; Gérard, B.; Tibouchi, M. Side-channel attacks on BLISS lattice-based signatures. In Proceedings of the ACM SIGSAC Conference, Dallas, TX, USA, 30 October–3 November 2017; pp. 1857–1874. [Google Scholar]
  45. Petzoldt, A.; Bulygin, S.; Buchmann, J. Selecting parameters for the rainbow signature scheme. Lect. Notes Comput. Sci. 2010, 6061, 218–240. [Google Scholar]
  46. Bernstein, D.J.; Hopwood, D.; Hülsing, A.; Lange, T.; Niederhagen, R.; Papachristodoulou, L.; Schneider, M.; Schwabe, P.; Wilcox-O’hearn, Z. SPHINCS: Practical stateless hash-based signatures. Lect. Notes Comput. Sci. 2015, 9056, 368–397. [Google Scholar]
  47. Bernstein, D.J.; Dobraunig, C.; Eichlseder, M.; Fluhrer, S.; Gazdag, S.I.; Kampanakis, P.; Lange, T.; Lauridsen, M.M.; Mendel, F.; Niederhagen, R.; et al. SPHINCS+: Submission to the NIST Post-Quantum Project. 2019. Available online: https://sphincs.org/data/sphincs+-round2-specification.pdf#page=6&zoom=100,0,254 (accessed on 15 July 2020).
  48. Tillich, J. Attacks in Code Based Cryptography: A Survey, New Results and Open Problems. 2018. Available online: www.math.fau.edu/april09.code.based.survey.pdf (accessed on 12 September 2020).
  49. Repka, M. McELIECE PKC calculator. J. Electr. Eng. 2014, 65, 342–348. [Google Scholar]
  50. Post-Quantum Cryptography: A Ten-Year Market and Technology Forecast. 2020. Available online: https://www.researchandmarkets.com/reports/4700915/post-quantum-cryptography-a-ten-year-market-and#relb0-5118342 (accessed on 5 July 2020).
  51. IBM. Post-Quantum Cryptography. Available online: https://researcher.watson.ibm.com/researcher/view_group.php?id=8231 (accessed on 11 November 2020).
  52. Microsoft. Post Quantum Cryptography. Available online: https://www.microsoft.com/en-us/research/project/post-quantum-cryptography/ (accessed on 11 November 2020).
  53. Avaya. Post-Quantum to Team on Identity-as-a-Service. Available online: https://www.avaya.com/en/about-avaya/newsroom/pr-us-1803012c/ (accessed on 28 February 2021).
  54. Envieta. Post Quantum Consulting. Available online: https://envieta.com/post-quantum-consulting (accessed on 28 February 2021).
  55. Google and Cloudflare are Testing Post-Quantum Cryptography. Available online: https://www.revyuh.com/news/hardware-and-gadgets/google-cloudflare-testing-post-quantum-cryptography/ (accessed on 9 February 2021).
  56. Infineon Technologies. Post-Quantum Cryptography: Cybersecurity in Post-Quantum Computer World. Available online: https://www.infineon.com/cms/en/product/promopages/post-quantum-cryptography/ (accessed on 9 February 2021).
  57. Security Innovation Announces Intent to Create OnBoard Security Inc. Available online: https://www.globenewswire.com/news-release/2017/02/14/917023/0/en/Security-Innovation-Announces-Intent-to-Create-OnBoard-Security-Inc.html (accessed on 10 February 2021).
  58. Post-Quantum Cryptography|CSRC. Available online: https://csrc.nist.gov/projects/post-quantum-cryptography/post-quantum-cryptography-standardization (accessed on 7 February 2021).
  59. Alagic, G.; Alperin-Sheriff, J.; Apon, D.; Cooper, D.; Dang, Q.; Kelsey, J.; Liu, Y.; Miller, C.; Moody, D.; Peralta, R.; et al. Status Report on the Second Round of the NIST Post-Quantum Cryptography Standardization Proces. Available online: https://csrc.nist.gov/publications/detail/nistir/8309/final (accessed on 3 February 2021).
  60. Qin, H. Standardization of Quantum Cryptography in ITU-T and ISO/IEC, Qcrypt. 2020. Available online: https://2020.qcrypt.net/slides/Qcrypt2020_ITU_ISO.pdf (accessed on 5 February 2021).
  61. ETSI ICT Standards. Available online: https://www.etsi.org/standards (accessed on 5 February 2021).
  62. Chen, L. Preparation of Standardization of Quantum-Resistant Cryptography in ISO/IEC JTC1 SC27. 2018. Available online: https://docbox.etsi.org/Workshop/2018/201811_ETSI_IQC_QUANTUMSAFE/TECHNICAL_TRACK/01worldtour/NICT_Moriai.pdf (accessed on 5 February 2021).
  63. Shinohara, N.; Moriai, S. Trends in Post-Quantum Cryptography: Cryptosystems for the Quantum Computing Era. 2019. Available online: https://www.ituaj.jp/wp-content/uploads/2019/01/nb31-1_web-05-Special-TrendsPostQuantum.pdf (accessed on 14 July 2020).
  64. Framework to Integrate Post-quantum Key Exchanges into Internet Key Exchange Protocol Version 2 (IKEv2). Available online: https://tools.ietf.org/id/draft-tjhai-ipsecme-hybrid-qske-ikev2-03.html (accessed on 15 October 2020).
  65. Paterson, K. Post-Quantum Crypto Standardisation in IETF/IRTF. Available online: www.isg.rhul.ac.uk/~kp (accessed on 3 February 2021).
  66. libpqcrypto: Intro. Available online: https://libpqcrypto.org/ (accessed on 18 January 2021).
  67. Quantum-Safe Security|Cloud Security Alliance. Available online: https://cloudsecurityalliance.org/research/working-groups/quantum-safe-security/ (accessed on 5 February 2021).
  68. Kratochvíl, M. Implementation of Cryptosystem Based on Error-Correcting Codes; Technical Report; Faculty of Mathematics and Physics, Charles University in Prague: Prague, Czech Republic, 2013. [Google Scholar]
  69. Stebila, D.; Mosca, M. Post-quantum key exchange for the Internet and the Open Quantum Safe project. Sel. Areas Cryptogr. 2016, 10532, 1–24. [Google Scholar]
  70. Minihold, M. Linear Codes and Applications in Cryptography. Master’s Thesis, Vienna University of Technology, Vienna, Austria, 2013. [Google Scholar]
  71. Londahl, C. Some Notes on Code-Based Cryptography. Ph.D. Thesis, Lund University, Lund, Sweden, 2015. [Google Scholar]
  72. Williams, F.M.; Sloane, N. The Theory of Error-Correcting Codes; North Holland Publishing Company: Amsterdam, The Netherlands, 1977. [Google Scholar]
  73. Gadouleau, M.; Yan, Z. Properties of Codes with Rank Metric. In Proceedings of the IEEE Globecom 2006, San Francisco, CA, USA, 27 November–1 December 2006. [Google Scholar] [CrossRef]
  74. Penaz, U.M. On The similarities between Rank and Hamming Weights and their Applications to Network Coding. IEEE Trans. Inf. Theory 2016, 62, 4081–4095. [Google Scholar] [CrossRef] [Green Version]
  75. Tim, L.; Alderson, S.H. On Maximum Lee Distance Codes. J. Discret. Math. 2013, 2013, 625912. [Google Scholar]
  76. Valentijn, A. Goppa Codes and Their Use in the McEliece Cryptosystems. 2015. Available online: https://surface.syr.edu/honors_capstone/845/ (accessed on 20 November 2020).
  77. Maliky, S.A.; Sattar, B.; Abbas, N.A. Multidisciplinary perspectives in cryptology and information security. In Multidisciplinary Perspectives in Cryptology and Information Security; IGI Global: Philadephia, PA, USA, 2014; pp. 1–443. [Google Scholar]
  78. Joachim, R. An Overview to Code Based Cryptography. 2016. Available online: hkumath.hku.hk/~ghan/WAM/Joachim.pdf (accessed on 7 June 2020).
  79. Roering, C. Coding Theory-Based Cryptopraphy: McEliece Cryptosystems in Sage. 2015. Available online: https://digitalcommons.csbsju.edu/honors_theses/17/ (accessed on 20 November 2020).
  80. Niederreiter, H. Knapsack-type Cryptosystems and algebraic coding theory. Probl. Control. Inf. Theory 1986, 15, 157–166. [Google Scholar]
  81. Heyse, S.; von Maurich, I.; Güneysu, T. Smaller Keys for Code-Based Cryptography: QC-MDPC McEliece Implementations on Embedded Devices. In Cryptographic Hardware and Embedded Systems, CHES 2013. Lecture Notes in Computer Science; Bertoni, G., Ed.; Springer: Berlin/Heidelberg, Germany, 2013; Volume 8086, pp. 273–292. [Google Scholar]
  82. von Maurich, I.; Güneysu, T. Lightweight code-based cryptography: QC-MDPC McEliece encryption on reconfigurable devices. In Proceedings of the 2014 Design, Automation Test in Europe Conference Exhibition (DATE), Dresden, Germany, 24–28 March 2014; pp. 1–6. [Google Scholar]
  83. Heyse, S.; Güneysu, T. Code-based cryptography on reconfigurable hardware: Tweaking Niederreiter encryption for performance. J. Cryptogr. Eng. 2013, 3, 29–43. [Google Scholar] [CrossRef]
  84. Courtois, N.T.; Finiasz, M.; Sendrier, N. How to Achieve a McEliece-Based Digital Signature Scheme. In Advances in Cryptology—ASIACRYPT 2001; Springer: Berlin/Heidelberg, Germany, 2001; pp. 157–174. [Google Scholar]
  85. Dallot, L. Towards a Concrete Security Proof of Courtois, Finiasz and Sendrier Signature Scheme. In Research in Cryptology WEWoRC 2007; Springer: Berlin/Heidelberg, Germany, 2008; pp. 65–77. [Google Scholar]
  86. Stern, J. A new identification scheme based on syndrome decoding. In Advances in Cryptology—CRYPTO’93; Springer: Berlin/Heidelberg, Germany, 1994; pp. 13–21. [Google Scholar]
  87. Jain, A.; Krenn, S.; Pietrzak, K.; Tentes, A. Commitments and efficient zero-knowledge proofs from learning parity with noise. In Proceedings of the International Conference on the Theory and Application of Cryptology and Information Security, Beijing, China, 2–6 December 2012; pp. 663–680. [Google Scholar]
  88. Cayrel, P.; Alaoui, S.M.E.Y.; Hoffmann, G.; Véron, P. An improved threshold ring signature scheme based on error correcting codes. In Proceedings of the 4th International Conference on Arithmetic of Finite Fields, Bochum, Germany, 16–19 June 2012; pp. 45–63. [Google Scholar]
  89. Roy, P.S.; Morozov, K.; Fukushima, K. Evaluation of Code-Based Signature Schemes. 2019. Available online: https://eprint.iacr.org/2019/544 (accessed on 21 January 2021).
  90. Gaborit, P.; Ruatta, O.; Schrek, J.; Zémor, G. New Results for Rank-Based Cryptography. In Progress in Cryptology—AFRICACRYPT 2014; Springer: Berlin/Heidelberg, Germany, 2014; pp. 1–12.
  91. Debris-Alazard, T.; Tillich, J. Two attacks on rank metric code based schemes: Ranksign and an IBE scheme. In Advances in Cryptology—ASIACRYPT 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 62–92. [Google Scholar]
  92. Fukushima, K.; Roy, P.S.; Xu, R.; Kiyomoto, K.M.S.; Takagi, T. RaCoSS: Random Code-Based Signature Scheme, 2017. Available online: https://eprint.iacr.org/2018/831 (accessed on 20 November 2020).
  93. Bernstein, D.J.; Hulsing, A.; Lange, T.; Panny, L. Comments on RaCoSS, a Submission to NIST’s PQC Competition, 2017. Available online: https://helaas.org/racoss/ (accessed on 20 November 2020).
  94. Roy, P.S.; Morozov, K.; Fukushima, K.; Kiyomoto, S.; Takagi, T.; Code-Based Signature Scheme Without Trapdoors. Int. Assoc. Cryptologic Res. 2018. Available online: https://eprint.iacr.org/2021/294 (accessed on 20 November 2020).
  95. Xagawa, K. Practical Attack on RaCoSS-R. Int. Assoc. Cryptologic Res. 2018, 831. Available online: https://eprint.iacr.org/2018/831 (accessed on 20 November 2020).
  96. Persichetti, E. Efficient one-time signatures from quasi-cyclic codes: A full treatment. Cryptography 2018, 2, 30. [Google Scholar] [CrossRef] [Green Version]
  97. Li, Z.; Xing, C.; Yeo, S.L. A New Code Based Signature Scheme without Trapdoors. Available online: https://eprint.iacr.org/2020/1250 (accessed on 20 November 2020).
  98. Santini, P.; Baldi, M.; Chiaraluce, F. Cryptanalysis of a one-time code-based digital signature scheme. In Proceedings of the IEEE International Symposium on Information Theory, Paris, France, 7–12 September 2019; pp. 2594–2598. [Google Scholar]
  99. Deneuville, J.; Gaborit, P. Cryptanalysis of a code-based one-time signature. Des. Codes Cryptogr. 2020, 88, 1857–1866. [Google Scholar] [CrossRef]
  100. Aragon, N.; Blazy, O.; Gaborit, P.; Hauteville, A.; Zémor, G. Durandal: A rank metric based signature scheme. In Advances in Cryptology—EUROCRYPT 2019; Springer: Berlin/Heidelberg, Germany, 2019; pp. 728–758. [Google Scholar]
  101. Debris-Alazard, T.; Sendrier, N.; Tillich, J. Wave: A new family of trapdoor one-way preimage sampleable functions based on codes. In Advances in Cryptology—ASIACRYPT 2019; Springer: Berlin/Heidelberg, Germany, 2019; pp. 21–51. [Google Scholar]
  102. Cayrel, P.; Gueye, C.T.; Ndiaye, O.; Niebuhr, R. Critical attacks in Code based cryptography. Int. J. Inf. Coding Theory 2015, 3, 158–176. [Google Scholar] [CrossRef]
  103. Niebuhr, R. Attacking and Defending Code-based Cryptosystems. Ph.D. Thesis, Vom Fachbereich Informatik der Technischen Universität Darmstadt, Darmstadt, Germany, 2012. [Google Scholar]
  104. Niebuhr, R.; Cayrel, P.L. Broadcast attacks against code-based encryption schemes. In Research in Cryptology; Springer: Berlin/Heidelberg, Germany, 2011. [Google Scholar]
  105. Niebuhr, R. Application of Algebraic-Geometric Codes in Cryptography. Technical Report; Fachbereich Mathematik der Technischen Universität Darmstadt: Darmstadt, Germany, 2006. [Google Scholar]
  106. Sidelnikov, V.; Shestakov, S. On the insecurity of cryptosystems based on generalized Reed-Solomon Codes. Discret. Math. 1992, 1, 439–444. [Google Scholar] [CrossRef]
  107. Kobara, K.; Imai, H. Semantically Secure McEliece Public-key Cryptosystems—Conversions for McEliece PKC. In Proceedings of the 4th International Workshop on Practice and Theory in Public Key Cryptosystems, Cheju Island, Korea, 13–15 February 2001; pp. 19–35. [Google Scholar]
  108. Wagner, D. A generalized birthday problem. In Advances in Cryptology—CRYPTO 2002; Springer: Berlin/Heidelberg, Germany, 2002; pp. 288–303. [Google Scholar]
  109. Sendrier, N. Finding the permutation between equivalent linear codes: The support splitting algorithm. IEEE Trans. Inf. Theory 2000, 46, 1193–1203. [Google Scholar] [CrossRef]
  110. Overbeck, R.; Sendrier, N. Code-Based Cryptography; Springer: Berlin/Heidelberg, Germany, 2008; pp. 95–146. [Google Scholar]
  111. Bernstein, D.J.; Buchmann, J.; Dahmen, E. Post-Quantum Cryptography; Springer: Berlin/Heidelberg, Germany, 2008. [Google Scholar]
  112. Repka, M.; Cayrel, P.L. Cryptography Based on Error Correcting Codes: A Survey; IGI Global: Philadephia, PA, USA, 2014. [Google Scholar]
  113. PQCRYPTO. Post-Quantum Cryptography for Long-Term Security; Technical Report; Project number: Horizon 2020 ICT-645622; Technische Universiteit Eindhoven: Eindhoven, The Netherlands, 2015. [Google Scholar]
  114. Bucerzan, D.; Dragoi, V.; Kalachi, H. Evolution of the McEliece Public Key Encryption Scheme. In International Conference for Information Technology and Communications SecITC 2017; Springer: Berlin/Heidelberg, Germany, 2017; pp. 129–149. [Google Scholar]
  115. Drăgoi, V.; Richmond, T.; Bucerzan, D.; Legay, A. Survey on Cryptanalysis of Code-Based Cryptography: From Theoretical to Physical Attacks. In Proceedings of the 7th International Conference on Computers Communications and Control (ICCCC), Oradea, Romania, 11–15 May 2018. [Google Scholar]
  116. Best Codes: 27 Steps—Instructables. Available online: https://www.instructables.com/Best-Codes/ (accessed on 18 November 2020).
  117. Hussain, U.N. A Novel String Matrix Modeling Based DNA Computing Inspired Cryptosystem. Ph.D. Thesis, Pondicherry University, Puducherry, India, 2016. [Google Scholar]
  118. Kumar, P.V.; Aluvalu, R. Key policy attribute based encryption (KP-ABE): A review. Int. J. Innov. Emerg. Res. Eng. 2015, 2, 49–52. [Google Scholar]
  119. Gentry, C. A Fully Homomorphic Encryption Scheme. Ph.D. Thesis, Stanford University, Stanford, CA, USA, 2009. [Google Scholar]
  120. Zhao, C.; Yang, Y.-T.; Li, Z.-C. The homomorphic properties of McEliece public-key cryptosystem. In Proceedings of the 2012 Fourth International Conference on Multimedia Information Networking and Security, Nanjing, China, 2–4 November 2012; pp. 39–42. [Google Scholar]
  121. Line Coding Techniques. Available online: https://technologyuk.net/telecommunications/telecom-principles/line-coding-techniques.shtml (accessed on 16 November 2020).
  122. Huffman, D.A. A method for the construction of minimum-redundancy codes. Proc. IRE 1952, 40, 1098–1101. [Google Scholar] [CrossRef]
  123. Carron, L.P. Morse Code: The Essential Language; Radio Amateur’s Library, American Radio Relay League: Newington, CT, USA, 1986; ISBN 0-87259-035-6. [Google Scholar]
  124. ASCII—Wikipedia. Available online: https://en.wikipedia.org/wiki/ASCII (accessed on 20 November 2020).
  125. Unicode—Wikipedia. Available online: https://en.wikipedia.org/wiki/Unicode (accessed on 18 November 2020).
  126. Number Systems (Binary, Octal, Decimal, Hexadecimal). Available online: https://www.mathemania.com/lesson/number-systems/ (accessed on 19 November 2020).
  127. HTML Codes. Available online: https://www.html.am/html-codes/ (accessed on 18 November 2020).
  128. QR Code—Wikipedia. Available online: https://en.wikipedia.org/wiki/QR{_}code (accessed on 18 November 2020).
  129. Barcode—Wikipedia. Available online: https://en.wikipedia.org/wiki/Barcode (accessed on 19 November 2020).
  130. Categorical Data Encoding Techniques to Boost Your Model in Python! Available online: https://www.analyticsvidhya.com/blog/2020/08/types-of-categorical-data-encoding/ (accessed on 16 November 2020).
  131. Categorical Feature Encoding in SAS (Bayesian Encoders)—Selerity. Available online: https://seleritysas.com/blog/2021/02/19/categorical-feature-encoding-in-sas-bayesian-encoders/ (accessed on 16 November 2020).
  132. Braille—Wikipedia. Available online: https://en.wikipedia.org/wiki/Braille (accessed on 20 November 2020).
  133. Peikert, C. A Decade of Lattice Cryptography; Now Foundations and Trends: Norwell, MA, USA, 2016. [Google Scholar]
  134. Campello, A.; Jorge, G.; Costa, S. Decoding q-ary lattices in the Lee metric. In Proceedings of the 2011 IEEE Information Theory Workshop, Paraty, Brazil, 16–20 October 2011. [Google Scholar] [CrossRef]
Figure 1. Sample Code Word of length n.
Figure 1. Sample Code Word of length n.
Cryptography 05 00038 g001
Figure 2. Relationship between codes.
Figure 2. Relationship between codes.
Cryptography 05 00038 g002
Figure 3. Relationship Between Codes.
Figure 3. Relationship Between Codes.
Cryptography 05 00038 g003
Figure 4. Special codes.
Figure 4. Special codes.
Cryptography 05 00038 g004
Table 1. Comparison of PQC algorithms.
Table 1. Comparison of PQC algorithms.
Serial NumberPost-Quantum
Algorithm Category
Post-Quantum
Algorithms available
in this Category
Name of Most
Prevalent Algorithm
Type of Algorithm -
Encryption/ Signature/
Key Exchange
Public Key SizePrivate Key SizeSignature SizeStrengthsWeaknessesincluded in open-source library LiboqsAttacksOther Detailed
Comparisons
1Lattice
Based
Crypto-
gaphy
1. Encr-
yption/
Decry-
ption
2. Signa-
ture
3. Key
Exch-
ange
(RLWE)
NTRU EncryptE6130 B6743 B1. More efficient encryption and decryption, in both hardware and software implementations
2. Much faster key generation allowing the use of disposable keys.
3. low memory use allows it to use in applications such as mobile devices and smart-cards.
1. Complexity is high in NTRU 2. There is the possibility of the occurrence of a decryption failure from a validly created ciphertextBrute Force attack, meet-in-middle attack, lattice reduction attack, chosen cipher text attackGaithru et al. 2014 [43]
BLISS II (Bimodal Lattice Signature Scheme)S7 KB2 KB5 KB Side Channel Attack, Branch tracing attack, Rejection Sampling, Scalar Product LeakageEspitau et al. 2017 [44]
2Multi-VariateSignature onlyRainbowS124 KB95 KB424 KBIt is based on the difficulty of solving systems of multivariate equationsOnly Signature Scheme is availableDirect Attack, Min Rank Attack, High Rank Attack, UOV Attack, UOV Reconciliation attack, Attacks against hash functionPetzoldt et al. 2010 [45]
3Hash
Based
Signature
Signature
Only
SPHINCSS1 KB1 KB41 KB1. Best alternative
to number theore-
tic signature
2. Small and medi-
um size signatures
3. Small Key size
Only
Signature
Scheme is
available
Speed
Subset Resilience, One-wayness, Second Pre-image resistance, PRG, PRF and undetectability, Fault Injection AttacksBernstein et al. 2015 [46]
SPHINCS+S32 B64 B8 KB Distinct-function multi-target second-preimage resistance, Pseudorandomness (of function families), and interleaved target subset resilience, timing attack, differential and fault attacksBernstein et al. 2019 [47]
4Super-singular elliptic curve isogeny cryptographyKey Exchange OnlySupersing-ular Isogeny Diffie Helman (SIDH)K751 B 564 (compressed SIDH)48 B 48 (compressed SIDH)Difficulty of computing isogenies between supersingular elliptic curves which is immune to quantum attacksCannot be used for non-interactive key exchange, can only be safely used with CCA2 protectionSide-channel attacks, Auxiliary points active attack, adaptive attackCostello et al. 2016 [41]
5Code-Based Cryptography1. Encryption/Decryption
2. Signature
Classic McEliece CryptosystemE1 MB11.5 KBOne of the cryptosystem which is successful till the third round of NIST Post-Quantum algorithm standardization processVery Large Key sizeStructural Attack, Key recovery attack, Squaring Attack, Power Analysis Attack, Side Channel attack, Reaction attack, Distinguishing attack, message recovery attackTillich 2018 [48], Repka et al. 2014 [49]
Table 2. Summary of Industry Initiatives in Post-Quantum Cryptography.
Table 2. Summary of Industry Initiatives in Post-Quantum Cryptography.
Serial NumberName of the OrganizationCountryType of PQC Work InvolvedAlgorithms UsedCollaboratorPQC Product Developed
1.Avaya [53]USAPQC based products Post-QuantumQuantum-safe messaging, voice calls and document sharing
2.Envieta Systems [54]USAPQC products, PQC Consulting Developed Hardware and Software Post-Quantum Implementation Cores including those for embedded systems as well
3.Google [55]USAPQC productsHRSS-SXY (variant of NTRU encryption) and SIKE (supersingular isogeny key exchange)CloudfarePost-quantum cryptography˙encryption and signature methods for chrome browser
4.IBM [51]Switzer-landR&DLattice-based CryptographyQuantum safe Cloud and Systems
5.Infineon [56]GermanyPQC productsVariant of New Hope AlgorithmImplemented a post-quantum key exchange scheme on a commercially available contactless smart card chip Post-Quantum security for Government Identity Documents, ICT technology, Automotive Security, Communication Protocols
6.Isara [56]CanadaPQC productsHierarchical Signature Scheme (HSS) and eXtended Merkle Signature Scheme (XMSS)Futurex, Post-QuantumISARA Radiate, Quantum-safe Toolkit is a high-performance, lightweight, standards-based quantum-safe software development kit, built for developers who want to test and integrate next-generation post-quantum cryptography into their commercial products
7.Microsoft Research [52]USAPQC productsFrodoKEM, SIKE, Picnic, QTeslaPost-Quantum SSH, TLS, VPN
8.Qualcomm/ OnBoard Security [57]USAPQC based productspqNTRUsignOnBoard SecurityOnBoard Security has developed a digital signature algorithm that can resist all known quantum computing attacks. pqNTRUsign will replace RSA and ECDSA, the most commonly used quantum-vulnerable signature schemes.
Table 3. NIST Standardization Efforts [59].
Table 3. NIST Standardization Efforts [59].
Post-Quantum
Algorithm Type
Third Round FinalistTechnologyAlternate
Candidates
Technology
Public Key
Encryption/
Key Encapsulation
Mechanisms  
Classic McElieceCodeBIKECode
CRYSTALS KYBERLatticeFrodoKEMLattice
NTRULatticeHQCCode
SABERLatticeSIKESupersingular
Isogeny
Digital Signature
Algorithms
CRYSTALS–
DILITHIUM
LatticeGeMSSMultivariate
Polynomial
FALCONLatticePICNICOther
RAINBOWMultivariate
Polynomial
SPHIMCS+Hash
Table 4. Summary of standardisation efforts in Post-Quantum Cryptography.
Table 4. Summary of standardisation efforts in Post-Quantum Cryptography.
ParameterCountryFocus Area of Standardization in PQCFunction Standards/QOS StandardsStatus
NISTUSAQuantum resistant Algorithm Standardization for Cryptography, Key encapsulation mechanism, digital signatureFunction StandardsRound 1 and Round 2 of standardization process is completed. Round 3 in progress
ETSIEuropeQuantum Safe Cryptography, Quantum Key DistributionFunction StandardsPublished standards
ISONA (A Non-Governmental Organization)Quantum Key DistributionFunction StandardsWork in progress—Standards yet to be published
ITUA specialized agency of United NationsQuantum Key DistributionFunction and QOS StandardsOnly two standards published. Others are work in progress
CRYPTRECJapanQuantum resistant Algorithm Standardization for Cryptography, Key encapsulation mechanism, digital signatureFunction StandardsList of standardized algorithms are expected to be published between 2022-2024
Table 5. Common Code-based Cryptosystems, Codes used and thier Application.
Table 5. Common Code-based Cryptosystems, Codes used and thier Application.
Serial NumberCode-Based CryptographyTechnique & Codes UsedApplied to
1Mc ElieceBinary Goppa Codes, GRS, Concatenated Codes, Product codes, Quasi- Cyclic, Reed muller codes, Rank matric ( Gadidulin) codes, LDPC, MDPC, Genaralized Shri vastava codesComputing Systems, Embedded Devices [81], FPGA systems [82]
2NiederreiterGRS Codes, Quasi-Cyclic Codes, Binary Goppa codesComputing Systems, FPGA Systems [83]
Table 6. Comparison of Code-based signature schemes.
Table 6. Comparison of Code-based signature schemes.
SternJain et al.Cayrel et al.
Keygen
Sign
Verify
0.0170 ms
31.5 ms
2.27 ms
0.0201 ms
16.5 ms
135 ms
0.339 ms
24.3 ms
9.81 ms
sk
pk
System prams
Signature
1.24 bits
512 bits
65.5 kB
245 kB
1536 bits
1024 bits
65.5 kB
263 kB
1840 bits
920 bits
229 kB
229 kB
Table 7. Requirements fulfillment of various codes.
Table 7. Requirements fulfillment of various codes.
Serial NumberPurpose of EncodingCodeType of CodeProperties to Be Fulfilled for EncodingSupport
Complete Character
Set Encoding
Representation, Manipulation
and Storage in Digital Systems
Data HidingSupport Multiple Data TypesDynamic EncodingVariable EncodingHard DecodingRandomness of EncodingPossible for Decoding
1To encode data for digital data communicationAll codes generated using Line coding techniques [121]Binary data encoded to digital signals
2To encode data for digital data communication with error correction capabilities [121]All codes generated using Block Codes and Convolutional Coding Techniques [72]Binary data encoded to linear or non-linear codes with error detection and correction capabilities
3To encode data in a compressed format for faster message communicationHuffman Codes [122],Alphabets/character set encoded to a compressed codeNA
Morse Code [123](a)Alphabets/ Character set encoded to a compressed codeNA
(b)Alphabets/ Character set encoded to image(s)/ symbols(s)/ patterns(s)
4To represent data in a digital systemAscii, Unicode, … [124,125]Alphabets/character set encoded to a number
5To store and manipulate data in a digital systemBinary, BCD, Hexadecimal, Octal [126]Alphabets/character set encoded to binary/BCD/Hexadecimal/Octal through ordinal encoding
6Programmatic Representation of Character SetHTML Code [127]Alphabets/character set encoded to a Hexadecimal number
7To communicate digital data confidentiallyAtbash, Caesar Cipher, Columnar Cipher, Combination cipher, Grid Transposition cipher, Keyboard Code, Phone code, Rot Cipher, Rout Cipher [116]Alphabets encoded to another alphabet
A1Z26Alphabets encoded to a number
QR Code, Bar Code,Alphabets/Character Set encoded to image(s)/symbol(s)/pattern(s)
Dice Cipher, Digraph cipher, Dorabella Cipher, Rosicrucian Cipher, Pigpen cipher [116,128,129]
Francis Bacon Code [116]Alphabets/character set encoded to a sequence of alphabets
DNA Code [117]Alphabets/character set encoded to a sequence of alphabets
8To represent data or data set features to be used in machine learningOrdinal Code or label codeData encoded to a non-numerical labelNANANANANANA
One Hot Encoding, Dummy Encoding, Effect Encoding, Binary Encoding, Base N Encoding, Multi-Label Binarizer, DictVectorizer [130]Data is converted to a vectorNANANANANANA
Hash Encoding [130]Data is converted to its hash valueNANANANANANA
Bayesian Encoding [131]Data is encoded to its average valueNA NANANANANANA
9To represent data in a format comprehensible for visually challenged personsBraille Code [132]Alphabets/character set encoded to image(s)/symbol(s)/pattern(s)NA
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Balamurugan, C.; Singh, K.; Ganesan, G.; Rajarajan, M. Post-Quantum and Code-Based Cryptography—Some Prospective Research Directions. Cryptography 2021, 5, 38. https://doi.org/10.3390/cryptography5040038

AMA Style

Balamurugan C, Singh K, Ganesan G, Rajarajan M. Post-Quantum and Code-Based Cryptography—Some Prospective Research Directions. Cryptography. 2021; 5(4):38. https://doi.org/10.3390/cryptography5040038

Chicago/Turabian Style

Balamurugan, Chithralekha, Kalpana Singh, Ganeshvani Ganesan, and Muttukrishnan Rajarajan. 2021. "Post-Quantum and Code-Based Cryptography—Some Prospective Research Directions" Cryptography 5, no. 4: 38. https://doi.org/10.3390/cryptography5040038

APA Style

Balamurugan, C., Singh, K., Ganesan, G., & Rajarajan, M. (2021). Post-Quantum and Code-Based Cryptography—Some Prospective Research Directions. Cryptography, 5(4), 38. https://doi.org/10.3390/cryptography5040038

Article Metrics

Back to TopTop