GnuPG – post-quantum crypto landing in mainline

(lists.gnupg.org)

117 points | by zdkaster 12 hours ago

10 comments

  • tokenhub_dev 6 hours ago
    For people wondering whether to migrate now: the practical question isn't "is a CRQC imminent" (it isn't), it's whether your encrypted messages have a useful lifetime longer than the optimistic deployment timeline.

    If you encrypt a one-off email with a 5-year confidentiality requirement, harvest-now-decrypt-later actually matters. If you're encrypting backups that get rotated every 90 days, it doesn't.

    The hybrid construction (Kyber/ML-KEM + X25519) is nice precisely because it's a no-regret move — you don't lose anything by adopting early. If Kyber turns out to have a structural flaw, X25519 still protects you. If a CRQC arrives, ML-KEM still protects you. The only real cost is key/ciphertext size, which for OpenPGP isn't a hot path anyway.

    The interesting question is what happens to long-lived smartcard/HSM-backed keys. Those typically have a 5–10 year lifecycle and most hardware won't grow ML-KEM support without a hardware refresh. That's where I'd expect the first real compatibility headaches.

    • BoppreH 5 hours ago
      Some Hardware Security Module manufacturers were smart enough to include FPGAs in their products, which they can now use to accelerate PQC algorithms without a hardware refresh.

      The trouble is that PQC already has inherent size/performance downsides, and it won't benefit from the decades of optimizations that classical algorithms had. Expect a hefty performance tax for some time.

  • utopiah 9 hours ago
    > introduction of Kyber (aka ML-KEM or FIPS-203) as PQC encryption algorithm

    Funny to read 1-liner changelog versus the plethora of articles just few years ago along the line of "Quantum computer, it might just change our entire lives and make privacy impossible!".

    The simple addition (of a not so simple algorithm) to the software (and few others, e.g. OpenSSL) and voila, me can move on with our daily lives. Cryptography and computational complexity are truly amazing.

    • BoppreH 5 hours ago
      It reminds me a lot of Y2K. The fix is simple, but finding the places where it's needed and doing it in a compatible way are absolutely non-trivial problems. The best we can hope is the same as Y2K: the plethora of articles convince businesses to invest large amounts of money to migrate algorithms, so that when a quantum computer arrives it won't be a big deal.
      • CatMustard 2 hours ago
        > it won't be a big deal.

        This isn't a space I know too much about, but even if we all start using quantum-safe encryption for everything today, won't the arrival of quantum computers that can break traditional encryption not still be a big deal?

        Given that intelligence agencies, tech companies and various bad actors have been storing encrypted data for a long time, hoping to decrypt when (if?) that day comes?

        • parsimo2010 1 hour ago
          Intelligence agencies and companies for which industrial espionage is an actual concern will re-encrypt their data storage, or have already done so. The only risk is on data that was already obtained with a vulnerable encryption. So there is some risk that a few secrets are lost, but it won’t be everything. And if you were to start now and quantum decryption isn’t viable for a decade then any secrets that do get exposed are surely less of a problem than if they were discovered today.
        • BoppreH 2 hours ago
          Definitely, but then the damage is limited to the encrypted data that those actors managed to intercept some years before. Compared to QC arriving to an unprepared world, that's a very limited impact.
  • aborsy 8 hours ago
    Does it implement the hybrid version ML-KEM-768 + X25519 or ML-KEM-768 only ?

    The X25519 key could remain in hardware keys for a while til manufactures catch up.

    • darkamaul 6 hours ago
      If I understood the code correctly, it always use the hybrid version.

      > Kyber is always used in a composite scheme along with a classic ECC algorithm.

  • zdkaster 12 hours ago
    GnuPG Version 2.5.19

    The 2.5 series are improvements for 64 bit Windows and the introduction of Kyber (aka ML-KEM or FIPS-203) as PQC encryption algorithm.

    The old 2.4 series reaches end-of-life in just two months.

  • growse 7 hours ago
    I don't know enough about either the technical nuance or the political drama, but some observers have noted that GnuPG's implementation is (deliberately?) incompatible with the IETF's standards. It's not clear why.

    https://floss.social/@hko/116459621169318785

    • upofadown 4 hours ago
      From the GnuPG prospective RFC-9580 is a deliberate fork away from what agreement could be achieved. Basically the faction that is now called RFC-9580 (mostly Sequoia and Proton) wanted to make a lot of changes to the existing standard but the faction that is now called LibrePGP (mostly GnuPG and RNP) was not convinced that those changes were necessary.

      Traditionally the OpenPGP standards process has been very conservative and minimalistic. GnuPG comes from that tradition. So the RFC-9580 faction created their own maximalist version of the standard and are actively promoting it as the standard.

      So from a user perspective, there are two incompatible proposals out there. It's a mess. So it is better to aggressively ignore them both and maintain interoperability by sticking with RFC-4880 (OpenPGP). That might be a problem if you for some reason are still concerned about a quantum attack against cryptography as the post quantum stuff has gotten caught in this schism. It is certainly something that the users need to keep in mind.

      • throw0101a 4 hours ago
        > […] and are actively promoting it as the standard.

        Well:

        > Category: Standards Track

        * https://datatracker.ietf.org/doc/html/rfc9580

        • upofadown 3 hours ago
          It is very hard to prevent a proposal from becoming a RFC. You have to generate ongoing opposition for longer than the supporters. FWIW, here is the LibrePGP proposal:

          * https://datatracker.ietf.org/doc/draft-koch-librepgp/

          Observing the OpenPGP schism mess I think I have gained some insight as to why some RFCs become so bloated. For example it has been recently pointed out that there are 60 RFCs for TLS (with 31 drafts in progress)[1]. The RFC process seems to be more optimal during the design phase. Once we have an established standard there should to be some way to force those that propose changes/extensions to provide appropriately strong justifications for those changes/extensions. Right now it is a popularity contest and there will always be more people out there in favour of changes/extensions than those willing to endlessly fight against those changes/extensions. Because cryptography is so specialized and obscure, the users tend to get left out of the discussion.

          [1] https://www.cs.auckland.ac.nz/~pgut001/pubs/bollocks.pdf

        • tovej 3 hours ago
          It is a standard proposal, which is why it's in the standards track. The point was that it is not the only (the) standard, and not the universally accepted one.
    • woodruffw 1 hour ago
      > It's not clear why.

      The situation is farcical, and stems from the double bind that PGP has been in for at least 20 years: the standards are bad and need modernization, but it’s impossible to modernize them because the single thing that retains “serious” users of PGP is backwards compatibility.

      The end result of this is a version of Weekend at Bernie’s where both GPG and OpenPGP are fighting over how to dress up the corpse, while the rest of the world has moved on.

    • capitol_ 6 hours ago
      As far as I understood it: GnuPG started to implement stuff from the standard before it was finished, the standard continued to improve and GnuPG refused to change code already written.

      Combined with some personal drama.

      • em-bee 3 hours ago
        it's not that simple. the new standard is a complete rewrite of the old one. they are not even compatible anymore. things the old standard used to support are not supported in the new standard. that makes any implementation of the new standard incompatible with implementations of the old one. GnuPG simply refused to stop supporting the old standard and decided to fork the standard itself. on the personal drama my interpretation is that it resulted from people backing the new standard being unhappy that GnuPG didn't go along.

        my opinion is that rewriting standards like that is the result of design by committee. everyone wants to put their mark on it. designing a new standard is fine, but the new standard should have also received a new name, or it should at least have been acknowledged that the old standard still needs to be supported until enough time has passed that the old standard is no longer in use. (which could take decades if not more if we want to be realistic and consider that encrypted data at rest could linger around pretty much forever unless actively re-encoded.)

        (source: i talked to a GnuPG developer)

  • trueno 9 hours ago
    been thinking about this a bit. someone just tell me what algo to use and ill start using it now. are the quantum-resistant cryptos significantly slower?
    • d1sxeyes 9 hours ago
      Basically the idea is use hybrid. AES-GCM-256 or ChaCha20-Poly1305 for symmetric encryption (which is already PQ-safe), and ML-KEM looks set to become the standard for key encapsulation.

      ML-KEM-768 is fast as an algorithm, faster than X25519 in terms of pure computation, but uses large keys, so has higher overheads on small payloads. Most of the time, they’re about equal, or the absolute time is so slow it doesn’t matter.

      Most folks now are doing hybrid ML-KEM and X25519 to guard against undiscovered flaws in ML-KEM.

    • purplehat_ 9 hours ago
      I believe ML-KEM is the standard algorithm for post-quantum asymmetric encryption. I think it's slower mainly because there's not good hardware support, but it shouldn't be a big deal because most encryption is hybrid where you only use the asymmetric crypto briefly to share a secret you can use for symmetric cryptography.

      ML-KEM based on a lattice problem called "Learning With Errors", and there are similar lattice-based algorithms which have no known quantum speedup. Most traditional asymmetric encryption algorithms are based on number-theoretic assumptions like the discrete logarithm problem or the RSA assumption, which are broken by Shor's algorithm.

      Symmetric cryptography (AES and SHA hash functions) are post-quantum resistant for now. Grover's algorithm technically cuts their asymptotic security in half, but that doesn't parallelize, so practically there is no known good quantum attack, and cryptographers and standards agencies tend to not worry about that. You can keep using those.

      [edit: according to the sister comment posted simulataneously ML-KEM is faster than X25519. good to know!]

    • upofadown 3 hours ago
      For something like PGP, any performance difference wouldn't matter. There is one message and the key agreement is done once. As long as things are fast enough to be imperceptible to the user we are fine.
  • immanuwell 8 hours ago
    cool, now my emails that nobody's reading anyway are safe from quantum computers that don't exist yet
    • PunchyHamster 7 hours ago
      it's mostly to make clowns repeating "It's not PQ secure therefore bad" happy I think
  • Hendrikto 6 hours ago
    [flagged]
    • em-bee 3 hours ago
      woha, this is totally twisted and the opposite of what really happened: https://news.ycombinator.com/item?id=47909640

      GnuPG did not unilaterally implement new non-OpenPGP formats. it kept supporting the old version of the OpenPGP standard. it unilaterally decided to NOT CHANGE its implementation. it's not trying to derail anything. the lack of engagement came from everyone else refusing to listen to the idea that you can't just break compatibility like that.

    • LtWorf 4 hours ago
      I see it as the usual push from sw companies to replace important copyleft projects with company directed ones with business-friendly (user unfriendly) licenses.
  • sipsi 7 hours ago
    [flagged]
  • dsecurity49 9 hours ago
    [dead]
    • Retr0id 6 hours ago
      Why is this AI-generated comment still at the top of the thread, after 3 hours? Is it finally time to give up on HN?
      • Pay08 3 hours ago
        It didn't read as AI-generated to me. Subsequent comments did, but this one seems perfectly normal.
    • jore 8 hours ago
      I haven’t heard this quote before, but I am copying it here because it makes so much sense:

      Arguing that you don't care about the right to privacy because you have nothing to hide is no different from saying you don't care about free speech because you have nothing to say. - Edward Snowden

    • pabs3 8 hours ago
      IIRC the GnuPG folks do a lot of consulting and sell additional software:

      https://gnupg.org/service.html https://gnupg.com/ https://g10code.com/

    • da_grift_shift 6 hours ago
      The [THING] has been living rent-free in my head since [YEAR]. Also the fact that [THING]. No [X]. No [Y]. No [Z]. Just: [A]. Absolute [HYPERBOLE] energy.

      At least this comment didn't have the double quotes left in ˙ ͜ʟ˙

    • snthpy 7 hours ago
      And they use SHA-1 for verification?
      • noosphr 7 hours ago

           If you already have a version of GnuPG installed, you can simply
           verify the supplied signature.  For example to verify the signature
           of the file gnupg-2.5.19.tar.bz2 you would use this command:
        
             gpg --verify gnupg-2.5.19.tar.bz2.sig gnupg-2.5.19.tar.bz2
        
           This checks whether the signature file matches the source file.
           You should see a message indicating that the signature is good and
           made by one or more of the release signing keys.  Make sure that
           this is a valid key, either by matching the shown fingerprint
           against a trustworthy list of valid release signing keys or by
           checking that the key has been signed by trustworthy other keys.
           See the end of this mail for information on the signing keys.
        
         * If you are not able to use an existing version of GnuPG, you have
           to verify the SHA-1 checksum.  On Unix systems the command to do
           this is either "sha1sum" or "shasum".  Assuming you downloaded the
           file gnupg-2.5.19.tar.bz2, you run the command like this:
      • dsecurity49 7 hours ago
        [dead]
    • gtsnexp 8 hours ago
      [flagged]