Doh, copy-paste failure, I missed the top of the mail I wrote. Trying again: I'm co-chair of QIRG, but speaking as an individual here. I was planning to let this slip by, since I don't think ietf@ietf is the right forum for deciding whether or not a minority view on the effectiveness of quantum computing is correct. But since we're here... I've known mohta-san for several years now, and we've had F2F discussions of this and relevant topics. He's asking a good question, and in fact I had similar ones when I began studying quantum computing in 2003. He certainly isn't the first person to ask these questions, but he doesn't accept the answer that the vast majority of the QC community does. Fair enough. Skepticism is largely healthy, but I expect that time will prove him wrong. I would say that his view diverges enough from the broad understanding of the community that it does not affect important aspects of the IRTF decision to charter QIRG and invest some effort (entirely voluntary, of course) of the IETF/IRTF community in the area. In fact, I would argue that time has *just recently* shown skeptics to be wrong. (Or, to be accurate, "is now showing," since the paper is apparently still under peer review.) Chris Monroe's group at UMD (moving to Duke) in September posted a paper to the arXiv preprint server claiming that their ion trap can create a fully error-protected logical qubit, using a scheme known as the 9-qubit Bacon-Shor code. I was so excited about this paper that I tweeted, "It's here! The logical qubit is here!" There are a few 't's to be dotted and a few 'i's to be crossed before it can be used in practice, but I think this will stand as a landmark paper. https://arxiv.org/abs/2009.11482 For 90% of ietf@ietf, that's probably enough; don't feel guilty if you stop reading right here. But I'll follow up with some references, then some specific points about mohta-san's concerns. ------- For people who are mildly interested in the topic of quantum error correction, we have a short discussion and a Javascript app that shows some of the basic ideas in our MOOC (which is now free). https://www.futurelearn.com/courses/intro-to-quantum-computing For people with serious interest, there are two topics to cover: decoherence, and quantum error correction. (We'll leave aside fault-tolerant execution of algorithms, which requires yet more work.) My favorite source on decoherence is "Introduction to decoherence and noise in open quantum systems," by Todd Brun and Daniel Lidar, which is the opening chapter in the compendium _Quantum Error Correction_ (Cambridge, 2013). Unfortunately, AFAIK, it's not publicly available in PDF. Other sources include John Preskill's lecture notes (which are freely available) at http://www.theory.caltech.edu/~preskill/ph219/ph219_2018-19 and many books on quantum computing, such as the classic Nielsen & Chuang (Cambridge, 2000). Be warned that this does take moderately good linear algebra chops, and if you want to really understand decoherence, some partial differential equations. If you want to learn about quantum error correction, my favorite source is Devitt, Nemoto & Munro, "Quantum Error Correction for Beginners," arXiv:0905.2794v3 [quant-ph]. Already over a decade old, but still outstanding. But "for beginners" assumes you have a decent quantum background, just nothing on QEC. ----- Okay, a little on mohta-san's concerns... There are multiple sources of errors and multiple types. * bit flip: a symmetric channel, with 0-->1 and 1-->0 errors equally likely * phase flip: ditto, but for a qubit's phase * amplitude damping: In this mode, 1-->0 errors are more likely than 0-->1 errors, because the former happens when a photon is emitted and thpe latter when a photon is absorbed; moving to the lower-energy state is more common. This is the harder one to understand. https://en.wikipedia.org/wiki/Amplitude_damping_channel http://www.thphys.nuim.ie/staff/jvala/Lecture_12.pdf We can also talk about various gate errors, and errors in measurement. One that particularly worried me is, in fact, one of the things that Peter Shor simplified away in his first QEC paper: what happens when the errors on individual qubits aren't independent? *ALL* of these are handled properly by quantum error correction. In fact, they can all be treated as types of memory errors, which (thankfully) simplifies the math. This is over-simplified, but QEC consists of several phases: 1. Encoding a logical state 1.5 <<errors happen>> 2. Extracting syndrome information a. calculating the syndromes in superposition b. measuring the syndrome qubits, which *forces the state into one with discrete errors* 3. Using that information to correct the state. Steps 2 & 3 are repeated over the lifetime of the algorithm execution, interleaved with the operations that execute logical gates. (Well, in codes known as topological codes, surface codes, toric codes and color codes, that's not quite true. Those codes also have no classical equivalent, whereas other codes derive very directly from classical error correction.) Step 2a involves quantum operations, and so *adds* errors into the state, meaning an interesting question is when the error correction process as a whole *removes* more errors than the extraction adds. Monroe's group claims to have exceeded that threshold. Mohta's concern seems to be two-fold; first, that errors aren't independent, and second, that forcing of the state (actually, a *partial* collapse of the quantum state) doesn't take you a state where applying correction works. As it happens, the first problem worried me enough that I wrote about it in my Ph.D. thesis. See Sec. 2.3.5 (p. 55) of https://arxiv.org/abs/quant-ph/0607065 Fundamentally, coherent noise over a group of qubits (e.g., some coherent radio noise from outside, or shared dependence on a slightly incorrect classical oscillator, or something) results in all of the suppressed *to second order* -- to the *square* of the induced rotation angle (since it's less than 1.0, that means reducing the error). A similar argument applies to the amplitude damping that is the harder error type mentioned above: the syndrome extraction forces us back to a state where, *with high probability*, the correction we choose to apply makes the state better (== closer to the original) rather than worse. Some QEC schemes operate by applying multiple recursive levels of the same (or, possibly, different) encoding schemes. Yes, that results in a big growth in resources used, but a) we're only talking about 2 or 3 levels, not some huge number; and b) the errors are actually suppressed at a rate that exceeds the resource consumption, so it all works out in the end. I won't go into the math of that here. Fundamentally, quantum error correction *works*, and over the course of 2020-2025, we will see it demonstrated increasingly clearly in experimental systems. Barring something really extraordinary, this is all I will have to say on the topic on ietf@ietf. People who are interested are welcome to attend the QIRG meeting (Monday afternoon Bangkok time during IETF) and ask about this during the open mic. Since we have other business to attend to on the already-published agenda, I will defer discussion of this draft and this topic to the very last thing, but I expect there will be some time available. See y'all online in a couple of weeks! —Rod