Advertisement · 728 × 90

Posts by Dulwich Quantum Computing

Video

If you're tired of the AI assistent in Excel, you can now connect it to a quantum computer instead!
multiversecomputing.com/singularity

1 hour ago 3 0 0 0

Fun twist in Google's zero knowledge proof of quantum circuit.

3 days ago 13 3 0 0
3 days ago 2 0 0 0

Ever wanted to be trapped in a room where the only way out is solving a problem in quantum mechanics? Apply for a PhD at the University of Zurich!

Just kidding... Check out this escape room!👇

3 days ago 7 0 0 0

Why bother with recommendation letters when you get a like from Peter Shor?

3 days ago 24 1 0 0
4 days ago 1 0 1 0
Preview
Dequantizing Short-Path Quantum Algorithms The short-path quantum algorithm introduced by Hastings (Quantum 2018, 2019) is a variant of adiabatic quantum algorithms that enables an easier worst-case analysis by avoiding the need to control the...

"On the positive side, our results provide a new ``quantum-inspired'' approach to designing classical algorithms for important classes of constraint satisfaction problems."
arxiv.org/abs/2604.12131

Positive to whom?

5 days ago 6 1 0 0
Advertisement

Since the whole world will now have access to it, I expect nothing less than the first demonstration of time travel, teleportation of soul or the simulation of God to come out of this. Let your imagination go wild!

6 days ago 9 0 1 0

It would be hugely disappointing if the device that was officially announced as "lending credence to the notion that quantum computation occurs in many parallel universes" and whose predecessor was used to "create a wormhole" would now be used by mere mortals to simulate some boring Hamiltonian.

6 days ago 27 0 1 1
The fact that no CPTP map can annihilate σz while
preserving σx and σy is known, in quantum information
folklore, as the “No-Pancake Theorem”.

Our central result might be thought of as a fully general
No-Pancake Theorem, since it rules out the dynamical
superselection of all such non-algebraic IPS

The fact that no CPTP map can annihilate σz while preserving σx and σy is known, in quantum information folklore, as the “No-Pancake Theorem”. Our central result might be thought of as a fully general No-Pancake Theorem, since it rules out the dynamical superselection of all such non-algebraic IPS

Are you telling me there is a “No-Pancake Theorem” in quantum information folklore and I have never heard about it?
arxiv.org/pdf/1006.135...

1 week ago 17 0 0 0

If you prayed that your paper gets into TQC but it didn't, you should pick up a fight with the pope.

1 week ago 3 0 0 0
Post image

I cannot decide what's my favorite slop: The single qubit CNOT gate or the tentative to make a graph for a bell pair or maybe the new notation ∣0^2 and ∣1 ^1 .

Source: a linked in post I saw while scrolling out of boredom

1 week ago 9 3 0 0

"I don't think anyone has ever heard anything like this. It's too crazy! Who brings a quantum guitar into a chapel?"

1 week ago 3 1 0 0

okay someone who actually knows what the fuck they're talking about WAS there actually a significant breakthrough in quantum computing recently or is everyone doing the thing they do where they hear the word 'quantum' and immediately get an erection that lasts so long they hafta call their doctor

1 week ago 50 7 9 2
Preview
Exponential quantum advantage in processing massive classical data Broadly applicable quantum advantage, particularly in classical data processing and machine learning, has been a fundamental open problem. In this work, we prove that a small quantum computer of polylogarithmic size can perform large-scale classification and dimension reduction on massive classical data by processing samples on the fly, whereas any classical machine achieving the same prediction performance requires exponentially larger size. Furthermore, classical machines that are exponentially larger yet below the required size need superpolynomially more samples and time. We validate these quantum advantages in real-world applications, including single-cell RNA sequencing and movie review sentiment analysis, demonstrating four to six orders of magnitude reduction in size with fewer than 60 logical qubits. These quantum advantages are enabled by quantum oracle sketching, an algorithm for accessing the classical world in quantum superposition using only random classical data samples. Combined with classical shadows, our algorithm circumvents the data loading and readout bottleneck to construct succinct classical models from massive classical data, a task provably impossible for any classical machine that is not exponentially larger than the quantum machine. These quantum advantages persist even when classical machines are granted unlimited time or if BPP=BQP, and rely only on the correctness of quantum mechanics. Together, our results establish machine learning on classical data as a broad and natural domain of quantum advantage and a fundamental test of quantum mechanics at the complexity frontier.

One remarkable thing about this paper is that Preskill @preskill.bsky.social no longer seems to be affiliated with AWS. If he could leave one of the largest companies in the world to join a tiny new startup, it means there's still hope for Dulwich!
scirate.com/arxiv/2604.0...

1 week ago 20 0 0 0
Advertisement
Preview
MQNN: Multi‐Block Quantum Neural Networks for Radar Echo Signal Recognition of Unmanned Aerial Vehicles This paper proposes a novel multi-block quantum neural network(MQNN).MQNN is applied to the real collected dataset of radar echo signals from UAVs. Compared with five quantum machine learning algorit...

This is so weirdly specific. They should have also included in the title that this quantum neural network is particularly good at detecting drones sent from country A to country B.
advanced.onlinelibrary.wiley.com/doi/abs/10.1...

1 week ago 4 0 0 0

When a big company announces their new quantum paper.

1 week ago 11 1 0 0

It's not us who should go to The Hague, it's you-know-who.

1 week ago 1 0 1 0
Definition 2.1 (Felinity of a quantum state). For a n-qubit state ρ, the felinity of ρ, F n(ρ) ∈ [0, 1] is defined as...

Definition 2.1 (Felinity of a quantum state). For a n-qubit state ρ, the felinity of ρ, F n(ρ) ∈ [0, 1] is defined as...

"We [...] introduce a new measure, felinity."
arxiv.org/abs/2604.02793

1 week ago 26 4 0 0
2 weeks ago 1 0 1 0
Preview
Declarative bespoke modelling: A new approach Modern numerical models are increasingly complex, opaque, and computationally expensive, yet frequently fail to predict even qualitative features of observed phenomena. We propose a new paradigm, Decl...

Not a ton of April Fools' submissions this year, but I'll do a quick thread anyhow.

Here's a charming little paper on 'declarative bespoke modeling' "in which the modeller explicitly declares the relationship between model inputs and outputs."

2 weeks ago 14 6 2 1
Post image

A new, very serious preprint is out from Veronika and Alberto. No, don't look at today's date it's 100% legit we promise. It's a genuine proof that P=NP...

(...if you believe in quantum many-worlds and are willing to destroy almost every observer in the multiverse 👀)

arxiv.org/abs/2603.28869

2 weeks ago 16 3 1 2
Preview
Remarks on "Further comments on "Rebuttal of "Refutation of "Comment on "Reply to "Comments on "A genuinely natural information measure" " " " " " " It's a bit tedious, but as John Doe and Jean Roe have insisted on offering further comments on our comprehensive refutation of the former's already tiringly obstinate advances, we feel compelled to re...

Let's make this the top paper on SciRate today!
scirate.com/arxiv/2603.2...

2 weeks ago 9 0 1 0
Advertisement
This is indeed a surprising result, provided that every piece can be confidently verified by a third party and that all assumptions can be further justified. Since this work appears to make more assumptions than prior work, that somewhat weakens the baseline.

Below, I list all 11 places in the paper that contain “assume” or related forms:

    Assumes non-local connectivity.

    Here, we assume a 1 ms stabilizer measurement cycle.

    We assume that each surgery gadget is implemented in (\tau_s \approx 2d/3) code cycles, where (d) denotes the distance of the processor code.

    We assume that each layer of surgery takes (\tau_{\mathrm{cult.s}} = 2d_s/3) code cycles, which often minimizes the total logical error rate.

    In the above, we assumed that for each (C_i), the involved (m_i) qubits are read from memory at the beginning and written back to memory at the end.

    For the balanced architecture, we assume that (2q_a < k_p), so that the processor can store the (2q_a) qubits needed for unary iteration while still leaving enough space to apply the word operators.

    For the ECC-256 algorithm, based on compilations in [116--120], we assume the following split in the Toffoli count among controlled adders, adders, and lookups:
        40% for 256-bit adders,
        50% for 256-bit controlled adders,
        10% for lookups with 16 address bits and word size 256.

    In addition, we assume the existence of logical gadgets capable of measuring many logically disjoint PPMs on high-rate codes in parallel.

    For (P < 600) ((P \geq 600)), we assume processor encoding rates of (r = 20\%) ((r = 30\%)), based on upcoming work [125].

    We assume that these codes have the same rate as the processor, except when (P < 100), where we assume a lower encoding rate of 4% to allow transversal coupling of smaller block sizes (e.g., the bb18 code).

    To account for larger codes and more complex PPMs, we consider (\gamma = 1\text{--}3) and plot results for (\gamma = 2…

This is indeed a surprising result, provided that every piece can be confidently verified by a third party and that all assumptions can be further justified. Since this work appears to make more assumptions than prior work, that somewhat weakens the baseline. Below, I list all 11 places in the paper that contain “assume” or related forms: Assumes non-local connectivity. Here, we assume a 1 ms stabilizer measurement cycle. We assume that each surgery gadget is implemented in (\tau_s \approx 2d/3) code cycles, where (d) denotes the distance of the processor code. We assume that each layer of surgery takes (\tau_{\mathrm{cult.s}} = 2d_s/3) code cycles, which often minimizes the total logical error rate. In the above, we assumed that for each (C_i), the involved (m_i) qubits are read from memory at the beginning and written back to memory at the end. For the balanced architecture, we assume that (2q_a < k_p), so that the processor can store the (2q_a) qubits needed for unary iteration while still leaving enough space to apply the word operators. For the ECC-256 algorithm, based on compilations in [116--120], we assume the following split in the Toffoli count among controlled adders, adders, and lookups: 40% for 256-bit adders, 50% for 256-bit controlled adders, 10% for lookups with 16 address bits and word size 256. In addition, we assume the existence of logical gadgets capable of measuring many logically disjoint PPMs on high-rate codes in parallel. For (P < 600) ((P \geq 600)), we assume processor encoding rates of (r = 20\%) ((r = 30\%)), based on upcoming work [125]. We assume that these codes have the same rate as the processor, except when (P < 100), where we assume a lower encoding rate of 4% to allow transversal coupling of smaller block sizes (e.g., the bb18 code). To account for larger codes and more complex PPMs, we consider (\gamma = 1\text{--}3) and plot results for (\gamma = 2…

Also, check out this comment on SciRate!
scirate.com/arxiv/2603.2...

2 weeks ago 1 0 0 0

Or perhaps as a new startup they need to generate publicity, and Nature lets you get away with it?

2 weeks ago 7 0 1 0

Then you owe me a qubit!

2 weeks ago 2 0 1 0

People often state their theorems as "there exists an algorithm" even though they explicitly provide it. But this time Google actually means it.
research.google/blog/safegua...

2 weeks ago 9 2 0 0

When you initialize your spin chain to all spins up.

2 weeks ago 9 0 0 0

So excited, I literally can't wait to see the result!

"To help readers evaluate this work more carefully: the proposed 10,000-qubit scheme appears to require on the order of 117 years to execute."
scirate.com/arxiv/2603.2...

2 weeks ago 6 0 1 0

We're making such good progress at reducing the number of qubits needed for factoring that we might accidentally overshoot and bring it down to zero, at which point we would have a classical factoring algorithm.
arxiv.org/abs/2603.28627

2 weeks ago 26 4 1 2