We know of only a few cases where quantum advantage is proven, Shor’s algorithm for factoring integers being the classic example. In most other cases, especially in machine learning, the situation is fuzzier: speedups are anticipated, but it’s not clear where the boundary lies between classical and quantum relevance. This matters most in areas like biology, where the data is huge, complex, and deeply interconnected, exactly the kind of setting where quantum resources should shine.
That’s the gap we set out to close. In our latest paper, “How many qubits does a machine learning problem require?”, we introduce the first framework that can tell you how many qubits are needed for a quantum model to reach a desired accuracy on a dataset of interest.
We find that medium-sized classical machine learning problems from OpenML require about 20 logical qubits to reach 100% training and testing accuracy – something within the reach of today’s classical simulators – suggesting no quantum advantage. However, when applied to subsets of the Tahoe-100M dataset, a transcriptomic dataset with 100 million samples, the required qubits quickly exceed the practical limit for classical simulation. But they are still well within the reach of the quantum computers that are expected to be viable in the next 5 years! These results support that some of the largest, most information-rich problems in biology and beyond may be prime candidates for quantum machine learning.
As a further technical result, we find that the number of qubits required to reach 100% train and test accuracy did not necessarily increase with the number of features of the Tahoe-100M subsets. The key lies in the bit-bit encoding technique we introduced earlier this year, which efficiently compresses classical data into a fixed qubit budget while preserving the information most relevant for learning.
In conclusion, by offering a concrete way to connect datasets to quantum hardware requirements, we mark a turning point in the field of quantum machine learning. Instead of speculating about where quantum advantage may appear, researchers can now calculate it. For companies exploring AI and quantum technology, that means clearer roadmaps, smarter allocation of resources, and a sharper sense of when quantum computing will make a meaningful impact.
To learn more about this research, check out the paper https://arxiv.org/pdf/2508.20992 .
Stay tuned for a convenient online tool to calculate whether or not your own datasets are likely to benefit from quantum advantage! If you would like alpha access to our tool or want to learn more about our other offerings, please feel free to reach out to us at [email protected].
]]>Let’s say we want to create a unitary operator acting on a large Hilbert space. This operator isn’t monolithic—it’s built from smaller sub-operators, each of which may act on its own local Hilbert space. Some might act on just a few qubits, others on the full space. And each of those sub-operators may, in turn, be composed of even smaller building blocks. As a result, the overall operator can be naturally organized as a tree, with the highest-level operator at the root and more granular components forming the branches and leaves.
In the diagram, time flows from left to right. That is, an operator placed to the left of another in this structure will be applied earlier. Up to this point, we haven’t mentioned quantum circuits explicitly—although circuits could be one way to implement such a structure, the construction we’re describing is more general.
Now consider adding an operator at the lowest level of the tree, say to the branch all the way on the left.
In quantum computing, a unitary operator can be defined concretely using a sequence of low-level gates, similar in spirit to a classical list of instructions. But here’s the key difference: in quantum mechanics, the same operator can be executed in a variety of contexts which may change how they relate to the overall operator. For example:
Crucially, we don’t want to define a new operator for each such context. Instead, we define the operator once and annotate the tree with the context in which it should be interpreted. Since the context is not intrinsic to the operator itself—but rather to how it is being used—it makes sense to store that context in the parent node of the operator, not within the operator’s own definition.
And here’s where things get especially interesting: these contexts can interact with each other, and these interactions can be exploited for compilation. For example, when an operator is both controlled and wrapped in a compute–uncompute structure,
Thus, once we’ve constructed the overall operator tree, the first step in compilation is to resolve the hierarchy of contexts. This means systematically combining and simplifying the contexts so that the final operator realizes the correct unitary evolution in the most efficient manner.
This approach opens up a rich set of questions. Are there a finite set of physically meaningful contexts? Can their interactions be classified or composed in some principled way? Understanding these possibilities could guide future research and help define the semantics of high-level quantum programming.Coherent Computing is pioneering high-performance and user-friendly software platforms that unlock the full potential of quantum hardware. Get in touch with us to learn more.
]]>At the top of the stack, we find the hardware-independent layers. These are the tools and abstractions that allow users — scientists, engineers, and domain experts — to design quantum solutions without worrying about the quirks of any particular device.
At a certain point, abstraction must give way to physical reality. This happens at the hardware-specific layers.
This layered structure serves a critical purpose: separation of concerns. By carefully dividing the software into hardware-independent and hardware-specific components, Coherent Computing aims to enable rapid development, broader accessibility, and easier hardware innovation. Users should be able to build complex quantum workflows without ever needing to manage the deep technical challenges of working with quantum devices. At the same time, hardware specialists can focus on improving qubit technologies without constantly rewriting higher-level software.
In short, it’s a design that acknowledges the unique challenges of quantum computing while opening the door to real-world adoption.Coherent Computing is pioneering software platforms that unlock the full potential of quantum hardware.
Get in touch with us to learn more.
]]>There are three major criteria any scalable quantum model must satisfy:
The paper develops techniques that directly address all of the above criteria for classification, and demonstrates it on the MNIST dataset.
Traditionally, quantum machine learning has struggled with the “data loading problem” — how to efficiently represent classical data with quantum states. Loading real-world datasets (like MNIST images) into quantum circuits without losing key information has seemed impractical.
This work introduces bit-bit encoding, a simple yet powerful idea:
Critically, this method ensures universal approximation: any function between the input bits and output bits can be represented by a unitary operation. Unlike amplitude or angle encoding — where expressivity is limited — bit-bit encoding allows the model to learn any relationship as quantum gate depth and qubit count increase.
It flips what one might intuitively assume, showing that sparsity in data loading enables better expressivity in quantum models, in contrast to dense loading.
Another major obstacle has been trainability. Classical machine learning relies heavily on gradient-based optimizers like Adam or SGD, which require careful tuning of learning rates and other hyperparameters. Worse, these optimizers can get stuck at saddle points which occur frequently in high-dimensional optimization.
Here, we propose a radical shift: exact coordinate updates. Instead of guessing step sizes or tuning learning rates:
Moreover, this method naturally avoids getting stuck at saddle points — an advantage that becomes more critical as models grow in size since large models are much more susceptible to getting stuck at saddle points than they are at bad local minima.
Large variational quantum circuits often suffer from barren plateaus on initialization — regions where gradients vanish and training grinds to a halt.
The solution here is sub-net initialization:
This avoids random initialization that causes barren plateaus and creates a “ladder” where each rung builds smoothly on the last. It’s like growing a deep neural network layer-by-layer — but with exact parameter transfer, something classical ML can’t easily replicate.
It also enables a powerful fail-fast strategy: If a small model gets stuck in a poor local minimum, retrain it (cheaply) before scaling up, rather than wasting time on huge models stuck in bad basins.
Using these three techniques, we demonstrated scalable quantum learning on the MNIST dataset (digits 0–3 and 0–9), scaling from 4 to 16 qubits on quantum simulators. We designed an ‘entanglement net’ architecture which consists of entangling nodes between all possible pairs of qubits, as well as a funnel-like architecture that systematically concentrates the relevant information until it is measured. This architecture can encode arbitrary correlations and can approximate any function between the inputs and outputs as the depth increases thus satisfying the universal approximation property.
The paper finds that
Most importantly, this framework gives near-term quantum computers — even ones that can’t yet beat the performance of classical systems on machine learning, but have gone beyond the latter’s simulation capabilities — a crucial role: Training smaller sub-models today prepares better initializations for bigger quantum models tomorrow. Thus, every generation of quantum hardware becomes a stepping stone toward quantum advantage.
This paper is a major step toward building practical, scalable quantum machine learning models.
It offers:
Unlike many approaches that mimic classical machine learning, these methods embrace quantum-native thinking — leveraging sparsity in data encoding, strongly-entangling unitaries, a training strategy that eschews classical optimizers and instead exploits the functional form of the quantum loss, and the fact that entangling nodes can be added to the entanglement net as an identity operator in order to do sub-net initialization. In contrast to earlier studies, it demonstrates that quantum models can be extremely expressive and trainable at the same time. And importantly, it demonstrates that you do not need QRAM (quantum random access memory) or multiple data reuploadings in order to get a model that has the property of universal approximation.Coherent Computing is pioneering high-performance and user-friendly software platforms that unlock the full potential of quantum hardware.
Get in touch with us to learn more.
]]>