The Intuition
An -qubit state has the form:
We typically encode information by choosing which basis states have non-zero amplitudes. But the amplitudes are continuous values—why not encode information in their structure?
Idea: Apply FFT to the amplitude vector, discretize the frequency-domain coefficients into bins , and encode data in this discretized pattern. Could this let us pack more information into fewer qubits?
The Mathematical Reality
Problem 1: The Born Rule
When we measure in the computational basis, we get outcome with probability:
We observe probabilities, not amplitudes. The phase information is lost. To reconstruct the amplitude distribution, we need:
- Many copies of (violates no-cloning for unknown states)
- Or quantum state tomography ( measurements for full reconstruction)
Problem 2: The Holevo Bound
Even if we could prepare a state with rich amplitude structure, Holevo's theorem tells us the accessible classical information is bounded:
For a pure state , we have . The accessible information comes from the ensemble of states we prepare, not from a single state's amplitude structure.
Translation: You can't extract arbitrarily many bits from a single quantum state by clever amplitude engineering.
Problem 3: State Preparation Complexity
Preparing a state with arbitrary amplitudes requires a circuit of depth in the worst case. If our goal is efficiency, this defeats the purpose.
Where This Does Work
Interestingly, similar ideas already exist:
1. Quantum Fourier Transform (QFT)
The QFT maps:
This is encoding information in the frequency domain of amplitudes! Used in Shor's algorithm, phase estimation, etc.
2. Amplitude Amplification
Grover's algorithm and amplitude amplification manipulate amplitudes to increase the probability of measuring the correct answer:
3. Quantum Signal Processing (QSP)
QSP can implement polynomial transformations on amplitude distributions. This is essentially what you're describing—structured manipulation of amplitudes.
A Possible Hybrid Approach?
Here's where it might get interesting:
Amplitude Discretization for Noise-Resilient Encoding
Instead of maximizing information density, what if we discretize amplitudes to create error-correcting structure?
Apply FFT, then constrain frequency-domain coefficients to discrete values. This might give:
- Reduced sensitivity to amplitude damping
- Natural compression (like JPEG for quantum states)
- Easier verification (finite precision arithmetic)
The trade-off: We sacrifice some information capacity for robustness.
Practical Implication: Vocabulary Encoding
Question: If we can't beat the information bounds, do we need as many qubits as classical bits to encode a vocabulary of size ?
Short answer: Yes, fundamentally.
To encode one of distinct tokens:
Classical: bits
Quantum: qubits
Here's why:
Distinguishability Requires Orthogonal States
To reliably distinguish between different tokens, we need orthogonal quantum states. In an -qubit system, we have at most orthogonal basis states. Therefore:
No amplitude tricks change this—orthogonality is a geometric constraint in Hilbert space.
Where Quantum Wins (Not in Storage)
The advantage of quantum computing is not in data compression. It's in:
-
Superposition: A token embedding can be in superposition:
This lets us process multiple possibilities in parallel.
-
Entanglement: Token relationships can be encoded non-locally:
With qubits, you can represent entangled states that have no classical equivalent description shorter than parameters.
-
Interference: Quantum algorithms use constructive/destructive interference to amplify correct answers—this is where speedup comes from, not from packing more data per qubit.
The Trap for Quantum NLP
If you're building a quantum transformer or language model, you might think: "Can I use fewer qubits than classical bits for my vocabulary?"
No. For a vocabulary of 50k tokens, you need:
- Classical: bits
- Quantum: qubits
But: The quantum version can process superpositions of tokens, entangle positional encodings with semantic embeddings in ways classical systems can't, and potentially offer speedups in attention mechanisms through Grover-like searches.
The win is in computation, not representation.
The Verdict
Can we use discretized amplitudes to encode more than bits on qubits?
No—Holevo bound and measurement constraints prevent extracting more classical information than the system's entropy allows.
But: We can use amplitude structure for:
- Algorithmic advantage (QFT, QSP already do this)
- Noise-resilient encoding (discretization as a form of "quantization")
- Quantum machine learning (amplitude patterns as features)
The 2^n scaling is fundamental to how much information we can extract, but how we structure amplitudes still matters for computation and error correction.
Next: Testing Discretized QSP
I'm curious whether variational circuits can learn to prepare "quantized amplitude" states that are more robust to noise. Might be worth implementing a small experiment:
- Parameterized circuit to prepare
- Constraint: amplitudes must lie in discrete bins
- Measure noise resilience vs. standard amplitude encoding
Worth exploring whether discretization helps or hurts in practice.