In quantum information theory, the classical capacity of a quantum channel is the maximum rate at which classical data can be sent over it error-free in the limit of many uses of the channel.
Background
Mixed states and quantum channels
A mixed quantum state is a unit trace,
positive operator known as a density operator, and is often denoted
by
,
,
, etc. The simplest model for a quantum channel
is a classical-quantum channel

which sends the classical letter
at the transmitting end to a quantum state
at the receiving
end, with noise possibly introduced in between. The receiver's task is to perform a measurement to determine the
input of the sender. If the states
are perfectly
distinguishable from one another (i.e., if they have orthogonal supports such
that
for
) and the channel is noiseless, then perfect decoding is trivially possible. If the states
all
commute with each other then the channel is effectively classical.
The situation becomes nontrivial only when the states
have overlapping support and do not necessarily commute.
Quantum measurements
The most general way to describe a quantum measurement is with a
positive operator-valued measure, whose elements are typically denoted as
. These operators should satisfy
positivity and completeness in order to form a valid POVM:


The probabilistic interpretation of quantum mechanics states that if someone
measures a quantum state
using a measurement device corresponding to
the POVM
, then the probability
for obtaining outcome
is equal to

and the post-measurement state is

if the person measuring obtains outcome
.
Classical communication over quantum channels
The above is sufficient to consider a classical classical communication scheme over a cq channel. The sender uses a cq channel to map a classical letter x to a quantum state
, which is then sent through some noisy quantum channel, and then measured using some POVM by the receiver, who obtains another classical letter.
Precise definition
The classical capacity can be defined as the maximum rate achievable by a coding scheme for classical information transmission, which can be defined as follows.[1]
Definition. (Coding scheme) A
-coding scheme for classical information transmission using a quantum channel
is given by pair of an encoding map
and a decoding POVM
such that
with respect to the Hilbert-Schmidt inner product for all
.
Definition. (Achievable rate) A rate
is achievable for the channel
if either
or
and for any
there exists a
-coding scheme such that
and
both hold.
Holevo-Schumacher-Westmoreland theorem
The Holevo information (also called the Holevo
quantity) of a quantum channel
can be defined as

where
is a classical-quantum state of the form

for some probability distribution
and density operators
which can be input to the given channel.
Schumacher and Westmoreland in 1997,[2] and Holevo independently in 1998,[3] proved that the classical capacity of a quantum channel can be equivalently defined as

Gentle measurement lemma
The gentle measurement lemma states that a measurement succeeding with high probability does not disturb the state too much on average.
Lemma. (Winter) Given an
ensemble
with expected
density operator
, suppose
that an operator
with
succeeds with
probability
on the state
:

Then the subnormalized state
is close
in expected trace distance to the original state
:
![{\displaystyle \mathbb {E} _{X}[\left\Vert {\sqrt {\Lambda }}\rho _{X}{\sqrt {\Lambda }}-\rho _{X}\right\Vert _{1}]\leq 2{\sqrt {\epsilon }}.}](./_assets_/eb734a37dd21ce173a46342d1cc64c92/779eb70143f74ce6fb080e187bfedf113b0fa108.svg)
The gentle measurement lemma has the following analog which holds for any operators
,
,
such that
:
 | | 1 |
The quantum information-theoretic interpretation of this inequality is
that the probability of obtaining outcome
from a quantum measurement
acting on the state
is bounded by the sum of the probability of obtaining
on
summed and the distinguishability of
the two states
and
.
Non-commutative union bound
Lemma. (Sen's bound)[4] For a subnormalized state
such that
and
, and for projectors
, ... ,
we have
Intuitively, Sen's bound is a sort of "non-commutative union
bound" because it is analogous to the union bound
from classical probability theory:
where
are events. The analogous quantum bound would be

if we think of
as a projector onto the intersection of
subspaces. However, this only holds if the projectors
,
...,
commute (choosing
,
, and
gives a counterexample). If the projectors are non-commuting, then one must use a non-commutative or quantum union bound.
Proof
We now prove the HSW theorem with Sen's non-commutative union bound. We
first describe how the code is chosen, then give the construction of Bob's POVM,
and finally analyze the error of the protocol.
Encoding map
We first describe how Alice and Bob agree on a
random choice of code. They have the channel
and a
distribution
. They choose
classical sequences
according to the IID distribution
.
After selecting them, they label them with indices as
. This leads to the following
quantum codewords:

The quantum codebook is then
. The average state of the codebook is then
![{\displaystyle \mathbb {E} _{X^{n}}[\rho _{X^{n}}]=\sum _{x^{n}}p_{X^{n}}(x^{n})\rho _{x^{n}}=\rho ^{\otimes n},}](./_assets_/eb734a37dd21ce173a46342d1cc64c92/287fbd9e39696358f615ce411a71e85568b11011.svg) | | 2 |
where
.
Decoding POVM construction
Sen's bound from the above lemma
suggests a method for Bob to decode a state that Alice transmits. Bob should
first ask "Is the received state in the average typical
subspace?" He can do this operationally by performing a
typical subspace measurement corresponding to
. Next, he asks in sequential order,
"Is the received codeword in the
conditionally typical subspace?" This is in some sense
equivalent to the question, "Is the received codeword the
transmitted codeword?" He can ask these
questions operationally by performing the measurements corresponding to the
conditionally typical projectors
.
Why should this sequential decoding scheme work well? The reason is that the
transmitted codeword lies in the typical subspace on average:
![{\displaystyle \mathbb {E} _{X^{n}}[\operatorname {Tr} \Pi _{\rho ,\delta }\rho _{X^{n}}]=\operatorname {Tr} \Pi _{\rho ,\delta }\ \mathbb {E} _{X^{n}}[\rho _{X^{n}}]}](./_assets_/eb734a37dd21ce173a46342d1cc64c92/17b33562beb7789ffe3c27fba337fb5421686294.svg)


where the inequality follows from (\ref{eq:1st-typ-prop}). Also, the
projectors
are "good detectors" for the states
(on average) because the following condition holds from conditional quantum
typicality:
![{\displaystyle \mathbb {E} _{X^{n}}[\operatorname {Tr} \Pi _{\rho _{X^{n}},\delta }\rho _{X^{n}}]\geq 1-\epsilon .}](./_assets_/eb734a37dd21ce173a46342d1cc64c92/fbbd7084f0177cbc646e992cfbfe88112a09a0e1.svg)
Error analysis
The probability of detecting the
codeword correctly under our sequential decoding scheme is equal to

where we make the abbreviation
. (Observe that we
project into the average typical subspace just once.) Thus, the probability of
an incorrect detection for the
codeword is given by

and the average error probability of this scheme is equal to

Instead of analyzing the average error probability, we analyze the expectation
of the average error probability, where the expectation is with respect to the
random choice of code:
![{\displaystyle 1-\mathbb {E} _{X^{n}}\left[{\frac {1}{M}}\sum _{m}\operatorname {Tr} \Pi _{\rho _{X^{n}(m)},\delta }{\hat {\Pi }}_{\rho _{X^{n}(m-1)},\delta }\cdots {\hat {\Pi }}_{\rho _{X^{n}(1)},\delta }\Pi _{\rho ,\delta }^{n}\rho _{x^{n}(m)}\Pi _{\rho ,\delta }^{n}{\hat {\Pi }}_{\rho _{X^{n}(1)},\delta }\cdots {\hat {\Pi }}_{\rho _{X^{n}(m-1)},\delta }\Pi _{\rho _{X^{n}(m)},\delta }\right].}](./_assets_/eb734a37dd21ce173a46342d1cc64c92/b45fe474f0f2658e2e5a6a31d3c1de0719007b5d.svg) | | 3 |
Our first step is to apply Sen's bound to the above quantity. But before doing
so, we should rewrite the above expression just slightly, by observing that
![{\displaystyle 1=\mathbb {E} _{X^{n}}\left[{\frac {1}{M}}\sum _{m}\operatorname {Tr} \rho _{X^{n}(m)}\right]}](./_assets_/eb734a37dd21ce173a46342d1cc64c92/0546277caca1172c78fd3f31c0716478c2729907.svg)
![{\displaystyle =\mathbb {E} _{X^{n}}\left[{\frac {1}{M}}\sum _{m}\operatorname {Tr} \Pi _{\rho ,\delta }^{n}\rho _{X^{n}(m)}+\operatorname {Tr} {\hat {\Pi }}_{\rho ,\delta }^{n}\rho _{X^{n}(m)}\right]}](./_assets_/eb734a37dd21ce173a46342d1cc64c92/97ca65803a383cb4b1853e2cb4e6fa86ee9ffaaa.svg)
![{\displaystyle =\mathbb {E} _{X^{n}}\left[{\frac {1}{M}}\sum _{m}\operatorname {Tr} \Pi _{\rho ,\delta }^{n}\rho _{X^{n}(m)}\Pi _{\rho ,\delta }^{n}\right]+{\frac {1}{M}}\sum _{m}\operatorname {Tr} {\hat {\Pi }}_{\rho \delta }^{n}\mathbb {E} _{X^{n}}[\rho _{X^{n}(m)}]}](./_assets_/eb734a37dd21ce173a46342d1cc64c92/a44c3a3e5de66de41a3b97f822bdc631283d42d6.svg)
![{\displaystyle =\mathbb {E} _{X^{n}}\left[{\frac {1}{M}}\sum _{m}\operatorname {Tr} \Pi _{\rho ,\delta }^{n}\rho _{X^{n}(m)}\Pi _{\rho ,\delta }^{n}\right]+\operatorname {Tr} {\hat {\Pi }}_{\rho ,\delta }^{n}\rho ^{\otimes n}}](./_assets_/eb734a37dd21ce173a46342d1cc64c92/808dbcb1eaea25ed002a376d480567ef576c6a66.svg)
![{\displaystyle \leq \mathbb {E} _{X^{n}}\left[{\frac {1}{M}}\sum _{m}\operatorname {Tr} \Pi _{\rho ,\delta }^{n}\rho _{X^{n}(m)}\Pi _{\rho ,\delta }^{n}\right]+\epsilon }](./_assets_/eb734a37dd21ce173a46342d1cc64c92/139a80ff07a83a236f29dd6aabfed74158d01cef.svg)
Substituting into (3) (and forgetting about the small
term for now) gives an upper bound of
![{\displaystyle \mathbb {E} _{X^{n}}\left[{\frac {1}{M}}\sum _{m}\operatorname {Tr} \Pi _{\rho ,\delta }^{n}\rho _{X^{n}(m)}\Pi _{\rho ,\delta }^{n}\right]}](./_assets_/eb734a37dd21ce173a46342d1cc64c92/8df9eb5725a110e6ccea1f2a1d6279c99153cea6.svg)
![{\displaystyle -\mathbb {E} _{X^{n}}\left[{\frac {1}{M}}\sum _{m}\operatorname {Tr} \Pi _{\rho _{X^{n}(m)},\delta }{\hat {\Pi }}_{\rho _{X^{n}(m-1)},\delta }\cdots {\hat {\Pi }}_{\rho _{X^{n}(1)},\delta }\Pi _{\rho ,\delta }^{n}\rho _{x^{n}(m)}\Pi _{\rho ,\delta }^{n}{\hat {\Pi }}_{\rho _{X^{n}(1)},\delta }\cdots {\hat {\Pi }}_{\rho _{X^{n}(m-1)},\delta }\Pi _{\rho _{X^{n}(m)},\delta }\right].}](./_assets_/eb734a37dd21ce173a46342d1cc64c92/21694264e67eb757a19cd628cc474ed9ae88c282.svg)
We then apply Sen's bound to this expression with
and the sequential
projectors as
,
, ...,
. This gives the upper bound
Due to concavity of the square root, we can bound this expression from above
by
![{\displaystyle 2\left(\mathbb {E} _{X^{n}}\left[{\frac {1}{M}}\sum _{m}\operatorname {Tr} (I-\Pi _{\rho _{X^{n}(m)},\delta })\Pi _{\rho ,\delta }^{n}\rho _{X^{n}(m)}\Pi _{\rho ,\delta }^{n}+\sum _{i=1}^{m-1}\operatorname {Tr} \Pi _{\rho _{X^{n}(i)},\delta }\Pi _{\rho ,\delta }^{n}\rho _{X^{n}(m)}\Pi _{\rho ,\delta }^{n}\right]\right)^{1/2}}](./_assets_/eb734a37dd21ce173a46342d1cc64c92/5696764f2824c8cdfa03b22cf8c3a12ea8dabc19.svg)
![{\displaystyle 2\left(\mathbb {E} _{X^{n}}\left[{\frac {1}{M}}\sum _{m}\operatorname {Tr} (I-\Pi _{\rho _{X^{n}(m)},\delta })\Pi _{\rho ,\delta }^{n}\rho _{X^{n}(m)}\Pi _{\rho ,\delta }^{n}+\sum _{i\neq m}\operatorname {Tr} \Pi _{\rho _{X^{n}(i)},\delta }\Pi _{\rho ,\delta }^{n}\rho _{X^{n}(m)}\Pi _{\rho ,\delta }^{n}\right]\right)^{1/2}}](./_assets_/eb734a37dd21ce173a46342d1cc64c92/2867df17b8a5131ef0647fab2291625373026f49.svg)
where the second bound follows by summing over all of the codewords not equal
to the
codeword (this sum can only be larger).
We now focus exclusively on showing that the term inside the square root can
be made small. Consider the first term:
![{\displaystyle \mathbb {E} _{X^{n}}\left[{\frac {1}{M}}\sum _{m}\operatorname {Tr} (I-\Pi _{\rho _{X^{n}(m)},\delta })\Pi _{\rho ,\delta }^{n}\rho _{X^{n}(m)}\Pi _{\rho ,\delta }^{n}\right]}](./_assets_/eb734a37dd21ce173a46342d1cc64c92/791163946b4ead05b1e51cc3879d668b46ecf386.svg)
![{\displaystyle \leq \mathbb {E} _{X^{n}}\left[{\frac {1}{M}}\sum _{m}\operatorname {Tr} (I-\Pi _{\rho _{X^{n}(m)},\delta })\rho _{X^{n}(m)}+\left\Vert \rho _{X^{n}(m)}-\Pi _{\rho ,\delta }^{n}\rho _{X^{n}(m)}\Pi _{\rho ,\delta }^{n}\right\Vert _{1}\right]}](./_assets_/eb734a37dd21ce173a46342d1cc64c92/0abfaf12ce258b12b2beeb17c3ebe14297c75fbc.svg)

where the first inequality follows from (1) and the
second inequality follows from the gentle operator lemma and the
properties of unconditional and conditional typicality. Consider now the
second term and the following chain of inequalities:
![{\displaystyle \sum _{i\neq m}\mathbb {E} _{X^{n}}[\operatorname {Tr} \Pi _{\rho _{X^{n}(i)},\delta }\Pi _{\rho ,\delta }^{n}\rho _{X^{n}(m)}\Pi _{\rho ,\delta }^{n}]}](./_assets_/eb734a37dd21ce173a46342d1cc64c92/264c03a037c839ac2add6becb201a851d31d04d2.svg)
![{\displaystyle =\sum _{i\neq m}\operatorname {Tr} \mathbb {E} _{X^{n}}[\Pi _{\rho _{X^{n}(i)},\delta }]\Pi _{\rho ,\delta }^{n}\mathbb {E} _{X^{n}}[\rho _{X^{n}(m)}]\Pi _{\rho ,\delta }^{n}}](./_assets_/eb734a37dd21ce173a46342d1cc64c92/93179f03d13ac614cfaacfb2fc9493c1ce1af9c0.svg)
![{\displaystyle =\sum _{i\neq m}\operatorname {Tr} \mathbb {E} _{X^{n}}[\Pi _{\rho _{X^{n}(i)},\delta }]\Pi _{\rho ,\delta }^{n}\rho ^{\otimes n}\Pi _{\rho ,\delta }^{n}}](./_assets_/eb734a37dd21ce173a46342d1cc64c92/7a95cf25543e5390748384957305d6a28468d190.svg)
![{\displaystyle \leq \sum _{i\neq m}2^{-n\lfloor H(B)-\delta \rfloor }\operatorname {Tr} \mathbb {E} _{X^{n}}[\Pi _{\rho _{X^{n}(i)},\delta }]\Pi _{\rho ,\delta }^{n}}](./_assets_/eb734a37dd21ce173a46342d1cc64c92/fddce85f33c0828c862c97a39ee6c8e247d182ab.svg)
The first equality follows because the codewords
and
are independent since they are different. The second
equality follows from (2). The first inequality follows from
(\ref{eq:3rd-typ-prop}). Continuing, we have
![{\displaystyle \leq \sum _{i\neq m}2^{-n\lfloor H(B)-\delta \rfloor }\mathbb {E} _{X^{n}}[\operatorname {Tr} \Pi _{\rho _{X^{n}(i)},\delta }]}](./_assets_/eb734a37dd21ce173a46342d1cc64c92/5f743b24e6c6682d698494bb3ebf8d85e05ddca7.svg)



The first inequality follows from
and exchanging
the trace with the expectation. The second inequality follows from
(\ref{eq:2nd-cond-typ}). The next two are straightforward.
Putting everything together, we get our final bound on the expectation of the
average error probability:
![{\displaystyle 1-\mathbb {E} _{X^{n}}\left[\operatorname {Tr} \Pi _{\rho _{X^{n}(m)},\delta }{\hat {\Pi }}_{\rho _{X^{n}(m-1)},\delta }\cdots {\hat {\Pi }}_{\rho _{X^{n}(1)},\delta }\Pi _{\rho ,\delta }^{n}\rho _{x^{n}(m)}\Pi _{\rho ,\delta }^{n}{\hat {\Pi }}_{\rho _{X^{n}(1)},\delta }\cdots {\hat {\Pi }}_{\rho _{X^{n}(m-1)},\delta }\Pi _{\rho _{X^{n}(m)},\delta }\right]}](./_assets_/eb734a37dd21ce173a46342d1cc64c92/548955161d319acefca17a8fb5d876410f081ad3.svg)

Thus, as long as we choose
, there exists a code with vanishing error probability.
Non-additivity of the classical capacity
The HSW theorem can be seen as expressing the classical capacity of a channel
in terms of a regularization of the Holevo
-quantity over multiple uses of
. An open problem in quantum information theory was to determine if the
-quantity is additive, which would imply that the classical capacity could be expressed using a single use of
.[5] However, channels giving counterexamples to this statement were eventually given by Matthew Hastings in 2009.[6] Follow-up work showed that this is a generic phenomenon, in the sense that a channel chosen randomly from a natural probability distribution will give a counterexample with high probability. (This stands in contrast to proofs using the probabilistic method, where random sampling is shown to give a counterexample only with nonzero probability.) A proof of this can be given using Dvoretzky's theorem.[5]
Minimal output entropy
Non-additivity of the classical capacity is closely related to non-additivity of the minimal von Neumann entropy of the output of a quantum channel. An easier problem is to consider the minimal output quantum Rényi entropy for
, for which simple counterexamples using the inherent entanglement of fermions were given by Grudka, Horodecki, and Pankowski.[7]
See also
References
- Wilde, Mark M. (2017), Quantum Information Theory, Cambridge University Press, arXiv:1106.1445, Bibcode:2011arXiv1106.1445W, doi:10.1017/9781316809976.001, S2CID 2515538
- Guha, Saikat; Tan, Si-Hui; Wilde, Mark M. (2012), "Explicit capacity-achieving receivers for optical communication and quantum reading", IEEE International Symposium on Information Theory Proceedings (ISIT 2012), pp. 551–555, arXiv:1202.0518, doi:10.1109/ISIT.2012.6284251, ISBN 978-1-4673-2579-0, S2CID 8786400.
Notes
- ^ "Lecture 11: The classical capacity of a quantum channel" (PDF).
- ^ Schumacher, Benjamin; Westmoreland, Michael (1997), "Sending classical information via noisy quantum channels", Phys. Rev. A, 56 (1): 131–138, Bibcode:1997PhRvA..56..131S, doi:10.1103/PhysRevA.56.131
- ^ Holevo, Alexander S. (1998), "The Capacity of Quantum Channel with General Signal States", IEEE Transactions on Information Theory, 44 (1): 269–273, arXiv:quant-ph/9611023, doi:10.1109/18.651037
- ^ Sen, Pranab (2012), "Achieving the Han-Kobayashi inner bound for the quantum interference channel by sequential decoding", IEEE International Symposium on Information Theory Proceedings (ISIT 2012), pp. 736–740, arXiv:1109.0802, doi:10.1109/ISIT.2012.6284656, S2CID 15119225
- ^ a b Aubrun, Guillaume; Szarek, Stanisław; Werner, Elisabeth (2011). "Hastings's Additivity Counterexample via Dvoretzky's Theorem". Communications in Mathematical Physics. 305 (1): 85–97. arXiv:1003.4925. doi:10.1007/s00220-010-1172-y. ISSN 0010-3616.
- ^ Hastings, M. B. (2009-03-15). "Superadditivity of communication capacity using entangled inputs". Nature Physics. 5 (4). Springer Science and Business Media LLC: 255–257. arXiv:0809.3972. doi:10.1038/nphys1224. ISSN 1745-2473.
- ^ Grudka, Andrzej; Horodecki, Michał; Pankowski, Łukasz (2010-10-22). "Constructive counterexamples to the additivity of the minimum output Rényi entropy of quantum channels for all p > 2". Journal of Physics A: Mathematical and Theoretical. 43 (42) 425304. arXiv:0911.2515. doi:10.1088/1751-8113/43/42/425304. ISSN 1751-8113. Retrieved 2025-12-15.