Download Quantum Probability - Institut Camille Jordan

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the work of artificial intelligence, which forms the content of this project

Document related concepts

Quantum computing wikipedia , lookup

Hilbert space wikipedia , lookup

Scalar field theory wikipedia , lookup

Relativistic quantum mechanics wikipedia , lookup

Copenhagen interpretation wikipedia , lookup

Quantum machine learning wikipedia , lookup

Theoretical and experimental justification for the Schrödinger equation wikipedia , lookup

Quantum teleportation wikipedia , lookup

History of quantum field theory wikipedia , lookup

Quantum entanglement wikipedia , lookup

Quantum decoherence wikipedia , lookup

Interpretations of quantum mechanics wikipedia , lookup

Coherent states wikipedia , lookup

Self-adjoint operator wikipedia , lookup

Quantum group wikipedia , lookup

Path integral formulation wikipedia , lookup

Measurement in quantum mechanics wikipedia , lookup

EPR paradox wikipedia , lookup

T-symmetry wikipedia , lookup

Bell's theorem wikipedia , lookup

Quantum key distribution wikipedia , lookup

Density matrix wikipedia , lookup

Bra–ket notation wikipedia , lookup

Hidden variable theory wikipedia , lookup

Symmetry in quantum mechanics wikipedia , lookup

Compact operator on Hilbert space wikipedia , lookup

Quantum state wikipedia , lookup

Quantum electrodynamics wikipedia , lookup

Canonical quantization wikipedia , lookup

Probability amplitude wikipedia , lookup

Transcript
Lecture 7
QUANTUM PROBABILITY
Stéphane ATTAL
Abstract Bell Inequalities and the failure of hidden variable approaches
show that random phenomena of Quantum Mechanics cannot be modeled
by classical Probability Theory. The aim of Quantum Probability Theory is
to provide an extension of the classical theory of probability which allows
to describe those quantum phenomena. Quantum Probability Theory does
not add anything to the axioms of Quantum Mechanics, it just emphasizes
the probabilistic nature of them. This lecture is devoted to introducing this
quantum extension of Probability Theory, its connection with the quantum
mechanical axioms. We also go a step further by introducing the Toy Fock
spaces and by showing how they hide very rich quantum and classical probabilistic structures.
For reading this chapter the reader should be familiar with the basic notions of Quantum Mechanics, but also of Probability Theory and discrete
time stochastic processes. If necessary read Lectures 4 and 5.
Stéphane ATTAL
Institut Camille Jordan, Université Lyon 1, France
e-mail: [email protected]
1
2
Stéphane ATTAL
7.1
Observables and Laws . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
7.1.1
States and Observables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
7.1.2
Observables vs Random Variables . . . . . . . . . . . . . . . . . . . . . . . . .
7.1.3
The Role of Multiplication Operators . . . . . . . . . . . . . . . . . . . . . .
7.1.4
Events . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
7.2
Quantum Bernoulli . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
7.2.1
Quantum Bernoulli Random Variables . . . . . . . . . . . . . . . . . . . . .
7.2.2
The Toy Fock Space . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
7.2.3
Probabilistic Interpretations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
7.2.4
Brownian Motion and Poisson Process . . . . . . . . . . . . . . . . . . . . .
7.3
Higher Level Chains . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
7.3.1
Structure of Higher Level Chains . . . . . . . . . . . . . . . . . . . . . . . . . .
7.3.2
Obtuse Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
7.3.3
Obtuse Random Variables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
7.3.4
Generic Character of Obtuse Random Variables . . . . . . . . . . . . .
7.3.5
Doubly-symmetric 3-tensors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
7.3.6
The Main Diagonalization Theorem . . . . . . . . . . . . . . . . . . . . . . . .
7.3.7
Back to Obtuse Random Variables . . . . . . . . . . . . . . . . . . . . . . . . .
7.3.8
Probabilistic Interpretations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
2
2
5
9
10
12
12
15
19
23
26
26
27
30
33
34
36
41
42
46
7.1 Observables and Laws
There exists plenty of different definitions of a quantum probability space in
the literature, with different levels of generality. Most commonly Quantum
Probability Theory is defined at the level of von Neumann algebras and normal states. In this lecture we consider a lower level of generality in order to
stick to the most usual axioms of Quantum Mechanics, that is, the Hilbert
space level, where states are density matrices and observables are self-adjoint
operators.
7.1.1 States and Observables
Definition 7.1. A quantum probability space is a pair (H, ρ) where H is a
separable Hilbert space and ρ is a density matrix on H. Such a ρ on H is
called a state on H. The set of states on H is denoted by S(H).
Proposition 7.2. The set of states S(H) is convex. The extremal points of
S(H) are the pure states ρ = |uihu| .
Proof. The convexity of the set S(H) is obvious, let us characterize the extremal points of this set. Recall that any state ρ on H admits a canonical
decomposition
X
ρ=
λn |φn ihφn | ,
n∈N
7 QUANTUM PROBABILITY
3
P
where the λn are positive with n λn = 1 and the φn ’s form an orthonormal
basis of H. From this representation it is clear that if ρ is extremal then
it is of the form |uihu|. Conversely, take ρ = |uihu| and assume that ρ =
λρ1 + (1 − λ)ρ2 is a convex combination of two states. Put P = I − |uihu|, so
that Tr (ρP
P) = 0 and thus Tr (ρ1 P) = Tr (ρ2 P) = 0. If for every i = 1, 2 we
put ρi = j λij |uij ihuij |, we then get
Tr (ρi P) =
X
λij huij , Puij i = 0 ,
j
P
that is, j λij |huij , ui|2 = 1. The only possibility is that only one of λij is not
null (and thus equal to 1) and the corresponding uij is equal to u. This gives
ρ1 = ρ2 . t
u
Definition 7.3. An observable on H is a self-adjoint operator on H. The
space of observables on H is denoted by O(H).
Let A be an observable on H. By the von Neumann Spectral Theorem
there exists a spectral measure ξ associated to A such that
Z
A=
λ dξ(λ) .
R
If ρ is a state on H then the mapping
µ : Bor(R) −→
[0, 1]
E
7−→ Tr (ρ ξ(E))
is a probability measure on R. This probability measure µ is called the law,
or distribution of A under the state ρ.
The following proposition gives another characterization of the distribution
µ of A.
Proposition 7.4. Let (H, ρ) be a quantum probability space and A be an
observable on H. Then the law of A under the state ρ is the only probability
measure µ on R such that, for all bounded Borel function f on R we have
Z
f (x) dµ(x) = Tr ρ f (A) .
(7.1)
R
Proof. Let µ be the law of A. We write
X
ρ=
λn |un ihun |
n∈N
and as a consequence
X
Tr ρ f (A) =
λn hun , f (A) un i
n∈N
4
Stéphane ATTAL
where the series is absolutely convergent. Let ξ denote the spectral measure of
A and let µn be the measure given by µn (E) = hun , ξ(E) un i = kξ(E) un k2 ,
for all n. We then have
Z
X
Tr ρ f (A) =
λn
f (x) dµn (x) .
n∈N
R
On the other hand, if µ is the law of A we have
X
X
µ(E) =
λn hun , ξ(E) un i =
λn µn (E) ,
n∈N
n∈N
P
that is, µ is the measure
n∈N λn µn . This proves that the relation (7.1)
holds when µ is the law of A.
Conversely, if Equation (7.1) holds for a measure µ, then automatically
µ(E) = Tr (ρ 1lE (A)) = Tr (ρ ξ(E)) ,
for all E ∈ Bor(R). This says exactly that µ is the law of A. t
u
Another nice characterization is obtained in terms of the Fourier transform.
Theorem 7.5. Let (H, ρ) be a quantum probability space and let A be an
observable on H. Then the function
f : R −→
R
t 7−→ Tr (ρ eitA )
is the Fourier transform of some probability measure µ on R. This measure
is the law of A under the state ρ.
Proof. For any λ1 , . . . , λn ∈ C and t1 , . . . , tn ∈ R we have
n
X
λk λj f (tk − tj ) =
k,j=1
=
n
X
k,j=1
n
X
λk λj Tr ρ ei(tk −tj )A
λk λj Tr ρ eitk A eitj A
∗ k,j=1

= Tr ρ
n
X
k=1
∗ 
! n
X
λk eitk A 
λj eitj w A   .
j=1
As ρ is positive, the quantity above is positive. By Bochner’s criterion this
proves that f is the Fourier transform of some probability measure µ on R.
By Proposition 7.4 and by uniqueness of the Fourier transform, this measure
is the law of A. t
u
7 QUANTUM PROBABILITY
5
We now make a remark concerning the action of unitary conjugation in
this setup.
Proposition 7.6. Let ρ be a state and A be an observable on H. If U is a
unitary operator on H, then U∗ A U is still an observable and U ρ U∗ is still
a state. Furthermore, the law of U∗ A U under the state ρ is the same as the
law of A under the state U ρ U∗ .
Proof. The facts that U∗ A U is still an observable and U ρ U∗ is still a state
are obvious and left to the reader. We have
Tr ρ U∗ A U = Tr U ρ U∗ A) .
Furthermore, for any bounded function f on R we have, by the functional
calculus
f (U∗ A U) = U∗ f (A) U ,
which gives finally
Tr (ρ f (U∗ A U)) = Tr (U ρ U∗ f (A)) .
By Proposition 7.4, this says exactly that the law of U∗ A U under the state
ρ is the same as the law of A under the state U ρ U∗ . t
u
7.1.2 Observables vs Random Variables
We stop for a moment our list of definitions and properties and we focus on
the relations between observables and usual random variables in Probability
Theory.
For the simplicity of the discussion we consider only the case of pure states,
the general case can be easily deduced by a procedure called the G.N.S.
representation, which is not worth developing for this discussion.
A Hilbert space H and a wave function ϕ being given, let us see that any
observable on H can be viewed as a classical random variable. Let A be an
observable on H, i.e. a self-adjoint operator on H. Consider the probability
measure
2
E 7→ µ(E) = k1lE (A) ϕk
on (R, Bor(R)). Consider the operator U from L2 (R, Bor(R), µ) to H given
by Uf = f (A) ϕ. This is a well-defined operator and an isometry, for the von
Neumann Spectral Theorem gives
Z
2
2
kf (A) ϕk =
|f (x)| dµ(x) .
R
We deduce easily from this definition
6
Stéphane ATTAL
U∗ f (A) ϕ = f
and
U∗ ϕ = 1l .
The operator U∗ A U is then clearly equal to the operator MX of multiplication by the function X(x) = x on R. Furthermore, we have
h ϕ , f (A) ϕ i = h U∗ ϕ , U∗ f (A)U U∗ ϕ i
= h 1l , Mf (X) 1l i
Z
=
f (X(x)) dµ(x) .
R
This means that the probability distribution associated to the observable A,
in the state ϕ, as described in the setup of Quantum Probability Theory, is
the same as the usual probability distribution of X as a random variable on
(R, Bor(R), µ).
Being given a single observable A on H together with a state ϕ is thus the
same as being given a classical random variable X on some probability space
(Ω, F, P). The observable A specifies the random variable X as a function on
Ω, the state ϕ specifies the underlying probability measure P and therefore
specifies the probability distribution of X.
In the converse direction, it is easy to see a classical probabilistic setup
as a particular case of the quantum setup. Let X be a real random variable
defined on some probability space (Ω, F, P). Put H = L2 (Ω, F, P) and ϕ = 1l.
The operator A = MX is then a self-adjoint operator on H. Let us compute
the probability distribution of the observable A in the state ϕ, in the sense
of Quantum Probability. We get
h ϕ , f (A) ϕ i = h 1l , Mf (X) 1l i
Z
=
f (X(ω)) dP(ω)
R
= E[f (X)] .
Thus, the distribution of the observable A, in the quantum sense is the same
as the probability distribution of the random variable X in the usual sense.
One could be tempted to conclude that there is no difference between
the probabilistic formalism of Quantum Mechanics and the usual Probability
Theory. But this is only true for one observable, or actually for a family
of commuting observables. Indeed, once considering two observables A and
B on H which do not commute, the picture is completely different. Each of
the observables A and B can be seen as concrete random variables X and Y
respectively, but on some different probability spaces. Indeed, if A and
B do not commute they cannot be diagonalized simultaneously and they are
7 QUANTUM PROBABILITY
7
represented as multiplication operators on some different probability spaces
(otherwise they would commute!).
Furthermore, their associated probability spaces have nothing to do together. They cannot be put together, as one usually does in the case of independent random variables, by taking the tensor product of the two spaces,
because the two random variables X and Y are not independent. As operators on H the observables A and B may have some strong relations, such
as [A, B] = λI, for example (this is the case for momentum and position
observables in Quantum Mechanics).
Is there a way, with the random variables X, Y and their associated probability spaces, to give an account of such a relation? Aspect’s experiment and
Bell’s inequalities prove that this is actually impossible! As an example, the
spin of an electron in two different directions gives rise to two Bernoulli random variables but each one on its own probability space. These two Bernoulli
random variables are not at all independent, they depend of each other in a
way which is impossible to express in classical terms. The only way to express
their dependency is to represent them as 2 × 2 spin matrices on C2 and work
with the quantum mechanical axioms!
As a conclusion, each single observable in Quantum Mechanics is like a
concrete random variable, but on its own probability space, “with its own
dice”. Two non-commuting observables have different associated probability
spaces, “different dices”; they cannot be represented by independent random
variables or anything else (even more complicated) in the classical language of
probability. The only way to express their mutual dependency is to consider
them as two self-adjoint operators on the same Hilbert space and to compute
their probability distributions as the quantum mechanical axioms dictate us.
In the quantum formalism it is possible to make operations on non commuting observables A, B, for example to add them and to consider the result
as a new observable. The distribution of A + B has then nothing to do with
the convolution of the distribution of A by the one of B, it depends in a complicated way on the relations between A and B as operators on H. We shall
see examples later on in this course.
The case of commuting observables is equivalent to usual case of a pair of
random variables. Let us here give some details about that.
Let A and B be two observables on H which commute (in the case of unbounded operators this means that their spectral measures commute). Then
there exists a spectral measure dξ(x, y) on R × R such that
Z
Z
A=
x dξ(x, y) and B =
y dξ(x, y) .
R2
R2
In other words, A and B can be diagonalized simultaneously, they can be
represented as multiplication operators on the same probability space.
The pair (A, B) admits a law in the quantum sense. Indeed, if ρ is a state
on H, then the mapping
8
Stéphane ATTAL
µ : Bor(R × R) −→
[0, 1]
E×F
7−→ Tr ρ ξ(E, F )
extends to a probability measure on R × R. This probability measure is called
the law of the pair (A, B) under the state ρ. Another way to understand this
law is to say that if A and B commute they admit a two variable functional
calculus and, for every bounded measurable function f on R × R, we have
Z
f (x, y) dµ(x, y) = Tr (ρ f (A, B)) .
R2
From the Fourier transform point of view, the law µ is obtained by noticing
that if A and B commute then the function
(t1 , t2 ) 7−→ Tr ρ ei(t1 A+t2 B)
satisfies Bochner’s criterion for two variable functions and hence is the Fourier
transform of some probability measure µ on R × R. Clearly, this measure is
the law of (A, B) under ρ as defined above.
At this stage, it is interesting to note a way to produce independent random
variables in the quantum probability context.
Proposition 7.7. Let A and B be two observables on H and K respectively.
Let µ and ν denote their laws under the states ρ and τ , respectively. On the
Hilbert space H ⊗ K consider the commuting observables
b =A⊗I
A
b = I ⊗ B.
B
and
b B)
b follows the law
Then, under the state ρ ⊗ τ , the pair of observables (A,
µ ⊗ ν, that is, they are independent observables with the same individual law
as A and B respectively.
b and B
b commute, hence they admit a law of pair, let us
Proof. Clearly A
compute this law. For every bounded real functions f and g we have
b = f (A) ⊗ I
f (A)
b = I ⊗ g(B) .
g(B)
and
Thus
b B)
b = Tr ρ f (A) ⊗ τ g(B)
Tr (ρ ⊗ τ ) f (A)g(
= Tr ρ f (A) Tr τ g(B)
Z
Z
=
f (x) dµ(x)
g(y) dν(y)
R
ZR Z
=
f (x)g(y) dµ(x) dν(y) .
R
R
7 QUANTUM PROBABILITY
9
b has the same law under ρ ⊗ τ
Taking g = 1l, the identity above shows that A
as the one of A under the state ρ. The corresponding property of course holds
for B. Taking general f and g, the identity above shows the independence of
b and B.
b t
A
u
7.1.3 The Role of Multiplication Operators
Together with the discussion we had above on the connection between observables and classical random variables, I want to put the emphasis here on
the particular role of multiplication operators in the quantum setup.
Consider a probability space (Ω, F, P) and a family of real random variables on (Xi )i∈I which are of interest for some reasons. Imagine we have
found a natural unitary isomorphism U : L2 (Ω, F, P) → H with some Hilbert
space H. The question we want to discuss here is the following: Where can
we read on H the random variables Xi , with all their probabilistic properties
(individual distributions, relations with each other such as independence or
dependence, joint distributions, functional calculus, etc.)?
This is certainly not by looking at the images hi = UXi of the Xi ’s in H.
Indeed, these elements hi of H could be almost any element of H by choosing
correctly the unitary operator U. They carry no probabilistic information
whatsoever on the random variables Xi .
The pertinent objects to look at are the operators
Ai = U MXi U∗ ,
that is, the push-forward of the operators of multiplications by the Xi ’s.
Indeed, these operators are a commuting family of self-adjoint operators on
H. The functional calculus for such families of operators gives easily
f (Ai1 , . . . , Ain ) = U Mf (Xi1 ,...,Xin ) U∗ .
If we put φ = U1l, then φ is a pure state of H and we have
Tr (|φihφ| f (Ai1 , . . . , Ain )) = hφ , f (Ai1 , . . . , Ain )φi
= h1l , U∗ f (Ai1 , . . . , Ain ) U 1li
D
E
= 1l , Mf (Xi1 ,...,Xin ) 1l
= E [f (Xi1 , . . . , Xin )] .
The law of the n-uplet (Ai1 , . . . , Ain ) as observables in the state |φihφ| is thus
the same as the usual law of the n-uplet (Xi1 , . . . , Xin ).
The operators Ai contain all the probabilistic informations on the Xi ’s:
individual and joint laws, functional calculus, etc. The operators Ai play
exactly the same role in H as the random variables Xi play in L2 (Ω, F, P).
10
Stéphane ATTAL
The only informations concerning the Xi ’s that one cannot recover from
the Ai ’s are the samples, that is, the individual values Xi (ω) for ω ∈ Ω.
We end this discussion with an extension to a situation which is very
common in Quantum Noise Theory. Together with the situation described
above, we have another Hilbert space K and an operator on K ⊗ H of the
form
n
X
T=
Bi ⊗ Ai ,
i=1
for some bounded operators Bi on K. On the other hand, consider the random
operator
S : Ω −→
B(K)
P
n
ω 7−→ S(ω) = i=1 Xi (ω)Bi .
We want to make clear the connections between T and S and show that they
are one and the same thing.
Let us define the “multiplication” operator associated to S:
b
S : L2 (Ω, F, P) ⊗ H −→ P
L2 (Ω, F, P) ⊗ H
n
f ⊗h
7−→ i=1 Xi f ⊗ Bi h .
In other words
b
S=
n
X
MXi ⊗ Bi .
i=1
Then the connection becomes clear:
T = (U ⊗ I) b
S (U∗ ⊗ I) .
Even more than just this unitary equivalence, the operator T contains all
the probabilistic informations of the random operator S; any kind of computation that one wants to perform on S can be made in the same way on
T. The functional calculus is the same, the expectations and the laws are
computed is the same way as described above.
7.1.4 Events
In this subsection we discuss the notion of event in Quantum Probability.
In classical probability theory an event E can be identified to the random
variable X = 1lE . That is, a random variable X such that X 2 = X = X.
Conversely, every random variable X which satisfies the above relation is
the indicator function of some event. On the space L2 (Ω, F, P), the indicator
function X = 1lE can be identified with its multiplication operator f 7→ 1lE f ,
7 QUANTUM PROBABILITY
11
it is thus an orthogonal projection of L2 (Ω, F, P), whose range is the subspace
of functions with support included in E.
Definition 7.8. Following the analogy with the classical case, we define an
event on a quantum probability space (H, ρ) to be any orthogonal projector
on H. We denote by P(H) the set of events of H. In particular, if A is an
observable on H, with spectral measure ξA and if E is a Borel set of R, the
projector ξA (E) can be interpreted as the event “A belongs to E”. Up to a
normalization constant it is the state one would obtain after a measurement of
the observable A, if one had observed “the value of the measurement belongs
to the set E”. In other words, the subspace Ran ξA (E) is the subspace of wave
functions for which a measure of A would give a result in E with probability
1; this subspace is indeed the subspace corresponding to the knowledge “A
belongs to E ” on the system.
This terminology coincides also with the definitions and results of previous
subsections, for by Proposition 7.4 we have
Z
Tr (ρ ξA (E)) = Tr (ρ 1lE (A)) =
1lE (x) dµ(x) = µ(E)
R
which says that Tr (ρ ξA (E)) is the probability that A belongs to E under the
state ρ.
Definition 7.9. Together with this quantum notion of event we make use
the following terminology:
– If P1 , P2 are two events and if P1 ≤ P2 , we say that P1 implies P2 .
– The operators 0 and I are the null and certain events.
– The complementary of the event P is the event I − P.
– If P1 , . . . , Pn are events then ∨ni=1 Pi is the orthogonal projector onto the
subspace generated by ∪ni=1 Ran Pi ; it is the event “occurrence of at least
one of the Pi ’s”. In the same way ∧ni=1 Pi is the orthogonal projector onto
∩ni=1 Ran Pi ; it is the event “simultaneous occurrence of all the Pi ’s”.
Definition 7.10. When a state ρ is given, the mapping α : P 7→ Tr (ρ P)
behaves like a probability measure on P(H) in the sense that
i) α(I) = 1
P
P
ii) α
i∈N Pi =
i∈N α(Pi ), for any family (Pi )i∈N of pairwise orthogonal
projections.
A mapping α : P(H) → R which satisfies i) and ii) is called a probability
measure on P(H). Note that every probability measure α on P(H) satisfies:
iii) α(P) ∈ [0, 1] for all P ∈ P(H).
iv) α(I − P) = 1 − α(P) for all P ∈ P(H).
v) α(0) = 0.
12
Stéphane ATTAL
One may naturally wonder what is the general form of all probability
measures on P(H). The answer is given by the celebrated Gleason’s theorem.
This theorem is very deep and has many applications, developments and
consequences in the literature. Its proof is very long and heavy. As we do not
really need it in the following (we just present it as a remark), we state it
without proof. One may consult the book [Par92], pp. 31–40, for a complete
proof.
Theorem 7.11 (Gleason’s theorem). Let H be a separable Hilbert space with
dimension greater than or equal to 3. A mapping α : P(H) → R is a probability measure on P(H) if and only if there exists a density matrix ρ on H
such that
α(P) = Tr (ρ P)
for all P ∈ P(H).
The correspondence ρ 7→ α between density matrices on H and probability
measures on P(H) is a bijection.
7.2 Quantum Bernoulli
We shall now enter into concrete examples of Quantum Probability spaces and
specific properties of this quantum setup. In this section we aim to explore
the simplest situation possible, the quantum analogue of a Bernoulli random
walk. We shall see that this simple situation in classical probability becomes
incredibly richer in the quantum context.
7.2.1 Quantum Bernoulli Random Variables
Before going to the Bernoulli random walks, we study the quantum analogue
of Bernoulli random variables. This is to say that we study now the simplest
non-trivial example of a quantum probability space: the space H = C2 .
Definition 7.12. Let (e1 , e2 ) be the canonical basis of C2 . The matrices
01
0 −i
1 0
σ1 =
, σ2 =
, σ3 =
10
i 0
0 −1
are the so-called Pauli matrices. There are also usually denoted by σx , σy , σz ,
respectively.
With the help of these matrices we have very useful parametrizations of
observables and states.
7 QUANTUM PROBABILITY
13
Proposition 7.13.
1) Together with the identity matrix I, the Pauli matrices form a real basis
of the space O(H). That is, any observable A on H can be written
A = tI +
3
X
xi σi
i=1
for some t, x1 , x2 , x3 ∈ R. In particular we get Tr (A) = 2t.
2) Putting
q
x21 + x22 + x23 ,
the eigenvalues of A are t − kxk , t + kxk .
kxk =
3) Any state ρ is of the form
ρ=
1
2
1 + x y − iz
y + iz 1 − x
with x, y, z ∈ R satisfying x2 + y 2 + z 2 ≤ 1. In particular, ρ is a pure state if
and only if x2 + y 2 + z 2 = 1.
Proof. 1) and 2) are immediate and left to the reader. Let us prove 3). A
state ρ on H is an observable which has trace 1 and positive eingenvalues.
Hence by 1) and 2), a state ρ on H is of the form
!
3
X
1
xi σi
ρ=
I+
2
i=1
with kxk ≤ 1. As a consequence the space S(H) identifies to B(0, 1), the
closed unit ball of R3 , with the same convex structure. As a consequence, the
extreme points of S(H), the pure states, correspond to the extreme points of
B(0, 1), that is, the unit sphere S 2 of R3 . t
u
The result 3) above means that one can always write a pure state as:


1 + cos(ϕ) e−iθ sin(ϕ)
1

ρ= 
2
eiθ sin(ϕ) 1 − cos(ϕ)

=
cos2
eiθ sin
ϕ
2
ϕ
2
e−iθ sin
cos
ϕ
2
ϕ
2
sin2
cos
ϕ
2
ϕ
2

,
for some θ ∈ [0, 2π] and ϕ ∈ [0, π]. We recognize the operator |uihu| where
14
Stéphane ATTAL
θ
u=
e−i 2 cos
θ
ei 2 sin
!
ϕ
2
ϕ
2
.
This way, one cleary describes all unitary vectors of C2 up to a phase factor,
that is, one describes all rank one projectors |uihu|.
We shall now compute the law of quantum observables in this context.
Proposition 7.14. Consider an orthonormal basis of C2 for which the state
ρ is diagonal:
p 0
ρ=
,
0q
with 0 ≤ q ≤ p ≤ 1 and p + q = 1. If A is any observable
A = tI +
3
X
xi σi ,
i=1
then the law of A under the state ρ is the law of a random variable which
takes the values t − kxk and t + kxk with probability
x3
1
(q − p)
1+
2
kxk
and
1
x3
(q − p)
1−
2
kxk
respectively.
Proof. The eigenvalues of A are t − kxk and t + kxk, as we have already seen.
The eigenvectors are easily obtained too, they are of the form
−x1 + ix2
x1 − ix2
v1 = λ
v2 = µ
,
kxk + x3
kxk − x3
λ, µ ∈ C, respectively.
Taking the normalized version u1 , u2 of these vectors, the probability of
measuring A to be equal to t − kxk is then
1
p(x21 + x22 ) + q(kxk + x3 )2
2 kxk (kxk + x3 )
p(kxk − x3 ) + q(kxk + x3 )
=
2 kxk
1
x3
=
1+
(q − p) .
2
kxk
hu1 , ρ u1 i =
This also gives immediately the other probability. t
u
In the particular case p = 1 and q = 0, i.e. if ρ = |e1 ihe1 |, we get all
probability distribution on R with (at most) 2 point support. Note that this
7 QUANTUM PROBABILITY
15
situation would be impossible in a classical probability context. Indeed, let
Ω = {0, 1} and P be a fixed probability measure on Ω with P(1) = p and
P(0) = q. Then a random variable X on Ω can take any couple of values (x, y)
in R but only with P(X = x) = p and P(X = y) = q. In quantum probability,
we observe that, under a fixed state ρ, the set of observables gives rise to the
whole range of two-point support laws. The space of observables is much
richer than in the classical context.
We end up this subsection with an illustration of non-commutativity in this
context. Let ρ = |e1 ihe1 |. The observables σ1 , σ2 both have the distribution
1
1
2 δ−1 + 2 δ1 under that state. But the observable
0 1−i
σ1 + σ2 =
1+i 0
√ √
has spectrum {− 2, 2} and distribution
1
1 √
δ
+ δ√ .
2 − 2 2 2
Adding two Bernoulli random variables with values {−1, √
+1},√we have ended
up with a Bernoulli random variable with values in {− 2, 2} ! This fact
would be of course impossible in the classical context, it comes from the fact
that σ1 and σ2 do not commute, they do not admit a pair law. Let us check
this fact explicitely. When computing
ϕ(t1 , t2 ) = he1 , ei(t1 σ1 +t2 σ2 ) e1 i
p
one easily finds ϕ(t1 , t2 ) = cos
t21 + t22 which is not the Fourier transform
of a positive measure on R2 . Indeed, Bochner’s positivity criterion
n
X
λi λ̄j ϕ(ti − tj , si − sj ) ≥ 0
i,j=1
fails when taking n = 3, λ1 = λ2 = λ3 = 1, t1 = s2 = t3 = s3 = 2π/3 and
s1 = t2 = 0.
7.2.2 The Toy Fock Space
In this subsection we go beyond the simple example studied above and we
focus on the very rich and fundamental structure of a chain of quantum
Bernoulli random variables. The constructions and results of this subsection
are fundamental in discrete time Quantum Noise Theory.
16
Stéphane ATTAL
Consider the space C2 with a fixed orthonormal basis which we shall denote
by {χ0 , χ1 }. Physically, one can think of this basis as representing a two-level
system with its ground state χ0 and excited state χ1 , or a spin state space
with spin down χ0 and spin up χ1 states etc.
Definition 7.15. We define the Toy Fock space 1 to be the countable tensor
product
O
C2 ,
TΦ =
N∗
associated to the stabilizing sequence (χ0 )n∈N∗ . In other words, it is the
Hilbert space whose Hilbertian orthonormal basis is given by the elements
e1 ⊗ e2 ⊗ · · · ⊗ en ⊗ · · ·
where for each i we have ei = χ0 or χ1 , but only a finite number of ei are
equal to χ1 .
Definition 7.16. Let us describe the above orthonormal basis in another
useful way. Let
P = Pf (N∗ )
be the set of finite subsets of N∗ . An element σ of P is thus a finite sequence
of integers σ = {i1 < . . . < in }. For every σ = {i1 < . . . < in } ∈ P we put
χσ to be the basis element e1 ⊗ · · · ⊗ en ⊗ · · · of TΦ given by
(
χ1 if j = ik for some k,
ej =
χ0 otherwise.
Note the particular case
χ∅ = χ0 ⊗ . . . ⊗ χ0 ⊗ . . .
Every element f of TΦ writes in unique way as
X
f=
f (σ) χσ ,
σ∈P
with
2
kf kTΦ =
X
2
|f (σ)| < ∞ .
σ∈P
In other words, with these notations, the space TΦ naturally identifies to the
space `2 (P).
The fundamental structure of TΦ is the one of a countable tensor product
of copies of C2 with respect to a given basis vector χ0 . Actually a more
1
The toy Fock space was called “bébé Fock” in french, by P.-A. Meyer. Exercise:
find the joke.
7 QUANTUM PROBABILITY
17
abstract definition for TΦ is possible as follows: TΦ is the separable Hilbert
space whose orthonormal basis is fixed and chosen to be indexed by P, the
set of finite subsets of N∗ . This is enough to carry all the properties we need
for the space TΦ.
Definition 7.17. Let n ≤ m be any fixed elements of N∗ . We denote by
TΦ[n,m] the subspace of TΦ generated by the χσ such that σ P
⊂ [n, m]. The
space TΦ[n,m] is thus the subspace of TΦ made of those f = σ∈P f (σ)χσ
such that f (σ) = 0 once σ 6⊂ [n, m].
Among those spaces TΦ[n,m] , we denote by TΦn] the space TΦ[1,n] . We
also write TΦ[n for the space TΦ[n,+∞[ , with obvious definition.
Following the same idea, we denote by P[n,m] the set of finite subsets of
N∗ ∩ [n, m] and we have the corresponding notations Pn] and P[n . We then
get the natural identification
TΦ[n,m] = `2 (P[n,m] ) .
For a σ ∈ P, we put σ[n,m] = σ ∩ [n, m]. In the same way we define σn] =
σ ∩ [1, n] and σ[n = σ ∩ [n, +∞[.
With these notations, the fundamental structure of TΦ is reflected by the
following result.
Proposition 7.18. Let n < m ∈ N∗ be fixed. The mapping
U : TΦn] ⊗ TΦ[n+1,m−1] ⊗ TΦ[m −→ TΦ
f ⊗g⊗h
7−→ k
where
k(σ) = f (σn] ) g(σ[n+1,m−1] ) h(σ[m ) ,
for all σ ∈ P, extends to a unitary operator.
Proof. The operator U is isometric for
hU(f ⊗ g ⊗ h) , U(f 0 ⊗ g 0 ⊗ h0 )iTΦ =
X
f (σn] ) g(σ[n+1,m−1] ) h(σ[m ) f 0 (σn] ) g 0 (σ[n+1,m−1] ) h0 (σ[m )
=
σ∈P
=
X
X
f (α) f 0 (α)
α∈Pn]
g(β) g 0 (β)
0
h(γ) h0 (γ)
γ∈P[m
β∈P[n+1,m−1]
0
X
0
= hf , f iTΦn] hg , g iTΦ[n+1,m−1] hh , h iTΦ[m
= hf ⊗ g ⊗ h , f 0 ⊗ g 0 ⊗ h0 iTΦn] ⊗TΦ[n+1,m−1] ⊗TΦ[m .
But U is defined on a dense subspace and thus extends to an isometry on the
whole of TΦn] ⊗ TΦ[n−1,m+1] ⊗ TΦ[m . Its range is dense in TΦ for it contains
all the χσ ’s. Thus U is unitary. t
u
18
Stéphane ATTAL
We have completely described the structure of the space TΦ. We now turn
to the operators on TΦ.
Definition 7.19. We shall make use of a particular basis, which is not the
usual Pauli matrix basis, as in the previous subsection. In the orthonormal
basis {χ0 , χ1 } of C2 we consider the basis of matrices
a00
=
10
00
01
00
0
1
1
, a1 =
, a0 =
, a1 =
.
00
10
00
01
In other words, we have
aij χk = δk,i χj
for all i, j, k = 0, 1.
From these basic operators on C2 , we built basic operators on TΦ by
considering the operators aij (n) which act as aij on the n-th copy of C2 and
as the identity on the other copies. This means, when acting on the natural
orthonormal basis of TΦ
a00 (n)χσ = 1ln6∈σ χσ
a01 (n)χσ = 1ln6∈σ χσ∪{n}
a10 (n)χσ = 1ln∈σ χσ\{n}
a11 (n)χσ = 1ln∈σ χσ .
In some sense the operators aij (n) form a basis of B(H). Of course, one
needs to be precise with the meaning of such a sentence, for now we deal
with an infinite dimensional space. This could be done easily in terms of von
Neumann algebras: “The von Neumann algebra generated by the operators
aij (n), i, j = 0, 1, n ∈ N∗ , is the whole of B(TΦ).” But this would bring us
too far and would need some knowledge on operator algebras which is not
required for this lecture.
Let us describe how random walks can be produced with the help of the
operators aij (n).
Theorem 7.20. Let A be an observable on C2 , with coefficients
X
A=
αji aij .
i,j=0,1
The observable A admits a certain law µ under the state χ0 . For all n ∈ N∗
define the observable
X
A(n) =
αji aij (n)
i,j=0,1
7 QUANTUM PROBABILITY
19
on TΦ. Then the sequence (A(n))n∈N in B(TΦ) is commutative and, under
the state χ∅ , it has the law of a sequence of independent random variables,
each of which has the same law µ.
Proof. This is an easy application of Proposition 7.7 and Proposition 7.18
when noticing that:
i) for every n ∈ N∗ , in the following splitting of TΦ:
TΦ ' TΦn−1] ⊗ TΦ[n,n] ⊗ TΦ[n+1
the operator A(n) is of the form I ⊗ A(n) ⊗ I.
ii) the state χ∅ on TΦ is the tensor product ⊗n∈N∗ Ω of the ground states of
each copy of C2 .
Details are left to the reader. t
u
Of course in the result above one can easily make the αji ’s depend on n
also; this allows to produce any sequence of independent random variables.
Summing up these random variables, this provides an easy way of realizing
any classical random walk, in law, on TΦ. We shall see in next subsection that
the Toy Fock space provides much more than that: it can realize the multiplication operators of any classical Bernoulli random walk on its canonical
space, by means of linear combinations of the operators aij (n).
7.2.3 Probabilistic Interpretations
In this section we explore the so-called probabilistic interpretations of TΦ,
that is, we shall realize on TΦ the multiplication operators associated to any
Bernoulli random walk, by means of only linear combinations of the operators
aij (n).
Definition 7.21. Let p ∈]0, 1[ be fixed and q = 1 − p. Consider a classical
Bernoulli sequence (νn )n∈N∗ , that is, a sequence of independent identically
distributed Bernoulli random variables νn with law pδ1 + qδ0 . We realize
∗
this random walk on its canonical space (Ω, F, µp ) where Ω = {0, 1}N , F is
the σ-field generated by finite-based cylinders and µp is the only probability
measure on (Ω, F) which makes the coordinate mappings
νn : Ω −→ {0, 1}
ω 7−→ ωn
being independent, identically distributed with law pδ1 + qδ0 .
Definition 7.22. We center and normalize νn by putting
νn − p
Xn = √
pq
20
Stéphane ATTAL
so that E[Xn ] = 0 and E[Xn2 ] = 1. The random variables Xn are then independent, they take the values
q
 q,
with probability p ,
p
q
p
−
q , with probability q .
For every σ ∈ P we put
(
Xi1 · · · Xin
Xσ =
1l
if σ = {i1 , . . . , in } ,
if σ = ∅ ,
where 1l is the deterministic random variable always equal to 1.
Proposition 7.23. The set {Xσ ; σ∈P} forms an orthonormal basis of the
Hilbert space L2 (Ω, F, µp ).
Proof. Let us first check that the set {Xσ ; σ∈P} forms an orthonormal set.
Let α, β ∈ P be fixed, we have
hXα , Xβ iL2 (Ω,F ,µp ) = Eµp [Xα Xβ ]
= Eµp [Xα\β (Xα∩β )2 Xβ\α ]
2
= Eµp [Xα\β ] Eµp [Xα∩β
] Eµp [Xβ\α ] ,
for the random variables Xi and Xj are independent once i 6= j. As we have
E[Xσ ] = 0 for all σ 6= ∅, we get that hXα , Xβ iL2 (Ω,F ,µp ) vanishes unless
α \ β = β \ α = ∅, that is, unless α = β. In the case where α = β we get
hXα , Xβ iL2 (Ω,F ,µp ) = Eµp [Xσ2 ] = 1 .
This proves the orthonormality.
Let us now prove the totality of the Xσ ’s. Had we replaced N∗ by
{1, . . . , N } in the above definition of Ω, we would directly conclude that
the Xσ form a basis for they are orthonormal and have cardinal 2N , the dimension of the space L2 (Ω, F, µp ) in that case. We conclude in the general
∗
case Ω = {0, 1}N by noticing that any f ∈ L2 (Ω, F, µp ) can be approached
by finitely supported functions. t
u
From this proposition we see that there is a natural isomorphism between
L2 (Ω, F, µp ) and the toy Fock space TΦ: it consists in identifying the orthonormal basis {Xσ , σ ∈ P} of L2 (Ω, F, µp ) with the orthonormal basis
{χσ , σ ∈ P} of TΦ. The space L2 (Ω, F, µp ) is now denoted by TΦp and
is called the p-probabilistic interpretation of TΦ. The space TΦp is an exact
reproduction of TΦ, but it is concretely represented as the space of L2 functionals of some random walk. The element Xσ of TΦ interprets in TΦp as a
concrete random variable Xσ .
7 QUANTUM PROBABILITY
21
All the spaces TΦp identify pairwise this way and they all identify to TΦ.
But this identification only belongs to the level of Hilbert spaces. As we have
already discussed in Subsection 7.1.3, the fact that the random variable Xn in
the p-probabilistic interpretation TΦp is represented by the basis vector χ{n}
in TΦ does not mean much. All the probabilistic properties of Xn such as its
law, its independence with respect to the other Xm ’s, ... all these informations are lost in this identification. The actual representation of the random
variable Xn of TΦp in TΦ should be the push forward of the multiplication
operator by Xn .
Let us be more precise.
Definition 7.24. Let Up : TΦp 7−→ TΦ be the unitary operator which realizes the basis identification between TΦp and TΦ. Let MXn be the operator
of multiplication by Xn in TΦp . We shall consider the operator
MpXn = Up MXn U∗p
on TΦ, which is the representation of the random variable Xn but in the
space TΦ.
We shall now prove a striking result which shows that these different random variables Xn , for different p’s, are represented on TΦ by means of a very
simple linear combination of the aij (n)’s.
Theorem 7.25. For all p ∈]0, 1[ and every n ∈ N∗ , the operator MpXn on
TΦ is given by
MpXn = a01 (n) + a10 (n) + cp a11 (n) ,
(7.2)
where
q−p
cp = √ .
pq
The mapping p 7→ cp is a bijection from ]0, 1[ to R.
Proof. Consider a f ∈ TΦp , the product Xn f is clearly determined by the
products Xn Xσ , σ ∈ P. If n does not belong to σ then by definition of our
basis we get
Xn Xσ = Xσ∪{n} ,
that is, there is nothing to compute. If n belongs to σ then
Xn Xσ = Xn2 Xσ\{n} .
In other words, the product Xn f on TΦp is determined by the Xn2 , n ∈ N∗ .
Lemma 7.26. On TΦp we have, for all n ∈ N∗
22
Stéphane ATTAL
Xn2 = 1l + cp Xn
where
(7.3)
q−p
cp = √ .
pq
Proof (of the lemma). We have on TΦp
2
νn − p
ν 2 − 2pνn + p2
= n
√
pq
pq
p
(1 − 2p)νn
+
=
pq
q
p
q−p √
=
pqXn + p1l + 1l
pq
q
q−p
= 1l + √ Xn .
pq
Xn2 =
This proves (7.3) and the lemma.
Coming back to the proof of the theorem and applying Lemma 7.26, we get
for all σ ∈ P
Xn Xσ = 1ln6∈σ Xσ∪{n} + 1ln∈σ (1l + cp Xn ) Xσ\{n}
= 1ln6∈σ Xσ∪{n} + 1ln∈σ Xσ\{n} + cp 1ln∈σ Xσ
This means that on TΦ we have
MpXn Xσ = 1ln6∈σ Xσ∪{n} + 1ln∈σ Xσ\{n} + cp 1ln∈σ Xσ .
We recognize the action of the operator
a01 (n) + a10 (n) + cp a11 (n)
on the basis of TΦ. This gives (7.2).
Finally, we have
1 − 2p
q−p
=p
cp = √
pq
p(1 − p)
and as a function of p it is easy to check that it is a bijection from ]0, 1[ to
R. t
u
What we have seen throughout this section is fundamental. On a very
simple space TΦ we have been able to reproduce a continuum of different
probabilistic situation: all the Bernoulli random walk with any parameter
p ∈]0, 1[. We are able to completely reproduce inside TΦ all these different
probability spaces, all these different laws. Even more surprising is the fact
that this representation is achieved only with the help of linear combinations
7 QUANTUM PROBABILITY
23
of three basic processes (a10 (n))n∈N∗ , (a01 (n))n∈N∗ , (a11 (n))n∈N∗ . These three
quantum processes play the role of three fundamental quantum Bernoulli
random walk from which every classical Bernoulli random walk can be constructed.
7.2.4 Brownian Motion and Poisson Process
The fact that one gets all classical Bernoulli random variables from the Toy
Fock space has many important consequences, some of them are fundamental
in the theory of quantum noises. We shall give here a taste of it by showing
that the toy Fock space gives rise to two fundamental stochastic processes:
the Brownian motion and the Poisson process.
First of all, as we wish to give some heuristic of a continuous-time limit, we
shall now have our Toy Fock indexed by hN∗ , for some small real parameter
h > 0, instead of N∗ .
In Theorem 7.25, consider the case p = 1/2. The random variables Xn
then take the values ±1 with probability 1/2. In that case we get c 21 = 0,
which means that the operators
Snh =
n
X
a01 (ih) + a10 (ih) , n ∈ N∗ ,
i=1
are a commutative sequence of observables on TΦ, their distribution in the
state X∅ is the one of a symmetric Bernoulli random walk indexed by hN∗ .
Even more, they are the multiplication operators by these random walks on
their canonical space. This means in particular that all the functional calculus
that can be usually derived from these random walks, can be obtained in the
same way with the sequence of operators (Snh ).
Renormalizing (Snh ), we consider the sequence
Qnh =
n √
X
h a01 (ih) + a10 (ih) , n ∈ N∗ ,
i=1
which is a random walk converging to a Brownian motion when h tends to
0. This is a convergence in law for the associated random walk, but when
regarding the operators Qnh this is far more. It is a convergence at the level
of multiplication operators, it is a convergence which respects the functional
calculus of the Brownian motion in the limit. For example, computing
24
Stéphane ATTAL
n
n−1
X
X
2
(Qih − Q(i−1)h )2 =
h a01 (ih) + a10 (ih)
i=2
i=1
2
n
X
01
=
h
1 0 ih
i=2
n
X
10
=
h
0 1 ih
i=2
= hI,
gives directly the quadratic variation of the Brownian motion:
(dWt )2 = dt
or more precisely
[W , W ]t = t .
Even the Ito formula is encoded in this matrix representation. For example,
the famous formula
Z t
2
Wt = 2
Ws dWs + t
0
appears as a simple continuous time limit of the following computation
Q2nh =
n
X
h a01 (ih) + a10 (ih)
a01 (jh) + a10 (jh)
i,j=1
=2
n
X


i−1
X

h a01 (jh) + a10 (jh)  a01 (ih) + a10 (ih)
i=2
+
j=1
n
X
2
h a01 (ih) + a10 (ih)
i=1
=2
n
X
Q(i−1)h Q(ih) − Q((i − 1)h) + h I .
i=2
Where the situation becomes even more striking is the way the Poisson
process enters into the game also, in the Toy Fock space.
Coming back to the random walks of Theorem 7.25, but with a particular
choice of the probability p now. Indeed, let us take p = h, where h is the
same parameter as the time step of the random walk. In order to make the
computation not too huggly we prefer to take
p=
so that
h
,
1+h
q=
1
,
1+h
7 QUANTUM PROBABILITY
25
1−h
cp = √ .
h
√
√
The associated random variable Xnh takes the values 1/ h and − h with
probability p and q respectively. The random walk
Snh =
n √
X
h Xih
i=1
is a Bernoulli random walk, with time step h, which takes the value −h very
often (with probability 1 − h, more or less) and the value 1 very rarely (with
probability h, more or less). In the continuous time limit it is a compensated
Poisson process.
From the point of view of the Toy Fock space representation we obtain the
operators
Xnh =
n √
X
h a01 (ih) + a10 (ih) + cp a11 (ih)
i=1
=
n √
X
h a01 (ih) + a10 (ih) + (1 − h)a11 (ih) .
i=1
This operator family is supposed to behave like a compensated Poisson
process, in the limit h → 0. As a consequence, the family
Nnh = Xnh + nhI =
n √
X
h a01 (ih) + a10 (ih) + a11 (ih) + ha00 (ih)
i=1
should represent a standard Poisson process in the limit. In the same way
as for the Brownian motion, one can recover from these operators all the
properties of the Poisson process, of its functional calculus. For example, the
matrix
√ √
h
h
0
1
1
0
A = h a1 (ih) + a0 (ih) + a1 (ih) + ha0 (ih) = √
h 1 (ih)
satisfies
A2 = (1 + h)A .
This relation gives in the continuous time limit the fundamental characterization of the Poisson process
(dNt )2 = dNt ,
or else
[N , N ]t = Nt .
26
Stéphane ATTAL
7.3 Higher Level Chains
The aim of this section is to present the natural extension of the Toy Fock
space structure, but for higher number of levels on each sites. From a quantum
mechanical point of view this is clearly a richer physical structure. From the
probabilistic interpretation point of view it gives access to all random walks
in RN . But the interesting point is that the probabilistic interpretations of
the N + 1-level chain gives rise to a very particular probabilistic structure:
the obtuse random variables.
7.3.1 Structure of Higher Level Chains
Let H be any separable Hilbert space. Let us fix a particular Hilbertian
basis (χi )i∈N ∪{0} for the Hilbert space H, where we assume (for notational
purposes) that 0 6∈ N . This particular choice of notations is motivated both
by physical interpretations (we see the χi , i ∈ N , as representing, for example,
different possible excited states of a quantum system, the vector χ0 represents
the “ground state” of the quantum system) and by a mathematical necessity
(for defining a countable tensor product of Hilbert space one needs to specify
one vector in each Hilbert space).
N
Definition 7.27. Let TΦ be the tensor product N∗ H with respect to the
stabilizing sequence χ0 . In other words, an orthonormal basis of TΦ is given
by the family {χσ ; σ ∈ P} where
– the set P = P(N∗ , N ) is the set of finite subsets {(n1 , i1 ), . . . , (nk , ik )} of
N∗ ×N such that the ni ’s are mutually different. Another way to describe the
set P is to identify it to the set of sequences (σn )n∈N∗ with values in N ∪ {0}
which take a value different from 0 only finitely many times;
– χσ denotes the vector
χ0 ⊗ . . . ⊗ χ0 ⊗ χi1 ⊗ χ0 ⊗ . . . ⊗ χ0 ⊗ χi2 ⊗ . . .
where χi1 appears in the n1 -th copy of H, where χi2 appears in the n2 -th
copy of H etc.
The physical signification of this basis is easy to understand: we have a
chain of quantum systems, indexed by N∗ . The space TΦ is the state space of
this chain, the vector χσ with σ = {(n1 , i1 ), . . . , (nk , ik )} represents the state
in which exactly k sites are excited: the site n1 in the state χi1 , the site n2
in the state χi2 etc, all the other sites being in their ground state χ0 .
Definition 7.28. This particular choice of a basis gives TΦ the same particular structure as the spin chain. If we denote by TΦn] the space generated by
the χσ such that σ ⊂ {1, . . . , n} × N and by TΦ[m the one generated by the
7 QUANTUM PROBABILITY
27
χσ such that σ ⊂ {m, m+1, . . .}×N , we get an obvious natural isomorphism
between TΦ and TΦn−1] ⊗ TΦ[n given by
[f ⊗ g](σ) = f (σ ∩ {1, . . . , n − 1} × N ) g (σ ∩ {n, . . .} × N ) .
Definition 7.29. Put {aij ; i, j ∈ N ∪ {0}} to be the natural basis of B(H),
that is,
aij (χk ) = δi,k χj
for all i, j, k ∈ N ∪ {0}. We denote by aij (n) the natural ampliation of the
operator aij to TΦ which acts on the copy number n as aij and which acts as
the identity on the other copies. That is, in terms of the basis χσ ,
aij (n)χσ = 1l(n,i)∈σ χσ\(n,i)∪(n,j)
if neither i nor j is zero, and
ai0 (n)χσ = 1l(n,i)∈σ χσ\(n,i) ,
a0j (n)χσ = 1ln6∈σ χσ∪(n,j) ,
a00 (n)χσ = 1ln6∈σ χσ ,
where n 6∈ σ actually means “for all i in N , (n, i) 6∈ N ”; we could equivalently
use the notation (n, 0) ∈ σ instead of n 6∈ σ, having in mind the interpretation
of σ as a sequence in N with finitely many non-zero terms.
In the case N = {1, . . . , N } we say that TΦ is Toy Fock space with multiplicity N + 1. In the case N = N∗ we say that TΦ is the Toy Fock space with
infinite multiplicity.
7.3.2 Obtuse Systems
The probabilistic interpretations of higher level chains is carried, in a natural
way, by particular random variables, the obtuse random variables. We shall
see in the next subsections that these random variables have very interesting
algebraic properties which characterize their behavior (and which encodes
the behavior of their continuous-time limit). Before entering into all those
properties, we start with obtuse systems.
Definition 7.30. Let N ∈ N∗ be fixed. An obtuse system in RN is a family
of N + 1 vectors v1 , . . . , vN +1 such that
hvi , vj i = −1
for all i 6= j. In that case we put
28
Stéphane ATTAL
vbi =
1
∈ RN +1 ,
vi
so that
hb
vi , vbj i = 0
for all i 6= j. They then form an orthogonal basis of RN +1 . We put
pi =
1
2
kb
vi k
=
1
1 + kvi k
2
,
for i = 1, . . . N + 1.
Lemma 7.31. With the notations above, we have
N
+1
X
pi = 1
(7.4)
p i vi = 0 .
(7.5)
i=1
and
N
+1
X
i=1
Proof. We have, for all j,
*N +1
+
X
1
2
pi vbi , vbj = pj kb
vj k = 1 =
, vbj .
0
i=1
As the vbj ’s form a basis, this means that
N
+1
X
pi vbi =
i=1
1
.
0
This gives the two announced equalities. t
u
Lemma 7.32. With the notations above, we also have
N
+1
X
pi |vi ihvi | = IRN .
(7.6)
i=1
√
Proof. As the vectors ( pi vbi )i∈{1,...,N +1} form an orthonormal basis of RN +1
we have
N
+1
X
IRN +1 =
pi |b
vi ihb
vi | .
i=1
Now, for all i = 1, . . . , N + 1, put
7 QUANTUM PROBABILITY
u=
29
1
0
and
vei =
0
,
vi
so that vbi = u + vei . We get
IRN +1 =
N
+1
X
pi |u + vei ihu + vei |
i=1
=
N
+1
X
pi |uihu| +
N
+1
X
i=1
pi |uihe
vi | +
N
+1
X
i=1
pi |e
vi ihu| +
N
+1
X
i=1
pi |e
vi ihe
vi | .
i=1
Using (7.4) and (7.5), this simplifies into
IRN +1 = |uihu| +
N
+1
X
pi |e
vi ihe
vi | .
i=1
In particular we have
N
+1
X
pi |vi ihvi | = IRN ,
i=1
that is, the announced equality. t
u
Let us consider two examples (that we shall meet again later in this section). On R2 , the 3 vectors
1
−1
−1
v1 =
,
v2 =
,
v3 =
0
1
−2
form an obtuse system of R2 . The associated pi ’s are then respectively
p1 =
1
,
2
p2 =
1
,
3
p3 =
1
.
6
We shall be interested also in another example. Let h > 0 be a parameter,
which shall be though of as small. On R2 , associated to the probabilities
1
,
2
p2 = h ,
v2 =
−1
1−2h 1/2
p1 =
p3 =
1
− h,
2
the 3 vectors
1
v1 =
,
0
form an obtuse system of R2 .
h
−1
!
v3 =
−2
h
1−2h
!
1/2
30
Stéphane ATTAL
We end this subsection with some useful properties of obtuse systems.
First, we prove an independence property for obtuse systems.
Proposition 7.33. Every strict sub-family of an obtuse family is linearly
free.
Proof. Let {v1 , . . . , vN +1 } be an obtuse family of RN . Let us show that
{v1 , . . . , vN } is free, which would be enough for our claim.
PN −1
If we had vN = i=1 λi vi then, taking the scalar product with vN we
P
2
N −1
get kvN k = i=1 −λi , whereas taking the scalar product with vN +1 gives
PN −1
−1 = i=1 −λi , whence a contradiction. t
u
Now we prove a kind of uniqueness result for obtuse systems.
Proposition 7.34. Let {v1 , . . . , vN +1 } be an obtuse system of RN having
{p1 , . . . , pN +1 } as associated probabilities. Then the following assertions are
equivalent.
i) The family {w1 , . . . , wN +1 } is an obtuse system on RN with same respective
probabilities {p1 , . . . , pN +1 }.
ii) There exists an orthogonal operator U on RN such that wi = U vi , for all
i = 1, . . . , N + 1.
Proof. One direction is obvious. If wi = U vi , for all i = 1, . . . , N + 1 and for
some orthogonal operator U, then the scalars products hvi , vj i and hwi , wj i
are equal, for each pair (i, j). This shows that {w1 , . . . , wN +1 } is an obtuse
system with the same probabilities.
In the converse direction, if v1 , . . . , vN +1 and w1 , . . . , wN +1 are obtuse
systems associated to the same probabilities p1 , . . . , pN +1 , then
hvi , vj i = hwi , wj i
for all i, j. The sub-families {v1 , . . . , vN } and {w1 , . . . , wN } are bases of RN ,
by Proposition 7.33, and their elements have same respective norms. The map
U : RN → RN , such that U vi = wi for all i = 1, . . . , N , is thus orthogonal,
by the conservation of scalar products. Finally, the vector vN +1 is a certain
linear combination of the vi ’s, i = 1, . . . , N , but wN +1 is the same linear
combination of the wi ’s, i = 1, . . . , N , by the conservation of scalar products.
Hence U vN +1 is equal to wN +1 and the proposition is proved. t
u
7.3.3 Obtuse Random Variables
Obtuse systems are strongly related to a certain class of random variables.
7 QUANTUM PROBABILITY
31
Definition 7.35. Consider a random variable X, with values in RN . We
shall denote by X 1 , . . . , X N the coordinates of X in RN . We say that X is
centered if its expectation is 0, that is, if E[X i ] = 0 for all i. We say that X
is normalized if its covariance matrix is I, that is, if
cov(X i , X j ) = E[X i X j ] − E[X i ] E[X j ] = δi,j ,
for all i, j = 1, . . . N .
Definition 7.36. Consider a random variable X, with values in RN , which
can take only N + 1 different non-null values v1 , . . . , vN +1 , with strictly
positive probability p1 , . . . , pN +1 , respectively. We consider the canonical
version of X, that is, we consider the probability space (Ω, F, P) where
Ω = {1, . . . , N + 1}, where F is the full σ-algebra of Ω, where the probability measure P is given by P ({i}) = pi and the random variable X is given
by X(i) = vi , for all i ∈ Ω. The coordinates of vi are denoted by vik , for
k = 1, . . . , N , so that X k (i) = vik .
In the same way as previously, we put
1
vbi =
∈ RN +1 ,
vi
for all i = 1, . . . , N + 1.
We shall also consider the deterministic random variable X 0 on (Ω, F, P),
e i be the random variable
which is always equal to 1. For i = 0, . . . , N let X
defined by
e i (j) = √pj X i (j)
X
for all i = 0, . . . , N and all j = 1, . . . , N + 1.
Proposition 7.37. With the notations above, the following assertions are
equivalent.
1) X is centered and normalized.
e i (j)
is an orthogonal matrix.
2) The (N + 1) × (N + 1)-matrix X
i,j
√
3) The (N + 1) × (N + 1)-matrix
pi vbij
is an orthogonal matrix.
i,j
4) The family {v1 , . . . , vN +1 } is an obtuse system with
pi =
1
,
1 + kvi k2
for all i = 1, . . . , N + 1.
Proof.
1) ⇒ 2): Since the random variable X is centered and normalized, each
component X i has a zero mean and the scalar product in L2 between two
components X i and X j is equal to δi,j . Hence, for all i in {1, . . . , N }, we get
32
Stéphane ATTAL
E[X i ] = 0
⇐⇒
N
+1
X
pk vki = 0 ,
(7.7)
k=1
and for all i, j = 1, . . . N ,
i
j
E[X X ] = δi,j
⇐⇒
N
+1
X
pk vki vkj = δi,j .
(7.8)
k=1
Now, using Eqs. (7.7) and (7.8), we get, for all i, j = 1, . . . , N
+1
E NX
D
e0 , X
e0 =
X
pk = 1 ,
k=1
+1
D
E NX
√ √
e0 , X
ei =
pk pk vki = 0 ,
X
k=1
+1
D
E NX
√
√
ei , X
ej =
X
pk vkj pk vki = δi,j .
k=1
The orthogonal character follows immediately.
e i (j)
2) ⇒ 1): Conversely, if the matrix X
is orthogonal, the scalar prodi,j
ucts of column vectors give the mean 0 and the covariance I for the random
variable X.
√
e i (j)
. Therepj vbij
is the transpose matrix of X
2) ⇔ 3): The matrix
i,j
i,j
fore, if one of these two matrices is orthogonal, its transpose matrix is orthogonal too.
√
3) ⇔ 4): The matrix
pj vbji i,j is orthogonal if and only if
√
pi vbi ,
√
pj vbj = δi,j ,
√
√
for all i, j = 1, . . . , N +1. On the other hand, the condition pi vbi , pi vbi =
√
√
pi vbi , pj vbj =
1 is equivalent to pi (1+kvi k2 ) = 1, whereas the condition
√ √
0 is equivalent to pi pj (1+hvi , vj i) = 0, that is, hvi , vj i = −1 . This gives
the result. t
u
Definition 7.38. Because of the equivalence between 1) and 4) above, the
random variables in RN which take only N + 1 different values with strictly
positive probability, which are centered and normalized, are called obtuse
random variables in RN .
7 QUANTUM PROBABILITY
33
7.3.4 Generic Character of Obtuse Random Variables
We shall here apply the properties of obtuse systems to obtuse random variable, in order to show that these random variables somehow generate all the
finitely supported probability distributions on RN .
First of all, an immediate consequence of Proposition 7.34 is that obtuse random variables on RN with a prescribed probability distribution
{p1 , . . . , pN +1 } are essentially unique.
Proposition 7.39. Let X be an obtuse random variable of RN with associated probabilities {p1 , . . . , pN +1 }. Then the following assertions are equivalent.
i) The random variable Y is an obtuse random variable on RN with same
probabilities {p1 , . . . , pN +1 }.
ii) There exists an orthogonal operator U on RN such that Y = U X in
distribution.
Having proved that uniqueness, we shall now prove that obtuse random
variables generate all the random variables (at least with finite support).
First of all, a rather simple remark which shows that the choice of taking
N +1 different values is the minimal one for centered and normalized random
variables in RN .
Proposition 7.40. Let X be a centered and normalized random variable in
Rd , taking n different values. Then we must have
n ≥ d + 1.
Proof. Let X be centered and normalized in Rd , taking the values v1 , . . . , vn
with probabilities p1 , . . . , pn and with n ≤ d, that is, n < d + 1. The relation
0 = E[X] =
n
X
pi vi
i=1
shows that Rank{v1 , . . . , vn } < n. On the other hand, the relation
ICd = E [|XihX|] =
n
X
pi |vi ihvi |
i=1
shows that Rank{v1 , . . . , vn } ≥ d. This gives the result. t
u
We can now state the theorem which shows how general, finitely supported,
random variables on Rd are generated by the obtuse ones. We concentrate
only on centered and normalized random variables, for they obviously generate all the others, up to an affine transform of Rd .
34
Stéphane ATTAL
Theorem 7.41. Let n ≥ d + 1 and let X be a centered and normalized random variable in Rd , taking n different values v1 , . . . , vn , with probabilities
p1 , . . . , pn respectively.
If Y is any obtuse random variable on Rn−1 associated to the probabilities p1 , . . . , pn , then there exists a partial isometry A from Rn−1 to Rd , with
Ran A = Rd , such that
X = AY
in distribution.
Proof. Assume that the obtuse random variable Y takes the values w1 , . . . , wn
in Rn−1 . The family {w1 , . . . , wn−1 } is linearly independent by Proposition
7.33, hence there exists a linear map A : Rn−1 → Rd such that A wi = vi for
all i < n. Now we have
!
X
X
X
pn vn = −
p i vi = −
pi A wi = A −
pi wi = pn A wn .
i<n
i<n
i<n
Hence the relation A wi = vi holds for all i ≤ n.
We have proved the relation X = A Y in distribution, with A being a
linear map from Rn−1 to Rd . The fact that X is normalized can be written
as E[XX ∗ ] = Id . But
E[XX ∗ ] = E[A Y Y ∗ A∗ ] = A E[Y Y ∗ ] A∗ = A In A∗ = A A∗ .
Hence A must satisfy A A∗ = Id , which is exactly saying that A is a partial
isometry with range Rd . t
u
7.3.5 Doubly-symmetric 3-tensors
Obtuse random variables are naturally associated to some 3-tensors with
particular symmetries. This is what we shall prove here.
Definition 7.42. A 3-tensor on Rn is an element of (RN )∗ ⊗ RN ⊗ RN , that
is, a linear map from RN to RN ⊗ RN . Coordinate-wise, it is represented by
N
a collection of coefficients (Sij
as
k )i,j,k=1,...,n . It acts on R
(S(x))ij =
n
X
k
Sij
kx .
k=1
We shall see below that obtuse random variables on RN have a naturally associated 3-tensor on RN +1 . Note that, because of our notation
choice X 0 , X 1 , . . . , X N , the 3-tensor is indexed by {0, 1, . . . , N } instead of
{1, . . . , N + 1}.
7 QUANTUM PROBABILITY
35
Proposition 7.43. Let X be an obtuse random variable in RN . Then there
exists a unique 3-tensor S on RN +1 such that
Xi Xj =
N
X
k
Sij
k X ,
(7.9)
k=0
for all i, j = 0, . . . , N . This 3-tensor S is given by
i
j
k
Sij
k = E[X X X ] ,
(7.10)
for all i, j, k = 0, . . . N .
Proof. As X is an obtuse random variable, that is, a centered and normalized random variable in RN taking exactly N +1 different values, the random
variables {X 0 , X 1 , . . . , X N } are orthonormal in L2 (Ω, F, P), hence they form
an orthonormal basis of L2 (Ω, F, P), for the latter space is N +1-dimensional.
These random variables being bounded, the products X i X j are still elements
of L2 (Ω, F, P), hence they can be written, in a unique way, as linear combinations of the X k ’s. As a consequence, there exists a unique 3-tensor S on
RN +1 such that
N
X
k
Sij
Xi Xj =
k X
k=0
for all i, j = 0, . . . , N . In particular we have
E[X i X j X k ] =
N
X
ij
Sij
l E[Xl Xk ] = Sk .
l=0
This shows Identity (7.10). t
u
This 3-tensor S has quite some symmetries, let us detail them.
Proposition 7.44. Let S be the 3-tensor associated to an obtuse random
variable X on RN . Then the 3-tensor S satisfies the following relations, for
all i, j, k, l = 0, . . . , N
Sij
(7.11)
0 = δij ,
N
X
Sij
k is symmetric in (i, j, k) ,
(7.12)
kl
Sij
m Sm is symmetric in (i, j, k, l) ,
(7.13)
m=0
Proof.
Relation (7.11) is immediate for
i
j
Sij
0 = E[X X ] = δij .
36
Stéphane ATTAL
Equation (7.12) comes directly from Formula (7.10) which shows a clear symmetry in (i, j, k).
By (7.9) we have
N
X
m
Xi Xj =
Sij
mX ,
m=0
whereas
Xk Xl =
N
X
n
Skl
n X .
n=0
Altogether this gives
N
X
kl
E Xi Xj Xk Xl =
Sij
m Sm .
m=0
But the left hand side is clearly symmetric in (i, j, k, l) and (7.13) follows.
t
u
7.3.6 The Main Diagonalization Theorem
We are going to leave for a moment the obtuse random variables and concentrate only on the symmetries we have obtained above. The relation (7.11)
is really specific to obtuse random variables, we shall leave it for a moment.
We concentrate on the relations (7.12) and (7.13) which have important consequences for the 3-tensor.
Definition 7.45. A 3-tensor S on RN +1 which satisfies (7.12) and (7.13) is
called a doubly-symmetric 3-tensor on RN +1 .
The main result concerning doubly-symmetric 3-tensors in RN +1 is that
they are the exact generalization for 3-tensors of normal matrices for 2tensors: they are exactly those 3-tensors which can be diagonalized in some
orthonormal basis of RN +1 .
Definition 7.46. A 3-tensor S on RN +1 is said to be diagonalizable in some
N
N
orthonormal basis (am )m=0 of RN +1 if there exist real numbers (λm )m=0
such that
N
X
S=
λm a∗m ⊗ am ⊗ am .
(7.14)
m=0
In other words
S(x) =
N
X
m=0
for all x ∈ RN +1 .
λm ham , xi am ⊗ am
(7.15)
7 QUANTUM PROBABILITY
37
Note that, as opposed to the case of 2-tensors (that is, matrices), the
“eigenvalues” λm are not completely determined by the representation (7.15).
Indeed, if we put e
am = sm am for all m, where the sm ’s are any modulus one
real numbers, then the e
am ’s still form an orthonormal basis of RN +1 and we
have
N
X
S(x) =
sm λm he
am , xi e
am ⊗ e
am .
m=0
Hence the λm ’s are only determined up to a sign, only their modulus is
determined by the representation (7.15).
Definition 7.47. Actually, there are more natural objects that can be associated to diagonalizable 3-tensors; they are the orthogonal families in RN .
Indeed, if S is diagonalizable as above, for all m such that λm 6= 0 put
vm = λm am . The family {vm ; m = 1, . . . , K} is then an orthogonal family
in RN +1 and we have
S(vm ) = vm ⊗ vm
for all m. In terms of the vm ’s, the decomposition (7.15) of S becomes
S(x) =
K
X
1
m=1
kvm k
2
hvm , xi vm ⊗ vm .
(7.16)
This is the form of diagonalization we shall retain for 3-tensors. Be aware
that in the above representation the vectors are orthogonal, but not normalized anymore. Also note that they only represent the eigenvectors of S
associated to non-vanishing eigenvalues.
We can now state the main theorem.
Theorem 7.48. A 3-tensor S on RN +1 is diagonalizable in some orthonormal basis if and only if it is doubly-symmetric.
More precisely, the formulas
V = v ∈ RN +1 \ {0} ; S(v) = v ⊗ v ,
and
S(x) =
X
v∈V
1
kvk
2
hv , xi v ⊗ v ,
establish a bijection between the set of orthogonal systems V in RN +1 \ {0}
and the set of complex doubly-symmetric 3-tensors S.
Proof. First step: let V = {vm ; m = 1, . . . , K} be an orthogonal family in
RN +1 \ {0}. Put
K
X
1
i
j
k
Sij
=
k
2 vm vm v m ,
kv
k
m
m=1
38
Stéphane ATTAL
for all i, j, k = 0, . . . , N . We shall check that S is a complex doubly-symmetric
3-tensor in RN . The symmetry of Sij
k in i, j, k is obvious from the definition.
This gives (7.12).
We have
N
X
kl
Sij
m Sm =
m=0
=
=
N
K
X
X
1
1
2
2
m=0 n,p=1 kvn k kvp k
K
X
1
1
2
2
n,p=1 kvn k kvp k
K
X
1
n=1 kvn k
2
vni vnj vnm vpm vpk vpl
vni vnj hvp , vn i vpk vpl
vni vnj vnk vnl
and the symmetry in i, j, k, l is obvious. This gives (7.13).
We have proved that the formula
S(x) =
X
v∈V
1
kvk
2
hv , xi v ⊗ v
(7.17)
defines a complex doubly-symmetric 3-tensor if V is any family of (nonvanishing) orthogonal vectors.
Second step: now given a complex doubly-symmetric 3-tensor S of the form
(7.17), we shall prove that the set V coincides with the set
b = {v ∈ CN \ {0} ; S(v) = v ⊗ v} .
V
Clearly, if y ∈ V we have by (7.17)
S(y) = y ⊗ y .
b Now, let v ∈ V.
b On one side we have
This proves that V ⊂ V.
S(v) = v ⊗ v ,
on the other side we have
S(v) =
X
z∈V
1
kzk
2
hz , vi z ⊗ z .
In particular, applying hy| ∈ V ∗ to both sides, we get
hy , vi v = hy , vi y
7 QUANTUM PROBABILITY
39
and thus either v is orthogonal to y or v = y. This proves that v is one of
the elements y of V, for it were orthogonal to all the y ∈ V we would get
v ⊗ v = S(v) = 0 and v would be the null vector.
We have proved that V coincides with the set
{v ∈ RN \ {0} ; S(v) = v ⊗ v} .
Third step: now we shall prove that all complex doubly-symmetric 3tensors S on RN +1 are diagonalizable in some orthonormal basis.
The property (7.12) indicates that the matrices
Sk = (Sij
k )i,j=0,...,N
are real symmetric hence they are all diagonalizable in some orthonormal
basis. The properties (7.12) and (7.13) imply
N
X
mk
Sim
=
j Sl
m=0
N
X
mk
Sim
l Sj .
m=0
In other words
(Sj Sl )ik = (Sl Sj )ik ,
for all i, k, all j, l. The matrices Sk commute pairwise. Thus the matrices
Sk can be simultaneously diagonalized: there exists an orthogonal matrix
U = (uij )i,j=0,··· ,N such that, for all k in {0, · · · , N },
Sk = U Dk U∗ ,
(7.18)
where the matrices Dk are diagonal: Dk = diag(λ0k , · · · , λN
k ). As a consequence, the coefficient Sij
can
be
written
as
k
Sij
k =
N
X
im jm
λm
u .
k u
m=0
Let us denote by am the mth column vector of U, that is, am = (ulm )l=0,··· ,N .
Moreover, we denote by λm the vector of λm
k , for k = 0, · · · , N . Since the
matrix U is orthogonal, the vectors am form an orthonormal basis of RN +1 .
We have
N
X
i
j
Sij
=
λm
k am am .
k
m=0
m
Our aim now is to prove that λ is proportional to am . To this end, we shall
use the symmetry properties of S. From the simultaneous reduction (7.18),
we get
Sj Sq = U Dj Dq U∗ .
40
Stéphane ATTAL
Therefore, we have
(Sj Sq )i,r =
N
X
mr
Sim
=
j Sq
m=0
N
X
m r
aim λm
j λq am .
m=0
In particular we have, for all p ∈ {0, . . . , N }
N
X
N
X
(Sj Sq )i,r aip λpj λpq arp =
ham , ap i hλm , λp i hλp , λm i hap , am i
m=0
i,j,q,r=0
4
= kλp k .
Note that
N
X
(7.19)
mr
Sim
j Sq
m=0
is also symmetric in (j, r). Applying this, the expression (7.19) is also equal
to
N
X
N
X
m j
i p p r
aim λm
r λq am ap λj λq ap =
i,j,q,r=0 m=0
=
N
X
hap , λm i hλm , λp i ham , λp i hap , am i
m=0
2
2
= |hap , λp i| kλp k .
This gives
|hap , λp i| = kλp k = kap k kλp k .
This is a case of equality in Cauchy-Schwarz inequality, hence there exists
µp ∈ R such that λp = µp ap , for all p = 0, . . . , N . This way, the 3-tensor S
can be written as
N
X
Sij
=
µm aim ajm akm .
(7.20)
k
m=0
In other words
S(x) =
N
X
µm ham , xi am ⊗ am .
m=0
We have obtained the orthonormal diagonalization of S. The proof is complete. t
u
7 QUANTUM PROBABILITY
41
7.3.7 Back to Obtuse Random Variables
The theorem above is a general diagonalization theorem for 3-tensors. For
the moment it does not take into account the relation (7.11). When we make
it enter into the game, we see obtuse systems appearing.
Theorem 7.49. Let S be a doubly-symmetric 3-tensor on RN +1 satisfying
also the relation
Sij
0 = δij
for all i, j = 0, . . . , N . Then the orthogonal system V such that
X
S(x) =
1
2
v∈V
kvk
hv , xi v ⊗ v
(7.21)
is made of exactly N + 1 vectors v1 , . . . , vN +1 , all of them satisfying vi0 = 1.
In particular the family of N + 1 vectors of RN , obtained by restricting the
vi ’s to their N last coordinates, forms an obtuse system in RN .
Proof. First assume that V = {v1 , . . . , vK }. By hypothesis, we have
Sij
k =
K
X
1
m=1 kvm k
2
i
j
k
vm
vm
vm
,
for all i, j, k = 0, . . . , N . With the supplementary property (7.11) we have in
particular
K
X
1
i
j
0
Sij
=
0
2 vm vm vm = δij
m=1 kvm k
for all i, j = 0, . . . , N .
Consider the orthonormal family of RN +1 made of the vectors em =
vm / kvm k. We have obtained above the relation
K
X
0
vm
|em ihem | = I ,
m=0
as matrices acting on RN +1 . The above is thus a spectral decomposition of
the identity matrix, this implies that the em ’s are exactly N + 1 vectors and
0
that all the vm
are equal to 1.
This proves the first part of the theorem. The last part concerning obtuse
systems is now obvious and was already noticed when we introduced obtuse
systems. t
u
In particular we have proved the following theorem.
42
Stéphane ATTAL
Theorem 7.50. The set of doubly-symmetric 3-tensors S on RN +1 which
satisfy also the relation
Sij
0 = δij
for all i, j = 0, . . . , N , is in bijection with the set of distributions of obtuse
random variables X on RN . The bijection is described by the following, with
the convention X 0 = 1l.
– The random variable X is the only (in distribution) random variable satisfying
N
X
k
X iX j =
Sij
kX ,
k=0
for all i, j = 1, . . . , N .
– The 3-tensor S is obtained by
i
j
k
Sij
k = E[X X X ] ,
for all i, j, k = 0, . . . , N .
In particular the different possible values taken by X in RN coincide with
the vectors wn ∈ RN , made of the last N coordinates of the eigenvectors vn
associated to S in the representation (7.21). The associated probabilities are
2
2
then pn = 1/(1 + kwn k ) = 1/kvn k .
7.3.8 Probabilistic Interpretations
Definition 7.51. Let X be an obtuse random variable in RN , with associated
3-tensor S and let (Ω, F, PS ) be the canonical space of X. Note that we have
added the dependency on S for the probability measure PS . The reason is that,
when changing the obtuse random variable X on RN , the canonical space Ω
and the canonical σ-field F do not change, only the canonical measure P does
change.
We have seen that the space L2 (Ω, F, PS ) is a N + 1-dimensional Hilbert
space and that the family {X 0 , X 1 , . . . , X N } is an orthonormal basis of that
space. Hence for every obtuse random variable X, with associated 3-tensor
S, we have a natural unitary operator
US : L2 (Ω, F, PS ) −→ CN +1
Xi
7−→ ei ,
where {e0 , . . . , eN } is the canonical orthonormal basis of CN +1 . The operator
US is called the canonical isomorphism associated to X.
Definition 7.52. On the space L2 (Ω, F, PS ), for each i = 0, . . . , N , we consider the multiplication operator
7 QUANTUM PROBABILITY
43
MX i : L2 (Ω, F, PS ) −→ L2 (Ω, F, PS )
Y
7−→
Xi Y ,
Definition 7.53. On the space CN +1 , with canonical basis {e0 , . . . , eN } we
consider the basic matrices aij , for i, j = 0, . . . , N defined by
aij ek = δi,k ej .
We shall see now that, when carried out on the same canonical space by
US , the obtuse random variables of RN admit a simple and compact matrix
representation in terms of their 3-tensor.
Theorem 7.54. Let X be an obtuse random variable on RN , with associated 3-tensor S and canonical isomorphism US . Then we have, for all
i, j = 0, . . . , N
N
X
Sijk ajk .
US MX i U∗S = a0i + ai0 +
(7.22)
j,k=1
for all i = 1, . . . , N .
Proof. We have, for any fixed i ∈ {1, . . . , N }, for all j = 0, . . . , N
US MX i U∗S ej = US MX i X j
= US X i X j
= US
N
X
k
Sij
kX
k=0
=
N
X
Sij
k ek .
k=0
Hence the operator US MX i U∗S has the same action on the orthonormal basis
{e0 , . . . , eN } as the operator
N
X
j
Sij
k ak .
j,k=0
Using the symmetries of S (Proposition 7.44) and the identity (7.11), we get
US M
Xi
U∗S
=
N
X
Si0
k
a0k
k=0
= ai0 + a0i +
+
N
X
Sij
0
aj0
+
j=1
N
X
j,k=1
This proves the representation (7.22). t
u
j
Sjk
i ak .
N
X
j,k=1
j
Sij
k ak
44
Stéphane ATTAL
Once again this is a remarkable point that we get here. All these different
obtuse random variables X of RN (and by the way, all the random variables in
RN with finite support) can be represented as very simple linear combinations
of the basic operators aij . As we have discussed above, all the probabilistic
properties of X (law, independence, functional calculus etc.) are carried by
these linear combinations of basic operators.
Let us check how this works with our examples. We first start with the
obtuse random variable X in R2 taking the values
1
−1
−1
v1 =
,
v2 =
,
v3 =
,
0
1
−2
with respective probabilities
p1 =
1
,
2
p2 =
1
,
3
p3 =
1
.
6
Adding a third coordinate 1 to each vector, we get the following orthogonal
system of R3 :
 
 
 
1
1
1
vb3 = −1 .
vb2 = −1 ,
vb1 = 1 ,
−2
1
0
We compute the matrices |b
vi ihb
vi | and get respectively


110
|b
v1 ihb
v1 | = 1 1 0
000


1 −1 1
|b
v2 ihb
v2 | = −1 1 −1
1 −1 1


1 −1 −2
|b
v3 ihb
v3 | = −1 1 2  .
−2 2 4
The matrices Sk , of multiplication by X k , are given by
X
i i i
Sij
p m vm
vj vk .
k =
m=02
This gives
7 QUANTUM PROBABILITY
45


100
S0 = 0 1 0
001


01 0
S1 = 1 0 0 
0 0 −1


0 0 1
S2 = 0 0 −1 .
1 −1 −1
These are the matrices associated to the multiplication operators by X 0 ,
X and X 2 , respectively. We recover facts that can be easily checked :
1
S0 = I,
(X 1 )2 = X 0 ,
X 1 X 2 = −X 2 ,
(X 2 )2 = X 0 − X 1 − X 2 .
As a consequence, the two operators
X1 = a10 + a01 − a22 ,
X2 = a20 + a02 − a12 − a21 − a22
have exactly the same probabilistic properties, in the sense of Quantum Probability on C3 in the state |e0 ihe0 |, as the pair of random variables (X 1 , X 2 ).
Let us now compute the 3-tensor associated to our second example, that
is the obtuse random X on R2 which takes the values
!
!
−1
−1
1
1/2
v3 =
v1 =
,
v2 =
h
1−2h 1/2
0
−2 1−2h
h
with respective probabilities
p1 =
1
,
2
p2 = h ,
p3 =
1
− h.
2
The same kind of computation as above gives


100
S0 = 0 1 0
001


01 0
S1 = 1 0 0 
0 0 −1


0 0
1

−1 
S2 = 0 0
.
1 −1 √ 1−4h
h(1−2h)
46
Stéphane ATTAL
At the level of this lecture we do not have the tools for computing rigorously the continuous-time limits of the random walks associated to the
two examples above, but we shall end this lecture by just describing these
continuous-time limits. In both cases consider the stochastic process
Ynh =
n √
X
h Xkh ,
k=0
where the Xkh are independent copies of the random variable X.
In the case of the first example, the process Y converges, when h tends to
0, to a two dimensional Brownian motion.
In the case of our second example, the limit process is a Brownian motion
in the first coordinate and a standard compensated poisson process in the
second coordinate.
Notes
Many parts of this chapter are inspired from the two reference books of
Parthasarathy [Par92] and Meyer [Mey93], and also from the course by Biane
[Bia95]. The two first references contain long developments concerning the
theory of quantum probability. Our approach in this chapter is closer to the
one of Parthasarathy.
The notion of Toy Fock space and its probabilistic interpretation as natural
space for all sequences of Bernoulli random variables, is due to P.-A. Meyer.
It first appeared in [Mey86]. The way we present it in this chapter is an
extended version, developed by Attal and Pautrat in [AP06].
Real obtuse random variables and their diagonalization theorem were introduced by Attal and Emery in [AÉ94].
References
[AÉ94] S. Attal and M. Émery. Équations de structure pour des martingales vectorielles. In Séminaire de Probabilités, XXVIII, volume 1583 of Lecture Notes
in Math., pages 256–278. Springer, Berlin, 1994.
[AP06] Stéphane Attal and Yan Pautrat. From repeated to continuous quantum
interactions. Ann. Henri Poincaré, 7(1):59–104, 2006.
[Bia95] Philippe Biane. Calcul stochastique non-commutatif. In Lectures on probability theory (Saint-Flour, 1993), volume 1608 of Lecture Notes in Math.,
pages 1–96. Springer, Berlin, 1995.
[Mey86] P.-A. Meyer. Éléments de probabilités quantiques. I–V. In Séminaire de
Probabilités, XX, 1984/85, volume 1204 of Lecture Notes in Math., pages
186–312. Springer, Berlin, 1986.
7 QUANTUM PROBABILITY
47
[Mey93] Paul-André Meyer. Quantum probability for probabilists, volume 1538 of
Lecture Notes in Mathematics. Springer-Verlag, Berlin, 1993.
[Par92] K. R. Parthasarathy. An introduction to quantum stochastic calculus, volume 85 of Monographs in Mathematics. Birkhäuser Verlag, Basel, 1992.