Download Quantum Memory in Atomic Ensembles - Oxford Physics

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Probability amplitude wikipedia , lookup

Renormalization wikipedia , lookup

Quantum field theory wikipedia , lookup

Wave function wikipedia , lookup

Matter wave wikipedia , lookup

Quantum machine learning wikipedia , lookup

Double-slit experiment wikipedia , lookup

Quantum entanglement wikipedia , lookup

Interpretations of quantum mechanics wikipedia , lookup

X-ray fluorescence wikipedia , lookup

Quantum teleportation wikipedia , lookup

EPR paradox wikipedia , lookup

Coherent states wikipedia , lookup

Path integral formulation wikipedia , lookup

Quantum group wikipedia , lookup

Atomic theory wikipedia , lookup

Renormalization group wikipedia , lookup

Bohr–Einstein debates wikipedia , lookup

Hidden variable theory wikipedia , lookup

Hydrogen atom wikipedia , lookup

Density matrix wikipedia , lookup

Quantum electrodynamics wikipedia , lookup

Relativistic quantum mechanics wikipedia , lookup

Tight binding wikipedia , lookup

Scalar field theory wikipedia , lookup

Quantum state wikipedia , lookup

Wave–particle duality wikipedia , lookup

T-symmetry wikipedia , lookup

History of quantum field theory wikipedia , lookup

Delayed choice quantum eraser wikipedia , lookup

Canonical quantization wikipedia , lookup

Quantum key distribution wikipedia , lookup

Symmetry in quantum mechanics wikipedia , lookup

Ultrafast laser spectroscopy wikipedia , lookup

Theoretical and experimental justification for the Schrödinger equation wikipedia , lookup

Transcript
Quantum Memory in Atomic
Ensembles
Joshua Nunn
St. John’s College, Oxford
Submitted for the degree of Doctor of Philosophy
Hilary Term 2008
Supervised by
Prof. Ian A. Walmsley
Clarendon Laboratory
University of Oxford
United Kingdom
For glory. I mean Gloria.
Abstract
This thesis is a predominantly theoretical study of light storage in atomic ensembles. The efficiency of ensemble quantum memories is analyzed and optimized
using the techniques of linear algebra. Analytic expressions describing the memory
interaction in both EIT and Raman regimes are derived, and numerical methods
provide solutions where the analytic expressions break down. A three dimensional
numerical model of off-axis retrieval is presented. Multimode storage is considered,
and the EIT, Raman, CRIB and AFC protocols are analyzed. It is shown that
inhomogeneous broadening improves the multimode capacity of a memory. Raman
storage in a diamond crystal is shown to be feasible. Finally, experimental progress
toward implementing a Raman quantum memory in cesium vapour is described.
Acknowledgements
I have been very lucky to work with a fantastic set of people. I owe a debt of
gratitude to my supervisor Ian Walmsley, whose implacable good humour always
transmutes frustration into comedy, and who has overseen my experimental failures
with only mild panic. Much of the theory was conceived in the course of meetings
with Karl Surmacz, who has been Dr. Surmacz for a year already. I fried my first
laser in the lab with Felix Waldermann, and his patience and subtle sense of humour
are missed — he is also qualified and long-gone! My current postdocs Virginia Lorenz
and Ben Sussman have been a continual source of exciting discussion, and North
American optimism. We’re gonna make it work guys! And no misunderstanding
can survive a keen frisking at the hands of Klaus Reim, who is currently completing
his D.Phil — and mine — within our group. KC Lee has made the transition
from theorist to experimentalist, without a blip in his coffee intake, and he remains
an inspiration. The help and encouragement of Dieter Jaksch, and more recently
Christoph Simon, are greatly appreciated. Thanks must also be due to my office
neighbours: Pete Mosley, who introduced me to the concept of a progress chart, and
with it a reification of inadequacy, and Dave Crosby, who go-karts better than he
sings.
The rest of the ultrafast group divides cleanly into those who drink tea and those
who do not. A great big thank you to the tea drinkers: you understand that a tea
break is more than a bathroom break with a drink. It lies at the heart of what
it means to prevaricate. Adam Wyatt knows this. He is a tea soldier. As for the
tea-less philistines (you know who you are), I have nothing to say to you. (dramatic
pause). Nothing.
Thanks to my Oxford massif, Andy Scott and Tom Rowlands-Rees, who know a
good lunch when they see one. And in that vein, thanks to Matthijs Branderhorst,
who along with Ben, introduced me to the burrito. There is a growing ultrafast
diaspora — good people in far-off places — and of these I should like to big-up
Daryl Achilles and Jeff Lundeen, who are awethome even when in errr-rr.
I could not have made it this far without the support of Sonia: there can be few
less attractive prospects than the unshaven maniac that is a D.Phil student writing
up. Thank you for keeping me sane!
And lastly my parents. Thanks mum and thanks dad! Next up: driving license.
Contents
1 Introduction
1
1.1
Classical Computation . . . . . . . . . . . . . . . . . . . . . . . . . .
3
1.2
Quantum Computation . . . . . . . . . . . . . . . . . . . . . . . . .
4
1.2.1
Qubits . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
4
1.2.2
Noise . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
6
1.2.3
No cloning . . . . . . . . . . . . . . . . . . . . . . . . . . . .
7
1.2.4
Universality . . . . . . . . . . . . . . . . . . . . . . . . . . . .
9
1.3
Quantum Memory . . . . . . . . . . . . . . . . . . . . . . . . . . . .
10
1.4
Linear Optics Quantum Computing
. . . . . . . . . . . . . . . . . .
11
1.5
Quantum Communication . . . . . . . . . . . . . . . . . . . . . . . .
13
1.6
Quantum Repeaters . . . . . . . . . . . . . . . . . . . . . . . . . . .
19
1.6.1
The Ekert protocol . . . . . . . . . . . . . . . . . . . . . . . .
21
1.6.2
Entanglement Swapping . . . . . . . . . . . . . . . . . . . . .
22
1.6.3
Entanglement Purification . . . . . . . . . . . . . . . . . . . .
24
1.6.4
The DLCZ protocol and number state entanglement . . . . .
24
CONTENTS
1.7
v
Modified DLCZ with Quantum Memories . . . . . . . . . . . . . . .
2 Quantum Memory: Approaches
31
35
2.1
Cavity QED . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
38
2.2
Free space coupling . . . . . . . . . . . . . . . . . . . . . . . . . . . .
39
2.3
Ensembles . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
40
2.3.1
EIT . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
41
2.3.2
Raman
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
46
2.3.3
CRIB . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
50
2.3.4
AFC . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
55
Continuous Variables . . . . . . . . . . . . . . . . . . . . . . . . . . .
58
2.4
3 Optimization
3.1
67
The Singular Value Decomposition . . . . . . . . . . . . . . . . . . .
70
3.1.1
Unitary invariance . . . . . . . . . . . . . . . . . . . . . . . .
75
3.1.2
Connection with Eigenvalues . . . . . . . . . . . . . . . . . .
75
3.1.3
Hermitian SVD . . . . . . . . . . . . . . . . . . . . . . . . . .
76
3.1.4
Persymmetry . . . . . . . . . . . . . . . . . . . . . . . . . . .
76
3.2
Norm maximization . . . . . . . . . . . . . . . . . . . . . . . . . . .
78
3.3
Continuous maps . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
79
3.3.1
Normally and Anti-normally ordered kernels. . . . . . . . . .
81
3.3.2
Memory Optimization. . . . . . . . . . . . . . . . . . . . . . .
81
3.3.3
Unitary invariance . . . . . . . . . . . . . . . . . . . . . . . .
82
CONTENTS
vi
3.4
Optimizing storage followed by retrieval . . . . . . . . . . . . . . . .
85
3.5
A Simple Example . . . . . . . . . . . . . . . . . . . . . . . . . . . .
86
4 Equations of motion
92
4.1
Interaction
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
93
4.2
Electric Field . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
94
4.3
Dipole Operator . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
97
4.3.1
98
Parity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
4.4
Hamiltonian . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 100
4.5
Linear approximation (1) . . . . . . . . . . . . . . . . . . . . . . . . 102
4.6
Rotating Wave Approximation . . . . . . . . . . . . . . . . . . . . . 103
4.7
Unwanted Coupling
4.8
Linear Approximation (2) . . . . . . . . . . . . . . . . . . . . . . . . 106
4.9
Propagation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 108
. . . . . . . . . . . . . . . . . . . . . . . . . . . 105
4.10 Paraxial and SVE approximations . . . . . . . . . . . . . . . . . . . 110
4.11 Continuum Approximation . . . . . . . . . . . . . . . . . . . . . . . 112
4.12 Spontaneous Emission and Decoherence . . . . . . . . . . . . . . . . 116
5 Raman & EIT Storage
120
5.1
One Dimensional Approximation . . . . . . . . . . . . . . . . . . . . 120
5.2
Solution in k-space . . . . . . . . . . . . . . . . . . . . . . . . . . . . 123
5.2.1
Boundary Conditions
. . . . . . . . . . . . . . . . . . . . . . 123
5.2.2
Transformed Equations . . . . . . . . . . . . . . . . . . . . . 124
CONTENTS
5.3
5.4
5.2.3
Optimal efficiency . . . . . . . . . . . . . . . . . . . . . . . . 125
5.2.4
Solution in Wavelength Space . . . . . . . . . . . . . . . . . . 129
5.2.5
Including the Control . . . . . . . . . . . . . . . . . . . . . . 134
5.2.6
An Exact Solution: The Rosen-Zener case . . . . . . . . . . . 137
5.2.7
Adiabatic Limit . . . . . . . . . . . . . . . . . . . . . . . . . . 145
5.2.8
Reaching the optimal efficiency . . . . . . . . . . . . . . . . . 152
5.2.9
Adiabatic Approximation . . . . . . . . . . . . . . . . . . . . 155
Raman Storage . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 159
5.3.1
Validity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 163
5.3.2
Matter Biased Limit . . . . . . . . . . . . . . . . . . . . . . . 167
5.3.3
Transmitted Modes. . . . . . . . . . . . . . . . . . . . . . . . 168
Numerical Solution . . . . . . . . . . . . . . . . . . . . . . . . . . . . 182
5.4.1
5.5
Dispersion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 187
Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 188
6 Retrieval
6.1
vii
190
Collinear Retrieval . . . . . . . . . . . . . . . . . . . . . . . . . . . . 190
6.1.1
Forward Retrieval . . . . . . . . . . . . . . . . . . . . . . . . 191
6.2
Backward Retrieval . . . . . . . . . . . . . . . . . . . . . . . . . . . . 197
6.3
Phasematched Retrieval . . . . . . . . . . . . . . . . . . . . . . . . . 207
6.4
6.3.1
Dispersion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 210
6.3.2
Scheme . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 211
Full Propagation Model . . . . . . . . . . . . . . . . . . . . . . . . . 213
CONTENTS
viii
6.4.1
Diffraction
. . . . . . . . . . . . . . . . . . . . . . . . . . . . 215
6.4.2
Control Field . . . . . . . . . . . . . . . . . . . . . . . . . . . 216
6.4.3
Boundary Conditions
6.4.4
Read out . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 220
6.4.5
Efficiency . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 222
. . . . . . . . . . . . . . . . . . . . . . 218
6.5
Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 223
6.6
Angular Multiplexing . . . . . . . . . . . . . . . . . . . . . . . . . . . 228
6.6.1
Optimizing the carrier frequencies . . . . . . . . . . . . . . . 229
6.6.2
Capacity
7 Multimode Storage
7.1
7.2
234
Multimode Capacity from the SVD . . . . . . . . . . . . . . . . . . . 235
7.1.1
Schmidt Number . . . . . . . . . . . . . . . . . . . . . . . . . 237
7.1.2
Threshold multimode capacity . . . . . . . . . . . . . . . . . 239
Multimode scaling for EIT and Raman memories . . . . . . . . . . . 241
7.2.1
7.3
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 231
A spectral perspective . . . . . . . . . . . . . . . . . . . . . . 242
CRIB . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 245
7.3.1
lCRIB . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 246
7.3.2
Simplified Kernel . . . . . . . . . . . . . . . . . . . . . . . . . 249
7.3.3
tCRIB . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 254
7.4
Broadened Raman . . . . . . . . . . . . . . . . . . . . . . . . . . . . 262
7.5
AFC . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 268
CONTENTS
ix
8 Optimizing the Control
276
8.1
Adiabatic shaping . . . . . . . . . . . . . . . . . . . . . . . . . . . . 277
8.2
Non-adiabatic shaping . . . . . . . . . . . . . . . . . . . . . . . . . . 279
9 Diamond
286
9.1
Diamond Scheme . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 286
9.2
Quantization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 288
9.3
Acoustic and Optical Phonons
9.4
9.5
9.6
. . . . . . . . . . . . . . . . . . . . . 290
9.3.1
Decay . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 291
9.3.2
Energy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 293
Raman interaction . . . . . . . . . . . . . . . . . . . . . . . . . . . . 294
9.4.1
Excitons . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 294
9.4.2
Deformation Potential . . . . . . . . . . . . . . . . . . . . . . 296
Propagation in Diamond . . . . . . . . . . . . . . . . . . . . . . . . . 299
9.5.1
Hamiltonian . . . . . . . . . . . . . . . . . . . . . . . . . . . . 300
9.5.2
Electron-radiation interaction . . . . . . . . . . . . . . . . . . 301
9.5.3
Electron-lattice interaction . . . . . . . . . . . . . . . . . . . 307
9.5.4
Crystal energy . . . . . . . . . . . . . . . . . . . . . . . . . . 309
Heisenberg equations . . . . . . . . . . . . . . . . . . . . . . . . . . . 310
9.6.1
Adiabatic perturbative solution . . . . . . . . . . . . . . . . . 311
9.7
Signal propagation . . . . . . . . . . . . . . . . . . . . . . . . . . . . 314
9.8
Coupling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 318
9.9
Selection Rules . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 321
CONTENTS
x
9.10 Noise . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 322
10 Experiments
323
10.1 Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 323
10.2 Thallium
10.3 Cesium
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 325
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 326
10.4 Cell . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 328
10.4.1 Temperature control . . . . . . . . . . . . . . . . . . . . . . . 329
10.4.2 Magnetic shielding . . . . . . . . . . . . . . . . . . . . . . . . 329
10.5 Buffer gas . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 331
10.6 Control pulse . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 333
10.6.1 Pulse duration . . . . . . . . . . . . . . . . . . . . . . . . . . 333
10.6.2 Tuning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 336
10.6.3 Shaping . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 337
10.7 Pulse picker . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 337
10.8 Stokes scattering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 338
10.9 Coupling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 342
10.9.1 Optical depth . . . . . . . . . . . . . . . . . . . . . . . . . . . 342
10.9.2 Rabi frequency . . . . . . . . . . . . . . . . . . . . . . . . . . 344
10.9.3 Raman memory coupling . . . . . . . . . . . . . . . . . . . . 346
10.9.4 Focussing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 347
10.10Line shape . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 349
10.11Effective depth . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 351
CONTENTS
xi
10.12Optical pumping . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 353
10.12.1 Pumping efficiency . . . . . . . . . . . . . . . . . . . . . . . . 355
10.13Filtering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 356
10.13.1 Polarization filtering . . . . . . . . . . . . . . . . . . . . . . . 358
10.13.2 Lyot filter . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 358
10.13.3 Etalons . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 359
10.13.4 Spectrometer . . . . . . . . . . . . . . . . . . . . . . . . . . . 362
10.13.5 Spatial filtering . . . . . . . . . . . . . . . . . . . . . . . . . . 362
10.14Signal pulse . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 364
10.15Planned experiment . . . . . . . . . . . . . . . . . . . . . . . . . . . 366
11 Summary
369
11.1 Future work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 372
A Linear algebra
374
A.1 Vectors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 375
A.1.1 Adjoint vectors . . . . . . . . . . . . . . . . . . . . . . . . . . 377
A.1.2 Inner product . . . . . . . . . . . . . . . . . . . . . . . . . . . 378
A.1.3 Norm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 379
A.1.4 Bases . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 380
A.2 Matrices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 381
A.2.1 Outer product . . . . . . . . . . . . . . . . . . . . . . . . . . 384
A.2.2 Tensor product . . . . . . . . . . . . . . . . . . . . . . . . . . 385
CONTENTS
xii
A.3 Eigenvalues . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 388
A.3.1 Commutators . . . . . . . . . . . . . . . . . . . . . . . . . . . 389
A.4 Types of matrices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 391
A.4.1 The identity matrix . . . . . . . . . . . . . . . . . . . . . . . 391
A.4.2 Inverse matrix . . . . . . . . . . . . . . . . . . . . . . . . . . 392
A.4.3 Hermitian matrices . . . . . . . . . . . . . . . . . . . . . . . . 393
A.4.4 Diagonal matrices . . . . . . . . . . . . . . . . . . . . . . . . 394
A.4.5 Unitary matrices . . . . . . . . . . . . . . . . . . . . . . . . . 396
B Quantum mechanics
399
B.1 Postulates . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 400
B.1.1 State vector . . . . . . . . . . . . . . . . . . . . . . . . . . . . 400
B.1.2 Observables . . . . . . . . . . . . . . . . . . . . . . . . . . . . 400
B.1.3 Measurements
. . . . . . . . . . . . . . . . . . . . . . . . . . 400
B.1.4 Dynamics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 401
B.2 The Heisenberg Picture . . . . . . . . . . . . . . . . . . . . . . . . . 403
B.2.1 The Heisenberg interaction picture . . . . . . . . . . . . . . . 405
C Quantum optics
407
C.1 Modes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 407
C.2 Quantum states of light . . . . . . . . . . . . . . . . . . . . . . . . . 410
C.2.1 Fock states . . . . . . . . . . . . . . . . . . . . . . . . . . . . 410
C.2.2 Creation and Annihilation operators . . . . . . . . . . . . . . 411
CONTENTS
xiii
C.3 The electric field . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 414
C.4 Matter-Light Interaction . . . . . . . . . . . . . . . . . . . . . . . . . 415
C.4.1 The A.p Interaction . . . . . . . . . . . . . . . . . . . . . . . 415
C.4.2 The E.d Interaction . . . . . . . . . . . . . . . . . . . . . . . 418
C.5 Dissipation and Fluctuation . . . . . . . . . . . . . . . . . . . . . . . 422
D Sundry Analytical Techniques
427
D.1 Contour Integration . . . . . . . . . . . . . . . . . . . . . . . . . . . 427
D.1.1 Cauchy’s Integral Formula . . . . . . . . . . . . . . . . . . . . 429
D.1.2 Typical example . . . . . . . . . . . . . . . . . . . . . . . . . 430
D.2 The Dirac Delta Function . . . . . . . . . . . . . . . . . . . . . . . . 432
D.3 Fourier Transforms . . . . . . . . . . . . . . . . . . . . . . . . . . . . 434
D.3.1 Bilateral Transform . . . . . . . . . . . . . . . . . . . . . . . 434
D.3.2 Unitarity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 434
D.3.3 Inverse . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 435
D.3.4 Shift . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 435
D.3.5 Convolution . . . . . . . . . . . . . . . . . . . . . . . . . . . . 436
D.3.6 Transform of a Derivative . . . . . . . . . . . . . . . . . . . . 436
D.4 Unilateral Transform . . . . . . . . . . . . . . . . . . . . . . . . . . . 437
D.4.1 Shift . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 438
D.4.2 Convolution . . . . . . . . . . . . . . . . . . . . . . . . . . . . 438
D.4.3 Transform of a Derivative . . . . . . . . . . . . . . . . . . . . 439
D.4.4 Laplace Transform . . . . . . . . . . . . . . . . . . . . . . . . 440
CONTENTS
xiv
D.5 Bessel Functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 440
D.5.1 Orthogonality . . . . . . . . . . . . . . . . . . . . . . . . . . . 441
D.5.2 Memory Propagator . . . . . . . . . . . . . . . . . . . . . . . 442
D.5.3 Optimal Eigenvalue Kernel . . . . . . . . . . . . . . . . . . . 445
E Numerics
447
E.1 Spectral Collocation . . . . . . . . . . . . . . . . . . . . . . . . . . . 449
E.1.1 Polynomial Differentiation Matrices . . . . . . . . . . . . . . 452
E.1.2 Chebyshev points . . . . . . . . . . . . . . . . . . . . . . . . . 453
E.2 Time-stepping . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 455
E.3 Boundary Conditions . . . . . . . . . . . . . . . . . . . . . . . . . . . 457
E.4 Constructing the Solutions . . . . . . . . . . . . . . . . . . . . . . . . 459
E.5 Numerical Construction of a Green’s Function . . . . . . . . . . . . . 462
E.6 Spectral Methods for Two Dimensions . . . . . . . . . . . . . . . . . 465
F Atomic Vapours
F.1 Vapour pressure
471
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 471
F.2 Oscillator strengths . . . . . . . . . . . . . . . . . . . . . . . . . . . . 474
F.3 Line broadening . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 476
F.3.1 Doppler broadening . . . . . . . . . . . . . . . . . . . . . . . 476
F.3.2 Pressure broadening . . . . . . . . . . . . . . . . . . . . . . . 477
F.3.3 Power broadening . . . . . . . . . . . . . . . . . . . . . . . . 480
F.4 Raman polarization
. . . . . . . . . . . . . . . . . . . . . . . . . . . 481
List of Figures
1.1
The state space of a qubit . . . . . . . . . . . . . . . . . . . . . . . .
5
1.2
The BB84 protocol . . . . . . . . . . . . . . . . . . . . . . . . . . . .
17
1.3
Entanglement swapping . . . . . . . . . . . . . . . . . . . . . . . . .
23
1.4
Single-rail entanglement swapping . . . . . . . . . . . . . . . . . . .
26
1.5
QKD with single-rail entanglement . . . . . . . . . . . . . . . . . . .
27
1.6
Λ-level structure of atoms for DLCZ . . . . . . . . . . . . . . . . . .
29
1.7
Generation of number state entanglement in DLCZ . . . . . . . . . .
31
1.8
Modification to DLCZ with photon sources and quantum memories .
33
2.1
The simplest quantum memory . . . . . . . . . . . . . . . . . . . . .
36
2.2
Adding a dark state . . . . . . . . . . . . . . . . . . . . . . . . . . .
37
2.3
Cavity QED . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
39
2.4
Confocal coupling in free space . . . . . . . . . . . . . . . . . . . . .
40
2.5
Atomic ensemble memory . . . . . . . . . . . . . . . . . . . . . . . .
41
2.6
EIT. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
42
2.7
Stopping light with EIT. . . . . . . . . . . . . . . . . . . . . . . . . .
45
LIST OF FIGURES
xvi
2.8
Raman storage. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
48
2.9
CRIB storage. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
52
2.10 tCRIB vs. lCRIB. . . . . . . . . . . . . . . . . . . . . . . . . . . . .
54
2.11 AFC storage. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
56
2.12 Wigner distributions. . . . . . . . . . . . . . . . . . . . . . . . . . . .
61
2.13 Atomic quadratures. . . . . . . . . . . . . . . . . . . . . . . . . . . .
64
2.14 QND memory. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
65
2.15 Level scheme for a QND memory. . . . . . . . . . . . . . . . . . . . .
66
3.1
Storage map. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
68
3.2
Linear transformation . . . . . . . . . . . . . . . . . . . . . . . . . .
72
3.3
Persymmetry. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
77
4.1
The Λ-system again. . . . . . . . . . . . . . . . . . . . . . . . . . . .
93
4.2
Time-ordering. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 103
4.3
Useful and nuisance couplings. . . . . . . . . . . . . . . . . . . . . . 105
5.1
Quantum memory boundary conditions. . . . . . . . . . . . . . . . . 124
5.2
Bessel zeros. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 132
5.3
Optimal storage efficiency. . . . . . . . . . . . . . . . . . . . . . . . . 133
5.4
The Rosen-Zener model. . . . . . . . . . . . . . . . . . . . . . . . . . 140
5.5
Raman efficiency. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 162
5.6
Raman storage as a beamsplitter. . . . . . . . . . . . . . . . . . . . . 181
5.7
Modified DLCZ protocol with partial storage. . . . . . . . . . . . . . 182
LIST OF FIGURES
5.8
xvii
Comparison of predictions for the optimal input modes in the adiabatic limit. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 184
5.9
Comparison of predictions for the optimal input modes outside the
adiabatic limit. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 185
5.10 Broadband Raman storage. . . . . . . . . . . . . . . . . . . . . . . . 187
5.11 Broadband EIT storage. . . . . . . . . . . . . . . . . . . . . . . . . . 188
6.1
Forward retrieval. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 197
6.2
Phasematching considerations for backward retrieval. . . . . . . . . . 200
6.3
Backward Retrieval. . . . . . . . . . . . . . . . . . . . . . . . . . . . 206
6.4
Non-collinear phasematching. . . . . . . . . . . . . . . . . . . . . . . 209
6.5
Efficient, phasematched memory for positive and negative phase mismatches. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 209
6.6
Focussed beams. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 216
6.7
Effectiveness of our phasematching scheme. . . . . . . . . . . . . . . 224
6.8
Comparing phasematched and collinear efficiencies. . . . . . . . . . . 225
6.9
Angular multiplexing. . . . . . . . . . . . . . . . . . . . . . . . . . . 230
6.10 Minimum momentum mismatch. . . . . . . . . . . . . . . . . . . . . 232
7.1
Bright overlapping modes are distinct. . . . . . . . . . . . . . . . . . 236
7.2
Visualizing the multimode capacity.
7.3
The appearance of a multimode Green’s function. . . . . . . . . . . . 239
7.4
Multimode scaling for Raman and EIT memories. . . . . . . . . . . . 244
. . . . . . . . . . . . . . . . . . 237
LIST OF FIGURES
xviii
7.5
Scaling of Schmidt number with broadening.
. . . . . . . . . . . . . 252
7.6
Comparison of the predictions of the kernels (7.23) and (7.18). . . . 253
7.7
Understanding the linear multimode scaling of lCRIB. . . . . . . . . 254
7.8
Multimode scaling for CRIB memories. . . . . . . . . . . . . . . . . . 261
7.9
The multimode scaling of a broadened Raman protocol. . . . . . . . 268
7.10 The multimode scaling of the AFC memory protocol. . . . . . . . . . 275
8.1
Adiabatic control shaping. . . . . . . . . . . . . . . . . . . . . . . . . 284
8.2
Non-adiabatic control shaping. . . . . . . . . . . . . . . . . . . . . . 285
9.1
The crystal structure of diamond. . . . . . . . . . . . . . . . . . . . . 287
9.2
Phonon aliasing. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 290
9.3
Phonon dispersion. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 292
9.4
Band structure. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 295
9.5
An exciton. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 296
9.6
The Raman interaction in diamond. . . . . . . . . . . . . . . . . . . 298
10.1 Observing Stokes scattering as a first step. . . . . . . . . . . . . . . . 325
10.2 Thallium atomic structure. . . . . . . . . . . . . . . . . . . . . . . . 326
10.3 Cesium atomic structure. . . . . . . . . . . . . . . . . . . . . . . . . 327
10.4 First order autocorrelation. . . . . . . . . . . . . . . . . . . . . . . . 334
10.5 Second order interferometric autocorrelation. . . . . . . . . . . . . . 336
10.6 Stokes scattering efficiency. . . . . . . . . . . . . . . . . . . . . . . . 342
10.7 Cesium optical depth. . . . . . . . . . . . . . . . . . . . . . . . . . . 345
LIST OF FIGURES
xix
10.8 Cesium D2 absorption spectrum. . . . . . . . . . . . . . . . . . . . . 350
10.9 Absorption linewidth. . . . . . . . . . . . . . . . . . . . . . . . . . . 352
10.10Equal populations destroy quantum memory. . . . . . . . . . . . . . 354
10.11Optical pumping. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 355
10.12Verifying efficient optical pumping. . . . . . . . . . . . . . . . . . . . 357
10.13Lyot filter. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 360
10.14Stokes filtering. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 361
10.15Backward Stokes scattering. . . . . . . . . . . . . . . . . . . . . . . . 364
10.16A possible design for demonstration of a cesium quantum memory. . 368
A.1 A vector. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 376
A.2 The inner product of two vectors. . . . . . . . . . . . . . . . . . . . . 380
A.3 A matrix acting on a vector. . . . . . . . . . . . . . . . . . . . . . . . 383
A.4 Eigenvectors and eigenvalues. . . . . . . . . . . . . . . . . . . . . . . 389
A.5 Non-commuting operations. . . . . . . . . . . . . . . . . . . . . . . . 390
A.6 A unitary transformation. . . . . . . . . . . . . . . . . . . . . . . . . 396
C.1 Symmetrized photons. . . . . . . . . . . . . . . . . . . . . . . . . . . 413
D.1 Contour integrals. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 429
D.2 Upper closure. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 432
D.3 Integration limits. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 439
D.4 Lower closure. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 444
E.1 The method of lines. . . . . . . . . . . . . . . . . . . . . . . . . . . . 449
LIST OF FIGURES
xx
E.2 Periodic extension. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 451
E.3 Chebyshev Points. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 455
E.4 Example solutions. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 462
E.5 A numerically constructed Green’s function. . . . . . . . . . . . . . . 464
E.6 Spectral methods in two dimensions. . . . . . . . . . . . . . . . . . . 470
F.1 Vapour pressure. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 473
F.2 The Doppler shift. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 477
F.3 Collisions in a vapour. . . . . . . . . . . . . . . . . . . . . . . . . . . 478
F.4 Polarization selection rules. . . . . . . . . . . . . . . . . . . . . . . . 483
F.5 Alternative scattering pathways. . . . . . . . . . . . . . . . . . . . . 486
Chapter 1
Introduction
The prospect of building a quantum computer, with speed and power far outstripping
the best possible classical computers, has motivated an enormous and sustained
research effort over the last two decades. In this thesis we explore a number of
candidates for the ‘memory’ that would be required by such a device. As we will
see, building a quantum memory is considerably harder than fabricating the RAM
chips used by modern computers. For instance, it is not possible to copy quantum
information, nor can quantum information be digitized. These facts make quantum
storage particularly vulnerable to noise, and loss — problems for which solutions
must be found before quantum computation can mature into a viable technology.
The bulk of this thesis is concerned with optimization of the efficiency and storage
capacity of a quantum memory. We focus on optical memories, in which a pulse of
light is ‘stopped’ for a controllable period, before being re-released.
The structure of the thesis is as follows. In this chapter we introduce the concepts
2
of quantum computing and quantum communication, and we discuss the context
and motivation for the present work on quantum memories. In Chapter 2 we survey
the various approaches to quantum memory, and we describe the principles behind
the memory protocols analyzed later. Chapter 3 introduces the mathematical basis
for our approach to analyzing and optimizing ensemble memories — the Green’s
function and its singular value decomposition. In Chapter 4 we derive the equations
of motion describing the quantum memory interaction in an ensemble of Λ-type
atoms. In Chapter 5 we apply the techniques of Chapter 3 to this interaction. Several
new results are derived, and connections are made with previous work. Chapter 6
is concerned with retrieval of the stored excitations from a Λ-ensemble. It is shown
that both forward and backward retrieval are problematic. A numerical model
is presented that confirms the efficacy of an off-axis geometry, which solves these
problems. In Chapter 7 we move on to consider multimode storage. Our formalism
provides a natural way to calculate the multimode capacity of a memory, and we
study the multimode scaling of all the memory protocols introduced in Chapter 2.
Chapter 8 describes how to optimize a Λ-type memory by shaping the ‘control pulse’.
In Chapter 9 we study the Raman interaction in a diamond crystal: we show that a
diamond Raman quantum memory is feasible. Finally in Chapter 10 we review our
attempts to implement a Raman quantum memory in the laboratory, using cesium
vapour.
But let us begin at the beginning.
1.1 Classical Computation
1.1
3
Classical Computation
Classical computers are conventional computers, like the one I am using to typeset
this document. Their importance as enablers of technological progress, as well as
their utility as a technology in their own right, attest to the fantastic potential of
classical computation. They are typified by the use of bit1 strings — sequences of
1’s and 0’s — to encode information. Information is processed by application of
binary logic to the bits. That is, Boolean operations such as OR, AND or not-AND
(NAND). This last operation is a universal gate, because any logic operation can be
constructed using only NAND gates. Such a gate can be implemented electronically
using a pair of transistors, millions of which can be combined on a single silicon
chip. The rest is history.
Computers have progressed in leaps and bounds over the last fifty years. In 1965
computers were developing so fast that Gordon Moore, a founder of the industrial
giant Intel, proposed a ‘law’, stipulating that the number of transistors comprising
a processor would double every year [2] . Incredibly, this exponential improvement in
computing power has persisted for over 40 years. But improvement by miniaturization cannot continue indefinitely. The reason for this is that the physics of electronic
components undergoes a qualitative change at small scales: classical physics becomes
quantum physics. In his 1983 lectures on computation [3] , Richard Feynman consid1
‘Bit’ first appeared in Claude Shannon’s 1948 paper on the theory of communication as a
contraction of ‘binary digit’ [1] ; the name is apposite, since one bit is the smallest ‘piece’ or ‘chunk’
of information there can be: one bit of information is one bit of information. Shannon attributes
the term to John Tukey, a creator of the digital Fourier transform, who is also credited with coining
the word ‘software’.
1.2 Quantum Computation
4
ers the fate of classical computation as shrinking dimensions bring quantum effects
into play. Two years later David Deutsch published the first explicitly quantum
algorithm [4] , demonstrating how quantum physics actually permits more powerful
computation than classical physics allows. A quantum computer, capable of harnessing this greater power, must process quantum information, encoded not with
ordinary bits, but with quantum bits.
1.2
1.2.1
Quantum Computation
Qubits
A quantum bit — a qubit 2 — is an object with two mutually exclusive states, 0
and 1, say. The only difference with a classical bit is that the object is described
by quantum mechanics. Accordingly, we label the two states by the kets |0i and |1i
(see Appendix B). These kets are to be thought of as vectors in a two dimensional
space: the state space of the qubit (see Figure 1.1). The classical property of mutual
exclusivity is manifested in the quantum formalism by requiring that |0i and |1i are
perpendicular to one another in the state space. In general, the state of the qubit
can be any vector, of length 1, in the state space. Since both the kets |0i and |1i
have length 1, and since they point in perpendicular directions, an arbitrary qubit
state |ψi can always be written as a linear combination of them,
|ψi = α|0i + β|1i,
2
(1.1)
The term ‘qubit’ first appears in a paper by Benjamin Schumacher in 1995 [5] ; he credits its
invention to a conversation with William Wootters.
1.2 Quantum Computation
5
where α and β are two numbers which must satisfy the normalization condition
|α|2 + |β|2 = 1. States like (1.1), which are a combination of the two mutually
exclusive states |0i and |1i, are called superposition states, or just superpositions.
What does it mean to say that a qubit is in a superposition between its two mutually
exclusive states? Somehow it is both 0 and 1 at the same time. Physically, this is
like saying that a switch is both ‘open’ and ‘closed’, or that a lamp is both ‘on’ and
also ‘off’. Already, for the simplest possible system, without any real dynamics —
no interactions, nothing happening — we see that the basic structure of quantum
mechanics does not sit well with our intuition. Despite these interpretational diffi-
stat
e
sp
ace
Figure 1.1
A visual representation of the state space of a qubit
culties, superposition is central to the success of quantum mechanics. Atomic and
molecular physics, nuclear and particle physics, optics and electronics all make use of
superpositions to successfully explain processes and interactions. From the point of
view of computation, the existence of states like (1.1) provides a clue to the greater
capabilities of a quantum computer. Each qubit has two ‘parts’, the |0i part and the
1.2 Quantum Computation
6
|1i part; logical operations on qubits act on both parts together, and the output of
a calculation also has these two parts. So there’s some sense in which a qubit plays
the role of two classical bits, stuck ‘on top of eachother’. David Deutsch coined the
term quantum parallelism for this property — he considers it to be the strongest
evidence for the existence of parallel universes. The B-movie-esque connotations
of this ‘many-worlds’ view make it generally unpopular among physicists, but the
appeal of quantum computing remains, independently of how it is understood.
1.2.2
Noise
A potential difficulty associated with quantum computing is also apparent from
(1.1): the numbers α and β can be varied continuously (subject to the normalization constraint). So the number of possible states |ψi of a qubit is infinite! This also
hints at their greater information carrying capacity, but it means that they must
be carefully protected from the influence of noise. Classical bits have exactly two
states; if noise introduces some distortions, it is usually possible to correct these
simply by comparing the distorted bit to an ideal one. Only very large fluctuations
can make a 0 look like a 1, so the discrete structure of classical bits makes them
very robust. By contrast, a perturbed qubit state is also a valid qubit state. In
this respect, the difference between bits and qubits can be likened to the difference
between digital and analogue musical recordings: The quality of music reproduced
by a CD does not degrade gradually with time, whereas old cassettes sound progressively worse as distortions creep into the waveform imprinted on the tape. In fact,
1.2 Quantum Computation
7
it is possible to correct errors by constructing codes involving bunches of qubits.
The invention of these codes in 1996 by Calderbank, Shor and Steane [6,7] was a
major milestone in demonstrating the practical viability of quantum computation.
Nonetheless these error correcting schemes currently require that noise is suppressed
below thresholds of a few percent, which makes techniques for isolating qubits from
noise a technological sine qua non.
1.2.3
No cloning
Another problematic aspect of quantum information is that it cannot be copied.
The proof of this fact is known as the no-cloning theorem [8] . Suppose that we have
a device which can copy a qubit. If we give it a qubit in state |ψi, and also a ‘blank’
qubit in some standard initial state |blanki, this machine spits out our original qubit,
plus a clone, both in the state |ψi. In Dirac notation, using kets, the action of our
qubit photocopier is written as
U |blanki|ψi = |ψi|ψi.
(1.2)
Here U is the unitary transformation implemented by our machine. Unitary transformations are those which preserve the lengths of the kets upon which they act.
Since all physical states have length 1, and any process must produce physical states
from physical states, it follows that all processes are described by length-preserving
— unitary — transformations (see §B.1.4 in Appendix B). Had we fed our machine
1.2 Quantum Computation
8
a different state, for example |φi, we would have
U |blanki|φi = |φi|φi.
(1.3)
The length of a ket |ϕi is defined by taking the scalar product hϕ|ϕi of |ϕi with
itself (see §A.1.2 in Appendix A). To prove the impossibility of cloning, we take the
scalar product of the first relation (1.2) with the second, (1.3).
hψ|hblank|U † U |blanki|φi = hψ|hψ||φi|φi.
(1.4)
The U acting on |blanki on the left hand side does not change its length, which is
just equal to 1, so the result simplifies to
hψ|φi = hψ|φi2 .
(1.5)
Clearly, this expression does not hold for arbitrary choices of |ψi and |φi, and therefore cloning an arbitrary qubit is impossible. In fact, (1.5) is only true when hψ|φi is
either 1 or 0. The first case corresponds to |ψi = |φi, which says that it is possible to
build a machine that can make copies of one particular, pre-determined state. The
second case occurs only when |ψi and |φi are perpendicular, as is the case for |0i and
|1i. This says that it is possible to clone mutually exclusive states. Indeed, this is
precisely what classical computers are doing when they copy digitized information.
An immediate consequence of the no-cloning theorem is that a quantum memory
1.2 Quantum Computation
9
must work in a qualitatively different way to a classical computer memory. To
store quantum information, that information must be transferred to the memory,
rather than simply copied to it. It is never possible to ‘save a back-up’, as we
routinely do with classical computers. To build a quantum memory, we must find
an interaction between information carrier and storage medium which ‘swaps’ their
quantum states, so that the storage medium ends up with all the information, with
nothing left behind. In this thesis we will examine various ways of accomplishing this
optically, by considering collections of atoms which essentially ‘swallow’ a photon
in a controlled way, completely transferring the quantum state of an optical field to
that of the atoms.
1.2.4
Universality
Any classical computation is possible provided that one is able to apply NAND gates
to pairs of bits. What is required to perform arbitrary quantum computations?
This question is not trivial, but the answer is fortuitously simple [9] . Any quantum
computation can be performed, provided that one is able to arbitrarily control the
state of any qubit (single-qubit rotations), and provided that one can make pairs of
qubits interact with one another (two-qubit gates). It is generally sufficient to have
only a single type of interaction, so long as the final state of both interacting qubits
depends in some way on the initial state of both qubits. Such a gate is known as an
entangling gate, and they are notoriously difficult to implement.
1.3 Quantum Memory
1.3
10
Quantum Memory
In the light of the preceding discussion, a quantum memory can be understood as a
physical system that is well protected from noise, and that can be made to interact
with information carriers so that their quantum state is transferred into, or out of,
the memory. Note how we have distinguished the system comprising the memory
from the information carriers. In many cases, this distinction is artificial. For instance, in ion trap quantum computing [10] , the hyperfine states of calcium ions are
used as qubits. These ions are isolated from their noisy environment by trapping
them with oscillating electric fields; the quantum states of the qubits therefore remain un-distorted for long periods (on the order of seconds), and so there is no need
to transfer these states into a separate memory. But there are other proposals for
quantum computing that make explicit use of quantum memories. An example is
the use of nitrogen-vacancy centres in diamond for quantum computing [11,12] . Here
a single electron from a nitrogen atom, lodged in a diamond crystal and surrounded
by carbon atoms, is used as a qubit. The electron qubit can be controlled with laser
pulses to perform computations, but this very sensitivity to light makes it susceptible to damage from noise. Therefore a scheme was devised to transfer the quantum
state of the electron to that of a nearby carbon nucleus. The carbon nucleus is
de-coupled from the optical field, and it can be used to store quantum information
for many minutes.
A common theme among such computation schemes is an antagonism between
controllability and noise-resilience. That is, systems which are easily manipulated
1.4 Linear Optics Quantum Computing
11
and controlled with external fields are susceptible to noise from those same fields,
while well-isolated systems that are not badly affected by noise are generally hard
to access and control in order to perform computations. This trade-off leads to a
natural division of labour between systems that are easily manipulated, but shortlived, and systems that are not easily controlled, but long-lived. Many quantum
computing architectures put both types of system to use, the former as quantum
processor, the latter as quantum memory.
1.4
Linear Optics Quantum Computing
Since James Clerk Maxwell wrote down the equations of electromagnetism in 1873,
physics has undergone profound upheavals at least twice, with the development of
both Relativity and Quantum Mechanics in the early twentieth century. Maxwell’s
equations have weathered these storms with astonishing fortitude, being both relativistically covariant and directly applicable in quantum field theory. They are
probably the oldest correct equations in physics. Implicit within them is a description of the photon, the quantum of the electromagnetic field. Photons come with
one of two polarizations, and superposition states of these polarizations are readily
prepared in the lab. In addition, they are themselves discrete entities, and it is possible to generate superpositions of different numbers of photons. Photons therefore
embody the archetypal qubit, and for this reason Maxwell’s equations remain as
central to the emerging discipline of quantum information processing as they were
to the pioneers of telegraphy and radio.
1.4 Linear Optics Quantum Computing
12
Photons occupy a frustrating territory on the balance sheet of usefulness for
quantum computation. They are ideal qubits, and arbitrary manipulation of their
polarization and number states can be accomplished with simple waveplates, beamsplitters and phase-shifters. That is, single-qubit rotations are ‘cheap’. Unfortunately, entangling gates between photons are much more difficult to realise. This
is unsurprising, since such a gate requires that two photons be made to interact
with one another, and it is well known that light does not generally interact with
light: torch beams do not ‘bounce off’ each other; rather they pass through each
other unaffected. In 2001 Emanuel Knill, Raymond Laflamme and Gerard Milburn showed how to overcome these difficulties by careful use of measurements [13] ,
making universal quantum computation possible with only ‘linear optics’. Further
developments [14,15] have cemented linear optics quantum computing (LOQC) as an
important paradigm for the future of quantum computation. However the two-qubit
gates proposed are generally non-deterministic. As the number of gates required in
a computational step increases, the probability that all gates are implemented successfully decreases exponentially, so that large computations must be repeated many
times for yielding reliable answers. This problem of scalability can be mitigated if
the photons output from successful gates can be stored until all the required gates
succeed. But photons generally have a short lifetime because they travel at the
speed of light: if they are confined in a laboratory they must be trapped by mirrors
(in a cavity) or by a waveguide (optical fibre), and absorption or scattering losses
are inevitable on time scales of milliseconds or greater [16] . Therefore the ability to
1.5 Quantum Communication
13
transfer the quantum state of a photon into a quantum memory would be a boon to
LOQC.
Another possibility for quantum computing with photons is to implement twoqubit gates inside a quantum memory. Single-qubit operations are easily performed
on the photons directly; when interactions are needed, the photons are transferred
to atomic excitations which can be manipulated with external fields to accomplish
the entangling gates [17–19] .
Applications such as these constitute the most ambitious motivation for the study
of optical quantum memories. In the next section we will see that quantum memories are also required in extending the range of so-called quantum communication
protocols, which provide guaranteed security from eavesdroppers.
1.5
Quantum Communication
Although practical quantum computing remains beyond the reach of current technology, another application of quantum mechanics has already made the leap into
the commercial sector. Quantum Key Distribution (QKD) is a technique which allows two communicating parties to be absolutely certain, in so far as the laws of
physics are known to be correct, that their messages have not been intercepted [20] .
It is possible to purchase QKD systems from two companies: MagiQ based in New
York, and ID Quantique in Geneva; many other businesses are incumbent, and the
market for such guaranteed-secure communication is estimated at around a billion
dollars annually. The idea behind QKD is simple: if Alice sends a message to Bob
1.5 Quantum Communication
14
in which she has substituted each letter for a different one, in a completely random way, neither Bob, nor anyone else, can decode the message, unless Alice tells
them how she did the substitution. This information is known as the key, and only
someone in possession of the full key has access to the contents of Alice’s message.
Encrypting messages in this way is the oldest and simplest method of encryption. It
is absolutely and completely secure, provided that only the intended recipient has
access to the key. Once the key has been used, it should not be used again, since
with repeated use an eavesdropper, conventionally called Eve, might start to see
patterns in the encrypted messages and begin to guess the substitution rule. For
this reason this encryption protocol is known as the one time pad. For each message
sent, a new, completely random key must be used by both Alice and Bob. How does
Alice send the keys to Bob? If these are encrypted, she will need to send another
key beforehand, and our perfect security is swallowed by an infinite regression. If
she sends the keys unencrypted, can she be sure that Eve has not intercepted them?
If she has, then Eve has access to all of Alice’s subsequent messages, and there’s no
way for Alice or Bob to know their code has been cracked until the paparazzi arrive.
These issues are eliminated by public key cryptography. Here, Bob tells Alice
the substitution rule she should use for her message. The encryption is done in
such a way that Alice’s message cannot be decoded using this rule, so it doesn’t
matter if Eve discovers it. Alice then sends her coded message to Bob, who knows
how to decrypt the message. An implementation of this idea using the mathematics of large prime numbers was developed in 1978 by Ron Rivest, Adi Shamir and
1.5 Quantum Communication
15
Leonard Adleman [21] . The RSA cryptosystem is the industry standard for secure
communication over the internet; much of modern finance relies on its security. But
unlike the one time pad, no-one has proved that it is secure. The RSA algorithm
relies on the empirical fact that it is computationally very demanding to find the
prime factors of a large number. That is, if two large prime numbers p and q are
multiplied together to give their product n, it is not practically possible to find p
and q, given knowledge of n alone. The best algorithm, the number field sieve, can
find the factors of n in roughly eN
1/3
2/3
log2
N
computational steps [22] , where N is
the number of bits needed to represent n. This exponential scaling means that it
is easy to make the calculation intractably long by making n just a little larger.
But this is not the whole story. In 1994 Peter Shor showed how a quantum computer could be used to perform this factorization much faster [23] . Shor’s algorithm
requires just N 2 (log2 N ) log2 (log2 N ) steps, an exponential improvement over the
best conventional methods. A quantum computer that can implement this algorithm efficiently does not yet exist, although proof-of-principle experiments using
LOQC have been performed [24–26] . But it is now known that the RSA cryptosystem
is living on borrowed time: if a practical quantum computer is ever made, modern
secure communications will be spectacularly compromised.
Enter QKD. QKD makes use of quantum mechanics to distribute the keys required for a one time pad protocol in a secure way, avoiding the infinite regress
arising from a classical protocol, and obviating the need to rely on the fatally flawed
RSA system. The goal is to provide both Alice and Bob with an identical string
1.5 Quantum Communication
16
of completely random bits, which they can use as keys to encrypt and decrypt a
message. The most widely known protocol used to do this is known as BB84, after
the 1984 paper by Bennett and Brassard [27] . Alice sends photons, one at a time, to
Bob. Alice can choose the polarization of each photon to point in one of four possible directions: horizontal, vertical, anti-diagonal or diagonal (|Hi, |V i, |Ai or |Di;
see Figure 1.2). These directions form a pair of perpendicular polarizations, with a
quarter-turn between them. Each of these pairs is known as a basis. The unrotated
basis contains the |Hi and |V i polarizations, and is known as the rectilinear basis.
The rotated basis contains the |Ai and |Di polarizations, and is referred to as the
45-degree basis. Bob can measure the polarization of the photons he receives from
Alice, but to do so he has to line up his detector with either the rectilinear or the
45-degree basis — he has to choose. The measurement gives one of two possible
results, either 0 or 1, but these results mean nothing unless the photon polarization belonged to the basis that Bob chose to measure. For example, a |Di photon
polarized in the 45-degree basis will give a completely random result, either 0 or 1
with equal probability, if Bob aligns his detector with the rectilinear basis. So if he
gets the basis wrong, his measurement results are useless. If he chooses correctly
however, and the photon is polarized in the same basis that he measures in, then the
0 or 1 results tell him to which of the two possible directions in that basis the photon
polarization belonged. So for instance if Bob aligns his detector with the 45-degree
basis, the |Di photon will always give a 1 result. An |Ai photon would give a 0 result
for this measurement, while either |Hi or |V i photons would give random results.
1.5 Quantum Communication
17
This strange property of photons, that measurements give useful or uncertain results
depending on the measurement basis, is a manifestation of Heisenberg’s uncertainty
principle [28] . It is uniquely quantum mechanical.
45
-d
eg
re
e
rectilinear
Figure 1.2 BB84 protocol. Alice sends photons to Bob with polarizations chosen randomly from the four possible directions |Hi/|V i
and |Ai/|Di, represented here as qubit states. To measure the polarization, Bob (or Eve) must choose a basis, rectilinear or 45-degree,
for their measurement. Only photons polarized in this basis will yield
useful information.
To proceed with QKD, Alice generates two completely random, unrelated, bit
strings. For each photon, she uses the first bit string to decide in which basis to
polarize her photon. For instance, 0 could signify rectilinear and a 1 would mean 45degree. Then she uses the second bit string to decide which of the two perpendicular
polarizations in that basis to use. When Bob receives Alice’s photons, he records
the results of his measurements, and the basis he used for each measurement. Alice
then sends Bob her first bit string. This tells him the bases each photon belonged
to. He knows that his results are useless every time he chose the wrong basis for
his measurement. So he discards these results. The remaining results tell him the
1.5 Quantum Communication
18
correct polarization of each photon. That is, Bob’s remaining results now tally
exactly with Alice’s second bit string. Alice and Bob now share a cryptographic
key they can use for a one-time pad. But what about Eve? Well, Eve may have
intercepted Alice’s first bit string, but this only contains information about which
results to discard, it tells Eve nothing about what those measurement results were,
so this does not help her in cracking Alice and Bobs’ code. She could also have
intercepted the photons that Alice sent, and tried to measure their polarizations.
But, just like Bob, she has to guess at which basis to measure in. She gets a result,
but she has no idea whether the basis she chose is correct. She has to send photons
on to Bob, otherwise he will receive nothing and get suspicious. But Eve does not
know what polarization to give her photons, because she doesn’t know whether her
measurements are reliable or not. Suppose she just decides to send photons polarized
according to her measurement results in the basis she chose to measure in. Bob
receives these photo ns and measures them, none the wiser. But after Alice sends
her first bit string, and Bob discards his unreliable measurements, Bob’s remaining
results may not tally perfectly with Alice’s anymore. This is because sometimes Eve
will have chosen a different basis to Alice, obtained useless measurement results, and
sent photons to Bob with the wrong polarization. So Bob and Alice can compare
their keys, or a small part of them, and if they do not match up, they know that Eve
has been tampering with their photons. There is no way for Eve to listen in without
Alice and Bob discovering her presence. In quantum mechanics, measurements affect
the system being measured, and the BB84 protocol exploits this fact to guarantee
1.6 Quantum Repeaters
19
secure communication.
Quantum memories are not needed in the above protocol, provided Alice’s photons survive to reach Bob. As mentioned in Section 1.4 in the context of LOQC,
photons generally do not survive for longer than around 1 ms in an optical fibre, so
Bob should not be further than around 200 miles from Alice, otherwise her photons
will be scattered or absorbed before they reach him. In order to extend the distance
over which QKD is possible, some kind of amplifier is needed, which can give the
photons a ‘boost’, while maintaining their quantum state — that is, their polarization. But the photons are qubits. They cannot simply be copied; we know this
from the no-cloning theorem (see Section 1.2.3). What is required is a modification
to the protocol described above, and a device known as a quantum repeater. Such
a device requires a quantum memory. If quantum memories can be made efficient,
with storage times on the order of 1 s, intercontinental quantum communication becomes possible. In the next section we introduce the quantum repeater, and discuss
the usefulness of quantum memories in this context.
1.6
Quantum Repeaters
A quantum repeater is a device designed to extend entanglement. Entanglement is
a purely quantum mechanical property that can be used as a resource to perform
QKD. In this section we will introduce entanglement, examine how it degrades over
long distances, and how quantum repeaters ameliorate this degradation.
Entanglement is a property of composite quantum systems. As an example,
1.6 Quantum Repeaters
20
consider two qubits. Classically, a system composed of two parts could be described
by the states of each part. In quantum mechanics this is not always true. Just as
a qubit can exist in a superposition of different states, so a system comprising two
qubits can exist in a superposition of different combined states. Such states arise
from a blend of correlation and indeterminism. To see this, suppose that we have a
machine that produces two photons, each with the same polarization. Now suppose
that the direction of this polarization is not fixed. It might polarize both photons
horizontally, we’ll label this polarization |0i, or vertically, |1i. We have no way of
knowing which of these two polarizations the machine uses, we only know that both
photons will have the same polarization. Such a state cannot be described by talking
about each photon in turn, as is clear from the language we used to describe the set
up. Using subscripts to denote the two photons, the state is written as
1
|ψi = √ (|0i1 |0i2 + |1i1 |1i2 ) .
2
(1.6)
√
The factor of 1/ 2 appears simply to fix the length of the state vector |ψi to 1. This
state is an entangled state, because there is no way to write it as a product of states
of the first photon with states of the second. It expresses the two properties of our
machine: first, that the two photons always have the same polarization, and second,
that it is not certain which of the two polarizations will be produced. In fact, because
the two possible states |0i and |1i are mutually exclusive, (1.6) is a maximally
entangled state, sometimes known as a Bell state. Bell states represent much of
1.6 Quantum Repeaters
21
what is counter-intuitive about quantum mechanics. Their name derives from John
Bell’s famous 1964 paper [29] in which he proves that these states are incompatible
with local realism. A ‘local’ world is one in which no effect can propagate faster
than the speed of light; a ‘real’ world is one in which all properties can be assigned
definite values at all times. That modern physics describes states which do not admit
a local realistic interpretation is intriguing and controversial. Below we will see that
these states are also a resource for quantum communication. If their use becomes
widespread, we will be in the awkward position of deriving practical benefits from
a technology based on a philosophical conundrum!
1.6.1
The Ekert protocol
In 1991 Artur Ekert proposed a modification of the BB84 QKD protocol based on
the use of Bell states like (1.6) [30] . In this protocol, our machine for generating
entangled photon pairs is used. One photon from each pair is sent to Alice, the
other to Bob. Now Alice and Bob both have polarization detectors; they each have
to choose a basis to measure their photons in. Sometimes they will choose the same
basis as eachother, sometimes they will choose different bases. When they choose
differently, their results are meaningless, but when they choose the same basis, their
results are perfectly correlated. This is obvious for the rectilinear basis by inspection
of the form of (1.6). A little algebra shows that the same perfect correlations also
hold if both Alice and Bob measure in the 45-degree basis. The QKD is accomplished
in the same way as for the BB84 protocol: Alice tells Bob the measurement bases
1.6 Quantum Repeaters
22
she used; Bob discards the results of all measurements where his basis differed from
Alice’s. Alice and Bob then compare part of the remaining results to check that
they are correlated, as they should be. Poor correlations signify the presence of an
eavesdropper.
The importance of this modified protocol is that entanglement is a transferrable
resource. Below we will see how entanglement can be swapped between photons to
extend the range of quantum communication.
1.6.2
Entanglement Swapping
Entanglement swapping allows one to entangle two photons that have never encountered eachother. The situation is sketched in Figure 1.3. Two sources each emit
a pair of entangled photons in the state (1.6). One photon from each pair is sent
into a polarizing beam splitter, which transmits horizontally polarized photons, and
reflects vertically polarized photons. Behind the beamsplitter are a pair of photon
detectors. The beamsplitter has the effect of limiting the information we can learn
about the photons from the detectors. For instance, if both photon detectors D1
and D2 fire together, it could be that photons (2) and (3) were both vertically polarized, or that they were both horizontally polarized. That is, a ‘coincidence count’
from D1 and D2 only tells us that photons (2) and (3) had the same polarization;
it reveals nothing about what that polarization was. But we know from the state
(1.6) that photon (1) has the same polarization as photon (2), and similarly that
photon (4) has the same polarization as photon (3). So if photons (2) and (3) have
1.6 Quantum Repeaters
23
the same polarization, so do photons (1) and (4). Their polarization is unknown,
but correlated. Therefore, after a coincidence count, the two remaining photons,
(1) and (4), are in a Bell state. The entanglement between photons (1)-(2) and
(3)-(4) has been swapped to photons (1)-(4). This procedure was first demonstrated
experimentally by Jian Wei-Pan et al. in 1998 [31] , and is now an essential tool for
LOQC.
&
D1
D2
PBS
1
2
S1
3
4
S2
Figure 1.3 Entanglement swapping. Two independent sources, S1
and S2, emit pairs (1)-(2) and (3)-(4) of polarization entangled photons. Photons (2) and (3) are directed into a polarizing beam splitter
(PBS). When both detectors D1 and D2 fire behind the PBS, photons
(1) and (4), which have never met, become entangled.
It’s clear from the above arguments that entanglement swapping is not much
more than a re-assignment of our knowledge regarding correlations, in the light of
a measurement carefully designed to reveal only partial information. Nonetheless,
a real resource — entanglement — has been extended over a larger distance by this
procedure. And QKD can now be performed using photons (1) and (4).
1.6 Quantum Repeaters
1.6.3
24
Entanglement Purification
So far we have shown how entanglement can be extended over large distances by
swapping perfect Bell states, each distributed over shorter distances. In practice,
however, propagating even over short distances can distort the polarizations of the
photons. Small distortions do not completely destroy the entanglement; rather there
is a smooth degradation in the usefulness of the photons for QKD as the distortions
become worse. Nevertheless, with each entanglement swap, these deleterious effects
are compounded, so that the entanglement vanishes after only a few swapping operations. However, it is possible to transform several poorly entangled photon pairs
into one photon pair with near-perfect entanglement using an ‘entanglement purification protocol’. Several of these exist [32–36] , generally they involve mixing and
measuring photons in a way that strengthens the correlations between the remaining
photons. Such procedures allow entanglement to be ‘topped up’, at the expense of
having to use more photons. QKD across distances much larger than those over
which distortions affect photons can then be implemented. Below we will introduce
a paradigm for constructing a quantum repeater which does not explicitly make
use of polarization entanglement, but which combines entanglement swapping and
entanglement purification into a single step.
1.6.4
The DLCZ protocol and number state entanglement
As mentioned previously, photons have several different degrees of freedom that can
be used to encode qubits. We have mostly focussed on polarization qubits for QKD,
1.6 Quantum Repeaters
25
but in the following protocol qubits are encoded in the number of photons occupying
a given optical mode, so-called single-rail encoding. Consider a machine similar to
the Bell state sources discussed above, which emits one photon either to the left, or
to the right. Using subscripts L and R for these directions, the state produced by
this machine is
1
|ψi = √ (|0iL |1iR + |1iL |0iR ) .
2
(1.7)
Note that the states |0i and |1i now refer to the number of photons, rather than the
photon polarization as before. This state is also a maximally entangled state. Like
(1.6) it is also a Bell state. It expresses the correlation that a photon in one mode
always signifies the absence of a photon in the other mode. And it expresses the
indeterminacy, built into our device, that there is no way of knowing whether the
emitted photon will be found in the left or the right mode. How does entanglement
swapping work on such a state? A slightly different set-up is used (see Figure 1.4).
The action of the measurement is particularly clear in this example: the beamsplitter
(BS) mixes the two modes (2) and (3), so that a detection at D1 or D2 tells us only
that one of those modes contained a photon, but not which one. A bit of epistemic
book-keeping reveals the entanglement swap: if (2) contained a photon but (3) did
not, that means (1) had no photon while (4) carries a photon. Similarly if (2) was
empty but (3) contained a photon, (1) must carry a photon while (4) does not.
There is no way to distinguish these possibilities, and therefore a single detection
behind the BS puts the modes (1) and (4) into the entangled state (1.7).
For single-rail encoding, the most damaging effect of propagation is the pos-
1.6 Quantum Repeaters
26
|
D1
D2
BS
1
2
S1
3
4
S2
Figure 1.4 Single-rail entanglement swapping. Two independent
sources, S1 and S2, emit single photons into modes (1)-(2) and (3)-(4)
in the state (1.7). Modes (2) and (3) are mixed on a beam splitter
(BS). When one of the detectors D1 or D2 (but not both) fires behind
the BS, modes (1) and (4) become entangled.
sibility of photon loss, through absorption or scattering. This has the effect of
introducing a third term into the state (1.7) of the form |0iL |0iR , corresponding to
no photons in either mode (they’ve all been lost!). With the addition of this term,
the quality of the entanglement is reduced. But this ‘vacuum’ component can never
cause any detection events. Therefore if either of the detectors D1 or D2 fire, signaling a successful entanglement swap, the vacuum component is removed, since the
detection of a photon renders it counterfactual. For this type of state, entanglement
swapping also accomplishes a degree of entanglement purification.
The ability to perform both swapping and purification using such a simple measurement makes this type of encoding attractive as a means of distributing entanglement over large distances for quantum communication. However, it is not
immediately obvious how to generalize the Ekert QKD protocol to states encoded in
this way. For example, to perform a measurement in the analogue of the 45-degree
basis, one would require detectors that are sensitive to superpositions of photon
1.6 Quantum Repeaters
27
number states. These issues are avoided by combining two states of the form (1.7);
the measurement scheme is shown below in Figure 1.5. Two entangled states of the
&
1
2
D2
D1
BS
BS
D3
D4
4
3
&
Figure 1.5 QKD with single-rail entanglement. Two entangled
states are distributed between Alice — detectors D1 and D3 on the
left — and Bob — detectors D2 and D4 on the right. The photons
on each side are mixed on a beamsplitter (BS). A ‘polarization measurement’ is made when a single detector fires on each side. Both D1
and D2 firing is a ‘0’. D3 and D4 firing is a ‘1’. Adjusting the phases
φA,B allows to select the measurement bases.
form (1.7), are distributed between Alice and Bob. Alice, on the left hand side,
has two detectors, D1 and D3. Bob, on the right, has detectors D2 and D4. If
Alice only records measurements when one of her detectors fires, she knows that one
photon came to her, while the other went to Bob. Similarly a single detection at
Bob’s side tells him that Alice received the other photon. The phases φA and φB
are independently chosen by Alice and Bob from the set {0, π/2}. These phases,
in combination with the beamsplitters, allow Alice and Bob to control the basis in
which their detectors measure, in direct analogy with the rectilinear and 45-degree
bases of the Ekert protocol. When they choose the same phases, their measurements
1.6 Quantum Repeaters
28
should be correlated, with D3 and D4 firing together for a ‘1’ result, and D1 and D2
firing together for a ‘0’ result. Their is no correlation if they choose different phases.
Alice and Bob publicly announce their basis choices, and then compare some of their
results to check for the presence of Eve.
The above discussion shows how number state entanglement can be used for
quantum communication over long distances. A specific proposal for implementing this protocol using atomic ensembles to generate the entangled states was first
made in 2001 by Lu-Ming Duan, Michael Lukin, Ignacio Cirac and Peter Zoller [37] .
The DLCZ protocol is exciting because it is firmly grounded in feasible technology. Several improvements have since been suggested [38–42] , which make the scheme
more robust to phase instability, photon loss, detector noise and inefficiency. Below
we briefly introduce the principle behind the original protocol, since this will serve
as a useful introduction to the uses of atomic ensembles in quantum information
processing technologies.
We consider two clouds of atoms L and R, each with internal energy levels arranged as depicted in Figure 1.6. This type of Λ-structure is ubiquitous in quantum
information protocols. We will encounter it many times in our survey of quantum
memory protocols. It also arises in all but the simplest quantum systems: not just
atoms, but crystals, quantum dots and molecules, as we will see. The atomic ensembles L and R will together play the role of the entangled photon source introduced
earlier. Here’s how it works. Both ensembles are pumped simultaneously by a laser
pulse. The pump pulse is tuned out of resonance with the atomic state |2i, so that
1.6 Quantum Repeaters
29
Readout
Generation
(a)
(b)
Figure 1.6 Λ-level structure of atoms used in the DLCZ protocol.
(a): A pump pulse excites an atom out of the ground state |1i, which
decays down to a long-lived metastable state |3i, emitting a ‘Stokes
photon’. (b): A readout pulse brings the excited atom back to the
ground state, which emits an ‘anti-Stokes photon’ in the process.
most of the time nothing happens. But provided that the number of atoms in each
cloud is large enough, there is a small probability that the pump pulse will cause a
two-photon Raman transition in one of the atoms (see §2.3.2 in Chapter 2), exciting
it to a long-lived metastable state |3i, as shown in Figure 1.6 (a). Such an excitation is always accompanied by the emission of a Stokes photon. The probability of
two atoms being excited this way is negligibly small. Therefore at most one Stokes
photon is emitted from the ensembles. Due to the extended, pencil-like geometry
of the ensembles, any Stokes photons tend to be emitted in the forward direction,
so that they can be captured and directed as required [43] . Detectors D1 and D2
are placed behind a beamsplitter in front of the ensembles, as shown in Figure 1.7.
The beamsplitter mixes the optical modes from the two ensembles, so that if one
of the detectors fires, we know that a Stokes photon was emitted from one of the
ensembles, but we don’t know which. After a detection then, the state of the atomic
1.6 Quantum Repeaters
30
ensembles is of precisely the form (1.7), where |1i and |0i now refers to the presence
or absence of an excited atom in an ensemble.
We have now generated the number state entanglement required for the repeater
protocol. A difference with the preceding discussion is that the entanglement is
between the atomic ensembles, rather than optical modes. But this is an advantage, because the atomic excitations are stationary, and can last for a long time,
while photons are generally short lived, as mentioned earlier. Since all the steps of
the protocol involve waiting for particular detectors to fire, it is essential that the
entanglement can be preserved until all the required steps are completed. When
the entanglement is ‘needed’, for instance to perform an entanglement swapping
operation, the atomic excitations can be converted back into photons. This is done
by applying a ‘readout’ pulse to the ensembles, which returns any excited atoms to
their ground states via an anti-Stokes transition, as shown in Figure 1.6 (b). The
anti-Stokes modes inherit the entanglement from the ensembles. If the anti-Stokes
mode from ensemble L is sent off in one direction (left), and the anti-Stokes mode
from the R ensemble is sent in the other direction (right), we now have a device
that can emit entangled states on-demand. The sources S1 and S2 appearing in
Figure 1.4 should now be thought of as each comprised of a pair of entangled atomic
ensembles, waiting to be ‘read-out’ when desired.
So far we have seen how atomic ensembles show promise for quantum communication protocols, but we have not encountered a specific need for quantum optical
memories. To conclude this section, we describe how the DLCZ protocol can be
1.7 Modified DLCZ with Quantum Memories
31
D1
L
BS
R
D2
Figure 1.7 Generation of number state entanglement using atomic
ensembles in the DLCZ quantum repeater protocol. A click in one
of the detectors D1 or D2 tells us that one atom, in either of the
ensembles L or R, is excited, but because of the beamsplitter (BS),
we don’t know to which ensemble this atom belongs. Therefore the
ensembles are left in the entangled state (1.7).
improved by using memories in combination with single photon sources [41] .
1.7
Modified DLCZ with Quantum Memories
Recall that we claimed the probability for two atoms in the ensembles L and R to be
excited by the pump pulses was so small as to be insignificant. If two atoms could
be excited, it would mean that sometimes one Stokes photon from L and one from
R is emitted simultaneously. Suppose that one of these photons is lost somehow.
Perhaps it is absorbed in an optical fibre. Then only one of the detectors D1 or
D2 behind the beamsplitter will fire, but both ensembles are in fact excited. To
account for this possibility, we should add a term to the state (1.7) of the form
|1iL |1iR — that is, the ensembles are no longer in a maximally entangled state. Of
course the probability of a double excitation can always be made arbitrarily small
by making the pump pulses weaker. But this also reduces the probability of even
1.7 Modified DLCZ with Quantum Memories
32
a single excitation, so that it is necessary to wait a long time before one of the
detectors fires. If the waiting time becomes too long, the entanglement stored in the
ensembles starts to degrade as the atoms drift and collide, so this limits the distance
over which entanglement can be distributed.
This issue can be resolved if the entanglement is generated using a single-photon
source and a beamsplitter. The modified set up is shown in Figure 1.8. The atomic
ensembles L and R are now used as quantum memories, which we label QML and
QMR. We will explore the details of atomic ensemble quantum memories in Chapters
4–8. For now we only need to know that a photon entering a quantum memory is
trapped until it is needed. We have assumed that we have access to single photon
sources (SL and SR) that each emit one and only one photon, on-demand. Sources
like this are actually rather difficult to make, but research into this technology has
advanced greatly over the last few years [44–47] , and including such sources into the
design is no less realistic than the inclusion of quantum memories. To generate
number-state entanglement, we start by triggering both sources SL and SR. Each
emits a photon, which then encounters a beamsplitter (BS). At the beamsplitters
the photons can either be reflected into a quantum memory, or transmitted, in which
case they are brought together on a final BS placed in front of detectors D1 and
D2. In the case that just one of the detectors fires, we know that only one photon
was transmitted at the first BS; the other photon must have been reflected, in which
case it is now stored in a quantum memory. But the final BS prevents us knowing
which quantum memory contains the photon. We therefore have a superposition
1.7 Modified DLCZ with Quantum Memories
33
state, with either QML excited and QMR empty, or vice versa. This is the desired
entangled state (1.7). What if both photons are transmitted, but one of them is
D1
BS
BS
QML
SL
D2
BS
QMR
SR
Figure 1.8 Modification to the DLCZ protocol using single photon
sources (SL and SR) and quantum memories (QML and QMR). The
number state entanglement is generated by beamsplitters.
somehow lost? In that case only one detector fires, but neither quantum memory
is excited. We therefore have to add a vacuum component of the form |0iL |0iR
to the state (1.7). This certainly degrades the entanglement in the same way as
the |1iL |1iR component did for the corresponding error in the DLCZ protocol. But
when we perform entanglement swapping to extend the entanglement, we always
wait until the appropriate detector fires (see Section 1.6.2), and this purifies the
entanglement, removing the vacuum term (see Section 1.6.3). Therefore vacuum
errors do not damage the protocol, and we are able to run it faster and further than
the original DLCZ scheme permits.
Hopefully the possible uses for quantum optical memories is now clear. Presum-
1.7 Modified DLCZ with Quantum Memories
34
ably future proposals will develop further applications for them. In the next chapter,
we will review the various techniques used for quantum storage.
Chapter 2
Quantum Memory: Approaches
The aim of a quantum optical memory is to convert a flying qubit — an incident
photon — into a stationary qubit — an atomic excitation. This conversion should
be reversible, so that the photon may be re-emitted some time later, at the behest of
the user. In this chapter we review a number of approaches for achieving this kind
of quantum storage. More detailed calculations are deferred until the next chapter.
The simplest system one could imagine for storing a photon would consist of
a single atom, coupled to a single optical mode via an electric dipole transition
(see Figure 2.1). An incident photon, resonant with the transition, is absorbed,
promoting the atom from its ground state |1i to its excited state |2i. The photon
is now ‘trapped’ as a stationary excitation of the atom, and the quantum storage
is complete. However, quantum mechanics is always invariant under time-reversal.
That is to say, the electric dipole interaction (see §C.4 in Appendix C), which allows
the photon to be absorbed by the atom, also causes the atom to re-emit the photon,
36
(b)
(c)
1
Excitation probability
(a)
0.5
0
0
time/
Figure 2.1 The simplest quantum memory. (a): A single atom
is coupled to a single optical mode. (b): An incident photon is absorbed, exciting the atom. (c): Unfortunately, time reversal symmetry requires that the photon is immediately re-emitted.
almost immediately afterwards. In fact, these two processes compete continuously,
so that over time the population of the excited state oscillates back and forth as
the atom absorbs and re-emits the photon; see Figure 2.1 (c). This behaviour is
known as Rabi flopping, after Isidor Isaac Rabi, who first used the phenomenon
in the context of nuclear magnetic resonance [48] . The Rabi frequency Ω of these
oscillations is proportional to the strength of the coupling between the atom and
the electromagnetic field.
The above scheme needs some modification if it is to store a photon in a controlled
way. One solution is to introduce a third state |3i — a dark state — that is not
coupled to the photon mode we want to store. We should have some control field
we can apply that transfers the atomic state from |2i to |3i once the storage is
complete. In this way the Rabi oscillations are ‘frozen’, and the photon remains
trapped as an excitation of the dark state for as long as is desired. Provided the
atom is well isolated, the dark state will persist for as long as is needed. To retrieve
37
the photon, the control field is applied again, transferring the atomic state from
|3i back to |2i. The Rabi flopping of Figure 2.1 (c) continues and the photon is
re-emitted. The scheme is shown in Figure 2.2. This typifies the approach taken in
(a)
(b)
storage
(c)
retrieval
input
control
Figure 2.2 Adding a dark state. (a): We address the atom with
an auxiliary control. (b): After the input photon is absorbed, the
control transfers the atom to |3i. (c): When the photon is needed,
the control is re-applied.
many quantum optical memories, although the details of the protocols differ widely.
While single atoms have been used, ensembles of many atoms, each with the Λ-type
structure of Figure 2.2, are also commonly employed. The atom(s) can be enclosed
in an optical cavity, trapped at the centre of a confocal microscope, or addressed
by collimated beams. The optical fields may be resonant, or off-resonant with the
atomic transitions; the shape and timing of the control fields can vary, and indeed
non-optical controls, such as magnetic or electric fields, may be used. We will briefly
review this menagerie of memory protocols in the following sections.
2.1 Cavity QED
2.1
38
Cavity QED
In the forerunning discussion we presumed that a single atom could be coupled to a
single optical mode. This is a highly unnatural state of affairs, since the electromagnetic field pervades all of space, so that atoms are generally surrounded by a bath of
electromagnetic field modes. The most dramatic effect of this ‘reservoir’ of modes
is that any atomic excitations coupled to the field tend to leak away rather quickly:
an emitted photon is very unlikely to couple back to the atom, so the atomic population does not exhibit Rabi flopping; rather it decays exponentially. This is known
as spontaneous emission — the stronger the coupling between an atomic state and
the field, the shorter the lifetime of that state. An associated consequence is that
an incident photon is very unlikely to couple to the atom. Therefore a single atom
in free space cannot be used for quantum storage.
In order to recover strong coupling between an atom and an optical mode, it
is necessary to suppress the interaction with all unwanted field modes. This can
be done interferometrically, but introducing a highly reflective cavity around the
atom (see Figure 2.3). Any fields inside the cavity are reflected back on themselves
by the cavity mirrors. Only those fields with a wavelength equal to a half integer
multiple of the cavity length add constructively when folded back on themselves; all
other wavelengths interfere destructively. The optical modes supported by the cavity
are therefore spaced regularly in frequency. If the resonant frequency of the atomic
transition |1i ↔ |2i is close to one of these supported modes, the atomic coupling will
be confined to this single mode. All other cavity modes being too far from resonance
2.2 Free space coupling
39
to contribute significantly. If the volume of the cavity is sufficiently small, and the
Figure 2.3 Cavity QED. An atom is confined in a high-finesse optical cavity, which supports only a discrete set of optical frequencies.
Only one optical mode, resonant with the atomic transition, couples
to the atom.
cavity mirrors sufficiently reflective, it is possible to bring an atom into the so-called
strong-coupling regime [49–53] , where the light matter interaction behaves broadly as
described in Figures 2.1 and 2.2. This so-called Cavity QED approach to light-matter
interfaces has been widely applied. Both trapped and moving beams of atoms are
used [16,45,54,55] , as well as quantum dots [56–59] and molecules [60] . However, cavities
are rather difficult to fabricate; approaches which do away with this requirement
would be easier to scale up.
2.2
Free space coupling
Another possibility is to focus an incident photon onto an atom in a way that
maximizes the probability of it being absorbed. The cavity is replaced with a pair of
microscope objectives, as shown in Figure 2.4. The rationale behind this approach
is that good coupling should be possible if the spatial shape of an incoming photon
matches the spatial pattern of an outgoing spontaneously emitted photon. This
time-reversal argument suggests that an incident photon should be converted into a
2.3 Ensembles
40
collapsing dipole pattern. The dipole pattern of an atom is very far from a narrow
collimated beam — it is nearly isotropic, covering almost 4π steradians. Therefore
the lenses should be as wide as possible: this in itself represents a technical challenge.
Only preliminary experiments have been done [61] , but theoretical work [62,63] suggests
that efficient coupling can be achieved if the numerical aperture of the objective
lenses is made large enough.
Figure 2.4 Confocal coupling in free space. An atom is trapped
at the focus of a pair of microscope objectives. If the solid angle
subtended by an incident photon is sufficiently large, the coupling
efficiency is predicted to approach unity.
2.3
Ensembles
In addition to the practical difficulties of building a high-finesse cavity, or indeed
a high numerical aperture confocal microscope, there are technological hurdles associated with trapping and placing single atoms. Another possibility for quantum
storage is to use ensembles of atoms, as shown in Figure 2.5. An incident photon may have a small probability with interacting with any given atom, but as the
number of atoms in the ensemble increases, the probability that the photon fails
to interact with all the atoms decreases exponentially. Therefore it is possible to
asymptotically approach unit interaction probability simply by adding more atoms.
2.3 Ensembles
41
This approach is the main focus of this thesis. As we will see, introducing many
atoms makes it possible to store more than one photon, or more than one optical
mode. Complicated optical arrangements and cavities are not required, and a wide
range of possible storage media — from atomic vapours to Bose-Einstein condensates, quantum dots, crystals and fibres — are well-suited for protocols of this kind.
We are primarily concerned with schemes based on the absorption, and subsequent
Figure 2.5 Atomic ensemble memory. If an incident photon encounters enough atoms, it is almost certain to be absorbed.
re-emission, of a freely propagating photon by an atomic ensemble. These schemes
divide broadly into four protocols, all closely related, which we will now introduce.
2.3.1
EIT
EIT stands for Electromagnetically Induced Transparency. It was first observed
by Boller et al. in 1991 [64] . In EIT, a strong laser — the control — is shone
into an atomic ensemble with a Λ-structure, as shown below in Figure 2.6 (a).
Ordinarily, a weak probe beam would be absorbed by the atoms, but the interaction
with the control laser causes the ensemble to become transparent to the probe.
This can be understood by considering the dressed states of the atom, under the
influence of the control [65,66] . The control field couples the states |2i and |3i, so the
Hamiltonian for the electric dipole interaction of an atom with the control is of the
(a)
42
(b)
(c)
Susceptibility (arb. units)
2.3 Ensembles
0
Detuning (arb. units)
Figure 2.6 EIT. (a): a weak probe beam propagates through an
ensemble of Λ-type atoms, while a strong control field couples the
excited and metastable states. (b): the control mixes states |2i and
|3i to produce an Autler-Townes doublet. (c): The imaginary part of
the probe susceptibility (solid line) exhibits a double resonance, with
a transparency window at the atomic |1i ↔ |2i transition frequency.
The real part (dotted line) changes quickly within this window, causing marked dispersion.
form H = Ω|2ih3| + h.c., where Ω is the Rabi frequency of the control laser on the
|2i ↔ |3i transition. This can be simply re-written as H = Ω (|+ih+| − |−ih−|),
√
where the dressed states are defined by |±i = (|2i ± |3i) / 2. These dressed states
are simply rotated versions of the natural atomic states |2i and |3i. In fact, although
we have changed our notation, the relationship between these states is the same as
between the 45-degree and rectilinear bases discussed in the context of QKD in
chapter 1. The Hamiltonian is diagonal in the dressed basis, and we can read off
the energies of the dressed states as ±Ω. That is, the combination of the control
with the atom produces a system with a double-resonance — known as an AutlerTownes doublet [67] — with a splitting between the dressed states set by Ω. This is
a manifestation of the dynamic Stark effect: the control brings the energy of state
|3i up to that of |2i, and the dipole interaction induces an anti-crossing.
From the above arguments, we might already expect the probe absorption spec-
2.3 Ensembles
43
trum to divide into two peaks, with a transparency window in between. Figure 2.6
(c) shows the results of a steady-state calculation of the linear susceptibility for the
probe field as a function of its detuning from the |2i ↔ |3i resonance. The probe
absorption is proportional to the imaginary part, which shows the expected doublet
structure. But the depth of the transparency window cannot be explained simply
by superposing two identical resonances. As is clear from the plot, the absorption
actually vanishes completely in the centre of the transparency window. This total
transparency is due to quantum interference: the contributions to the susceptibility
from the two dressed states are of opposite sign, because one is shifted above, and
the other below the original resonance. There is therefore an exact cancellation at
this resonance, and the susceptibility is identically zero at this point.
In addition to propagating without absorption, the probe beam also propagates
extremely slowly. More precisely, the group velocity of the probe is reduced by
the strong dispersion within the transparency window. The refractive index n of
the atomic ensemble is given by n =
p
1 + Re(χ), where χ is the probe suscep-
tibility. Inspection of part (c) of Figure 2.6 shows that the real part of χ varies
extremely rapidly across the transparency window; therefore the refractive index is
also changing quickly in this spectral region. To see why this slows the group velocity, consider a pulsed probe. A pulse is composed of a range of frequencies, covering
a spectral bandwidth δω, all interfering constructively, so that the phase variation
over the pulse bandwidth is roughly zero. Each frequency component of the pulse
accumulates an optical phase kδz − ωδt over spatial and temporal increments δz,
2.3 Ensembles
44
δt, where k = nω/c is the wavevector of each component. The trajectory of the
pulse is the locus of those points at which all the components of the pulse remain
in phase, so that δkδz = δωδt, with δk the range of wavevectors spanned by the
pulse. The group velocity — the velocity of the pulse — is therefore defined by
vg = dz/dt = dω/dk = c/(n + ωn0 ), where n0 = dn/dω. The steep increase in Re(χ)
across the transparency window makes n0 large, and therefore vg is small. Group
velocities as low as 17 ms−1 have been demonstrated in the laboratory [68] .
The use of EIT as a method for storing light is an elegant application of these
effects. It was first described by Misha Lukin and Michael Fleischhauer in 2000 [69] ,
and has since been demonstrated experimentally many times [70–74] . The protocol
works as follows. An atomic ensemble of Λ-type atoms is illuminated by a control
field, preparing a transparency window. A probe pulse — to be stored — is now
directed into the ensemble, tuned to the centre of the transparency window. As
described above, it propagates slowly, but without loss. Even if the spatial extent of
the pulse is much longer than the ensemble initially, the sudden slowing of the pulse
as it enters the ensemble causes it to bunch up, so that it fits within the ensemble
as it propagates (see Figure 2.7). As soon as the entire pulse is inside the ensemble,
the control beam is attenuated. That is, the Rabi frequency Ω is reduced, so that
the splitting of the Autler-Townes doublet decreases. The transparency window gets
smaller, and the variation in Re(χ) becomes steeper, so the group velocity of the
probe falls. This process continues until the control is switched off entirely, at which
point the transparency window completely collapses; the dispersion diverges, and
2.3 Ensembles
45
the group velocity vanishes: the probe has been brought to a complete stop! In
fact, the quantum state of the optical field has been transferred to the atoms, and
the state can be stored for as long as the coherence of the atoms survives without
distortions. If the control field is switched back on, the probe field is re-accelerated,
and emerges — hopefully — unchanged, as if the ensemble had not been present.
Figure 2.7 Stopping light with EIT. A long probe pulse bunches
up as it enters the EIT medium, due to the slow group velocity at
the centre of the transparency window.
The arguments just given provide a useful physical picture for the mechanism
behind light-stopping by EIT. In isolation they don’t provide a satisfactory explanation for why the probe is not simply absorbed as the transparency window is
made narrower, nor for precisely how the probe light is transferred to an atomic
excitation. A more complete discussion can be found in the paper by Lukin and
Fleischhauer [69] . Nonetheless we can still draw some conclusions about the circumstances under which this procedure will work. The probe spectrum should not be
wider than the transparency window, otherwise it will be partially absorbed, so the
control must be sufficiently intense when the probe enters the ensemble. The control
cannot be turned off too quickly, since the dressed states producing the transparency
2.3 Ensembles
46
window are only a good approximation in the steady, or nearly steady state. The
control intensity must be reduced adiabatically; we’ll examine this condition more
closely in Chapter 5 (see §5.2.9). Finally, the control should be turned off before
the probe escapes from the ensemble, so the initial group velocity should be sufficiently slow — this means the ensemble density should not be too low. These
considerations tell us that the bandwidth of probe pulses that can be stored via
EIT, and the efficiency of the storage, is limited by the density of the ensemble, and
the available control intensity. We now introduce a related protocol, with the aim
of circumventing some of these limitations. As more detailed calculations show, this
attempt is only partially successful, but the flexibility of the new protocol makes it
an attractive alternative.
2.3.2
Raman
In 1928 Chandrasekhara Venkata Raman was the first to observe the weak inelastic
scattering of light from the internal excitations of vapours and liquids. No lasers
were available, and he used a focussed beam of sunlight to generate the required
intensity [75] . His work earned him the Nobel Prize in 1930; the eponymous Raman scattering is now used routinely in spectroscopy, industrial sensing and indeed
quantum optics.
Raman scattering can be understood rather simply in the context of an atomic
Λ-system. It is represented schematically in Figure 1.6 (a) from Chapter 1 — the
interaction used in the DLCZ repeater protocol is precisely Raman scattering. It
2.3 Ensembles
47
is a two-photon process — that is, second-order in the electric dipole interaction
— in which a pump photon (green arrow in the diagram) is absorbed, and at the
same time a Stokes photon (blue wavy arrow) is emitted. Energy conservation is
satisfied if the frequency difference between the pump and Stokes photons is equal
to the frequency splitting between the states |1i and |3i in the atoms. The optical
fields need not be tuned into resonance with the |1i ↔ |3i transition: the likelihood
of Raman scattering decreases as the fields are tuned further away from resonance,
but given a sufficiently intense pump field, and sufficiently many atoms, the Raman
interaction can be made rather strong (see §§10.8, 10.9 in Chapter 10).
In a Raman quantum memory, the Raman interaction is turned on its head: a
signal photon is absorbed, and a strong control field stimulates the emission of a
photon into the control beam (see Figure 2.8). The signal and control fields are
tuned into two-photon resonance, that is, the difference in their frequencies is equal
to the frequency splitting between |1i and |3i.
The process is conceptually very similar to that outlined at the beginning of
this chapter (c.f. Figure 2.2). But in a Raman memory, no atoms are ever excited
into the state |2i. This is because the fields are tuned out of resonance with this
state, with a common detuning ∆. Instead, the control field creates a virtual state
— represented by the dotted lines in Figure 2.8 — and an atom is excited into
this virtual state by the signal photon. The control field then transfers the atom
into the state |3i for storage. Of course, all possible time-orderings for this process
contribute to the interaction, but from this perspective it is clear that no storage is
2.3 Ensembles
48
Storage
(a)
Retrieval
(b)
Figure 2.8 Raman storage. Both fields are off-resonant, with a
detuning ∆ from the excited state. (a): A signal photon is absorbed,
and at the same time a photon is emitted into the control beam. This
two-photon Raman transition promotes one atom in the ensemble
from |1i to |3i. (b): The strong control is applied again, and the
inverse Raman process returns the excited atom to its ground state
|1i, re-emitting the signal photon.
possible without the presence of the control field. Hence the name.
We might expect the physics of Raman storage to relate very closely to that of
EIT storage, and indeed they are confusingly similar. To connect Raman storage
with our previous discussion of EIT, we can consider the dressed states of the atoms
in a Raman memory, under the influence of the control. In fact, we have already
done so. The virtual state into which the signal photon is absorbed is precisely the
dressed state produced from the atomic state |3i when the control is present. Just
as in the case of the |−i state of EIT, the virtual state is really a combination of
the states |2i and |3i. The difference is that it contains a very small amount of
the excited state |2i, because of the large detuning ∆. The other dressed state, the
equivalent of |+i for the Raman memory, is not shown in Figure 2.8, since it is so
close to the bare atomic state |2i as to be indiscernible. Again, the large detuning
means that this state is made up almost entirely of |2i, with almost no contribution
2.3 Ensembles
49
from |3i. We can therefore frame the difference between EIT and Raman storage in
the following way. In an EIT memory, the photon to be stored is tuned between the
dressed states; it is brought to a halt by turning the control field off. In a Raman
memory, the signal photon is tuned into resonance with one of the dressed states,
which is made up almost entirely of the storage state |3i. Once it has been absorbed,
the control field is turned off, and the state becomes ‘dark’ — decoupled from the
electromagnetic field.
One motivation for studying Raman storage is the possibility of broadband storage. That is, the storage of temporally short photons, which necessarily comprise a
large range of frequencies. EIT storage requires that the spectral bandwidth of the
input photon should fit within the transparency window. Raman storage does not
rely on a transparency effect, so this limitation does not pertain. In addition, the
spectral width of the virtual state into which the signal photon is absorbed is set by
the spectral width of the control field. We might therefore expect that broadband
photons can be stored with a broadband control. We should ensure that the excited
state |2i is never populated, since this state will eventually decay and our photon
will be lost. This means that the detuning ∆ must greatly exceed the bandwidth of
the signal photon. Nonetheless, given sufficiently many atoms, the large detuning
need not weaken the interaction, and efficient broadband storage should be possible.
In fact, as discussed in §5.2.9 of Chapter 5, the Raman memory protocol is ideally
suited to broadband storage, although its advantages over the EIT protocol in this
respect are not entirely clear-cut.
2.3 Ensembles
50
A Raman memory was first proposed by Kozhekin et al. in 2000 [76] . They did
not explicitly address the problem of storing a single photon; rather they considered
the transfer of quadrature squeezing — a phenomenon we’ll describe in the next
section on continuous variables memories — from light to atoms. The theoretical
treatment of a Raman memory for photon storage was published in 2007 [77] , and
forms part of this thesis. Raman storage is yet to be implemented experimentally,
but we are currently attempting to demonstrate the protocol in cesium vapour;
details are given in Chapter 10.
2.3.3
CRIB
CRIB stands for Controlled Reversible Inhomogeneous Broadening. In this protocol,
a spatially varying electric or magnetic field is applied to the ensemble. This shifts
the resonant frequency of the |1i ↔ |2i transition in the atoms, with a frequency
shift proportional to the strength of the field. Therefore, depending on their position, some atoms experience a large shift; others a smaller shift, or a negative shift.
The net effect is to produce an inhomogeneous broadening of the atomic resonance.
That is, each atom has a narrow resonance, but the ensemble as a whole absorbs
light over a broad range of frequencies, as determined by the applied field (see Figure
2.9). Storage is accomplished via the general procedure sketched in Figure 2.2 (b).
A signal photon is tuned into resonance with the broadened |1i ↔ |2i transition,
and is completely absorbed. An important feature of the scheme is that broadband
photons can be stored, because the ensemble resonance covers a wider spectral range
2.3 Ensembles
51
than the unbroadened transition. In particular, photons with a temporal duration
much shorter than the spontaneous emission lifetime of the state |2i can be stored.
Therefore the absorption process is finished long before the atoms have had time to
decay, and losses due to spontaneous emission are minimal. Note that so far in the
protocol we have not applied any optical control fields, so there are no dressed states
— the signal photon has simply been resonantly absorbed. When the absorption is
complete, a single atom in the ensemble has been excited into state |2i. The broadening field is now switched off, so that the atoms recover their natural resonance
frequencies. Then a control field — a laser pulse tuned to the |2i ↔ |3i transition
— transfers the excited atom to the storage state |3i.
Storage has now been completed, but the stored excitation is rather mixed up.
To understand why, it is easiest to drop our consideration of photons for a moment,
and consider the memory as it would be described classically. The atoms can be
thought of as a collection of dipoles (separated charges) that are set in motion by an
impinging field — the signal field. Due to the applied inhomogeneous broadening,
each dipole is oscillating at a slightly different frequency. Therefore, although they
all begin oscillating together, driven by the signal pulse, they soon drift out of phase
with one another. When the control pulse is applied, the dipoles are essentially
frozen, since their motion is transferred to the dark state |3i. In being stored, the
signal pulse has had its phase scrambled. That is, information about the time-ofarrival of the signal pulse has been lost. The crucial step in retrieving the signal is to
recover this information, by reversing the dephasing process. This is the vital ‘R’ in
2.3 Ensembles
52
CRIB. Because the inhomogeneous broadening is man-made, we are able to switch
it around, by changing the polarity of the external field. When the broadening
is flipped in this way, every atom that was shifted to a higher frequency is now
shifted to a correspondingly lower frequency, and vice versa. This provides us with
a way to completely reverse the dynamics of the memory. Retrieval works like this:
the control pulse is sent into the ensemble. This unfreezes the stored excitation,
transferring the atoms from |3i to |2i. The atomic dipoles are still out of phase, but
then the inhomogeneous broadening is re-applied, this time with the reverse polarity;
see Figure 2.9 (b). The atoms that were red-shifted at storage are now blue-shifted,
and those previously blue-shifted are now red-shifted. The atomic dipoles therefore
eventually re-phase. The collective oscillation of the entire ensemble, in phase, then
acts as a source for the electric field, and the signal pulse is re-emitted.
Retrieval
Storage
(a)
(b)
Figure 2.9 CRIB storage. (a): A signal photon is absorbed resonantly by the broadened ensemble, exciting an atom to state |2i.
Then a strong control pulse transfers the excited atom to the storage state. (b): To retrieve the stored photon, the control is applied
again, and the inhomogeneous broadening is switched on, this time
with reversed polarity. The optical dipoles eventually re-phase, and
the signal photon is re-emitted.
CRIB storage was first proposed by Nilsson and Kröll [78] , following analyses of
2.3 Ensembles
53
a generalized photon echo by Moiseev [79–81] ; the acronym appeared in a proposal
by Kraus et al. [82] . The protocol has been implemented experimentally by several
groups, all using rare-earth ions doped into solids [83–85] .
The reasons for choosing these materials as storage media are manifold. First,
use of atoms in the solid state eliminates limitations to the storage time arising from
atomic collisions, or from atoms drifting out of the interaction region (see §10.5 in
Chapter 10). Second, the rare-earth elements all share a rather peculiar feature in
their electronic structure, namely that the radius of the 4f shell is smaller than the
radii of the (filled) 5s and 5p shells [86] . The optically active electrons in the 4f shell
are therefore shielded by the 5s and 5p shells. These filled shells are spherically
symmetric, and can be thought of as metallic spheres that isolate the f electrons
from external fields, much as would a Faraday cage. Therefore, even when doped
into a solid, rare-earths maintain much of their electronic structure. Perturbations
due to the surrounding ‘host’ material may induce frequency shifts, but noise and
fluctuations, which might reduce the possible storage time, are effectively eliminated.
Optical transitions between different states within the f shell are generally forbidden,
since all these states have the same parity (see §4.3.1 in Chapter 4). This means that
spontaneous emission from these states is greatly suppressed, making them ideal
for use in a quantum memory. Spontaneous lifetimes of several hours have been
measured [87] , and photon storage times of up to 30 s are realistic [88] . Of course, if
no transitions are allowed at all, no incident photons can ever be absorbed. But
one effect of the host is to alter the atomic potential so that the f shell acquires an
2.3 Ensembles
54
admixture of d orbital states, making electric dipole transitions possible [89] .
Finally, the 4f 4 I15/2 ↔ 4 I13/2 transition in Erbium has a wavelength of 1.5 µm,
which matches the wavelength at which optical fibres used for telecommunications
have minimal absorption. Particular attention has therefore been paid to the possibilities for building a solid state quantum memory based on Erbium ions, since it
would integrate extremely well with existing telecoms systems.
We can distinguish two categories of CRIB, based on the direction of variation
of the external field with respect to the propagation direction of the signal (see
Figure 2.10). If these directions are perpendicular, so that the atomic frequencies
are broadened across the ensemble, we call this tCRIB (transverse CRIB). If they
are parallel, with the atoms broadened along the ensemble, we call this lCRIB
(longitudinal CRIB). The latter of these is sometimes referred to as GEM (gradient
echo memory [84,90] ). As we will see in Chapter 7, the performance of these two
schemes for photon storage is essentially the same.
(a)
(b)
Figure 2.10 tCRIB vs. lCRIB. (a): In tCRIB, an external field
broadens the ensemble resonance in a direction transverse to the
propagation direction of the light to be stored. (b) In lCRIB, the
broadening field is applied parallel to the propagation direction.
2.3 Ensembles
2.3.4
55
AFC
The atomic frequency comb memory protocol was proposed recently [91] by Afzelius
et al.. Their research group in Geneva have focussed on the implementation of CRIB,
and AFC takes a number of cues from their experience in this connection. In the
AFC protocol, we suppose that it is possible to prepare an atomic ensemble with
a large number of absorption lines equally spaced in frequency (see Figure 2.11).
This atomic frequency comb plays a similar role to the inhomogeneous broadening
in CRIB. It increases the bandwidth of the absorptive resonance, so that a temporally short signal is efficiently absorbed. As with CRIB, this means that the entire
absorption process can be completed long before the excited state |2i has any time
to decay. Again, for long-term storage, the excitation is mapped to the dark state
|3i. The ‘trick’ in the design of the AFC protocol becomes clear at retrieval. Recall
that, after transferring the stored excitation back to |2i, the atomic dipoles must be
brought back into phase with one another before the signal can be re-emitted. In
CRIB this is done by reversing the atomic detunings, and this would certainly work
for AFC. But even if the atomic resonances are not altered in any way, the AFC
memory still re-emits the signal! The reason is that the atomic frequency comb has
a discrete spectral structure. Therefore, the optical polarization undergoes periodic
revivals — re-phasings — at a rate given by the frequency of the beat note associated with the comb frequencies. If the frequency separation between adjacent comb
teeth is ∆, the time between re-phasings is roughly 1/∆.
Having introduced the principle behind AFC storage, a number of comments
2.3 Ensembles
56
(a)
Storage
(b)
Retrieval
Figure 2.11 AFC storage. (a): A broadband signal photon, with
a bandwidth covering many comb teeth, is absorbed by the atomic
frequency comb. The excitation is then transferred to state |3i by
a control pulse. (b) To retrieve the signal, the control is re-applied
to return the excitation to the frequency comb. The atomic dipoles
re-phase, due to the discrete structure of the comb, and the signal is
re-emitted.
are in order. First, it is not obvious that an ensemble with a comb-like resonance
will smoothly absorb the signal photon. One might expect that only those parts of
the signal spectrum overlapping with the comb teeth would be absorbed, with the
intervening frequencies simply transmitted and lost. In fact, the absorption between
the comb teeth never vanishes completely. Provided the ensemble contains enough
atoms, the combined absorption over the whole ensemble is enough to store all the
frequencies in the signal.
Second, it is not obvious how the control field can transfer all the atoms in the
frequency comb from state |2i to state |3i and back again. This must be done
in AFC so as to prevent the periodic revivals in polarization from re-emitting the
signal too early. A sufficiently bright and short control pulse will accomplish the
state transfer — the control pulse bandwidth should span the full spectral width of
the comb. So-called coherent control can be used to shape the control pulse so as to
2.3 Ensembles
57
maximize the transfer efficiency [92–98] .
Third, it is not obvious how to prepare an atomic frequency comb. The procedure suggested by Afzelius et al. is based around an implementation with rare-earths
doped into solids. In these materials, natural variations in the position within the
host occupied by the rare-earths causes the resonant frequencies to be shifted randomly over a broad spectral range. This natural inhomogeneous broadening is not
useful for CRIB, since it cannot be ‘reversed’. Therefore in the CRIB protocol, it is
necessary to remove all the atoms except those with the desired resonant frequency,
before applying the external field to artificially broaden the resonances of only these
atoms. To ‘remove’ undesired atoms, an optical pump is employed [99] (see §10.12 in
Chapter 10). This is a series of laser pulses with frequencies tuned so as to transfer
unwanted atoms from the ground state |1i into a new state |4i (not shown in any
diagrams so far), where they are ‘shelved’ for the duration of the memory protocol.
The shelf state can be chosen to have an extremely long lifetime (several hours, as
mentioned above in Section 2.3.3); this is why we can consider the atoms as having
been simply removed from the ensemble. To prepare the frequency comb for the
AFC protocol, a similar optical pumping procedure is used: all atoms with resonant
frequencies in between the comb teeth are shelved. A significant practical advantage
of AFC over CRIB is now clear. CRIB requires that we pump out all but a single
narrow resonance within the ensemble: we ‘throw away’ a lot of atoms. In AFC,
we pump out all but N narrow resonances, where N is the number of comb teeth,
which may be quite large. Therefore we throw away fewer atoms, and this allows for
2.4 Continuous Variables
58
a much stronger absorption — a much more efficient memory — using an ensemble
with the same doping concentration as a less efficient CRIB protocol. In Chapter
7 we will show that AFC is also well-suited to the parallel storage of multiple signal fields, making it attractive for use in quantum repeaters. AFC storage has not
yet been demonstrated in its entirety, but proof-of-principle experiments have been
performed by de Riedmatten et al. [100] .
2.4
Continuous Variables
So far, our discussion of quantum memories has focussed on the storage of single
photons. For the purposes of QKD and computation discussed in Chapter 1, the
quantum information encoded into these photons is of a discrete nature. Either the
number of photons, or their polarization, can be used to represent the two basis
states of a qubit. This is not the only paradigm for quantum information processing, however. Photons possess other degrees of freedom that are not discrete. The
position of a photon, or its momentum, for example. These quantities don’t lend
themselves to representation in terms of qubit states. But it is still possible to use
such continous variables to encode information (as is done in classical analogue computing). And the specifically quantum features of qubits that confer their increased
computational power — superposition; entanglement — all carry over to continuous
variables. Therefore quantum computing protocols [101,102] , and indeed QKD protocols [103–106] , that take advantage of the continuous degrees of freedom possessed
by photons, have all been developed. In general their relationship to the equivalent
2.4 Continuous Variables
59
qubit-based protocols is similar to that between analogue and digital classical computation. On the one hand, the continuous versions are robust to noise, in the sense
that information is not completely erased in the presence of distortions. On the
other hand, below a certain threshold, digital algorithms can be made essentially
impervious to noise, producing ‘perfect’ outputs, whereas analogue computations
are always subject to fluctuations — they never work perfectly. Anyone who has
ever witnessed the catastrophic failure of ‘crisp and clear’ digital television with a
poor signal will appreciate the fuzzy watchability of analogue television in the same
conditions.
In this section we briefly introduce the concepts of continuous variables, so as
to understand a class of memory based on continuous variables storage. The most
commonly used variables are the field quadratures. These are defined in terms of the
electric field E associated with an optical mode,
E = x cos(ωt) + p sin(ωt),
(2.1)
where ω is the optical carrier frequency, and t is the time. The coefficients x and
p for the amplitude of the field are known as the in-phase and out-of-phase field
quadratures, respectively [107] . By convention, the same symbols as would normally
be associated with position and momentum are used, and this is motivated by their
formal similarity. Like the position and momentum of a classical pendulum, the
quadratures are continuous variables that describe the oscillation of the field. Even
2.4 Continuous Variables
60
more strikingly, in quantum mechanics, x and p are complementary in the same
way as are the position and momentum of a harmonic oscillator. That is to say,
a measurement of one quadrature ‘disturbs’ the other, so that it is impossible to
precisely measure both simultaneously. In appropriately scaled units, the Heisenberg
uncertainty principle applies [103] :
∆x∆p ≥ 1,
(2.2)
where ∆x, ∆p are the precisions with which the quadratures are simultaneously
known. Just as in classical physics, the quantum state of an optical field can be
completely described in terms of x and p. Due to the uncertainty principle (2.2),
the state is not a single point in (x,p)-space, or phase space as it is more often called,
but rather a kind of ‘blob’, whose spread represents the uncertainties ∆x and ∆p.
This blob is known as the Wigner distribution [108] , and it is the representation of
choice for states parameterized by continuous variables. The most ‘normal’ state
of an optical field — the coherent state — is a monochromatic beam, like that
produced by a laser. It is completely classical, in the sense that the formalism
of quantum mechanics is not required to describe it. Classical electromagnetism
is sufficient. Its Wigner distribution is a Gaussian ‘hill’, with equal uncertainties
in x and p (see Figure 2.12 (a)). Clearly this is not the only type of distribution
compatible with 2.2. Figure 2.12 (b) shows an example of a squeezed state, with
a small uncertainty in x. The price for this increased precision in x is that the
2.4 Continuous Variables
61
uncertainty in p grows, but states like this can be extremely useful for reducing
the noise on measurements associated with just one of the quadratures. Squeezed
states of light would appear to be rather exotic, but they arise naturally in any
process that converts one frequency into another. Such non-linear processes only
require that the potential in which optically active electrons move is not exactly
quadratic1 in the electrons’ displacement. This happens to some extent in nearly all
materials, and the technology for generating efficient squeezing is now rather well
developed [109–112] .
(a)
(b)
Figure 2.12 Wigner distributions in quadrature phase space for
two different states of an optical mode. (a): A classical coherent
state, with equal uncertainties in x and p. (b) A squeezed state, with
∆x halved, but ∆p doubled. The lengths of the dotted lines give the
brightness of the states; their angles give their relative phases.
Research into squeezed light is being actively pursued as a means to improve the
sensitivity of gravitational wave detectors [113,114] , and to toughen the noise tolerance
of communication systems [115] . As might be expected from the ubiquity of uncer1
If the potential is not quadratic (i.e. anharmonic), the ‘restoring force’ on the electrons is
non-linear.
2.4 Continuous Variables
62
tainty relations like (2.2) in quantum mechanics, the phenomenon of squeezing is
not confined to light, and their are metrological benefits that accrue if atoms can be
squeezed [116] . Increased sensitivity to magnetic fields [117] , enhanced spectroscopic
precision [118] , and better atomic clocks [119–121] , as well as quantum computation and
communication, are all enabled by the ability to manipulate non-classical states of
light and matter.
One of the most successful demonstrations of quantum memory has grown out
of research in this area. The mechanism by which storage is accomplished is most
transparent when couched in the language of continuous variables. The memory
works by transferring the quadratures of a light beam into an ensemble of atoms,
where the quadratures X, P associated with the atoms are their ‘coronal’ angular
momenta Jx = X, Jy = P (see Figure 2.13). The storage interaction involves two
steps. In the first, a control and signal beam — both tuned far off-resonance —
are directed through an atomic ensemble. A strong magnetic field is applied to the
ensemble, which aligns the atomic spins, so that all the atoms are initially in state
|1i (see Figure 2.15). The control is polarized parallel to the field, so that it does not
induce a turning moment. On the other hand, the signal is polarized perpendicular to
the field. Two-photon Raman transitions, involving both the control and the signal,
can therefore change the z-component of the collective atomic angular momentum,
transferring atoms from |1i to |3i or vice versa. In terms of the atomic quadratures,
2.4 Continuous Variables
63
the passage of the optical fields through the ensemble induces the transformation
e = X + p,
X→X
(2.3)
where the tilde denotes the quadrature value at the end of the interaction. At the
same time, the x quadrature of the light is also shunted — a manifestation of the
Faraday effect,
x→x
e = x + P.
(2.4)
The other quadratures p, P of both the signal and the atoms are unchanged. For this
reason the interaction is known as a quantum non-demolition (QND) measurement,
since information about the p quadrature of the signal is transferred to the atoms,
without altering it.
In the second step of the storage protocol, the x quadrature of the signal is measured using balanced homodyne detection [107] . As shown in Figure 2.14, a polarizing
beamsplitter, aligned in the 45-degree basis, mixes the control and signal beams.
The difference in the intensities detected at the two output ports of the beamsplitter is directly proportional to x. This measurement result is then used to determine
the strength of a radio frequency pulse that applies a controlled torque to the atoms,
producing a shift
P → Pe = P − x
e = P − (x + P ) = −x.
(2.5)
The two maps (2.3) and (2.5) taken together almost constitute an ideal memory.
2.4 Continuous Variables
64
Figure 2.13 Atomic quadratures. In the presence of a strong magnetic field, the atomic spins in an ensemble align with the z-axis. The
quadratures X and P for the atoms are then given by small displacements of their angular momenta in the coronal plane, normal to the
z-axis.
Apart from an unimportant minus sign in (2.5), the only difficulty is the presence
of the initial atomic quadrature X in (2.3). The average value of X can be made to
vanish, but the finite spread of the initial atomic Wigner distribution still introduces
unwanted fluctuations. This spread can be reduced by squeezing the Wigner distribution of the atoms, so that ∆X → 0, before attempting the memory, and schemes
for doing this are in development [116] .
The above type of continuous variables memory, based on a QND interaction
followed by measurement and feedback, was first implemented by Julsgaard et al.
at the Niels Bohr intitute in Copenhagen [122] using an ensemble of cesium atoms.
The same research group, led by Eugene Polzik, has since refined and extended
the technology, and have demonstrated quantum teleportation of light onto atoms,
deterministic entanglement generation, efficient spin squeezing of atomic ensembles
2.4 Continuous Variables
65
Figure 2.14 QND memory. A magnetic field aligns the atomic
spins in an ensemble. A strong control, polarized parallel, and a
weak signal, polarized perpendicular to the spins, are sent through
the atoms. The control cannot rotate the atomic spins, but the signal
does. A homodyne measurement extracts the resulting x quadrature
of the signal, and a radio frequency pulse, generated by the coils
shown, applies a final rotation.
and many other continuous variables protocols. An excellent review of their progress
in this area can be found in the recent review article by Hammerer et al [123] .
The conceptual shift between the above description in terms of continuous variables, and the rather intuitive picture of ‘reversible absorption’ we have used for all
the other memory protocols, makes comparisons difficult. Certainly the optimization of this protocol does not fit into the general scheme we apply to the optimization
of the other memory protocols in this thesis. The scheme has enjoyed considerable
success, despite its technical complexity, probably due in large part to the experience
and prowess of those working at the Niels Bohr institute. Nonetheless, it does not
perform as well as the other protocols as a component of a DLCZ-type quantum
repeater. This is because, as described in Chapter 1, the entanglement purification
2.4 Continuous Variables
66
Figure 2.15 Level scheme for a QND memory. The atoms begin
with their spins aligned with the external magnetic field, in state
|1i. State |3i cannot be reached by interaction with the control alone
(green arrows), since it is π-polarized — parallel to the B-field —
and it cannot induce a spin flip (see Figure F.4 in Appendix F).
Raman transitions involving both the control and signal (blue wavy
arrows) can transfer atoms to state |3i. The intermediate excited
states are collectively labelled |2i. Both types of Raman interaction
are involved: Stokes scattering, as in the DLCZ protocol (see Figure
1.6 in Chapter 1), and the anti-Stokes scattering used in the Raman
memory protocol (see Figure 2.8).
in these repeaters is effective against photon loss, but not against photon addition,
and the Raman transitions shown in Figure 2.15 can produce extra photons that
contaminate the number state entanglement. Other applications for which photon
loss is particularly damaging would benefit from a continuous variables memory,
since loss can be essentially eliminated by increasing the power of the RF pulses
applied in the feedback step, at the expense of distorting the stored state. In any
case, we will not discuss continuous variables memory further.
In the next chapter we introduce the optimization scheme relevant for the absorptionbased memories we have discussed.
Chapter 3
Optimization
Here we introduce the optimization scheme that we will apply to the Raman, EIT,
CRIB and AFC ensemble memory protocols. In all these protocols, an input field,
the signal field, is transferred to a stationary excitation inside an atomic ensemble.
The aim of the optimization is to maximize the efficiency of this transfer, which
amounts to maximizing the amount of stored excitation, given a fixed input. Fortunately, a technique borrowed from the mathematical toolbox of linear algebra makes
this optimization extremely easy to perform.
Suppose that the signal field is a pulse, with a time-dependent amplitude A(τ ),
where τ is the time. The action of the memory is to absorb this input pulse, and
convert the incident energy into some kind of long-lasting excitation within the
ensemble (see Figure 3.1). We’ll denote the amplitude of this excitation by B(z),
where z is the position along the ensemble. Here the z dependence allows for the
possibility that the excitation may be distributed over the length of the ensemble
68
in a non-uniform way. This kind of collective, de-localized excitation is typically
referred to as a spin wave, since in many cases the atomic states involved are states
of different spin angular momentum (as in the QND memory discussed at the end
of Chapter 2). We’ll use this term indiscriminately, to describe any distributed
excitation relevant to quantum storage, regardless of the nature of the quantum
numbers associated with the atomic states. In the next chapter, we’ll define the spin
wave more precisely. For the purposes of optimization, it is sufficient to understand
the spin wave as the stationary counterpart of the propagating signal field. That
is, the storage process is a mapping A → B, and the retrieval process is the inverse
map B → A.
Figure 3.1 Storage map. An incident signal field A(τ ) is mapped
to a stationary spin wave B(z) within an atomic ensemble.
If the signal field initially contains NA = N photons, we would ideally like the
spin wave at the end of the storage process to contain the same number of excited
atoms NB = N , so that all the input light has been stored. In practice, some of
the input light will pass through the ensemble and be lost, and some of the excited
atoms will decay back down to their ground state, or drift out of the interaction
region. The efficiency η of the storage interaction is simply the ratio of the number
69
of excited atoms to the number of input photons, η = NB /NA , where the number of
quanta, either material of optical, is found by integrating the squared norm of the
relevant amplitude,
Z
∞
NA =
|A(τ )|2 dτ,
−∞
Z
NB =
L
|B(z)|2 dz.
(3.1)
0
Here L denotes the length of the ensemble. We note that in this type of memory
there is no process that can produce extra excitations. At least in principle, there
is no source of background noise. The failure mode of the memory is photon loss:
photons directed into the memory are not stored, and so are not recovered in the
retrieval process. This differs from the QND memory mentioned in the previous
chapter, which need not suffer from photon loss, but which may introduce noise at
retrieval. Therefore the performance of this type of memory is optimal when the
efficiency is maximized. No other figure of merit, involving the suppression of atomic
fluctuations for instance, is relevant.
A key requirement of a quantum memory is linearity. That is, suppose that we
store a signal field A = αA1 + βA2 built from two contributions A1 and A2 . This is
a superposition state, like (1.1) in Chapter 1, and to preserve its encoded quantum
information, the coefficients α, β should not be altered by the storage process. The
resulting spin wave should be of the form B = αB1 +βB2 , where B1 , B2 are the spin
waves generated by storage of A1 , A2 only. This property, required to faithfully store
superpositions, restricts the storage map A → B to be a linear map (and similarly
3.1 The Singular Value Decomposition
70
for the retrieval process). Fortunately, as we will see, all the memory protocols we
will analyze are indeed linear. In general, then, we can always write the storage map
in the following way
Z
∞
B(z) =
K(z, τ )A(τ ) dτ.
(3.2)
−∞
The integral kernel K is known as the Green’s function, or the propagator for the
storage interaction. It contains all the information about how the memory behaves.
In the next chapter we will show how to derive expressions for K in some cases.
Generally, it is possible to construct the form of K numerically, as we will show in
Chapter 5 (see §5.4). For the moment, we suppose that we are able to write the
memory interaction in the form (3.2). It is clear that the efficiency of the memory
depends on achieving a good ‘match’ between K and the shape, in time, of A. For
instance, no storage is possible if K ∼ 0 during the arrival time of the signal. K
and A should ‘overlap’. Optimizing a quantum memory involves finding the shape
of A that maximizes this overlap, so that all of A ends up in B. In the next section
we introduce the singular value decomposition, a valuable analytical tool that can
be used to find this optimal shape, and more besides.
3.1
The Singular Value Decomposition
The SVD is most commonly encountered in the context of matrices. It is one of
the most useful results in linear algebra, and it finds applications from face recognition [124] to weather prediction [125] . Under the name Schmidt decomposition it is
3.1 The Singular Value Decomposition
71
of critical importance in quantum information theory as a tool for the analysis of
bi-partite entanglement [126] . It seems to have been discovered independently several
times in the 19th century [127–129] . Erhard Schmidt applied the decomposition to
integral operators (an example of which is (3.2)) in 1907 [130] , and Bateman coined
the term ‘singular values’ in 1908 [131] . The proof of the decomposition for arbitrary
matrices was given in 1936 by Eckart and Young [132] .
At the heart of the SVD is a geometric interpretation for the action of a linear
operator. A linear operator, or linear map, takes some vector a as input, and spits
out another vector b as output. Representing the linear operator as a matrix M , we
have
b = M a.
(3.3)
Looking at Figure 3.2 (a), it is clear that such a transformation is equivalent to the
following procedure. (i) rotate a until it lies along one of the coordinate axes (ii)
re-scale the axes so that the length of the rotated version of a matches the length of
b (iii) rotate this re-scaled vector until it sits on top of b. This suggests that it should
be possible to decompose an arbitrary linear transformation by combining rotations
and coordinate re-scaling. The SVD is nothing more than this representation of a
general linear map as a rotation, a re-scaling, and a final rotation. An example is
shown in Figure 3.2 (b), where the action of M is shown on the set {a} of all vectors
with a certain length. The tips of these vectors trace out the surface of a sphere.
The effect of M is to ‘squish’ this sphere into an ellipsoid. This has to be done in
such a way that the black dot on the sphere ends up as the red dot on the ellipsoid.
3.1 The Singular Value Decomposition
(a)
72
(b)
Figure 3.2 Linear transformation. (a): a vector a is mapped onto
a vector b by M , which may be seen as implementing a rotation of
a onto the x-axis, followed by a coordinate re-scaling to increase the
length of a until it matches the length of b, followed by a rotation
onto b. (b): The action of M on the set of all initial vectors {a}
(the grey sphere) produces a set of vectors {b} (the red ellipsoid). As
an example, the red dot is the image of the black dot under under
M . The final ellipsoid can be generated from the initial sphere by:
rotating the sphere, re-scaling the axes and then rotating again.
To produce the ellipsoid from the sphere, it is necessary to first rotate the sphere
until the black dot is placed appropriately. Then the x, y and z axes are re-scaled
to deform the rotated sphere into the ellipsoid ‘shape’. Finally, a second rotation
puts the ellipsoid in the correct orientation, with the black dot sitting on top of the
red dot.
The above discussion is limited to the intuitive case of a real vector in three
dimensions being mapped to another real vector in three dimensions. In fact, the
SVD exists for all matrices M — all linear transformations. That is, M could be a
complex rectangular matrix of any size, that maps a complex n-dimensional vector
to a complex m-dimensional vector. In general, M can always be written in the
form
M = U DV † .
(3.4)
3.1 The Singular Value Decomposition
73
Here U and V are both unitary matrices, and D is a real, positive, diagonal matrix.
The properties of these types of matrices are reviewed in Appendix A. In terms
of the discussion above, V † represents a rotation into a new coordinate system,
D represents a re-scaling of this new coordinate system, and U represents a final
coordinate rotation. The elements of D, lying along its diagonal, are the factors
by which the coordinates are re-scaled in the second step. They are known as the
singular values of M , and they contain a great deal of useful information about the
transformation represented by M . Geometrically, the singular values correspond
to the lengths of the semi-axes of the ellipsoid in Figure 3.2 (b). In this thesis,
singular values are generally denoted by the symbol λ, the same as eigenvalues. It
should always be clear from the context what is meant. By convention, the singular
values, which are all positive, real numbers, are ordered in descending magnitude,
so that D11 = λ1 is the largest singular value, D22 = λ2 is smaller, and so on. It is
sometimes useful to visualize the structure of the matrices in the SVD, and so for
reference we provide the following tableau,



 


   λ1

 



 








M =   |u1 i  . . .  |um i 
 
 




 


|
{z
U
}|

..
.
λm
{z
D










}|

hv1 |
hvm |




.




{z
}
..
.
V†
(3.5)
3.1 The Singular Value Decomposition
74
Performing the matrix multiplications explicitly, we obtain
M=
X
λ j Mj ,
(3.6)
j
where each of the matrices Mj = |uj ihvj | is an outer product of the j th columns
of U and V (see Section A.2.1 in Appendix A). This last representation provides
a natural way to interpret the action of M , as a sum of independent mappings
from the column-space of V to the column-space of U . To understand why this is
useful, recall that the sets of column vectors {|vj i} and {|uj i} of U and V are both
orthonormal bases, because U and V are unitary (see Section A.4.5 in Appendix
A). Therefore any vector |ai to which M is applied can be written in the coordinate
system defined by the {|vj i}, and the result |bi can always be written in terms of
the {|uj i},
a = |ai = a1 |v1 i+a2 |v2 i+. . .+an |vn i,
and
b = |bi = b1 |u1 i+b2 |u2 i+. . .+bm |um i.
(3.7)
Applying (3.6) to |ai, and making use of the orthonormality of the u’s and v’s,
hui |uj i = hvi |vj i = δij , we find that
b1 = λ1 a1 ,
b2 = λ2 a2 ,
etc . . . .
(3.8)
So M can be viewed as a set of mappings between two special coordinate systems,
each with a different ‘fidelity’, given by the singular values. M may seem like
3.1 The Singular Value Decomposition
75
a complicated transformation, but as long as we use the {|vj i} to define our input
coordinate system, and the {|uj i} to define our output coordinate system, the action
of M is always extremely simple. It just maps each coordinate from the input onto
the corresponding output coordinate, re-scaled by the corresponding singular value.
3.1.1
Unitary invariance
If the input coordinate system is rotated, before performing the SVD, this cannot
change the singular values, but only the input basis that we should use. The same
is true if the output coordinate system is rotated. More generally, consider applying
some unitary transformation W to M . The SVD of this compound operator is then
e DV † ,
W M = W U DV † = U
(3.9)
e = W U is a unitary matrix, since both W and U are unitary. The output
where U
basis vectors are modified by W , but the singular values — the elements of D —
are unchanged. By the same token, the product M W also has the same singular
values as M , but V must be replaced by W † V . Sometimes it is only possible to find
rotated versions of a matrix M , in which case this property of the SVD is useful.
3.1.2
Connection with Eigenvalues
The SVD is connected with the eigenvalues of the ‘square’ of M . More precisely, consider the normally and antinormally ordered products KN = M † M and KA = M M † .
†
†
These two products are both Hermitian, since KN
= KN and KA
= KA . Therefore,
3.1 The Singular Value Decomposition
76
they both have real eigenvalues, and their eigenvectors each form orthonormal bases
(see Section A.4.3 in Appendix A for a derivation of this fact). And inserting the
decomposition (3.4), we see that
KN = V D2 V † ,
and
KA = U D2 U † ,
(3.10)
where we have used the relations U † U = V † V = I, which must hold for unitary
matrices. Therefore, the eigenvalues of KN and KA are both given by the squares
of the singular values.
3.1.3
Hermitian SVD
Note also that if M is a Hermitian matrix, with M = M † , we must have that U = V .
That is,
M = U DU † ,
(3.11)
which is precisely the spectral decomposition of M , with the eigenvalues of M lying
along the diagonal of D, and the eigenvectors of M making up the columns of U . So
the SVD and the spectral decomposition of M are identical for Hermitian matrices.
3.1.4
Persymmetry
Another case, which we will encounter in our treatment of Raman storage, is that
of persymmetry. Suppose that M is a real, square matrix, with n = m. Then M is
persymmetric if it is symmetric under reflection in its anti-diagonal — the diagonal
3.1 The Singular Value Decomposition
77
running from bottom left to top right, as shown in Figure 3.3. M = M P , where
(M P )ij = Mm−j+1 m−i+1 . This is rather an unusual type of symmetry, and it is
Figure 3.3 Persymmetry. The upper left and lower right portions
of a real persymmetric matrix are mirror images of eachother under
reflection in the anti-diagonal, represented as a grey stripe.
not often discussed in textbooks. But the action of a persymmetric matrix can be
viewed as very similar to that of a real Hermitian matrix, the only difference being
that the result is ‘flipped around’. To see the implications of persymmetry for the
SVD, let us write M = XH, where H is a real Hermitian matrix (i.e. symmetric
under reflection in its main diagonal), and where X is a ‘flip’ matrix, with ones
along its anti-diagonal, and zeros everywhere else (left blank for clarity below),






X=






1
.
..
1
1 




.





(3.12)
The action of X when multiplying a matrix is to flip it around a horizontal axis, so
that its last row becomes its first, and vice versa. For every persymmetric matrix M ,
there must always be some Hermitian matrix H such that M = XH. The property
3.2 Norm maximization
78
of persymmetry can also be easily written in terms of X. A little mental acrobatics
will verify that M P = XM † X, so that persymmetry requires M = XM † X. Inserting
the spectral decomposition of H, we obtain
e DU † ,
M = XM † X = X(XH)† X = XH † X † X = XHX 2 = XU DU † = U
(3.13)
where we used the Hermiticity of H and X, along with the fact that X 2 = I (two
e = XU is a unitary matrix whose
horizontal flips cancel each other out), and where U
columns have been flipped round. Therefore the SVD of a persymmetric matrix is
such that the columns of U , the basis for the output coordinate system, are flipped
versions of the columns of V , the input basis. This result will be of use to us in
Chapter 5.
3.2
Norm maximization
Suppose that we would like to know how to choose a in order to maximize the norm
(the length) of b. Since M is a linear transformation we can always increase the
norm of b just by increasing the norm of a: if we double the length of a, the length
of b also doubles. But this is not interesting. Clearly the direction of a matters;
some directions will result in a larger norm for b. If the norm of a is fixed, how
should we choose its direction? This question is easily answered if we are able to
compute the SVD of M . Without losing generality, let ||a|| = a = 1. That is,
3.3 Continuous maps
79
|a1 |2 + |a2 |2 + . . . + |an |2 =1. Using (3.8), the norm of b is then given by
b2 = |b1 |2 + |b2 |2 + . . . + |bm |2
= λ21 |a1 |2 + λ22 |a2 |2 + . . . λ2n |an |2 .
(3.14)
But by definition, λ1 is the largest of the singular values, so that
b2 < λ21 (|a1 |2 + |a2 |2 + . . . + |an |2 )
= λ21 ,
(3.15)
That is, the largest possible norm of b is b = λ1 . This maximum norm is obtained
by choosing a1 = 1, with all other components vanishing. So the ‘optimal’ vector
a, with regard to maximizing b, is a = v1 (or |ai = |v1 i). From the perspective
of Figure 3.2 (b), this amounts to choosing a so that the resulting b lies along the
largest semi-axis of the ellipsoid generated by M .
3.3
Continuous maps
The preceding discussion of matrices can be extended, without essential modification, to the continuous map (3.2) describing the storage interaction in a quantum
memory. The Green’s function K(z, τ ) has the same basic structure as a matrix,
except that it has two continuous arguments, instead of two discrete indices. And
it is also amenable to the SVD. The expression (3.6) for a matrix becomes, in the
3.3 Continuous maps
80
continuous case
K(z, τ ) =
X
λj ψj (z)φ∗j (τ ).
(3.16)
j
As before, the λ’s are the singular values. The functions ψj and φj are the continuous
analogues of the input and output basis vectors |uj i and |vj i. We will refer to them
as modes; φ1 is the first input mode, ψ2 is the second output mode, and so on.
Alternatively, since the signal field and spin wave both appear in (3.2), it may be
physically more meaningful to talk of the φ’s as the signal modes, and the ψ’s as
spin wave modes. The inner product between two vectors a, b takes the form of an
overlap integral, for continuous functions a(x), b(x),
†
a b = ha|bi =
X
a∗i bi
Z
−→
a∗ (x)b(x) dx.
(3.17)
i
The orthonormality conditions hui |uj i = hvi |vj i = δij for basis vectors are therefore
replaced by the relations
Z
0
L
ψi∗ (z)ψj (z) dz =
Z
∞
−∞
φ∗i (τ )φj (τ ) dτ = δij .
(3.18)
These conditions tend to produce sets of ‘wiggly’ functions, so that the product of
two different modes alternates between positive and negative values, and integrates
to zero. Sets of oscillating modes, satisfying orthonormality conditions like (3.18),
are common in harmonic analysis and acoustics; they are also the bread and butter
of much of quantum physics — quantum memories are no exception. In general, the
3.3 Continuous maps
81
first modes, associated with the largest singular value λ1 , are slowly varying. They
represent the basic ‘shape’ of the Green’s function, with broad brushstrokes. Higher
modes tend to oscillate faster, and they represent smaller corrections, at increasingly
fine levels of detail.
3.3.1
Normally and Anti-normally ordered kernels.
The continuous analogues of the matrices KN and KA in (3.10) are found by integrating the product of two K kernels over one of their arguments,
0
KN (τ, τ ) =
KA (z, z 0 ) =
Z
L
K ∗ (z, τ )K(z, τ 0 ) dz,
Z0 ∞
K(z, τ )K ∗ (z 0 , τ ) dτ.
(3.19)
−∞
These two kernels share the same eigenvalues; they satisfy the eigenvalue equations
Z
∞
−∞
Z L
KN (τ, τ 0 )φj (τ 0 ) dτ 0 = λj φj (τ 0 ),
KA (z, z 0 )ψj (z 0 ) dz 0 = λj ψj (z 0 ).
(3.20)
0
Sometimes these kernels are more convenient to work with than the kernel K.
3.3.2
Memory Optimization.
Now it is clear how to optimize the performance of a quantum memory. We want
the largest efficiency η = NB /NA . But from the expressions in (3.1), we see that
NB is just the continuous analogue of b2 , the squared norm of the output spin wave.
3.3 Continuous maps
82
Suppose that we fix NA = 1, just as we fixed a2 = 1 in Section 3.2. How should
we choose the signal field A(τ ) to maximize η? The answer carries over directly
from our considerations of matrices. We should choose A(τ ) = φ1 (τ ). With this
choice, the resulting spin wave is B(z) = λ1 ψ1 (z), and the optimal storage efficiency
is η = λ21 .
In practice, the SVD of the Green’s function is almost always computed numerically, and to do this, the continuous function K is discretized on a fine grid: it is
converted into a matrix. Therefore our treatment of matrices applies directly when
calculating the singular values, and the mode functions, for a quantum memory.
Note that, since we must of course have η ≤ 1, any physical Green’s function will
have λ1 ≤ 1, and this condition is a useful check that K has been sampled with a
sufficiently fine grid.
3.3.3
Unitary invariance
When deriving a form for the Green’s function K, it will sometimes be easier to
work in terms of a transformed coordinate system. Suppose that we introduce a
new variable y = y(z), where y(z) is some monotonic single-valued (i.e. invertible)
function. The Green’s function, expressed in terms of y, has the same singular values
as in the original coordinates, provided that we make the transformation unitary by
including a Jacobian factor:
1
K(y, τ ) = K [z(y), τ ] × p
,
J(y)
(3.21)
3.3 Continuous maps
83
where z(y) is the inverse transformation, relating z to y, and where J(y) = ∂z y is
the Jacobian, relating the line elements dy and dz. Including the factor involving J
ensures that the norm of K is preserved in the new coordinate system,
Z
|K(y, τ )|2 dy =
Z
|K [z(y), τ ] |2 ×
Z
|K(z, τ )|2 dz.
=
∂z
dy
∂y
The output modes transform accordingly,
ψj [z(y)]
.
ψj (y) = p
J(y)
(3.22)
An identical procedure is used when transforming the time coordinate.
Instead of re-parameterizing the kernel, we may wish to transform to a frequency
representation, instead of a temporal one. That is, we might be able to work out an
e
expression for the Fourier transformed Green’s function K
e ω) = √1
K(z,
2π
Z
K(z, τ )eiωτ dτ.
(3.23)
Inserting the SVD expansion (3.16) into (3.23), we obtain
e ω) =
K(z,
X
λj ψj (z)φe∗j (ω),
(3.24)
j
where the modes φej are Fourier transforms of the φj . These Fourier transformed
3.3 Continuous maps
84
modes also form an orthonormal set,
Z
Z Z
Z
1
∗
−iωτ
0 iωτ 0
0
φi (τ )e
dτ φj (τ )e
dτ dω
2π
Z Z
=
δ(τ − τ 0 )φ∗i (τ )φj (τ 0 ) dτ 0 dτ
φe∗i (ω)φej (ω) dω =
= δij ,
(3.25)
where in the penultimate step we used the plane-wave expansion of the Dirac delta
function (see §D.2 in Appendix D),
1
δ(x) =
2π
Z
e±ixy dy.
(3.26)
e
The expression (3.24) is therefore precisely the SVD of the transformed kernel K.
e are the same as those of K. In fact the temporal Fourier
The singular values of K
transform we applied is just an example of a unitary transformation; unitary because,
by Parseval’s theorem, it is norm-preserving. That is to say,
Z
2
|f (τ )| dτ =
Z
|fe(ω)|2 dω,
(3.27)
for any function f and its Fourier transform fe. As described in Section 3.1.1, the
singular values of a mapping are never altered by a unitary transformation.
Other useful possibilities include Fourier transforming over the spatial variable,
3.4 Optimizing storage followed by retrieval
85
to find the Green’s function in so-called k-space,
1
e
K(k,
τ) = √
2π
Z
K(z, τ )eikz dz.
(3.28)
Again the singular values are the same as those for K, but the output modes are
k-space versions of the ψj .
Generally the integration limits in these Fourier transforms would be [−∞, ∞],
but when boundary conditions are needed, it will sometimes be useful instead to
implement a unilateral transform, where the integration runs from 0 to ∞. The
utility of these techniques will become clear when we examine the memory interaction more closely in the next chapter. A brief review of both the bilateral and
unilateral Fourier transforms can be found in Appendix D.
3.4
Optimizing storage followed by retrieval
Suppose that we are not interested in the efficiency of storage alone, but rather
the combined efficiency of storage into, followed by retrieval from the memory. In
many situations it is this combined efficiency that is most experimentally relevant.
The same techniques as outlined above are directly applicable. The entire memory
interaction can be characterized as a map between the input and the output signal
fields,
Z
∞
Aout (τ ) =
−∞
Ktotal (τ, τ 0 )Ain (τ 0 ) dτ 0 .
(3.29)
3.5 A Simple Example
86
The total efficiency of the memory is given by the ratio of the norms of Aout and
Ain . The input mode that maximizes this efficiency is therefore found from the
SVD of the Green’s function Ktotal , and the resulting optimal efficiency is given by
ηcombined = λ21 , where λ1 is the largest singular value of Ktotal .
If we neglect any decoherence of the spin wave (as we do throughout this thesis),
the kernel Ktotal can be constructed from the kernels describing the storage and
retrieval processes individually. Under certain circumstances — when the retrieval
process is precisely the time-reverse of the storage process [133] — the combined
kernel Ktotal is equal to the normally ordered kernel KN defined in (3.19), and
ηcombined = η 2 . This is the optimal situation, and in general a mismatch between the
storage and retrieval processes reduces the memory efficiency so that ηcombined < η 2 .
These issues are explored in more detail in Chapter 6.
3.5
A Simple Example
Before embarking on a detailed derivation of the equations of motion for an ensemble
quantum memory, we run through a simple example that contains many of the
features that emerge from a more rigorous analysis.
We consider a classical optical signal pulse propagating through an ensemble of
identical atoms. We use a classical Lorentz model for the atoms [134] , in which the
electric field of the light pulse ‘pulls’ on an electron in each atom, while a harmonic
restoring force ‘pulls back’, keeping the electrons bound around their equilibrium
positions. Let the average displacement at time τ , away from equilibrium, of an
3.5 A Simple Example
87
electron located at position z, be given by x(z, τ ). The average is taken over all
the atoms with the same position coordinate z, all of which behave identically. The
restoring force on each electron is given by Frestore = −kx, for some constant k.
This is a good approximation for any kind of restoring force, provided the typical
displacement x is small enough. The force on an electron due to the electric field
E of the signal pulse is Flight = eE, where e is the electronic charge. The classical
equation of motion for x is then given by Newton’s second law, Ftotal = ma, where
m is the electronic mass and a = ∂τ2 x is the acceleration of the electron. Putting
this together yields the equation
m∂τ2 + k x = eE.
(3.30)
We can write the signal field as E = Aeiωs τ , where ωs is the optical carrier frequency
of the signal, and where A is a slow modulation describing the temporal profile of
the pulse. Irradiating the atoms will produce a response with a similar temporal
structure, so we write x = Beiωs τ , where B is some slowly varying envelope. Substituting this into the equation of motion, and neglecting the term ∂τ2 B, gives the
equation
∂τ B = −iαA,
where α =
e
2mωs .
(3.31)
Here for simplicity we have assumed that the signal field frequency
is tuned perfectly into resonance with the atoms, so that ωs =
p
k/m. This equation
describes the response of the atoms to the field; to complete the picture, we would
3.5 A Simple Example
88
like to understand how the field responds to the atoms. As a first step, we apply a
(bilateral) Fourier transform from τ −→ ω, so that (3.31) becomes
e = α A,
e
B
ω
(3.32)
e
where the tildes denote the transformed variables. The electronic displacement B
acts as a source for the signal field. More precisely, each oscillating electron generates
e
an electric field E(r)
= ae/4π0 c2 r in proportion to its acceleration1 . Here 0 as
usual denotes the permittivity of free space; r is the distance from the electron.
Summing the contributions from all the atoms in a thin slice of the ensemble with
thickness δz, we find the total field generated by the dipoles is
etot = −iδz ωs en x
e,
E
4π0 c
(3.33)
where n is the atomic number density [137] . If we consider the propagation of the
signal through a thin slice of the ensemble of thickness δz, we therefore have
e + δz, ω) = A(z,
e ω) − iδz ωs en B(z,
e ω)
A(z
4π0 c
βδz e
=
1−i
A(z, ω),
ω
(3.34)
1
This can be derived rather neatly from relativistic equivalence. The electrostatic potential
energy V = e2 /4π0 r of two electrons separated by r produces a relativistic mass increase M = V /c2
of the pair of electrons. The extra weight 12 gM of one of the electrons, in a gravitational field g,
must have its origin in a vertical force eE exerted by the electric field E of the other electron. The
equivalence principle demands that we cannot tell if the field g is swapped for an acceleration a.
Putting this all together, we find the field due to an accelerating charge is E ∼ ae/4π0 c2 r [135,136] .
3.5 A Simple Example
where β =
e2 n
8π0 mc .
89
Taking the limit as δz −→ 0, we derive Beer’s law of exponential
absorption for the signal field, with absorption coefficient β,
βδz z/δz e
1−i
Ain (ω)
δz→0
ω
e ω) =
A(z,
lim
ein (ω),
= e−iβz/ω A
(3.35)
ein (ω) = A(z
e = 0, ω) is the initial spectrum of the signal field, at the start of
where A
the ensemble [138] . Substituting this into (3.32) gives an expression for the average
electronic displacement in the Fourier domain,
−iβz/ω
e
e
B(ω,
z) = α
ω
ein (ω).
×A
(3.36)
Using the result (D.40) from §D.5 in Appendix D, along with the convolution theorem (D.21), we can take the inverse Fourier transform of this to obtain the map
Z
Bout (z) = −iα
T
h p
i
J0 2 βz(T − τ ) Ain (τ ) dτ 0 ,
(3.37)
0
where Bout (z) = B(τ = T, z) is the electronic displacement at time τ = T , with
T some time chosen to define the end of the interaction, after the signal pulse has
passed through the ensemble. Here J0 is a zero’th order ordinary Bessel function
of the first kind — we will encounter this function, through similar inverse Fourier
transforms, frequently in Chapter 5. We have arranged for the notation in (3.37) to
appear suggestive of the storage map described at the start of this Chapter. If we
3.5 A Simple Example
90
identify the electronic displacement Bout as the amplitude of a spin wave excited by
the signal field, then the storage kernel K in (3.2) can be identified with the Bessel
function appearing in the integrand of (3.37). A numerical SVD of the Green’s
function
h p
i
K(z, τ ) = −iαJ0 2 βz(T − τ )
(3.38)
would reveal the temporal shape of Ain that maximizes the degree of atomic excitation. We have not taken care to normalize the signal field A, or the electronic
displacement B, and so (3.1) is not quite true, using the current definitions. But
up to some constant normalization, the optimal efficiency of a quantum memory
described by the above model would be provided by squaring the largest singular
value of (3.38).
Our purpose in the above exposition has not been to describe a real quantum
memory in any quantitative detail. But this simple example contains many of the
ingredients we will encounter in the following Chapters. We have used a one dimensional propagation model, combined with some atomic physics, to obtain two
equations of motion — the first describing the atomic response to the field, and the
second describing the influence on the field due to the induced atomic polarization.
The solution was found using a Fourier transform, and the result took the form of
the storage map (3.2), with the kernel given by a Bessel function. This story applies
equally to the fuller treatment given shortly. Finally, note that we have made no
use of the formalism of quantum mechanics so far. Of course, a correct description
of the atoms at least requires that we quantize the electronic energies. But the
3.5 A Simple Example
91
propagation is simply Maxwellian electrodynamics. It may help to keep in mind
that, although we will treat the signal field quantum mechanically for the sake of
completeness, the quantum memories we consider behave essentially classically. Or,
we might say that their efficiencies may be derived classically, since efficiencies do
not depend on correlation functions of the optical fields, and it is only in the photon
statistics revealed by these correlation functions that non-classicality is manifest.
Chapter 4
Equations of motion
In this chapter we derive the dynamical equations describing the interaction of light
with the atoms of an ensemble quantum memory. We focus on EIT and Raman storage, which can be treated together; inhomogeneously broadened memories, such as
CRIB and AFC, are covered in Chapter 7. We borrow techniques from the treatment
of Stokes scattering by Mostowski et al. [139] , and some notation from the later treatment of quantum memories by Gorshkov et al. [140] . The derivation divides broadly
into three parts. First, we write down the Hamiltonian describing the interaction
of a single atom with the signal and control fields, and we obtain Heisenberg equations for the atomic evolution. Next, we introduce Maxwell’s equations, describing
the coupling of the signal field to the macroscopic atomic polarization as it propagates through the ensemble. Finally, we add up the contributions from many atoms
to form the macroscopic variables describing the atomic polarization and the spin
wave. Having derived the dynamical equations in a convenient form, we investigate
4.1 Interaction
93
various methods of solution, in order to optimize the performance of the memory,
in the next chapter.
4.1
Interaction
We consider the propagation of a weak signal field through an ensemble of Λ-type
atoms, in the presence of a bright control field, tuned into two-photon resonance
with the signal, as depicted in Figure (4.1).
Figure 4.1 The signal (blue) and control (green) fields involved in
a Λ-type ensemble quantum memory.
As shown in Section C.4 in Appendix C, the interaction of a light beam with an
atom is well described by the electric dipole Hamiltonian HED = −E.d, where E
is the electric field associated with the light at the atomic position, and where d is
the dipole moment associated with an optically active electron in the atom.
4.2 Electric Field
4.2
94
Electric Field
The electric field is composed of two parts, a bright classical control field, and a
weak signal field that we wish to store, both propagating along the z-axis,
E = Ec + Es .
(4.1)
The control field is sufficiently intense that it is not affected by its interaction with
the atoms of the memory, so that we do not need to treat it as a dynamical variable
in the Hamiltonian. We therefore represent it as a classical field,
Ec (t, z, ρ) = vc Ec (t, z, ρ)eiωc (t−z/c) + c.c.,
(4.2)
where vc is the control polarization vector1 , ωc is the central frequency of the control
field, and Ec (t, z, ρ) is the slowly varying envelope of the control, describing the
spatio-temporal profile of the control pulse. Here ρ = xx̂ + y ŷ is a transverse
position vector, as shown in Figure 4.1.
The signal field is much weaker than the control, and in general it may be in
a non-classical state (for example, a Fock state — see Appendix C). Therefore we
treat the signal field quantum mechanically. The signal field, when well-collimated,
1
Note that if the polarization is not linear (circular, for instance), the polarization vector is
complex, satisfying v ∗ .v = v † v = 1.
4.2 Electric Field
95
can be written in the form
Z
Es (z, ρ) = ivs
g(ω)a(ω, ρ)e−iωz/c dω + h.c.,
where vs is the signal polarization vector, g(ω) =
(4.3)
p
~ω/4π0 c is the mode amplitude,
and a(ω, ρ) is an annihilation operator for a signal photon with frequency ω, and
transverse position ρ, which satisfies the equal-time commutation relation
[a(ω, ρ), a† (ω 0 , ρ0 )] = δ(ω − ω 0 )δ(ρ − ρ0 ).
(4.4)
The form of (4.3) is very similar to the expression (C.10) in Appendix C, the only
difference being the inclusion of the transverse position ρ, which allows us to treat
diffraction (this is covered in Chapter 6). In the next chapter we will drop the
transverse coordinate and work with a one dimensional propagation model. Note
that all these operators have no time-dependence in the Schrödinger picture (see
Appendix B). In the Heisenberg picture, commutation with the optical free-field
Hamiltonian (see (4.14) below) gives the annihilation operators the simple timedependence
a(ω, ρ, t) = a(ω, ρ)eiωt .
(4.5)
It will be useful to factorize the signal field into a carrier wave, and a slowly varying envelope, as we did with the control field in (4.2). To do this, we make use of
the approximation that the bandwidth of the signal field will be very small in com-
4.2 Electric Field
96
parison to its central frequency ωs . Therefore, only terms with frequencies rather
close to ωs will be important in the integral in (4.3). Since the dependence of the
mode amplitude g(ω) on frequency is quite weak (only ‘square-root’), we make the
replacement g(ω) −→ g(ωs ). We are then able to perform the frequency integral
explicitly, to obtain
Es (t, z, ρ) = ivs gs A(t, z, ρ)eiωs (t−z/c) + h.c.,
where gs =
√
(4.6)
2πg(ωs ), and where we have defined the slowly varying time-domain
annihilation operator A according to the relation
−iωs (t−z/c)
A(t, z, ρ) = e
1
×√
2π
Z
a(ω, ρ, t)e−iωz/c dω.
(4.7)
That it retains the property of a photon annihilation operator, albeit for spatiotemporal, rather than spectral modes, can be seen from its commutator with its
Hermitian adjoint. Inserting (4.5) into (4.7), we find
[A(t, z, ρ), A† (t0 , z 0 , ρ0 )] = δ(t − t0 − (z − z 0 )/c)δ(ρ − ρ0 ).
(4.8)
Aside from the leading phase factor, which removes the rapid time-dependence due
to the carrier frequency ωs , the action of A(t, z, ρ) in free space can be understood
as the annihilation of a signal photon in a spatio-temporal mode centred at position
ρ and retarded time τ = t − z/c.
4.3 Dipole Operator
4.3
97
Dipole Operator
Having developed a convenient notation for the signal and control fields, we now
consider the atomic variables. The Coulomb interaction between the atomic nucleus and the electrons is of course rather complex in general, but the formalism
of quantum mechanics comes to the rescue. The energy levels labelled |1i, |2i, |3i
are eigenstates of the atomic Hamiltonian, and therefore they form an orthonormal
basis for the Hilbert space of quantum states of the atom (see Appendices B and C).
That is, any state of the atom can be described in terms of these states. And any
operator acting on the atomic states can be expressed using the coordinate system
defined by these states. In particular, the electric dipole operator d can be written
in the following way,
d=
X
djk σjk ,
(4.9)
j,k
where the coefficients djk = hj|d|ki are the matrix elements of the dipole operator, the σjk = |jihk| are ‘flip operators’ (sometimes known as transition projection
operators), and where the double summation runs over the three atomic states.
To preempt possible confusion, we should clarify that d is a three-dimensional
vector in space, whose elements are quantum mechanical operators acting on the
three-dimensional Hilbert space of the Λ-level atom. The Dirac notation (|1i, |2i,
etc...) refers to vectors and/or operators in/on this Hilbert space, while the bold
font notation (d, E, v, etc...) refers to vectors in ordinary space (whether their
elements are numbers, or operators).
4.3 Dipole Operator
4.3.1
98
Parity
So far, we have not used any properties of the dipole operator specifically — (4.9) is
an identity that holds for any atomic operator. But now we can use the parity of the
dipole operator to remove some terms from the sum in (4.9). Parity refers to the way
a quantity transforms under the operation of inversion, when all spatial coordinates
are reflected in the origin. That is, r −→ −r, where r is any position vector. As
discussed in Appendix C, the dipole operator is simply given by d = −er, where now
r is the position, with respect to the atomic centre-of-mass, of the optically active
electron. Therefore under inversion, we have d −→ −d. The atomic dipole operator
has negative parity. This means that all the diagonal dipole matrix elements must
vanish, djj = 0. To see this, we can express the matrix element djj in terms of the
wavefunction ψj of the state |ji,
djj
= hj|d|ji
Z
=
ψj∗ (r)dψj (r) d3 r.
(4.10)
The integral runs over all space, but we can divide it into a pair of integrals: the
first over half of all space, with positive coordinates r (+), and the second over the
remaining half of space, with negative coordinates −r (−). Since the dipole operator
4.3 Dipole Operator
99
changes sign under inversion, the second integral exactly cancels with the first,
Z
djj
2
3
Z
|ψj (r)| d d r +
=
Z
=
|ψj (r)|2 d d3 r
−
+
|ψj (r)|2 (d − d) d3 r
+
= 0.
(4.11)
Here we used the fact that |ψj |2 has positive parity (i.e. is unaffected by inversion).
This must be true since |ψj |2 describes the electronic charge density associated with
the state |ji, and this must be spherically symmetric: there is no interaction to
break the spherical symmetry of the bare atom.
The electric dipole interaction is therefore completely off-diagonal, meaning that
it only couples different states together, never the same state to itself. Furthermore,
we require that the state |3i is long-lived, in order that it can serve as an effective
storage state. Therefore we neglect any dipole coupling between the states |1i and
|3i, so that no direct transitions between these states are mediated by the dipole
operator: our goal, instead, is to implement an indirect transition, mediated by the
control field. In the light of these arguments, we arrive at the somewhat pared-down
expression
d = d12 σ12 + d23 σ23 + h.c..
(4.12)
4.4 Hamiltonian
4.4
100
Hamiltonian
We are now in a position to write down the Hamiltonian for the atom-light system,
H = HA + HL + HED .
(4.13)
Here HL is the free-field energy of the light field,
Z
HL =
Z
ω
a† (ω, ρ)a(ω, ρ) d2 ρ dω,
(4.14)
A
where the transverse integral runs over the transverse area A of the signal field.
Here we have neglected the zero-point energy, and also the fixed energy associated
with the control field. HA is the Hamiltonian for the bare atom. The atomic states
|1i, |2i, |3i are by definition eigenstates of HA , and therefore when written in terms
of these states, HA is purely diagonal (See Section A.4.3 in Appendix A),
HA =
X
ωj σjj ,
(4.15)
j
where ωj is the resonant frequency of state |ji.
We can now use the Heisenberg equation (see B.10 in Appendix B) to find the
time evolution of the atomic flip operators. Since these operators always commute
with the optical free-field Hamiltonian HL , we drop this from the Hamiltonian (it
4.4 Hamiltonian
101
has no effect on the atoms), and we work with the equation
∂t σjk = i[σjk , HA + HED ].
(4.16)
The flip operators satisfy the following multiplicative identity,
σij σkl = σil δjk ,
(4.17)
†
and under Hermitian conjugation we have σjk
= σkj . Using these relations we obtain
five independent atomic equations; two for the atomic populations,
∂t σ11 = −iE. (d12 σ12 − h.c.) ,
∂t σ33 = iE. (d23 σ23 − h.c.) ,
(4.18)
and three for the atomic coherences,
∂t σ12 = iω21 σ12 − iE. [d∗12 (σ11 − σ22 ) + d23 σ13 ] ,
∂t σ13 = iω31 σ13 − iE. [d∗23 σ12 − d∗12 σ23 ] ,
∂t σ23 = iω32 σ23 − iE. [d∗23 (σ22 − σ33 ) − d12 σ13 ] ,
(4.19)
where we have defined ωjk = ωj − ωk as the frequency difference between the states
|ji and |ki. Note that
P
j
σjj = I, the identity, so that the sum of the populations
commutes with the Hamiltonian, and therefore has no time-dependence. This simply
4.5 Linear approximation (1)
102
expresses the fact that the atom remains in one of the states |1i, |2i, |3i, at all times.
Therefore ∂t σ22 = −∂t (σ11 + σ33 ).
The coupled equations (4.18) and (4.19) constitute a rather complex system,
and it is not possible to extract an analytic solution, in general. Fortunately, the
equations simplify considerably in the linear regime.
4.5
Linear approximation (1)
Provided that we store a small number of photons in the quantum memory, such
that most of the atoms remain in their ground states, with only very few atoms
excited, we can ignore the dynamics of the atomic populations. We replace the
operators σjj by their expectation values on the atomic ground state,
σ11 −→ 1,
σ22 −→ 0,
σ33 −→ 0.
(4.20)
This leaves us with just the three equations for the coherences,
∂t σ12 = iω21 σ12 − iE. [d∗12 + d23 σ13 ] ,
∂t σ13 = iω31 σ13 − iE. [d∗23 σ12 − d∗12 σ23 ] ,
∂t σ23 = −iω23 σ23 + iE.d12 σ13 .
(4.21)
4.6 Rotating Wave Approximation
4.6
103
Rotating Wave Approximation
The leading terms in each of the equations in (4.21), of the form iωkj σjk , simply describe rapid oscillations. The electric field E is also oscillating rapidly; the combined
dynamics of the coherences will therefore contain components oscillating at both the
sum and difference frequencies of these oscillations. Physically, these contributions
give rise to different time-orderings of the two-photon Raman transition between
states |1i and |3i, as shown in Figure 4.2. When the detuning is small compared to
optical frequencies, the sum frequencies are many orders of magnitude higher than
the difference frequencies, and so the sum frequencies average out to zero: they can
be neglected. This is the content of the rotating wave approximation (RWA).
rotating
counter-rotating
Figure 4.2 Time-ordering. The rotating wave approximation neglects counter-rotating terms, which correspond to strongly suppressed time-orderings for the Raman process.
To implement the RWA, we define rotating coherences by the ansatz
σ
ejk = σjk eiωjk τ ,
(4.22)
4.6 Rotating Wave Approximation
104
where τ = t − z/c is the retarded time (see the text following (4.8)). Including the
dependence on z here will be useful when we consider propagation. Inserting (4.22)
into (4.21) yields the equations
∂t σ
e12 = −iE. d∗12 e−iω21 τ + d23 σ
e13 e−iω23 τ ,
∂t σ
e13 = −iE. d∗23 σ
e12 eiω23 τ − d∗12 σ
e23 e−iω21 τ ,
∂t σ
e23 = iE.d12 σ
e13 eiω21 τ .
(4.23)
Inserting the expressions (4.2) and (4.6) for Ec and Es into E, and multiplying out
the resulting expressions, we find terms multiplying rapidly varying exponentials,
like e−i(ω21 +ωs )τ , for instance, and also terms multiplying slowly varying exponentials, like e−i(ω21 −ωs )τ = e−i∆τ . Neglecting the fast oscillating terms, we obtain the
equations
∂t σ
e12 = −id∗12 . vc Ec e−i∆+ τ + ivs gs Ae−i∆τ − id23 . vc Ec e−i∆τ + ivs gs Ae−i∆− τ σ
e13 ,
i
h
∂t σ
e13 = −id∗23 . vc∗ Ec∗ ei∆τ − ivs∗ gs A† ei∆− τ σ
e12 + id∗12 . vc Ec e−i∆+ τ + ivs gs Ae−i∆τ σ
e23 ,
i
h
e13 ,
∂t σ
e23 = id12 . vc∗ Ec∗ ei∆+ τ − ivs∗ gs A† ei∆τ σ
where we have defined ∆+ = ω21 − ωc and ∆− = ω23 − ωs (see Figure 4.3).
(4.24)
4.7 Unwanted Coupling
(a)
105
(b)
(c)
Figure 4.3 Useful and nuisance couplings. (a) The desired quantum memory coupling. (b) The control field couples to the ground
state, initiating spontaneous Stokes scattering. (c) The signal field
couples to the storage state: it is very weak, and there is no significant
population in this state, so the effect of this coupling is negligible.
4.7
Unwanted Coupling
We have already succeeded in dramatically simplifying the dynamical equations, but
it is still not obvious how the behaviour described by this system of equations allows
for the implementation of a quantum memory. Other physical processes obscure
the useful features of the system. For instance, the term involving d∗12 .vc represents
the coupling of the control field to the |1i ↔ |2i transition. If this term is strong,
the control field can initiate spontaneous Stokes scattering, as shown in Figure 4.3
(b). Aside from complicating the equations, this process can be problematic, since it
generates excited atoms in the state |3i that are not correlated with the signal field.
When we attempt to retrieve the signal field, some of these uncorrelated atoms may
contribute a noisy background emission. In practice, it is possible to distinguish
the signal from the noise using phasematched retrieval, which we discuss in §6.3.2
in Chapter 6. In any case, a description of the memory interaction requires that we
can eliminate this unwanted coupling. As can be seen from Figure 4.3, the Stokes
scattering process is detuned further from resonance, with detuning ∆+ > ∆. If
4.8 Linear Approximation (2)
106
∆+ is sufficiently large (this requires that the splitting ω31 is large compared to ∆),
the term representing Stokes scattering will oscillate quickly enough to be neglected.
Alternatively, it may be possible to choose the polarization of the control such that
the product d12 .vc vanishes due to a selection rule, although strict selection rules
usually require the application of an external magnetic field. Polarization selection
rules are discussed in the context of cesium in §F.4 of Appendix F. Regardless of
the justification, in the following analysis we set d12 .vc = 0. By the same token, we
set d23 .vs = 0, thereby disregarding any coupling of the signal field to the |2i ↔ |3i
transition, as depicted in Figure 4.3 (c). The equations of motion are now given by
∂t σ
e12 = d∗12 .vs gs Ae−i∆τ − id23 .vc Ec e−i∆τ σ
e13 ,
∂t σ
e13 = −id∗23 .vc∗ Ec∗ ei∆τ σ
e12 − d∗12 .vs gs Ae−i∆τ σ
e23 ,
∂t σ
e23 = d12 .vs∗ gs A† ei∆τ σ
e13 .
4.8
(4.25)
Linear Approximation (2)
Now that we have arrived at the more transparent set of equations (4.25), we can
identify some terms in these equations that are ‘small’, in the sense that ‘weakly
excited’ operators are involved. To be more concrete, we analyse the size of each
term perurbatively. We attach a parameter to each of the coherences σ
ejk , and also
to the signal field A. This parameter just labels these quantities as ‘small’; in the
case of the coherences, they are initially vanishing, and in the case of the signal field,
4.8 Linear Approximation (2)
107
only a very few signal photons are sent into the memory. The equations become
∂t σ
e12 = d∗12 .vs gs Ae−i∆τ − id23 .vc Ec e−i∆τ e
σ13 ,
∂t σ
e13 = −id∗23 .vc∗ Ec∗ ei∆τ e
σ12 − d∗12 .vs gs 2 Ae−i∆τ σ
e23 ,
∂t σ
e23 = d12 .vs∗ gs 2 A† ei∆τ σ
e13 .
(4.26)
There are two terms proportional to 2 . They correspond to ‘second-order’ perturbative corrections to the dynamics, and they can be neglected. Dropping the label
, the linearized equations of motion for the atomic coherences are
∂t σ
e12 = d∗12 .vs gs Ae−i∆τ − id23 .vc Ec e−i∆τ σ
e13 ,
∂t σ
e13 = −id∗23 .vc∗ Ec∗ ei∆τ σ
e12 .
(4.27)
These equations contain the important physics of the quantum memory interaction.
They describe how the coherence σ
e12 is directly excited by the signal field A, and
how this excitation is then coupled to the coherence σ
e13 through the control field
Ec . Macroscopically, when many identical atoms are involved, we can identify σ
e12
with the atomic polarization, and σ
e13 with the spin-wave. Before considering the
collective dynamics of the ensemble as a whole, we first consider the propagation of
the signal field.
4.9 Propagation
4.9
108
Propagation
Heisenberg’s equations describe evolution in time, but it is not obvious how to treat
the spatial propagation of the signal field in this formalism. In fact, it is possible to
do this, but a more transparent derivation utilizes Maxwell’s equations. The atomic
ensemble behaves as a dielectric medium in the presence of the signal field, for which
the appropriate formulation is as follows,
∇.D = ρfree ,
∇ × E = −∂t B,
∇.B = 0,
∇ × H = Jfree + ∂t D.
(4.28)
Here D is the displacement field, H is the magnetic field, E is the electric field and
B is the magnetic induction. The distinction between B and H will not be very
important for us, since all the materials we are concerned with are non-magnetic.
We simply take B = µ0 H, where µ0 is the permeability of free space, and refer to
B as the magnetic field. The quantities ρfree and Jfree are the free charge density
and current, respectively. The designation ‘free’ refers to the fact that they are not
induced by the fields: they are charges and currents that are not associated with
any material dielectric properties. In any case we only deal with materials in which
there are no free charges or currents, and therefore we set ρfree = Jfree = 0. A wave
equation for the electric field is found by differentiating the equation for H, and
4.9 Propagation
109
substituting in the equation for E,
∇ × ∂t H = ∂t2 D,
⇒ −∇ × (∇ × E) = µ0 ∂t2 D.
(4.29)
The double curl derivative on the left can be simplified using the vector calculus
identity
∇ × (∇ × E) = ∇ (∇.E) − ∇2 E.
(4.30)
The signal field is a transverse propagating wave, which is divergence free2 . This
is evident from the definition of the Coulomb gauge (see Appendix C), in which
∇.A = 0, with E = ∂t A, where A is the magnetic vector potential. We therefore
drop the first term in (4.30), to obtain
∇2 E = µ0 ∂t2 D.
(4.31)
The displacement field is formed from the sum of the electric field, and the material
polarization P ,
D = 0 E + P ,
(4.32)
where P is the polarization density, defined as the dipole moment per unit volume.
Substituting (4.32) into (4.31), and using the relation 0 µ0 = 1/c2 , we arrive at the
2
These arguments apply to a homogeneous and isotropic dielectric containing no source
charges [141]
4.10 Paraxial and SVE approximations
110
wave equation
1 2
2
∇ − 2 ∂t E = µ0 ∂t2 P .
c
(4.33)
This equation relates the propagation of the optical fields to the atomic polarization
in the memory. Since the control field is so intense, it is not significantly affected by
its interaction with the ensemble, and so we do not consider its propagation further.
For the signal field, we use the equation
1 2
∇ − 2 ∂t Es = µ0 ∂t2 Ps ,
c
2
(4.34)
where Ps is the component of the atomic polarization which acts as a source for the
signal field. That is, Ps is the component of the polarization oscillating at the signal
carrier frequency ωs . Further simplification is accomplished by making use of the
fact that the signal and control fields are collimated beams.
4.10
Paraxial and SVE approximations
The slowly varying envelope approximation (SVE), and the paraxial approximation,
are both implicit in the decomposition (4.6). We assume that the amplitude A (its
status as an operator is not important for this discussion) is a smooth, slowly varying
function of time and space. The exponential factor then represents the optical carrier
wave, oscillating in time with frequency ωs , and oscillatiing along the z-axis with
wavevector ks = ωs /c.
The paraxial approximation allows us to treat the signal field as a beam traveling
4.10 Paraxial and SVE approximations
111
along the z-axis, with negligible divergence. This approximation is satisfied as long
as the transverse spatial profile of A is much larger than the signal wavelength
λs = 2π/ks .
The SVE approximation allows us to treat the propagation of the signal field
purely in terms of the envelope of the signal pulse, represented by the temporal
shape of A, without having to explicitly model the very fast time-dependence of the
carrier wave, which oscillates much faster. To make this approximation successfully,
we should have that the temporal duration of the signal field is much longer than
the optical period 2π/ωs .
To see how these approximations simplify the theoretical description, we insert
the signal field (4.6) into the wave equation (4.34). We consider only the positive
frequency component of the signal field, since only this component is coupled to
the atoms through the system (4.27). We also define the slowly varying atomic
polarization Pes by factorizing out the signal frequency,
Ps = Pes eiωs τ .
(4.35)
i
h
i
1 2 h
2
∇ − 2 ∂t ivs gs Aeiωs (t−z/c) = µ0 ∂t2 Pes eiωs (t−z/c) .
c
(4.36)
The resulting wave equation is
We take the scalar product of both sides with the polarization vector vs∗ (this is the
same as taking the inner product with vs† , see Appendix A), and apply the chain
4.11 Continuum Approximation
112
rule for the derivatives to obtain
1 2
ωs
1
ωs 2
µ0
1 2
2
2
∇⊥ + ∂z − 2 ∂t − 2i
∂z + ∂t −
A = −i vs∗ . ∂t2 + 2iωs ∂t − ωs2 Pes ,
− 2 ωs
c
c
c
c
c
gs
(4.37)
where we have divided out the exponential factor eiωs τ . The transverse Laplacian
∇2⊥ = ∂x2 + ∂y2 describes diffraction of the signal field as it propagates. Note that the
term in square brackets on the left hand side vanishes. Concerning the remaining
terms, we observe that according to the SVE approximation
ks ∂z + 1 ∂t A ∂z2 − 1 ∂t2 A .
2
c
c
(4.38)
We therefore drop the smaller term. For the same reason, we drop all but the last
term on the right hand side, and we arrive at the equation
1
i
µ0 ωs2 ∗ e
2
∇⊥ + ∂z + ∂t A = −
v .Ps .
2ks
c
2gs ks s
(4.39)
This equation describes the coupling of the atoms to the signal field amplitude.
It now remains for us to connect the atomic evolution equations (4.27) with the
macroscopic polarization Ps .
4.11
Continuum Approximation
We have in mind an ensemble of sufficient density that the atoms form an effective
continuum. Consider a small region within the ensemble at position r = (z, ρ),
4.11 Continuum Approximation
113
with volume δV . We’ll call this a voxel. To make the continuum approximation,
we should have many atoms in each voxel; nδV 1, where n is the number density of atoms in the ensemble. Each voxel should be ‘pancake-shaped’, so that its
thickness δz along the z-axis satisfies δz λs , while its transverse area δA satisfies δA λ2s . The condition on δz ensures that the longitudinal optical phase ks z
is roughly constant throughout the voxel. The condition on δA ensures that the
typical interatomic separation is much larger than the signal wavelength λs , so that
dipole-dipole interactions between the atoms can be neglected, and we can treat the
atoms as isolated from one another. The macroscopic polarization at position r is
then found by adding up the dipole moments in the voxel located at r,
P =
1 X β
d ,
δV
(4.40)
β(r)
where the index β runs over all the atoms in the voxel at position r. Using the
expression (4.12) for each atom, and recalling that the coherence σ23 is negligible
(see Section 4.8), we have
P =
1 X
β
d12 σ12
+ h.c. .
δV
(4.41)
β(r)
We now define macroscopic variables involving sums over the coherences σjk , in order
to ‘tie up’ the system of equations (4.27) with the propagation equation (4.39). For
4.11 Continuum Approximation
114
the macroscopic polarization, we define the operator
P =√
1 X β i∆τ
σ
e12 e .
nδV
(4.42)
β(r)
Note that P is not simply the magnitude of the vector P . They are closely related
(see (4.47) below), but some book-keeping is required to ensure that we keep track of
all the relevant constants. In the same vein, for the spin wave we define the operator
B=√
1 X β
σ
e13 .
nδV
(4.43)
β(r)
These definitions are motivated by analogy with the slowly varying photon annihilation operator A. For instance, the equal-time commutator of B with its Hermitian
adjoint is given by
h
i
B(z, ρ, t), B † (z 0 , ρ0 , t) =
=
XXh β γi
1
e31
σ
e13 , σ
(δV )2 n
β(r) γ(r 0 )





 (δV1)2 n × nδV (σ11 − σ33 )




0
if r and r 0 label the same voxel,
(4.44)
otherwise.
Using the linear approximation (4.20), we find, in the continuum limit δV −→ 0,
h
i
B(z, ρ, t), B † (z 0 , ρ0 , t) = δ(z − z 0 )δ(ρ − ρ0 ).
(4.45)
4.11 Continuum Approximation
115
Identical arguments yield the commutator
h
i
P (z, ρ, t), P † (z 0 , ρ0 , t) = δ(z − z 0 )δ(ρ − ρ0 ),
(4.46)
for the polarization. Therefore both the operators P and B satisfy bosonic commutation relations. We interpret them as annihilation operators for an atomic excitation
at position r. That is, P (z, ρ) annihilates a distributed excitation of the excited
state |2i at position (z, ρ). And B(z, ρ) annihilates a distributed excitation of the
storage state |3i.
Substituting the definition (4.42) into (4.41), and taking the positive frequency
component (i.e. the component oscillating at a frequency +ωs ), we find
Pes =
√
nd12 P,
(4.47)
for the slowly varying macroscopic polarization. We can now write down equations
governing both the propagation of the signal field, and the atomic dynamics. These
equations are,
1
i
2
∇ + ∂z + ∂t A = −κ∗ P,
2ks ⊥
c
∂t P
= i∆P + κA − iΩB,
∂t B = −iΩ∗ P,
(4.48)
4.12 Spontaneous Emission and Decoherence
116
where we have defined the control field Rabi frequency
Ω=
d23 .vc
Ec ,
~
(4.49)
and the coupling constant
d∗ .vs √
d∗ .vs
κ = 12
× ngs = 12
~
~
r
~ωs n
.
20 c
(4.50)
The equations (4.48) describe the propagation and diffraction of a weak signal field
through an ensemble of ideal atomic Λ-systems, prepared initially in their ground
states. We have not yet included any description of decay processes, such as spontaneous emission from the excited state, or collisional de-phasing of the spin wave.
These two processes are expected to happen on very different time-scales, but they
can be treated in exactly the same way, which we now introduce.
4.12
Spontaneous Emission and Decoherence
In Section (C.5) in Appendix (C), we discuss the description of Markovian decay
in the Heisenberg picture using Langevin equations. We used a model in which a
bosonic system was coupled to a large reservoir, also composed of bosons. In the
present discussion, all the operators we consider — A, B and P — are bosonic in
character (see Eqs. (4.8), (4.45) and (4.46) for their commutators). In the case of
spontaneous emission, these operators are coupled to the electromagnetic field, which
is a large reservoir of bosons, and so the model of Appendix C is applicable. Using
4.12 Spontaneous Emission and Decoherence
117
this model, spontaneous emission at the rate γ is described by incorporating an appropriate decay term into the dynamical equation for P , along with a Langevin noise
operator FP , which introduces fluctuations that preserve the bosonic commutation
relations of P . The noise operator is delta-correlated, meaning that no correlations
exist between its values at different times. And since all the atoms in the ensemble
are subject to independent fluctuations, no correlations exist between the noise at
different positions. These properties are summarized by the relations
hFP† (t, z, ρ)FP (t0 , z 0 , ρ0 )i = 2γ n̄P × δ(t − t0 )δ(z − z 0 )δ(ρ − ρ0 ),
hFP (t, z, ρ)FP† (t0 , z 0 , ρ0 )i = 2γ(n̄P + 1) × δ(t − t0 )δ(z − z 0 )δ(ρ − ρ0 ), (4.51)
where the expectation value is taken on the initial state of both the ensemble and
the reservoir (i.e. the electromagnetic field to which the atoms are coupled). Here
the number n̄P is the initial number of atoms thermally excited into state |2i, on
average. When dealing with optical transitions, we typically have n̄P = 0, of course.
Similarly, to treat decoherence of the spin wave B at a rate γB , we add a decay
term and a noise operator FB , which satisfies identical relations to (4.51), again with
n̄B = 0. The dynamical equations including these dissipative processes are then
i
1
2
∇ + ∂z + ∂t A = −κ∗ P,
2ks ⊥
c
∂t P
= −γP + i∆P + κA − iΩB + FP ,
∂t B = −γB B − iΩ∗ P + FB .
(4.52)
4.12 Spontaneous Emission and Decoherence
118
Note that these decay rates are for the atomic coherences. The atomic populations
decay at twice these rates. For instance, the number of spin wave excitations is
given by
Z
LZ
NB =
0
B † (z, ρ)B(z, ρ) d2 ρ dz.
(4.53)
A
In the absence of any optical fields (with Ω = 0), we find that
∂t hNB i = −2γB hNB i,
(4.54)
where the expectation value is taken on an arbitrary state. Any terms involving the
noise operator FB vanish when taking the expectation value because its fluctuations
average to zero. The relation (4.54) shows that the number of spin wave excitations
decays at the rate 2γB . The same argument applied to P shows that the number
of excited atoms, in the state |2i, decays at the rate 2γ, so that the spontaneous
lifetime of the state |2i is 1/2γ.
Now that we have introduced losses and decoherence with a degree of formal
rigour, we see that in fact the noise operators FB , FP may be neglected when
optimizing the performance of a quantum memory. The reason is that the efficiency
of a quantum memory depends only on the ratio of stored to input excitations.
That is, only the number operators for the signal field and spin wave are involved.
These number operators involve only normally ordered products, of the form A† A
or B † B, and therefore only normally ordered products of the F operators enter into
the efficiency. Since we have n̄ ∼ 0, any products of the form F † F vanish when an
4.12 Spontaneous Emission and Decoherence
119
expectation value is taken, as is clear from (4.51). Therefore the Langevin operators
can be safely dropped from the system of equations (4.52). Of course the decay
terms are important!
Chapter 5
Raman & EIT Storage
Here we use the equations of motion derived in the last chapter to study the optimization of the EIT and Raman quantum memory protocols.
5.1
One Dimensional Approximation
The analysis is greatly simplified if we use a one dimensional model, so that we
only consider propagation along the z-axis. This can always be made a good approximation by using laser-beams with low divergence. The effects of diffraction
are considered in Chapter 6. In the following, we will average over the transverse
coordinate ρ, to produce a one dimensional propagation model. We re-define the
5.1 One Dimensional Approximation
121
variables A, P and B by integrating over the transverse area A of the signal field,
A(t, z, ρ) −→ A(t, z) =
P (t, z, ρ) −→ P (t, z) =
B(t, z, ρ) −→ B(t, z) =
Z
1
√
A(t, z, ρ) d2 ρ,
A A
Z
1
√
P (t, z, ρ) d2 ρ,
A A
Z
1
√
B(t, z, ρ) d2 ρ.
A A
(5.1)
Having averaged the variables in this way, we now drop the transverse Laplacian
∇2⊥ from the propagation equation for A in (4.52). Further simplifications follow.
For instance, in the absence of any transverse structure, the control field envelope
Ec , propagating undisturbed at the speed of light along the z-axis, can be written
as a function of the retarded time τ = t − z/c only. We therefore make a change
of variables from (t, z) to (τ, z), which enables us to write the control field Rabi
frequency as Ω = Ω(τ ). Furthermore, the mixed derivative in the propagation
equation for A is simplified, since
∂z
1 + ∂t = ∂z ,
c
t
z
τ
∂t
z
= ∂τ
z
,
(5.2)
where the subscripted parentheses indicate the variables held constant. In this new
coordinate system, the one dimensional equations of motion for a Λ-type quantum
5.1 One Dimensional Approximation
122
memory are therefore given by
∂z A(z, τ ) = −κ∗ P (z, τ ),
∂τ P (z, τ ) = −ΓP (z, τ ) + κA(z, τ ) − iΩ(τ )B(z, τ ),
∂τ B(z, τ ) = −iΩ∗ (τ )P (z, τ ),
(5.3)
where Γ = γ−i∆ is the complex detuning. Note that we have dropped the decay term
associated with the decoherence of the spin wave B: by assumption this is negligible
on the time-scale of the storage process, which is what we seek to optimize. We have
also dropped the Langevin noise operator FP associated with spontaneous decay of
the polarization P , since this noise does not affect the efficiency (see the end of
Section 4.12 in the previous chapter). The equations (5.3) are mercifully rather
simple. Certainly they are easier on the eye than any of their previous incarnations
in Chapter 4! The elimination of the Langevin noise operators means that these
equations are now entirely classical in nature: we can treat (5.3) as a system of
coupled partial differential equations in three complex-valued functions A, P and
B. Since the equations are linear, the solutions will be linear, and we need not
worry about issues involving commutators or operator ordering. Our aim is to
find an expression for the Green’s function K(z, τ ), relating the input signal field
Ain (τ ) = A(z = 0, τ ) to the final spin wave Bout (z) = B(z, τ −→ ∞). In the
absence of decoherence of the spin wave, the limit τ −→ ∞ is simply a mathematical
shorthand for “the end of the storage interaction”, when the control and signal fields
5.2 Solution in k-space
123
have fallen away to zero. As described in (3.2) in Chapter 3, taking the SVD of K
will tell us how to optimize the memory efficiency. We now attempt a solution of
the system (5.3).
5.2
Solution in k-space
5.2.1
Boundary Conditions
We must solve three first order partial differential equations in three functions, and
therefore there must be three boundary conditions. For the storage process, we
begin with no excitations of the atomic polarization, and no spin wave excitations,
so the boundary conditions for the functions P and B are simply
Pin (z) = P (z, τ −→ −∞) = 0,
Bin (z) = B(z, τ −→ −∞) = 0.
(5.4)
As mentioned above, the boundary condition for the signal field is set by the initial
temporal profile of the signal envelope A, as it impinges on the front face of the
ensemble at z = 0,
Ain (τ ) = A(z = 0, τ ).
(5.5)
Our analysis will tell us the shape for Ain that maximizes the memory efficiency.
These boundary conditions are represented by the tableau in Figure 5.1.
5.2 Solution in k-space
124
Figure 5.1 Quantum memory boundary conditions. Example solutions for the functions A, P and B are shown in each panel, with the z
coordinate running from top to bottom and the τ coordinate running
from left to right. The red lines indicate the boundary conditions
that must be specified to generate the solutions.
5.2.2
Transformed Equations
To proceed with solving the equations of motion, it will be useful to reduce them to
a system of coupled ordinary differential equations. This can be done by applying
a unilateral Fourier transform over the z coordinate (see Appendix D). We define
Fourier transformed variables accordingly,
Z ∞
1
e
A(z, τ )eikz dz,
A(k, τ ) = √
2π 0
Z ∞
1
Pe(k, τ ) = √
P (z, τ )eikz dz,
2π 0
Z ∞
1
e
B(k, τ ) = √
B(z, τ )eikz dz.
2π 0
(5.6)
5.2 Solution in k-space
125
Using the result (D.27) for the transform of the spatial derivative ∂z , we obtain
e − √1 Ain = −κ∗ Pe,
−ik A
2π
e − iΩB,
e
∂τ Pe = −ΓPe + κA
e = −iΩ∗ Pe.
∂τ B
(5.7)
We remark that the independence of Ω from z is critical to the usefulness of this
transformation. The spatial propagation has now been reduced to an algebraic
e
equation, which we can solve for A.
5.2.3
Optimal efficiency
Even given unlimited energy for the control pulse, the storage efficiency is limited
by spontaneous emission from the excited state |2i. The storage into the dark state
|3i, which is not affected by spontaneous emission, is always mediated via coupling
to |2i. Even with perfect transfer between states |2i and |3i, we can never store
more efficiently into |3i than we can couple to |2i. Therefore the storage efficiency
is bounded by the efficiency with which we can transfer population into |2i. To
evaluate this upper bound, we simply neglect the spin wave, and solve the equation
e we obtain
for Pe with Ω = 0. Including the solution for A,
|κ|2 e
iκ
e
∂τ P = − Γ + i
Ain .
P+√
k
2πk
(5.8)
5.2 Solution in k-space
126
This equation can be integrated directly to give
iκ
2
Pe(k, τ ) = Pein (k)e−(Γ+i|κ| /k)τ + √
2πk
Z
τ
e−(Γ+i|κ|
2 /k)(τ −τ 0 )
Ain (τ 0 ) dτ 0 .
(5.9)
−∞
Since we are concerned with storage, we can set the initial polarization to zero.
We next assume that we are somehow able to transfer all the excitations from Pe
e with no loss, at some time τ = T , which marks the end of
to the spin wave B,
eout , and the
the storage interaction. We can then make the substitution Peout → B
optimal storage process is described by the map
Z
∞
e
K(k,
T − τ )Ain (τ ) dτ,
eout (k) =
B
(5.10)
−∞
where the k-space storage kernel is given by
iκ −(Γ+i|κ|2 /k)τ
e
e
.
K(k,
τ) = √
2πk
(5.11)
e = 0, so that no storage takes place after τ = T .
Note that for times τ > T , we set K
Now, some comments are warranted. First, the optimal storage efficiency does not
depend on the detuning ∆. To see this, note that e−Γτ = e−γτ × ei∆τ . The latter
factor, involving the detuning, represents a pure phase rotation. We could absorb it
into the definition of Ain without altering its norm. Therefore we can drop it from
the kernel — it has no effect on its singular values, and no effect on the optimal
efficiency. Second, the optimal efficiency only depends on the resonant optical depth,
5.2 Solution in k-space
127
defined by
d=
|κ|2 L
.
γ
(5.12)
To see this, we normalize the time and space coordinates, along with the spin wave
and signal field amplitudes, to make them dimensionless. The spontaneous decay
rate γ and the ensemble length L provide natural time and distance scales for this
normalization. We denote the normalized variables by an overbar,
τ = γτ,
e
A
e= √
A
,
γ
k = kL,
e
e = √B .
B
L
(5.13)
This notation is rather clumsy, but it serves to clarify the re-scaling. With these
definitions, the storage map becomes
e out k =
B
Z
∞
e k, T − τ Ain (τ ) dτ ,
K
(5.14)
−∞
where the kernel has been converted into the dimensionless form
√
i d −(1+id/k)τ
e
K k, τ = √ e
.
k 2π
(5.15)
Here we have assumed for simplicity that κ is real, (i.e. κ = κ∗ ), and we have
dropped the detuning for the reason mentioned above. It is now clear that the
optical depth d is the only parameter associated with the interaction that plays any
role in determining the efficiency of the storage process. This was first shown by
5.2 Solution in k-space
128
Gorshkov et al [133] . Their explanation is that, regardless of the memory protocol
used, the branching ratio between loss — spontaneous emission — and storage —
absorption — is fixed by the optical depth. We have derived the result by arguing
that we cannot do better than is possible through direct, linear absorption into the
state |2i. Clearly detuning from resonance cannot improve matters, and so from
this perspective it is unsurprising that the best possible efficiency is only limited
by the resonant coupling, parameterized by d. Below we examine the quantitative
behaviour of the optimal efficiency in more detail.
e
The optimal efficiency is given by the square of the largest singular value of K.
Note that it makes no difference whether the argument τ or the ‘flipped’ argument
T − τ is used; this just flips the input modes without affecting the singular values.
e has a singularity at k = 0, and this means we cannot approxiUnfortunately, K
mate it as a finite matrix, in order to compute the singular values numerically. To
proceed further, we need to transform to a different coordinate system to remove
the singularity. One possibility is to apply an inverse Fourier transform, and this
works well. Before doing this, however, we first introduce another way to analyse
the kernel (5.15), which gives a degree of insight into its structure. In what follows,
we will drop the overbar notation for the normalized parameters, as a concession to
legibility.
5.2 Solution in k-space
5.2.4
129
Solution in Wavelength Space
e are unaffected by a unitary transformation.
Recall that the singular values of K
Consider the coordinate transformation k −→ λ = 2π/k, from k-space to ‘wavelength space’. The kernel (5.15) is no longer singular in this coordinate system.
To guarantee unitarity, the transformed kernel must include a Jacobian factor of
√
√
1/ ∂k λ = −i 2π/λ (see §3.3.3 in Chapter 3). We obtain the result
√
e
K(λ,
τ) =
d −(1+idλ/2π)τ
e
.
2π
(5.16)
e A , as described in (3.19) in ChapWe now form the anti-normally ordered product K
ter 3, which takes the form
1
e A (λ, λ0 ) = 1 ×
.
K
4π
2πi λ − d i − λ0
(5.17)
e A give the singular values of K,
e so we should try to solve the
The eigenvalues of K
following eigenvalue problem,
1
2πi
Z
∞
−∞
ψej (λ0 )
dλ0 = ηj ψej (λ).
0
λ − 4π
i
−
λ
d
(5.18)
The integrand has a singularity at λ0 = λ − 4π
d i. We consider extending the integral
into the complex plane, and integrating along a semicircle-shaped contour, closed
in the lower half of the complex plane (as depicted in Figure D.4 in Appendix D),
so that the contour encloses the singularity. We assume that the mode functions
5.2 Solution in k-space
130
ψej fall away to zero along the curved portion of the contour, in the limit that its
radius is made infinitely large. Under this assumption, the only contribution to the
integral comes from the straight portion of the contour, along the real line, which
is precisely the integral in (5.18). We can now use Cauchy’s integral theorem (see
§D.1.1 in Appendix D) to evaluate the left hand side of (5.18),
ψej λ −
4π
d i
= ηj ψej (λ).
(5.19)
By inspection, a possible form for the modefunctions is
ψej (λ) ∝ e±iαj λ .
(5.20)
That is, plane waves in λ-space, each with some eigenfrequency αj . We should choose
the minus sign in (5.20), so that the modefunctions are exponentially damped in the
lower half of the complex plane, as we assumed above. The storage efficiencies ηj
are then given by
ηj = e−4παj /d .
(5.21)
This form is reassuring, since the ηj −→ 1 in the limit d −→ ∞, which makes sense.
However, it is not obvious that there is any constraint on the eigenfrequencies αj .
We need a ‘quantization condition’ on the modes. One possibility is to transform
back into ordinary z-space, and look for a physically reasonable condition. First, we
5.2 Solution in k-space
131
transform back in to k-space. Remembering to include the Jacobian factor, we get
1
ψej (k) ∝ × e−2πiαj /k .
k
(5.22)
The proportionality symbol reflects the fact that the modes should be properly
normalized; at the moment we are simply concerned with their functional form.
Transforming back into z-space requires taking the inverse Fourier transform of
(5.22). The method is described in §D.5.2 of Appendix D. The result is
ψj (z) ∝ J0 2
p
2παj z ,
(5.23)
where J0 is a zero’th order ordinary Bessel function of the first kind. In the limit of
large d, we expect that the signal field is completely absorbed, so that there is no
signal left at the end of the ensemble. We are working in normalized coordinates, so
the end of the ensemble is located at z = 1. With no remaining signal field, no spin
wave excitations can be excited, so we should expect ψj (1) ∼ 0. This is satisfied if
we choose αj = .23, 1.21, 2.98, 5.53, etc..., as shown in Figure 5.2. This choice of
quantization is consistent with the orthogonality requirement on the modes,
Z
1
ψi (z)ψj∗ (z) dz = 0,
if
i 6= j,
(5.24)
0
which follows from the orthogonality condition (D.34) of the Bessel functions (see
§D.5 in Appendix D).
5.2 Solution in k-space
132
1
0
0
0.5
1
1.5
2
2.5
3
3.5
4
4.5
5
√
Figure 5.2 The first three zeros of the function J0 (2 2πx). The
condition that the spin wave should vanish at z = 1 picks out these
zeros as the eigenfrequencies αj .
The optimal storage efficiency is then given by
η1 = e−4πα1 /d ≈ 1 − 2.9/d,
(5.25)
where the approximation holds in the limit of large d.
The above quantization procedure used to derive the optimal efficiency (5.25)
was rather ad hoc, and we should check it against a numerical SVD. We therefore
return to the k-space kernel (5.15). As noted previously the singularity at k = 0
makes this form of the kernel inconvenient. Fortunately it is easy to take the inverse
Fourier transform from k-space back into ordinary space (again using the method
described in §D.5.2 of Appendix D). The result is
√
K (z, τ ) =
√
de−τ J0 2 dτ z .
(5.26)
We could perform a numerical SVD on this kernel directly. The numerical problem
5.2 Solution in k-space
133
converges better, however, if we form the anti-normally ordered product kernel KA ,
which is given by
√ d
0
KA z, z 0 = e−d(z+z )/2 I0 d zz 0 .
2
(5.27)
Here I0 is a zero’th order modified Bessel function of the first kind (see §D.5.3 in
Appendix D for a clue as to how to perform the required integral). The optimal
efficiency η1 is the largest eigenvalue of this kernel. In Figure 5.3, we plot the analytic
prediction (5.25) alongside the numerical result, over a range of optical depths. The
analytic formula is an excellent approximation for optical depths larger than ∼ 50.
This scaling of the optimal storage efficiency was first noted by Gorshkov et al. [133] .
They also provided an elegant proof of the optimality of the kernel (5.27), that does
not rely on the heuristic assertion that ‘we cannot map more efficiently to B than
we can to P ’. Nonetheless, the result is the same.
1
10
0
10
−1
10
−2
10
−3
10
0
400
800
1200
1600
2000
Figure 5.3 Optimal storage efficiency. The plot shows the difference between the optimal efficiency η1 and unity, on a logarithmic
scale, as a function of the optical depth d. The analytic formula
1 − η1 ≈ 2.9/d derived above (green), is in excellent agreement with
the numerical result (blue), found by diagonalizing the kernel (5.27).
Having identified the best possible storage efficiency, we now continue with our
5.2 Solution in k-space
134
analysis of the equations of motion, including the control field. This analysis will
reveal how the temporal profile of the optimal input mode depends on the control
pulse, and in what regimes it is possible to effectively shape the optimal input mode
by shaping the control.
5.2.5
Including the Control
e and substituting
We return to the system (5.7). Solving the first equation for A,
the result into the second equation for Pe, yields a pair of coupled linear differential
equations in time only. For each spatial frequency k, we need to solve for the
e To do this, we define a vector |ψi whose
temporal dynamics of the system (Pe, B).
e
two elements are the functions Pe and B,
e )| ↓i,
|ψ(τ )i = Pe(τ )| ↑i + B(τ
(5.28)
where the basis kets | ↑i and | ↓i are given by


 1 

| ↑i = 
 ,
0

and

 0 

| ↓i = 
 .
1
(5.29)
We have suppressed the dependence on the spatial frequency k, which is no longer
a dynamical variable, but which of course must not be forgotten! The equation of
motion for |ψi is found to be
∂τ |ψi = −iM |ψi + |α0 i,
(5.30)
5.2 Solution in k-space
135
where the time-dependent matrix M is given by


M (τ ) = 

|κ|2
k

− iΓ Ω(τ ) 
,

Ω∗ (τ )
0
(5.31)
and where the time-dependent vector |α0 i includes the signal field boundary condtion,
|α0 (τ )i = i √
κ
Ain (τ )| ↑i.
2πk
(5.32)
Using the normalized variables introduced in the previous section, in which all
lengths are scaled by L and all frequencies by γ, we simply replace κ with
√
d. The
structure of (5.30) is very similar to Schrödinger’s equation (B.6) for a two-level
system, except that the evolution is not unitary, because M is not quite Hermitian, and because of the ‘driving term’ |α0 i. Nonetheless, techniques applied to the
solution of Schrödinger’s equation remain useful. First, we write down the formal
solution. Suppose that we are able to construct a propagation matrix V (τ ) such
that ∂τ V = iV M . We then have that
∂τ (V |ψi) = (∂τ V ) |ψi + V ∂τ |ψi
= iV M |ψi + V (−iM |ψi + |α0 i)
= V |α0 i.
(5.33)
5.2 Solution in k-space
136
Integrating this gives
|ψ(τ )i = V
−1
−1
(τ )Vin |ψin i + V
Z
τ
(τ )
V (τ 0 )|α0 (τ 0 )i dτ 0 ,
(5.34)
−∞
e and where Vin is the
where |ψin i contains the boundary conditions for Pe and B,
propagation matrix at the start of the interaction. Clearly we must have Vin = I,
the identity operator. To find the spin wave at the end of the storage process, for
which |ψin i = 0, we take the limit τ −→ ∞ to obtain
eout = h↓ |V −1
B
out
Z
∞
V (τ )|α0 (τ )i dτ,
(5.35)
−∞
where Vout = V (τ −→ ∞). The k-space storage kernel is then given by
i
e
K(k,
τ) =
k
r
d
−1
h↓ |Vout
(k)V (k, τ )| ↑i,
2π
(5.36)
where we have now included the dependence of the V matrices on k explicitly, lest
we forget it. If we can find an expression for V , we can construct the storage kernel,
and so find the optimal input mode by means of its SVD. Here we comment that if
the matrix M (τ ) were replaced by an ordinary function, we would simply have
Z
V (τ ) = exp i
τ
0
0
M (τ ) dτ .
(5.37)
−∞
And in fact this is still true whenever M is a diagonal matrix, since then [M (τ ), M (τ 0 )] =
0 (diagonal matrices always commute). Alternatively, if M is constant in time, it
5.2 Solution in k-space
137
can be pulled out of the integral in (5.37), and there is no issue with commutation
at different times. However, in general, when M is non-diagonal and time-varying,
as we have whenever Ω is not simply a constant, the solution (5.37) is not correct1 .
This is an example of what is sometimes known as the great matrix tragedy: the simple fact that eA eB 6= eA+B when [A, B] 6= 0 is responsible for most of the difficulty
arising in quantum mechanical calculations!
5.2.6
An Exact Solution: The Rosen-Zener case
It is possible to find an exact solution for V , in the particular case that the shape
of the control field envelope is given by a hyperbolic secant,
Ω0
sech
Ω(τ ) =
Tc
τ
Tc
,
(5.38)
where Tc sets the duration of the control pulse, and Ω0 is a dimensionless constant
that sets the strength of the pulse (see Figure 5.4 (a)). The method of solution
is due to Rosen and Zener [142–144] . We include details of the derivation here for
completeness, but in §5.4 below we introduce a numerical approach that is faster,
more accurate and more general. The analytic solutions presented here do provide
useful points of comparison, of course.
We first transform to the interaction picture, to remove the rapid oscillations
generated by the diagonal elements of M . To do this, we define the diagonal and
1
Sometimes the formal solution is written like this, but in general it is understood that the
exponential must be time ordered
5.2 Solution in k-space
138
off-diagonal matrices



M0 = 

2β
Tc
0

0 
,

0

 0 Ω 
,
MX = 


Ω∗ 0
(5.39)
where for later convenience we have defined 2β/Tc = d/k − iΓ. Clearly we have that
M = M0 + MX . The interaction picture evolution operator VI is then defined by
VI = V V0−1 , where V0 satisfies ∂τ V0 = iV0 M0 . We can write out V0 explicitly as
V0 (τ ) = eiM0 τ since M0 is a constant. The equation of motion for VI is found to be
∂τ VI
= (∂τ V ) V0−1 + V ∂τ V0−1
= (iVI V0 M ) V0−1 + VI V0 −iM0 V0−1
= iVI MI ,
(5.40)
where MI = V0 MX V0−1 is the operator that generates time evolution in the interaction picture, given by


MI = 


0
Ω∗ e−2iβτ /Tc
Ωe2iβτ /Tc 
.

0
(5.41)
As described by Pechukas and Light [143] , the matrix elements of VI each satisfy a
second order differential equation, which we find by differentiating the equation of
5.2 Solution in k-space
139
motion,
∂τ (∂τ VI ) = i (∂τ VI ) MI + iVI (∂τ MI )
= −VI MI2 + iVI (∂τ MI ) .
(5.42)
The square of MI is simply given by MI2 = |Ω|2 I, where I is the identity matrix. We
can express the derivative of MI as ∂τ MI = MI G, where G is the diagonal matrix


G=

∂τ Ω∗
Ω∗

−
2i Tβc
0
∂τ Ω
Ω
0
+
2i Tβc

.

(5.43)
Inserting this into (5.42) gives
∂τ2 VI − (∂τ VI ) G + |Ω|2 VI = 0.
(5.44)
The boundary conditions are
VI τ →−∞ = I,
∂τ VI τ →−∞ = iMI τ →−∞ .
(5.45)
The equation (5.44) is solved by making a temporal coordinate transformation.
We define the normalized integrated Rabi frequency by
1
ω(τ ) =
W
Z
τ
−∞
|Ω(τ 0 )|2 dτ 0 ,
(5.46)
5.2 Solution in k-space
140
where W is a normalization related to the total energy in the control pulse,
Z
∞
W =
|Ω(τ )|2 dτ.
(5.47)
−∞
The coordinate ω runs from 0 to 1, as τ runs from −∞ to ∞ (see Figure 5.4 (b)). ω
can be thought of as the time coordinate marked out by a clock that is powered by
the control field. Using the control field profile (5.38), we can evaluate the integral
in (5.46) explicitly, to get
ω(τ ) = 12 tanh
(a) 1
0
-6 -4 -2 0 2 4 6
τ T
+ 12 .
(5.48)
(b)
-6 -4 -2 0 2 4 6
Figure 5.4 The Rosen-Zener model. The control field, shown in
(a) with Ω0 = Tc = 1, takes the form of hyperbolic secant. (b): the
integrated Rabi frequency ω marks out time at a rate given by |Ω|2 .
Under the transformation τ −→ ω, the temporal derivative ∂τ transforms as
5.2 Solution in k-space
141
follows,
∂τ
= (∂τ ω) ∂ω
=
=
1
τ
2
sech
∂ω
2Tc
Tc
ΩTc 2
1
∂ω ,
2Tc Ω0
(5.49)
where we used the control shape (5.38). The second derivative is then given by
∂τ2 =
=
∂τ2 ω ∂ω + (∂τ ω)2 ∂ω2
"
#
1
ΩTc 2
1 ΩTc 4 2
(1 − 2ω)
∂ω +
∂ω .
Tc2
Ω0
4 Ω0
(5.50)
Putting all this together, and using
ΩTc
Ω0
2
= 4ω(1 − ω),
(5.51)
we find the equation
ω(1 −
ω)∂ω2 VI
1
+ (∂ω VI )
+ iβZ − ω + Ω20 VI = 0,
2
(5.52)
where Z is the Pauli matrix


 1 0 
.
Z=


0 −1
(5.53)
5.2 Solution in k-space
142
The boundary conditions, in terms of the new variable ω, are given by

VI ω→0 = I,
where θ± =
1
2

∂ω VI ω→0 = iΩ0 


0
ω −θ−
ω −θ+
0

,

(5.54)
± iβ. The equation (5.52) is known as a hypergeometric differential
equation. The solutions are known as hypergeometric functions, denoted2 by the
symbol F , and parameterized by the coefficients appearing in the equation. The
hypergeometric functions are special functions that can be evaluated using a mathematics application such as Matlab or Mathematica. Matching the general solutions
to the boundary conditions (5.54), the solution for VI is given by


VI (ω) = 


iΩ0 θ+
θ+ ω F (θ+
F (Ω0 , −Ω0 , θ+ , ω)
iΩ0 θ−
θ− ω F (θ−
+ Ω0 , θ+ − Ω0 , 1 + θ+ , ω) 
.

F (Ω0 , −Ω0 , θ− , ω)
+ Ω0 , θ− − Ω0 , 1 + θ− , ω)
(5.55)
The properties of these special functions can be used to show that at the end of the
storage process we have


VI ω→1 = 


Γ2 (θ+ )
Γ(θ+ +Ω0 )Γ(θ+ −Ω0 )
i sin(πΩ0 )
cosh(πβ)
i sin(πΩ0 )
cosh(πβ)
Γ2 (θ− )
Γ(θ− +Ω0 )Γ(θ− −Ω0 )

,

(5.56)
where Γ(x) is the Euler Gamma function. We substitute these matrices into (5.36) to
obtain an expression for the storage kernel in terms of k and ω. Some manipulations
2
Sometimes the symbol 2 F1 is used, and the designation Gauss hypergeometric function then
distinguishes this from the generalized hypergeometric functions p Fq .
5.2 Solution in k-space
143
reveal that the determinant of the matrix (5.56) is 1, which simplifies forming the
−1
inverse Vout
. The exponential factor ei2βτ /Tc that enters when transforming from VI
back to V = VI V0 translates into the factor
e
i2βτ /Tc
=
ω
1−ω
iβ
.
(5.57)
After some algebra, we arrive at the result
√ iβ
i d
ω
iQ(k, ω)
e
×p
,
K(k, ω) = √
k 2π 1 − ω
2ω(1 − ω)/Tc
(5.58)
where we’ve defined the function Q in the following way
Q(k, ω) =
Ω0 θ−
Γ2 (θ+ )
ω F (θ− + Ω0 , θ− − Ω0 , 1 + θ− , ω)
Γ(θ+ + Ω0 )Γ(θ+ − Ω0 ) θ−
sin(πΩ0 )
−
F (Ω0 , −Ω0 , θ+ , ω).
(5.59)
cosh(πβ)
It helps to keep in mind that θ± is a function of k through β. Note we have included
the Jacobian factor
p
√
2ω(1 − ω)/Tc = ∂τ ω in the denominator of (5.58), to make
the transformation from τ to ω a unitary one (see §3.3.3 in Chapter 3). We would
now like to extract the optimal input mode and its associated optimal efficiency by
applying the SVD. However, just as in the case of (5.15), the kernel in (5.58) has a
singularity at k = 0. As before, we remove this by transforming form k-space into
λ-space, where λ = 2π/k is the wavelength of the spin wave excitation. Including
5.2 Solution in k-space
144
e in terms of λ and ω is,
the Jacobian factor, the expression for K
i
e
K(λ,
ω) =
2π
r
Tc d −θ−
(1 − ω)−θ+ Q(λ, ω),
ω
2
(5.60)
where Q is given, as before, by (5.59), the only difference being that in λ-space, the
parameter β takes the form 2β/Tc =
dλ
2π
− iΓ. There is a minor pathology associated
with the points ω = 0 and ω = 1, which blow up, but in practice this is easily
addressed by introducing a small regularization that shifts the singularities into
the complex plane,
ω → ω + i.
(5.61)
After these steps, we have a non-singular kernel that is amenable to a numerical SVD.
From this we obtain the singular values {λj }, and a set of input modes {φej (ω)}. The
optimal storage efficiency is already given by η1 = λ21 . To find the temporal mode
of the signal field that is stored with this optimal efficiency, we need to transform
the mode φe1 (ω) back into the temporal domain. Including the Jacobian factor, we
have
1/2
2
ω(τ ) [1 − ω(τ )]
φe1 [ω(τ )]
Tc
sech Tτ e 1
τ
1
√
φ1 2 tanh
+2 .
Tc
2Tc
φ1 (τ ) =
=
(5.62)
e in (5.60) provides a check on the numerical optimizaThe analytic solution for K
tions we present in §5.4 (see Figures 5.8 and 5.9). In fact, evaluating the function
5.2 Solution in k-space
145
F (a, b, c, ω) with complex a, b or c can be time-consuming, since these values are
generated by analytic continuation of F into the complex plane. This procedure
is not always accurate, and so the direct numerical optimizations presented at the
end of this chapter are both faster and more reliable. Finally, the analytic solution
(5.60) is, of course, only valid for the particular control (5.38). It would be more
convenient if we could derive an expression that holds for a range of control field
profiles. We now show how to construct an approximation to V that holds in the
adiabatic limit, which is essentially the limit of a slowly varying control.
5.2.7
Adiabatic Limit
The idea behind adiabatic evolution is to adjust Ω sufficiently slowly that at each
moment we can neglect the time dependence of M , and treat the problem as if
it were time-stationary. In this limit, the state |ψi remains in an instantaneous
eigenstate of M at all times. As M changes, the eigenstates of M slowly evolve,
and we arrange for the populated eigenstate at the end of the storage interaction
to overlap with the | ↓i state; in this way excitations are transferred into the spin
wave. To see how this works, we re-cast the equation of motion for V in terms of
the adiabatic basis, which is the basis formed by the instantaneous eigenstates of M .
Suppose M has the following eigenvalue decomposition (see §A.4.4 in Appendix A),
M = RDR−1 .
(5.63)
5.2 Solution in k-space
146
We define the operator Vad = V R, and differentiate it to obtain its equation of
motion,
∂τ Vad = (∂τ V ) R + V ∂τ R
= iV M R + V ∂τ R
= iV RDR−1 R + V R R−1 ∂τ R
= iVad Mad ,
(5.64)
where Mad = D − iR−1 ∂τ R generates the time evolution in the adiabatic basis. The
content of the adiabatic approximation is to neglect the term R−1 ∂τ R in Mad , so
that Mad is a purely diagonal matrix. This allows us to solve the equation of motion
for Vad , using the result (5.37). That is,
Z
Vad (τ ) = exp i
τ
0
0
Mad (τ ) dτ .
(5.65)
−∞
Armed with this solution, we can construct the propagation matrix V = Vad R−1 , and
therefore the storage kernel (5.36). We now implement this programme explicitly,
after which the conditions under which the adiabatic approximation is justifiable
will become clearer.
Here we remark that corrections to the adiabatic approximation can be generated, in the current formalism, by making use of the Magnus expansion [143,145] ,
or Salzman’s expansion [146] , which provide approximations to the propagator V
5.2 Solution in k-space
147
when Mad is non-diagonal and time-dependent. These corrections quickly become
unwieldy however, and so we do not present them here: the numerical approach
presented in §5.4 obviates the need for them.
To find the adiabatic kernel, we start by finding the instantaneous eigenvalues
of M (τ ), by solving the equation |M − λI| = 0, where the vertical bars denote the
determinant (see §A.4.2 in Appendix A). The resulting eigenvalues are
λ± = b ±
p
b2 + |Ω|2 ,
(5.66)
where we have defined 2b = d/k − iΓ (thus replacing the notation 2β/Tc defined in
the previous section). Solving the equation M |±i = λ± |±i for the elements of the
eigenvectors |±i, we find


 λ+ 
,
|+i ∝ 


∗
Ω


 λ− 
.
|−i ∝ 


∗
Ω
(5.67)
The diagonalizing transformation R is the matrix with |+i as its first column and |−i
as its second. But there is some freedom as to the normalization of the vectors |±i.
This is fixed by requiring that limτ →−∞ R = I, the identity. This just codifies our
knowledge that Ω = 0 at the start of the storage interaction, so that M is initially
diagonal, which means that no transformation need be applied to diagonalize it at
5.2 Solution in k-space
148
τ −→ −∞. A suitable form for R, that satisfies this boundary condition, is


 1
R=

Ω∗
λ+
λ−
Ω∗
1

.

(5.68)
Note that limτ →−∞ λ− = 0. The inverse transformation is then given by

R−1 =
1
1−
λ−
λ+



1
−Ω∗
λ+
−λ−
Ω∗
1


.

(5.69)
We have Mad = diag(λ+ , λ− ), so that Vad is given by


i
h R
τ
0
0
0

 exp i −∞ λ+ (τ ) dτ

Vad (τ ) = 
i .
h R

τ
0
exp i −∞ λ− (τ 0 ) dτ 0
(5.70)
Combining these results together and substituting them into (5.36), we find there is
only a single non-vanishing term contributing to the storage kernel,
i
e
K(k,
τ) = −
k
r
R∞
Ω∗ (τ ) exp −i τ λ− (τ 0 ) dτ 0
d
×
.
2π
λ+ (τ ) − λ− (τ )
(5.71)
We can further simplify this result if we make the assumption that
|b| |Ω|
(5.72)
at all times. This does not have anything to do with the rate at which we change
the control field, but it is usually considered as part of the adiabatic approximation,
5.2 Solution in k-space
149
as discussed in §5.2.9 below. With this approximation, we can write
λ+ ≈ 2b +
|Ω|2
,
2b
and
λ− ≈ −
|Ω|2
.
2b
(5.73)
Inserting these expressions into (5.71), and making the replacement 2b = d/k − iΓ,
we find, after a little algebra,
" R∞
#
√
Z ∞
id
0 )|2 dτ 0
|Ω(τ
1
1
d
2
e
K(k,
τ) = √ ×
|Ω(τ 0 )|2 dτ 0 ×
exp Γ τ
.
Ω∗ (τ ) exp − Γ1
k + i Γd
k + i Γd
2π Γ
τ
(5.74)
The first exponential factor represents the accumulation of phase due to the dynamic
Stark effect, in which the strong control field ‘dresses’ the atoms and causes a time
dependent shift in the |2i ↔ |3i transition frequency. The second exponential factor
represents the ‘meat’ of the interaction: the response of the atoms to the incident
signal field. The kernel (5.74) is already amenable to a numerical SVD. However, to
make a connection with previous work, we take the inverse Fourier transform. We
use the method detailed in §D.5.2 in Appendix D, and apply the shift theorem, to
get the adiabatic z-space kernel
√
Z ∞
d ∗
1
0 2
0
|Ω(τ )| dτ + dz ×J0
K(z, τ ) = −i
Ω (τ ) exp −
Γ
Γ τ
!
√ sZ ∞
d
0
2
0
2i
|Ω(τ )| dτ z .
Γ
τ
(5.75)
Note the additional contribution dz to the exponential factor. This contribution
represents the change in refractive index experienced by the signal field as it propagates through the ensemble. On resonance (with Γ = 1 in normalized units), this
5.2 Solution in k-space
150
term describes exponential attenuation i.e. absorption, with an absorption coefficient of d. This explains why the quantity d is known as the optical depth: it directly
quantifies the optical thickness of the ensemble on resonance.
A numerical SVD could also be applied to the kernel (5.75) to obtain the optimal
temporal input mode φ1 (τ ), and its associated optimal efficiency η1 = λ21 . This
adiabatic solution appears in the work of Gorshkov et al. [133] , although the method
of derivation differs slightly. They provided a method to find the optimal input
mode, in the limit of large control field energy, but they did not use the SVD: The
SVD is a more direct method, and it does not require the assumption of large control
energy. We will discuss the differences between the SVD method and the method of
Gorshkov et al. shortly. First, we make a coordinate transformation that removes
the explicit dependence of K on the shape of the control field. As in the Rosen-Zener
solution above, we transform from τ to ω, where ω is the normalized integrated Rabi
frequency, defined by (5.46). We do not assume a particular shape for the control:
its profile can be arbitrary in the present case (within the limits of the adiabatic
approximation, to be discussed below). By removing the dependence on the control
field with this coordinate transformation, we only need to perform the SVD once, in
the transformed coordinate system, in order to obtain the optimal input mode for
any control field shape. To make the coordinate transformation unitary, we include
a Jacobian factor of
√
W /Ω∗ (τ ) in the transformed kernel (see §3.3.3). This rather
conveniently cancels with the factor of Ω∗ (τ ) in (5.75), so the transformed kernel
5.2 Solution in k-space
151
can be written as
√
K(z, ω) = −i
dW −(W −W ω+dz)/Γ √d p
e
J0 2i Γ (W − W ω)z ,
Γ
(5.76)
A final, cosmetic simplification is achieved by flipping the ω coordinate, ω −→
1 − ω. This has no effect on the singular values of the kernel, but simply flips the
input modes around. The optimal mode for adiabatic storage in a Λ-type quantum
memory, with an arbitrary control field profile, is now found by taking the SVD of
√
K(z, ω) = −i
dW −(W ω+dz)/Γ √dW √ e
J0 2i Γ
ωz .
Γ
(5.77)
We note that both ω and z run from 0 to 1. The optimal temporal input mode, is
then found from the mode φ1 (ω) by the relation
Ω(τ )
φ1 (τ ) = √ φ1 [1 − ω(τ )] .
W
(5.78)
√
The factor of Ω/ W arises from the Jacobian relating the coordinates ω and τ . We
now have a simple prescription for finding the optimal input mode for a quantum
memory. Given a fixed value for W , which essentially quantifies the total energy in
the control pulse, and given values for the detuning and the optical depth, we form
the kernel (5.77) and take the SVD. Then, for any arbitrary shape of the control,
we can construct the optimal input mode φ1 , using the transformation (5.78) above.
Some examples of the optimal input modes predicted using this approach can be
5.2 Solution in k-space
152
found in Figures 5.8, 5.9, 5.10 and 5.11 in §5.4 at the end of this chapter.
In general, the storage efficiency depends on the optical depth, the detuning
(through Γ), and the control pulse energy, through W , since the kernel (5.77) depends on all of these quantities. We also know however, from the discussion in
§5.2.3, that the best possible storage efficiency only depends on the optical depth.
Below we connect these two results together.
5.2.8
Reaching the optimal efficiency
In §5.2.3 we derived an expression for the optimal storage efficiency possible in a
Λ-type ensemble memory. In fact, it is possible to reach this optimal efficiency in the
adiabatic limit. As was first shown by Gorshkov et al. [133] , the anti-normally ordered
kernel formed from the adiabatic storage kernel (5.77) is equal to the optimal kernel
(5.27), in the limit of large control pulse energy. To see this, we substitute (5.77) into
the expression (3.19) for the anti-normally ordered kernel, and perform the integral
over ω. In the limit W −→ ∞, we can evaluate the integral analytically. After some
leg work — see §D.5.3 in Appendix D — we find that the result is exactly (5.27).
That is,
Z
lim
W →∞ 0
1
∗
0
Z
K(z, ω)K (z , ω) dω =
0
∞
K(z, ω)K ∗ (z 0 , ω) dω =
√
d −d(z+z 0 )/2
e
I0 (d zz 0 ).
2
(5.79)
This shows that it is possible to saturate the upper bound on the storage efficiency,
even in the adiabatic limit. The adiabatic limit is not only useful because we can
5.2 Solution in k-space
153
construct the optimal input mode explicitly. It is also useful because it is possible
to shape the input mode by shaping the control. From the form of (5.78), it is
clear that, by an appropriate choice of Ω(τ ), we can choose the shape of φ1 (τ ). In
particular, we can choose the control so that φ1 (τ ) matches the temporal profile of
some ‘given’ input field. This is of considerable practical importance, since it may
be experimentally much easier to shape the bright control field, than to shape the
weak input field (this is discussed in Chapter 8). The combination of these two facts
— that adiabatic storage can be optimal, and also that it enables one to shape the
input mode — makes the adiabatic limit an important regime for the operation of
a quantum memory.
Under what circumstances is the equality in (5.79) achieved? The limit of W −→
∞ is not really required. Examining the form of (5.77), it is clear that we just
need to make W large enough that the exponential factor e−W ω/Γ , evaluated at the
limit ω = 1, is sufficiently small that extending the integral further would make no
difference. We should certainly have that W |Γ| then. In addition, we should
ensure that any contribution from the Bessel function is negligible at ω = 1. Using
√
the approximation J0 (ix) = I0 (x) ∼ ex / x for x 1, we see that we should also
have that W d. To summarize, adiabatic storage is optimal if the control pulse
is sufficiently energetic that the conditions
W max (|∆|, d)
(5.80)
5.2 Solution in k-space
154
are satisfied. Recall that ∆ is the common detuning of the signal and control fields
from resonance. We can connect W with the total energy Ec in the control pulse.
The total energy is
Ec = A
Z
∞
Ic (τ ) dτ,
(5.81)
−∞
where Ic = 20 c|Ec |2 is the cycle-averaged control pulse intensity. From the definition
of the Rabi frequency (4.49), we then find
20 cA
Ec = γW.
d23 .vc 2
~ (5.82)
Here the presence of the factor of γ indicates that we have converted back into
ordinary units (rather than normalized units). From the definition of W (5.47), we
find it has the dimensions of frequency, and so in ordinary units it is accompanied
by the factor γ. Let us define the number of photons in the control pulse by Nc =
Ec /~ωc . We also define Na as the number of atoms in the ensemble addressed by
the optical fields, Na = nLA. Using (4.50) and (5.12), we can express the optical
depth as
∗
d12 .vs 2 ~ωs
× Na .
d=
~ 20 cAγ
(5.83)
With these definitions, we see that the condition W d in (5.80) amounts, essentially, to the condition Nc Na (we have used ωs ≈ ωc and |d∗12 .vs | ≈ |d23 .vc |).
We might therefore describe this condition as describing a ‘light-biased’ interaction,
where the number of control photons dominates over the number of atoms; this is
5.2 Solution in k-space
155
why it is the latter quantity that limits the efficiency.
In the work of Gorshkov et al. [133] , a method is presented for optimizing the
adiabatic storage efficiency in this light-biased limit. The method works by combining the optimal anti-normally ordered kernel (5.27), whose eigenfunctions are the
optimal spin waves, with the adiabatic storage kernel (5.75), which connects the
control field profile to the signal field profile. Combining these two kernels together
is possible only when Nc Na , and also W ∆. In this limit their method works
extremely well. One advantage of the SVD method however, is that we can apply
it directly to the kernel (5.77), without making this approximation, and we can
therefore find the optimal input modes for arbitrary values of W .
5.2.9
Adiabatic Approximation
We have employed several approximations in the name of adiabaticity. We now
examine the physical content of these approximations. The first assumption we
made, in the text following (5.64), was to neglect the term R−1 ∂τ R in the adiabatic
generator Mad . Clearly this term vanishes if the control field is held constant, and
so the size of this term is set by the rate of variation of the control field. To make
the adiabatic approximation, we must therefore limit the bandwidth of the control
pulse. To find this limit, we introduce the second approximation we made in (5.72),
namely that
|b| |Ω|.
(5.84)
5.2 Solution in k-space
156
With this approximation, we can write the diagonalizing transformations R, R−1 in
the form


 1
R=

Ω∗
2b
Ω
− 2b
1

;



R−1 = 


1
∗
− Ω2b
Ω
2b
1

.

(5.85)
For adiabatic evolution, we should have that ||R−1 ∂τ R|| ||D||, where D =
diag(2b + |Ω|2 /2b, −|Ω|2 /2b). Here the double bars represent the Frobenius norm,
which is found by adding in quadrature the magnitudes of all the elements in a
matrix. We neglect terms of order |Ω/b|, which are small by assumption. We then
arrive at the condition
∂τ Ω 2
2
2b |Ω| .
(5.86)
Recall that b = 21 (d/k − iΓ) depends on the wavevector k, and that Ω varies with
time, so we should be careful to satisfy this condition for all the values of these two
parameters that play a role in the storage process.
We can distinguish two regimes. First, the EIT regime (see §2.3.1 in Chapter
2), in which the optical fields are tuned into resonance with the excited state. In
this case, ∆ = 0 and so Γ = 1 (in normalized units). Since d 1 for reasonably
efficient storage, the contribution to b from the complex detuning Γ is small, and we
have |b| ∼ d/2k. Therefore the adiabatic conditions (5.84) and (5.86) vary strongly
with k, and we must take some care to identify the range of wavevectors that are
important for the storage process.
The second regime is the Raman regime (see §2.3.2 in Chapter 2), in which the
detuning is large compared to the excited state linewidth; ∆ 1 in normalized
5.2 Solution in k-space
157
units. In this regime |b| ∼ 21 (d/k − ∆). In this case the large contribution to b from
the detuning makes the adiabatic conditions less dependent on k.
What range of wavevectors are important in the storage process? One answer is
provided by inspection of the k-space kernel (5.74). Using the expansion (k+i Γd )−1 =
−i Γd +
Γ2
k
d2
3
+ i Γd3 k 2 + . . ., we can re-write the k-space map in the form
i
e
K(k,
ω) ≈ − √
2π
r
ΓW
W
2
W
× ei d ωk × e− d2 ωk ,
d
(5.87)
where we have introduced the integrated Rabi frequency ω, and ‘flipped’ the kernel,
as we did in deriving (5.77). This expression is a good approximation if |d/Γ| is
large, which is not guaranteed in the Raman regime, but which is generally true in
the EIT regime. The final exponential factor describes a Gaussian profile in k-space,
with a characteristic width given by
δk = δk(ω) = √
d
,
Wω
(5.88)
in the EIT regime with Γ = 1. In this case the most restrictive form of the adiabatic
condition (5.84) can be written as (dropping an unimportant factor of two)
Ωmax d
,
δk(ωmax )
(5.89)
where Ωmax = max(Ω) is the peak Rabi frequency of the control, and where ωmax is
the value of the integrated Rabi frequency when this peak occurs. For a symmetric
5.2 Solution in k-space
158
control pulse, we would have ωmax = 21 ; in general ωmax will be some fraction that
for our purposes we may approximate as ∼ 1. Substituting (5.88) into (5.89), we
find that this condition in fact restricts the bandwidth δc of the control pulse:
δc 1.
(5.90)
Here we made the approximation W ∼ Ω2max /δc . The condition (5.90), in ordinary
units, is δc γ. That is, the bandwidth of the control should not exceed the natural
linewidth of the |2i ↔ |3i transition in the ensemble, to achieve adiabatic storage on
resonance. This suggests that EIT is a memory protocol best suited for the storage of
narrowband fields. However, the analysis by Gorshkov et al. [133] reaches a different
conclusion: that the adiabatic restriction on the bandwidth of the control is δc dγ
(in ordinary units). This much less stringent condition is derived by evaluating the
adiabatic condition (5.86) using b ∼ d/max(δk) with max(δk) ∼ 1. Identifying the
maximum of the quantity |∂τ Ω/Ω| with the bandwidth δc , one arrives at their result.
One justification for using δk ∼ 1 is that for optimal storage, we only need to access
the optimal spin wave mode; higher modes are irrelevant. Since the optimal mode
is that mode which is most slowly varying in space, its width in k-space is limited
to a relatively small region, and the adiabatic approximation need only be satisfied
within this range. It is difficult to argue rigorously about these approximations, but
as we will see, numerics reveal that the reality lies somewhere between these two
cases: the adiabatic approximation breaks down rather quickly in the EIT regime
5.3 Raman Storage
159
as the bandwidth approaches the natural linewidth, although it is true that this is
mitigated somewhat by increasing the optical depth.
In the Raman regime, the analysis is simpler. The adiabatic condition is simply
that
Ωmax |∆|,
(5.91)
independent of k. We might comment that there may be some particular value of k
such that there is a cancellation, and b becomes small, but the effect of this isolated
point is generally negligible. The limitation on the control bandwidth comes from
(5.86), which yields the condition
δc |∆|.
(5.92)
Therefore adiabatic evolution is guaranteed in the Raman case whenever the detuning ∆ is the dominant frequency involved in the interaction, with both the Rabi
frequency and the bandwidth of the control field small by comparison.
5.3
Raman Storage
So far we have studied the properties of storage in a Λ-type ensemble for arbitrary
values of the detuning. We now specialize to the case of large detunings, ∆ γ
(or ∆ 1 in normalized units). In this Raman regime the storage kernel simplifies further, and an interesting connection between the input and spin wave modes
emerges. The following treatment forms the basis of our Rapid Communication on
5.3 Raman Storage
160
Raman storage [77] .
Taking the limit ∆ 1, we can write Γ ≈ −i∆, and the storage kernel (5.77)
becomes
√
K(z, ω) = C × e−i(W ω+dz)/∆ × J0 (2C ωz),
(5.93)
where we have defined the Raman memory coupling by
√
C=
Wd
≈
|Γ|
√
Wd
.
∆
(5.94)
The exponential factor in (5.93) represents only phase rotations applied to the ω
and z coordinates. We can absorb these phases into the signal field and spin wave,
and therefore we can drop them from the kernel (we will be careful to ‘put them
back’ when we write down the optimal modes). We can now write down a very
simple recipe for constructing the optimal input mode in the Raman regime. First,
we form the kernel
√
K(z, ω) = CJ0 (2C ωz).
(5.95)
This depends only on the memory coupling C, therefore this parameter uniquely
determines the efficiency of the memory in the Raman limit [77] . We note that
the kernel K is real, and symmetric under the exchange of ω and z. That is, K
is Hermitian (see §A.4.3 in Appendix A). Therefore, its SVD is the same as its
spectral decomposition (see §3.1.3 in Chapter 3). The singular values of K are also
its eigenvalues, and the input modes {φj }, as functions of ω, have the same form as
5.3 Raman Storage
161
the spin wave modes {ψj }, as functions of z. Their phases differ though, because of
the phase rotations we ‘absorbed’. To be precise, let us define the functions {ϕj } as
the eigenfunctions of the kernel (5.95). That is to say, ϕj satisfies
Z
1
√
CJ0 (2C xy)ϕj (x) dx = λj ϕj (y).
(5.96)
0
The optimal input mode for the signal field, including the correct phase rotation
and transforming back from ω to τ , is given by
1
φ1 (τ ) = √ Ω(τ ) × exp
W
iW [1 − ω(τ )]
∆
× ϕ1 [1 − ω(τ )].
(5.97)
The optimal output mode for the spin wave, to which this optimal input mode is
mapped by the storage process, is given by
ψ1 (z) = e−idz/∆ ϕ1 (z).
(5.98)
The optimal storage efficiency is given by the square of the largest eigenvalue in
(5.96); η1 = λ21 . Figure 5.5 shows the variation of this optimal efficiency with C.
In taking the Raman limit, we have neglected spontaneous emission. This means
that the predicted efficiency can approach unity, as long as C is large enough, even
if the optical depth is low. That is, a smaller d can be ‘compensated’ by a larger
W — a more energetic control. Of course the optimal efficiency derived in §5.2.3
remains correct, even in the Raman limit: the best achievable efficiency is always
5.3 Raman Storage
162
limited, through spontaneous emission, by the optical depth. If the efficiency predicted by (5.96) is larger than the upper limit (5.25), then we have reached a regime
where spontaneous emission dominates over other losses. However, generally a Raman memory requires a large optical depth to operate efficiently, as we discuss
below. Therefore the dominant loss mechanism in a Raman memory is not spontaneous emission, but insufficient coupling. That is, the large detuning from resonance
makes the interaction weak, so that the biggest problem in getting a Raman memory
to work is to make the coupling strong enough that the signal field is completely
absorbed. The utility of the kernel (5.95), is that it provides a simple way to analyze
the Raman limit before spontaneous emission becomes a limitation. Examples of
the optimal modes predicted by the Raman kernel are shown in Figures 5.8 and
5.10 in §5.4 at the end of this chapter. What is notable about the behaviour of the
(a)
(b)
1
10
0.8
10
0.6
10
0.4
10
0.2
10
0
0 0.5 1 1.5 2 2.5 3 3.5 4
0
−1
−2
−3
−4
−5
10
0 0.5 1 1.5 2 2.5 3 3.5 4
Figure 5.5 Raman efficiency. (a) the optimal storage efficiency
η1 = λ21 predicted by the kernel √
(5.95) in the Raman limit ∆ 1,
versus the memory coupling C = W d/|Γ|. The efficiency ‘saturates’
at around C ∼ 2. (b) A logarithmic plot of the difference 1 − η1
between the predicted efficiency and unity.
Raman efficiency is that it rises steeply for small values of C, before ‘saturating’
5.3 Raman Storage
163
at C ∼ 2. Physically, this saturation point coincides with the stimulated scattering
regime. For ordinary Stokes scattering, the scattering process becomes efficient as
the coupling is increased beyond this point (see Figure 10.6 in Chapter 10). In the
case of a Raman quantum memory, the transmission of the signal field through the
ensemble drops sharply, and the efficiency of the memory becomes limited by spontaneous emission, rather than insufficient coupling. Therefore in designing a Raman
memory, one need only ensure that C & 2, in order that the scheme is viable (see
§9.8 in Chapter 9 and §10.9 in Chapter 10).
5.3.1
Validity
As mentioned above, in deriving the Raman kernel we neglected spontaneous emission. We did this tacitly when we dropped the real part of Γ in the exponential
factor appearing in the storage kernel. This is valid so long as neither W nor d is
too large. To see this, we define the balance R according to the relation
r
R=
W
.
d
(5.99)
Note that the balance has no relation to the matrix R introduced in §5.2.7 above.
Using the arguments employed in §5.2.8, we see that the balance really expresses
the extent to which the interaction is dominated by light, or matter:
R2 ∼
Nc
.
Na
(5.100)
5.3 Raman Storage
164
That is, the case R 1 corresponds to a light-biased interaction — as described earlier, this is the limit in which the adiabatic kernel saturates the upper bound on the
storage efficiency — and the case R 1 describes a matter-biased interaction, with
a weak control, but a large/dense ensemble. Using the balance, and the definition
C=
√
W d/|Γ|, we can re-express the adiabatic storage kernel (5.77) as follows,
√
K(z, ω) = Ce−iθ × e−iC cos θ(Rω+z/R) × eC sin θ(Rω+z/R) × J0 (2Ce−iθ ωz), (5.101)
where we have defined the phase angle θ such that tan θ = 1/∆; the complex detuning being given by
Γ = −i|Γ|eiθ .
(5.102)
The Raman limit is the limit of large detuning, which corresponds to the limit θ 1.
The exponential factor involving cos θ in (5.101) corresponds to the phase rotations
we dealt with previously. On the other hand, the exponential factor involving sin θ
comes from the real part of Γ, and represents spontaneous emission. It cannot
be removed by a unitary transformation of the modes, and it reduces the storage
efficiency. This is the term we neglect in the Raman limit. For this approximation to
hold for both the optical and spin wave modes together, both terms in the exponent
should be small. This is true provided we satisfy the conditions
C sin θR 1,
and
C sin θ ×
1
1.
R
(5.103)
5.3 Raman Storage
165
Using sin θ ≈ tan θ, setting C ∼ 1 (as we should have for a reasonably efficient
memory), and using (5.100), we find that the Raman kernel is a good approximation
whenever
∆ γ,
and
γ2
Nc
∆2
,
∆2
Na
γ2
(5.104)
using ordinary units. That is to say, the interaction should be roughly ‘balanced’,
with broadly equal contributions to the coupling originating from atoms and light.
The larger the detuning, the more ‘leeway’ there is to bias the interaction one way
or the other.
Suppose that we set R ∼ 1, so that d ∼ W . For reasonable efficiency, we should
have C ∼ 1, or thereabouts. Squaring this, and using R ∼ 1 then, we find d ∼ ∆. In
a Raman memory, ∆ 1, and therefore d 1. That is, a Raman memory described
by (5.95) generally requires a large optical depth. This explains why spontaneous
emission is not as important a limiting factor as is the issue of sufficiently strong
coupling.
Finally, we comment that since ∆ ∼ d, the adiabatic condition on the bandwidth
of the control field can be written as δc d, or in ordinary units δc dγ. This
is the same as the condition derived by Gorshkov et al. for the limitation on the
control bandwidth for resonant EIT storage. It is therefore arguable that a Raman
memory does not allow for more broadband storage than an EIT memory does.
But the adiabatic condition in the Raman case is rather more robust that it is in
the EIT case, since it does not depend on identifying an ‘important’ region in kspace. Numerics show that the adiabatic storage kernel is a better approximation
5.3 Raman Storage
166
in the Raman case, for more broadband control pulses, than it is in the EIT case.
Broadband storage provided the motivation for studying the Raman memory, but
its advantages in this respect are not clear cut.
Other considerations that may favour detuning from resonance include the possibility of dealing with a more complex excited state manifold. Suppose that, instead
of a single excited state |2i, there are a host of states, perhaps resulting from spinorbit or hyperfine splitting (see, for example, Figure 10.3 in Chapter 10). Tuning
into resonance with one of these states could make the dynamics rather complicated.
There may be some direct absorption of the signal field into the nearby states, followed by spontaneous emission and loss. Detuning away from all of the states puts
the contribution from each state on an equal footing, so that the dynamics can be
treated just as we did the simple three level system, where we need only swap the
coupling C to a single state for an equivalent coupling that includes the scattering
amplitudes for all the states (see §F.4 in Appendix F). And by detuning we eliminate
the possibility of absorption losses. In addition, a Raman memory is tunable, since
if we tune further from resonance, we can maintain strong coupling by increasing
the control pulse energy. The Raman memory is also affected less adversely by inhomogeneous broadening of the excited state than an EIT memory might be. Again,
this is because the coupling to the ensemble is not dominated by resonance with a
single frequency.
5.3 Raman Storage
5.3.2
167
Matter Biased Limit
We have already shown how the anti-normally ordered product formed from the
adiabatic storage kernel tends to the optimal kernel (5.27) in the limit of large
control pulse energy. This is the light-biased limit, with R 1. To reach this
limit, the balance R should exceed the upper limit in (5.104), so we should have
R > ∆ (in normalized units). Now, there is a degree of symmetry to the structure
of (5.101): The kernel is unchanged when we swap z and ω, if at the same time we
send R −→ 1/R. It therefore follows that in the matter-biased limit R 1 — that
is Na > ∆Nc /γ in ordinary units — the normally ordered kernel tends to a limit
defined not by the optical depth, but by the control pulse energy:
Z
lim
d→∞ 0
1
K ∗ (z, ω)K(z, ω 0 ) dz =
√
W −W (ω+ω0 )/2
e
I0 (W ωω 0 ).
2
(5.105)
The kernel on the right hand side has precisely the same form as (5.27), except that
the efficiency is limited by W rather than by d. Therefore, if we are limited by the
energy of the control, the optimal efficiency achievable is given by η1 = 1 − 2.9/W .
Of course, extremely high energy lasers are readily available, whereas the size of the
ensemble is generally not easily varied. Nonetheless, in cases where the ensemble
may be damaged by a high energy laser pulse — as may be the case for a solid state
memory — it may be that the control energy becomes a limitation.
5.3 Raman Storage
5.3.3
168
Transmitted Modes.
In this section we describe a connection between the modes that are stored in the
memory, and the modes that are transmitted through it when the efficiency is not
perfect. This connection holds in the Raman limit, and it is clearest when we
form the equations of motion for the signal field and the spin wave in the adiabatic
approximation. This is the way the adiabatic approximation is most commonly
introduced, and so it is informative to run through the procedure. We start with
the equations of motion (5.3), which we reproduce below in normalized units,
√
∂z A(z, τ ) = − dP (z, τ ),
∂τ P (z, τ ) = −ΓP (z, τ ) +
√
dA(z, τ ) − iΩ(τ )B(z, τ ),
∂τ B(z, τ ) = −iΩ∗ (τ )P (z, τ ).
(5.106)
The adiabatic approximation is made by setting ∂τ P = 0 on the left hand side of
the second equation. This is reasonable when the natural dynamics of the optical
polarization P are overwhelmed by the motion driven by the signal and control fields
— known as adiabatic following. In this situation, 1/|Γ| is the shortest timescale
in the problem, so that when far-detuned we must have (δc , Ωmax ) |∆|, precisely
the adiabatic conditions (5.91), (5.92) derived above. We can then solve the second equation for P algebraically, and substitute the result into the first and third
5.3 Raman Storage
169
equations. The result is
√
Ω d
A = i
B,
Γ
√
|Ω|2
Ω∗ d
∂τ +
B = −i
A.
Γ
Γ
d
∂z +
Γ
(5.107)
We then switch coordinates from (τ, z) to (ω, z), and we define new variables α and
β for the signal field and spin wave as follows
α(z, ω)e−(W ω+dz)/Γ =
√
W
A(z, τ )
,
Ω(τ )
β(z, ω)e−(W ω+dz)/Γ = B(z, τ ).
(5.108)
These new variables incorporate the Jacobian factor associated with the coordinate
transformation, and also the phase rotations associated with the dynamic Stark shift
and the ensemble refractive index. In general, the transformation linking A and B to
α and β is not quite unitary, because Γ is not strictly imaginary, so the exponential
factors on the left hand side of (5.108) make the norms of {A, B} different to the
norms of {α, β} — they are not pure phase rotations. However, taking the Raman
limit ∆ 1, we find Γ ≈ −i∆, and in this case, the transformation is unitary. This
will be important shortly. For now, observe that the equations of motion simplify
greatly when they are cast in terms of the transformed variables {α, β}. We have
removed the control field, and also the homogeneous terms on the left hand side, so
5.3 Raman Storage
170
that we obtain the system
∂z α = −Cβ,
∂ω β = Cα.
(5.109)
Note the symmetry of this system of equations. If we swap α for β, and β for −α,
and then swap z and ω, the system of equations is unchanged. As we will see below,
this symmetry simplifies the form of the solution, so that there are only two different
Green’s functions, instead of a potential four.
We solve these equations by applying a unilateral Fourier transform. Having
eliminated the control field, we can choose either to apply the transform over the ω
coordinate or the z coordinate. We will apply the transform over the z coordinate,
as we have done previously, and solve the equations in k-space. Using tildes to
denote the transformed variables, we have
α0
−ik α
e− √
2π
e
= −C β,
∂ω βe = C α
e,
(5.110)
where α0 = α(z = 0, ω) represents the boundary condition for the signal field.
Solving the first equation for α
e yields
α
e = −i
Ce
i
β+√
α0 .
k
2πk
(5.111)
5.3 Raman Storage
171
Substituting this result into the second equation, and integrating, we obtain the
solution for β,
e ω) = e−iC 2 ω/k βe0 + i √C
β(k,
2πk
Z
ω
e−iC
2 (ω−ω 0 )/k
α0 (ω 0 ) dω 0 ,
(5.112)
0
where βe0 is the Fourier transform of the spin wave boundary condition β0 = β(z, ω =
0). Finally, substituting this back into (5.111) gives the solution for α,
C
i
C2
2
α
e(k, ω) = −i e−iC ω/k βe0 + √
α0 (ω) + √
k
2πk
2πk 2
Z
ω
e−iC
2 (ω−ω 0 )/k
α0 (ω 0 ) dω 0 .
0
(5.113)
These k-space solutions are singular at k = 0, but fortunately it is possible to take
the inverse Fourier transform analytically. We use the results described in §D.5 in
Appendix D, along with the convolution theorem for the unilateral Fourier transform
(see §D.4.2 in Appendix D), to obtain the rather formidable-looking solution
Z
ω
r
h p
i
z
0 ) α (ω 0 ) dω 0
J
2C
z(ω
−
ω
α(z, ω) = α0 (ω) − C
1
0
ω − ω0
0
Z z h p
i
−C
J0 2C (z − z 0 )ω β0 (z 0 ) dz 0 ,
0
Z zr
i
h p
ω
0 )ω β (z 0 ) dz 0
β(z, ω) = β0 (z) − C
J
2C
(z
−
z
1
0
z − z0
0
Z ω h p
i
J0 2C z(ω − ω 0 ) α0 (ω 0 ) dω 0 .
+C
(5.114)
0
We re-write this in terms of Green’s functions, or propagators, in order to bring out
5.3 Raman Storage
172
its structure.
Z
1
0
0
0
L(ω, ω )αin (ω ) dω −
0
Z
Z 1
L(z, z 0 )βin (z 0 ) dz 0 +
βout (z) =
αout (ω) =
0
1
Z
K(ω, z)βin (z) dz,
0
1
K(z, ω)αin (ω) dω,
(5.115)
0
where αout (ω) = α(z = 1, ω) is the transmitted signal field, and βout = β(z, ω = 1)
is the spin wave at the end of the interaction. We have introduced the Green’s
functions L and K, defined as follows,
L(x, y) = δ(x − y) − CΘ(x − y) × √
h p
i
K(x, y) = CJ0 2C x(1 − y) .
√
1
J1 (2C x − y),
x−y
(5.116)
Note that only two distinct Green’s functions are required, because of the symmetry
between α and β in the adiabatic equations of motion (5.109). The Green’s function
K(x, 1−y) is precisely the adiabatic Raman storage kernel (5.95), and this is why we
have used the same notation. The kernel L describes the relation of the transmitted
signal field to the input field, or equivalently it relates the final to the initial spin
wave. We are considering storage, so that βin = 0, but note that the same solutions
can be used to describe retrieval from the memory; retrieval is dicussed in Chapter
6. The Heaviside step function Θ(x) in L makes the kernel causal, so that the
transmitted signal field is never influenced by future values of the input field.
We now show that there is a connection between the SVD of K, which tells
us about the optimal input mode and its associated efficiency, and the SVD of L,
5.3 Raman Storage
173
which tells us about the transmitted fields. Note that it is only correct to associate
the singular values of the Green’s function K with storage efficiencies when the
transformation connecting {A, B} to {α, β} is unitary, and this is only true in the
Raman limit ∆ 1. Therefore the following analysis applies only in this limit, and
we will assume that we are tuned far from resonance in the remainder of this section.
To see the connection between the SVDs of K and L, it will help us to use matrix
notation, since it is much more compact. As discussed in Chapter 3, the solutions
5.115 may be considered as the infinite-dimensional limit of the following matrix
equations,
|αout i = L|αin i − K|βin i,
|βout i = L|βin i + K|αin i.
(5.117)
To be more precise, we define the vector |αout i as a discretized version of the continuous function αout ,

 αout (0)


 α ( 1 )
 out N −1


|αout i = 
 αout ( N 2−1 )


..

.



αout (1)








,







(5.118)
where N is the number of discretization points. The other vectors are defined
5.3 Raman Storage
174
similarly. The matrices K and L are given in terms of the continuous kernels by
Kjk = K
j−1 k−1
,
N −1 N −1
1
×
,
N −1
Ljk = L
j−1 k−1
,
N −1 N −1
×
1
.
N −1
(5.119)
The continuous equations are recovered by taking the limit N −→ ∞, when matrix
multiplication is replaced by integration, and the factors 1/(N − 1) become the
increments dz, dω. It is useful to be able to swap between the continuous and
discretized representations, in order to use both the machinery of calculus and linear
algebra in their ‘natural habitats’.
The relationship between the SVDs of the matrices K and L is fixed by the
equations of motion (5.109), which are unitary. To see this, observe that these
equations imply the identity
∂z (α† α) + ∂ω (β † β) = 0.
(5.120)
We have used the † notation for Hermitian conjugation, rather than the ∗ notation
for complex conjugation, since this equation holds whether or not we treat α, β as
ordinary functions, or as quantum mechanical annihilation operators3 . Integrating
this identity over all z and ω gives the condition
Nα,out + Nβ,out = Nα,in + Nβ,in ,
3
(5.121)
Strictly, the Langevin noise operators introduced in §4.12 of Chapter 4 contribute unless an
expectation value is taken, but in the Raman limit we neglect spontaneous emission, and these
operators along with it.
5.3 Raman Storage
175
where Nα,out , Nα,in are the numbers of transmitted and incident signal photons,
respectively, and where Nβ,out , Nβ,in are the numbers of final and initial spin wave
excitations:
1
Z
Nα,out =
0
Z
Nβ,out =
0
†
αout
(ω)αout (ω) dω,
1
†
βout
(z)βout (z) dz,
Z
1
Nα,in =
Z01
Nβ,in =
0
†
αin
(ω)αin (ω) dω,
†
βin
(z)βin (z) dz. (5.122)
The condition (5.121) fixes the combined transformation of signal field and spin
wave, implemented by the memory interaction, as unitary, meaning that the total
number of ‘particles’ is conserved. This unitarity holds in the Raman limit approximately, since we have neglected spontaneous emission, which process would scatter
particles into modes other than the signal field or the spin wave, thus violating the
conservation law (5.121). As discussed in §5.3.1, this approximation is generally a
good one for a Raman memory, since large optical depths are required, making spontaneous emission losses negligible. The conservation condition (5.121) has the same
form as that for a beamsplitter, which mixes a pair of input modes without loss, to
produce a pair of output modes. Indeed, this is a helpful perspective from which to
view the action of the Raman memory (see Figure 5.6 below). The difference with
a conventional beamsplitter, as we will see, is that the Raman interaction couples
multiple modes together in a pairwise fashion, with each pair of modes ‘seeing’ a
different reflectivity. We derive this fact by combining the solutions for α and β into
5.3 Raman Storage
176
a single transformation,
|xout i = U |xin i,
(5.123)
where the vectors |xout i and |xin i are defined by










|αout i












|xout i = 
,










|βout i

















|αin i












|xin i = 
,










|βin i







(5.124)
and where the matrix U is given by


 L −K 
.
U =


K L
(5.125)
In this notation, the conservation condition (5.121) is written as
hxout |xout i = hxin |xin i.
(5.126)
Substituting the transformation (5.123) into the above condition, we find that U † U =
I, where I is the identity matrix. Using this result, we multiply (5.123) by U † from
5.3 Raman Storage
177
the left, to obtain the inverse transformation
|xin i = U † |xout i.
(5.127)
Substituting this into the conservation condition, we then find U U † = I. So we see
that U is indeed a unitary transformation (see §A.4.5 in Appendix A). To see the
implications of this for the matrices K and L, we substitute the form (5.125) for U
into the conditions U † U = U U † = I, and perform the matrix multiplications. This
yields the conditions
L† L + K † K = LL† + KK † = I.
(5.128)
In terms of the normally and antinormally ordered products, these conditions are
written as
LN + KN = I;
LA + KA = I.
(5.129)
Therefore we must have that LN and KN commute:
[LN , KN ] = [LN , I − LN ] = [LN , I] − [LN , LN ] = 0.
(5.130)
And similarly for the antinormally ordered products, [LA , KA ] = 0. As discussed
in §3.1.2 of Chapter 3, the eigenvectors of KA are the output modes of K, and
the eigenvectors of KN are its input modes. That is, if we write the SVD of K as
†
2 U † , and K
2
K = UK DK VK† , then KA = UK DK
N = VK DK VK . The fact that KA
K
commutes with LA implies that the eigenvectors of LA are the same as those of KA
5.3 Raman Storage
178
(see §A.3.1 in A). So we can write LA as
2 †
LA = UK DL
UK ,
(5.131)
2 is a diagonal matrix of positive eigenvalues. The same argument applies
where DL
for the normally ordered products, so that we have
2 †
LN = VK DL
VK .
(5.132)
Note that the eigenvalues of LN are necessarily the same as those of LA . To summarize, the unitarity of the memory interaction constrains the matrices K and L
such that their SVDs are built from a common set of input and output modes:
L = UK DL VK†
K = UK DK VK† .
(5.133)
And substituting these expressions into the conditions 5.129, we find the following
relationship between the singular values,
2
2
DL
+ DK
= I.
(5.134)
Our analysis is nearly complete. Let us review what we have discovered. A general
transformation from |xin i to |xout i could involve four Green’s functions — that is,
there could have been four matrices to deal with, one for each of the elements of
U in (5.125). But the symmetry of the equations of motion (5.109) allows us to
5.3 Raman Storage
179
express the transformation in terms of just two Green’s functions, L and K. Each
of these has an SVD, so there are potentially four sets of orthogonal modes to deal
with, associated with the input and output modes of the two Green’s functions L
and K. The further condition of unitarity on U , however, allows us to reduce the
number of orthogonal sets from four down to two, since L and K must share the
same input modes, and also the same output modes, as one another. We now show
that in fact the output modes are just flipped versions of the input modes, so that
the entire interaction can be fully described using just a single set of orthogonal
modes. The final step is accomplished by noticing that the matrices L and K are
persymmetric. That is, they are symmetric under reflection in their anti-diagonal
(see §3.1.4 in Chapter 3). This can be seen by examining the functional forms
(5.116) of the Green’s functions. For example, the value of L(x, y) only depends on
the difference x − y. Therefore the contours of L all lie parallel to the line y = x,
which corresponds to the main diagonal of the matrix L. L is accordingly unchanged
by a reflection in the main anti-diagonal, that is to say, L is persymmetric. That
K is also persymmetric follows from the fact that K(x, 1 − y) is Hermitian. As
shown in §3.1.4 in Chapter 3, the input and output modes associated with the SVD
of a persymmetric matrix are simply ‘flipped’ versions of one another. Putting this
last result together with our previous analysis, we can express the Green’s functions
5.3 Raman Storage
180
entirely in terms of a single set of modefunctions {ϕj },
K(x, y) =
X
ϕj (x)λj ϕj (1 − y),
j
L(x, y) =
X
ϕj (x)µj ϕj (1 − y),
(5.135)
j
where the singular values add in quadrature to unity,
µ2j + λ2j = 1.
(5.136)
The modes ϕj can be found from a numerical SVD of the kernel K. Or, equivalently,
they are given by the eigenvalue equation (5.96) introduced previously in §5.3.
The procedure used to connect the SVDs of K and L through the unitarity of
U is known as the Bloch-Messiah Reduction [147–149] . The resulting decomposition
makes the assertion that Raman storage may be understood by analogy with a
beamsplitter into a rigorous correspondence. If we define aj , (bj ) as the annihilation
operator for a photon (spin wave excitation) in the j th input mode ϕj , and if Aj
(Bj ) annihilates a photon (spin wave excitation) in the j th output mode, then for
each mode, the memory interaction can be written as
Aj
= µj aj − λj bj ,
Bj
= µj bj + λj aj .
(5.137)
These relations are precisely those arising in the quantum mechanical description of
5.3 Raman Storage
181
an optical beamsplitter, coupling input modes aj , bj , to output modes Aj , Bj , with
reflectivity R = ηj = λ2j (see Figure 5.6). Optimal storage corresponds to the case
R = 1, so that the incident signal field is entirely ‘reflected’ into a spin wave mode.
Figure 5.6 Raman storage as a beamsplitter. Optical and spin
wave modes are mixed pairwise by the storage interaction, as light
beams are on a beamsplitter. The ideal quantum memory, with unit
storage efficiency, would see the beamsplitter replaced by a perfect
mirror.
A possible use of a Raman quantum memory with non-unit storage efficiency
is as part of the modified DLCZ quantum repeater protocol described in §1.7 in
Chapter 1. Instead of using a 50 : 50 beamsplitter in combination with an ideal
quantum memory to generate number state entanglement, a single Raman quantum
memory with η1 = 50% can be used, as shown in Figure 5.7. The quality of the
entanglement generated relies on good overlap of the transmitted field modes on
the final beamsplitter, so the preceding theoretical characterization of the temporal
structure of these modes simplifies the analysis of this type of protocol.
5.4 Numerical Solution
182
D1
QML
BS
SL
SR
D2
QMR
Figure 5.7 Modified DLCZ protocol with partial storage. When
a single detector fires behind the final beamsplitter, number state
entanglement is generated between the memories. The protocol is
explained in §1.7 in Chapter 1. The only difference is that here,
a single memory with 50% efficiency replaces the combination of a
beamsplitter and an ideal memory.
5.4
Numerical Solution
So far we have succeeded in deriving a form for the storage kernel in the adiabatic
limit, or for arbitrary bandwidths in the case of a hyperbolic secant control. The
general problem of finding the storage kernel for both arbitrary bandwidths and
arbitrary control pulse profiles has not been solved analytically. But it is possible
to construct the kernel numerically. This can be done by integrating the system
of coupled equations (5.106) multiple times, each time with a different boundary
condition. Provided the boundary conditions form a set of orthogonal functions, the
Green’s function can be reconstructed. The easiest set of boundary conditions to
implement is the set of ‘impulses’ — delta functions. To see why this works, recall
5.4 Numerical Solution
183
the definition of the storage kernel,
Z
∞
Bout (z) =
K(z, τ )Ain (τ ) dτ,
(5.138)
−∞
where we used normalized units for z. Now, if we insert a delta function δ(τ − τj )
as the signal field boundary condition Ain (τ ), where τj is some particular time slot,
the resulting spin wave Bout,j is
Bout,j (z) = K(z, τj ).
(5.139)
We can therefore reconstruct an approximation to the entire Green’s function by
numerically solving for Bout,j repeatedly, with the times τj chosen from a finite
grid. The grid should range over a sufficient range of times that all of the Green’s
function is sampled, and we should make the grid sufficiently fine that no features
of the Green’s function are missed. So long as these requirements can be met while
keeping the computation reasonably fast, this is a convenient way to find the storage
kernel without requiring the adiabatic approximation, and without imposing any
restriction on the temporal profile of the control field. This method was previously
used to reconstruct the adiabatic Green’s functions describing stimulated Stokes
scattering in a dispersive ensemble by Wasilewski and Raymer [148] . In this thesis
we solve the equations of motion numerically using Chebyshev spectral collocation
for the spatial derivatives, and a second order Runge-Kutta (RK2) method for the
time stepping. The method of solution is explained in detail in Appendix E.
5.4 Numerical Solution
184
In the figures below we compare the optimal input modes predicted by the various
methods presented in this chapter. In Figures 5.8 and 5.9 we plot, side by side, the
optimal input modes found from numerically constructed Green’s functions, the
Rosen-Zener kernel (5.60), the adiabatic kernel (5.77) and the Raman kernel (5.95).
In Figure 5.8 the adiabatic approximation is well satisfied; the Raman approximation only poorly so, and therefore there is good agreement between all predictions,
save for the Raman prediction, which deviates slightly. But note that the adiabatic
and Raman kernels do slightly overestimate the phase due to the dynamic Stark-
1
Rosen Zener
Numerical
Adiabatic
Raman
6
4
0.5
2
0
−5
0
5
−5
0
−5
5
Time
0
5
−5
0
Figure 5.8 Comparison of predictions for the optimal input modes
in the adiabatic limit. Here we used a hyperbolic secant control, given
by (5.38) with Tc = 1 and Ω0 = Ωmax = 3, an optical depth of d = 300
and a detuning ∆ = 15, in normalized units. The control intensity
profile |Ω(τ )|2 is indicated by the dotted lines, scaled for clarity. The
blue lines show the predicted optimal intensity profiles |Ain (τ )|2 =
|φ1 (τ )|2 , and the red lines show the variation of the temporal phase
of the mode φ1 in radians, referred to the axes on the right hand
side. The storage efficiency is ∼ 90% in all cases, and the predictions
are in good agreement generally. But both the adiabatic and Raman
kernels overestimate the phase shift due to the dynamic Stark shift.
And the detuning is not quite large enough to render the Raman
kernel correct.
In Figure 5.9 the fields are tuned into resonance, the optical depth is reduced,
5
0
Phase
Intensity
(arb. units)
shift.
5.4 Numerical Solution
185
and the control intensity is increased, so that the adiabatic approximation is no
longer satisfied. There is rough agreement between the numerical and Rosen-Zener
predictions, which do not rely on the adiabatic approximation, although there is
a slight discrepancy that we attribute to the accumulation of numerical errors in
the evaluation of the Hypergeometric functions in the Rosen-Zener kernel (5.60).
The optimal modes predicted by these two methods exhibit oscillations, as can be
seen from the phase jumps indicating sign changes. These oscillations are missing
from the predictions of the adiabatic kernel; the prediction of the Raman kernel is
1
Rosen Zener
Numerical
Adiabatic
Raman
6
4
0.5
2
0
−5
0
5
−5
0
5
−5
Time
0
5
−5
0
Figure 5.9 Comparison of predictions for the optimal input modes
outside the adiabatic limit. Again, the control takes the form of a
hyperbolic secant, this time with Tc = 1 and Ω0 = Ωmax = 5. We
reduce the optical depth down to d = 10, and we tune into resonance,
putting ∆ = 0. The numerical and Rosen-Zener predictions roughly
coincide, and both the predictions exhibit oscillations characteristic of
non-adiabatic ‘ringing’. Numerical errors in the Rosen-Zener kernel
amplify the size of these oscillations slightly. The adiabatic kernel
does not correctly reproduce the oscillations, while the Raman kernel
is totally inappropriate for modelling this resonant interaction. The
optimized storage efficiency predicted by the numerics is around 82%.
The numerically constructed Green’s functions yield the most reliable predictions. In addition, they take less time to calculate then the Rosen-Zener predictions
5
0
Phase
Intensity
(arb. units)
catastrophically wrong, as we would expect on resonance.
5.4 Numerical Solution
186
— the Matlab code for the numerical method runs in ∼ 40 s on a 3 GHz machine;
the Rosen-Zener kernel takes several hours to construct, simply because the hypergeometric functions are so difficult to evaluate efficiently. Also, the numerical method
is more flexible, since arbitrary control profiles can be used. In general, then, the
direct numerical approach is the method of choice for optimizing a Λ-type quantum
memory.
In the adiabatic and Raman limits, however, it is faster to use the analytic
kernels derived using these approximations. And of course, the optimization can
then be trivially generalized to arbitrary control pulse shapes using the transformations (5.78) and (5.97). In Figures 5.10 and 5.11 below we compare the numerical
predictions with those of the adiabatic and Raman kernels in the resonant and
Raman limits, using much more broadband control pulses. As expected, there is
good agreement between all three predictions in the Raman limit. In the resonant
case, describing EIT storage, the Raman kernel of course fails completely, but the
adiabatic kernel remains reasonably reliable. That said, its agreement with the numerical prediction is not as good as in the Raman limit, which is symptomatic of
the fragility of the adiabatic approximation on resonance.
When far into the Raman limit, the numerical, adiabatic and Raman methods
all predict a significant temporal phase variation due to the dynamic Stark shift.
Whereas in the EIT case, the optimal input mode has a flat phase. Depending on
the flexibility of the technology used to shape the optical pulses before storage, it
may be easier to implement EIT storage for this reason. On the other hand, Raman
5.4 Numerical Solution
Numerical
1
Raman
Adiabatic
6
4
0.5
2
0
−0.2
0
0.2
−0.2
0
Time
0.2
−0.2
0
0.2
Phase
Intensity
(arb. units)
187
0
Figure 5.10 Broadband Raman storage. Here the control is a
−(τ /Tc )2
Gaussian pulse Ω(τ ) = Ωmax ep
, where Tc = 0.1 and W =
302.5, so that Ωmax = (2/π)1/4 W/Tc = 49.1. The optical depth is
d = 300, and the detuning is ∆ = 150. As usual all these quantities
are in normalized units. These parameters give a Raman memory
coupling of C = 2.0 and a balanced interaction with R = 1.0. Note
that the control pulse duration is roughly one tenth of the spontaneous emission lifetime of the excited state |2i. Nonetheless, due to
the large detuning, the adiabatic and Raman approximations are well
satisfied, and the agreement between the numerical and analytic predictions is clear. The optimized storage efficiency is predicted by all
the methods shown to be ∼ 98%.
storage allows for more freedom in the carrier frequencies of the signal and control
fields.
5.4.1
Dispersion
All of the optimal modes shown in the figures above are slightly delayed in time
with respect to the control pulse. This is due to the characteristic dispersion associated with an absorption process, which produces a superluminal group velocity. Of
course, there is no question of violating causality, it is simply that as the trailing
edge of the signal field is absorbed, the ‘centre of mass’ of the signal pulse advances,
giving the appearance of superluminal propagation. The efficiency of the memory
5.5 Summary
Numerical
Adiabatic
Raman
6
4
0.5
2
0
−0.2
0
0.2
−0.2
0
Time
0.2
−0.2
0
0.2
Phase
Intensity
(arb. units)
1
188
0
Figure 5.11 Broadband EIT storage. All parameters are the same
as above in Figure 5.10, except that the storage is performed on
resonance, with ∆ = 0. The Raman kernel fails utterly, but the
adiabatic prediction fairs better, comparing well with the numerical
prediction. That the agreement is poorer on resonance is a signature
that the adiabatic approximation is less robust on resonance. The
optimized storage efficiency predicted by the numerical method is
∼ 99%.
is maximized by ‘pre-compensating’ for this effect, and this explains the time shift
common to the optimal input modes. In the case of resonant storage, this view conflicts with our characterization of EIT in §2.3.1 of Chapter 1 as working by slowing
the group velocity of the signal. Perhaps a better perspective is therefore that on
resonance the control must precede the signal in order to ‘prepare’ the transparency
window.
5.5
Summary
We have covered rather a lot of material in this chapter. Here we review the main
results.
1. The analysis of the storage process is greatly simplified if we use a one di-
5.5 Summary
189
mensional propagation model. All the results listed below make use of this
model.
2. The best possible storage efficiency is limited by the optical depth, being given
by η = 1 − 2.9/d.
3. The method of Rosen and Zener yields an analytic expression for the storage
kernel for a hyperbolic secant control. It is, however, difficult to evaluate this
efficiently and accurately (at least using Matlab).
4. In the adiabatic approximation, an analytic expression for the storage kernel
can be derived that holds for all control pulse profiles. This is quick to evaluate.
5. When far detuned, the expression further simplifies, yielding the Raman kernel,
which only depends on the Raman memory coupling C. The Raman memory
can be decomposed as a set of beamsplitter transformations between light and
matter, using just a single set of modefunctions.
6. If none of the above methods are appropriate, the storage kernel can be directly
constructed by repeated numerical integration of the equations of motion for
the memory. This method yields the correct optimal input mode for arbitrary
control profiles and detunings, for all values of the optical depth.
In the next chapter, we consider retrieval from a Λ-type memory.
Chapter 6
Retrieval
So far we have considered the optimization of storage in a Λ-type quantum memory. In some circumstances this optimization is sufficient to maximize the combined
efficiency of storage into, followed by retrieval from the memory. Specifically, this is
true when the retrieval process is the time reverse of the storage process [133] . When
this is not the case, the optimization of this combined efficiency is a distinct problem. In this Chapter, we discuss various strategies for retrieval, and we present the
results of a numerical analysis of their effectiveness.
6.1
Collinear Retrieval
In order to convert the stored excitation back into a propagating optical signal, we
send a second control pulse — a read out pulse — into the ensemble. Since the
control mediates coupling between the signal field and the spin wave, the rationale
is that the spin wave excitations will transfer back to the optical field, emerging
6.1 Collinear Retrieval
191
as a collimated pulse traveling in a well-defined direction. We have met with some
success in analyzing the storage process using a one dimensional propagation model,
and so it is natural to consider read out from this perspective. We will see later
that certain advantages accrue if a small angle is introduced between the signal
and control fields. For the moment, suppose that we have stored a signal photon
collinearly, as described in the previous chapter. Confining ourselves to the same
one dimensional model, there are two possible read out geometries: forward, and
backward retrieval.
6.1.1
Forward Retrieval
We can describe forward retrieval using the same equations of motion as we used
for storage — (5.3), or (5.106) in normalized units. For the retrieval process, the
signal field boundary condition is set to zero, and the spin wave boundary condition
is set by the coherence generated in the medium during the storage process. Let us
denote quantities associated with the retrieval process with a superscript r. In this
notation the boundary conditions at read out are
Arin (τ r ) = 0,
r
Bin
(z) = Bout (z).
(6.1)
The second condition assumes that there is no loss of coherence during the time that
the excitations are stored. This kind of loss is homogeneous in space, and so it would
r . The optimization
appear only as a constant factor reducing the amplitude of Bin
of the memory is therefore unaffected by the neglect of decoherence.
6.1 Collinear Retrieval
192
The first step in analyzing the efficiency of the retrieval process is to examine
the form of the map from the stored excitations to the retrieved signal field,
Arout (τ r ) =
Z
1
r
K r (τ r , z)Bin
(z) dz.
(6.2)
0
In fact, in the adiabatic limit, the retrieval kernel K r is identical in form to the
storage kernel K, when expressed in terms of the integrated Rabi frequency ω r ,
rather than the time τ r . This is guaranteed by the symmetrical form of the equations
of motion in the adiabatic limit, which is discussed in §5.3.3 in Chapter 5. For
completeness, we verify this by explicit calculation. In the notation of §5.2.5 in
Chapter 5, the retrieved signal field is expressed in k-space as
√
d
r
r
r
e
ein
A (k, τ ) = −i √
h↑ |V r−1 (k, τ r )Vinr (k)| ↓iB
(k).
2πk
(6.3)
Using Vinr = I, along with the adiabatic approximations (5.70) and (5.73), we obtain
√
Z τr
1
d r r
1
r 0 2
0
A (k, τ ) = − √
Ω
(τ
)
exp
−
|Ω
(τ
)|
dτ
Γr −∞
2π Γr
" id R τ r
#
r (τ 0 )|2 dτ 0
|Ω
1
r2
e r (k).
×
exp Γ −∞
B
in
k + i Γdr
k + i Γdr
r
r
(6.4)
Here Ωr is the Rabi frequency describing the temporal profile of the read out control
pulse, which may differ from that of the storage pulse, and Γr = γ − i∆r allows for
the possibility that the detuning is changed for the readout process. We transform
the time coordinate from τ r to ω r , and take the inverse Fourier transform from k
6.1 Collinear Retrieval
193
to z-space. Using the result (D.40), along with the shift and convolution theorems
(D.18), (D.26) in Appendix D, we find that the retrieval kernel is given by
Ωr∗ (τ r )
K r (τ r , z) = √
× K r [ω r (τ r ), z] ,
r
W
(6.5)
where
√
i
dW r −[W r ωr +d(1−z)]/Γr h √dW r p r
K (ω , z) = i
e
J0 2i Γr
ω (1 − z) .
Γr
r
r
(6.6)
This kernel has precisely the same form as (5.76), except that ω has been switched
for z, and z has been switched for ω r . Now we can write down the adiabatic map
describing the combined processes of storage followed by forward retrieval,
Arout (τ r ) =
Z
∞
Ktotal (τ r , τ )Ain (τ ) dτ.
(6.7)
−∞
The kernel Ktotal describes the entire memory interaction (see §3.4 in Chapter 3),
and is given by
Ktotal (τ r , τ ) =
Ωr∗ (τ r )Ω(τ )
√
× Ktotal [ω r (τ r ), ω(τ )] ,
r
W W
(6.8)
where
r
Z
Ktotal (ω , ω) =
1
K r (ω r , z)K(z, ω)dz.
(6.9)
0
Here the kernel K is the adiabatic storage kernel given in (5.76) in Chapter 5, and
K r , given in (6.6), is related to K by the symmetry connecting the storage and
6.1 Collinear Retrieval
194
retrieval kernels just derived above.
The input mode that optimizes the total memory efficiency is found from the
SVD of Ktotal in (6.9). The kernel looks superficially similar in form to the normally
ordered kernel KN formed from the storage kernel K. But Ktotal 6= KN because
K r 6= K ∗ . If these two kernels were equal to one another, then the optimal input
mode found from the SVD of Ktotal would be equal to the mode found from the
SVD of K; this follows from the properties of KN (see §3.3.1 in Chapter 3). Since
these two kernels are not equal to each other, the optimization of storage followed
by forward retrieval is different to the optimization of storage alone.
Note that the optimal input mode does not depend on the shape Ωr of the read
out control; it only depends on the total energy in the read out pulse, parameterized
by W r . But changing the shape of Ωr changes the temporal profile of the signal field
retrieved from the memory. This is useful: one can optimize the memory efficiency
once, for a fixed energy W r , and then changing the shape of Ωr allows one to produce
an output signal pulse of any shape, within the limits of the adiabatic approximation.
Novikova et al. have already demonstrated optimal storage, followed by shaped
retrieval, experimentally, on resonance [74,150,151] , using the theory of Gorshkov et
al. [133] , which applies in the light-biased limit (see the end of §5.2.8 in Chapter 5).
In Figure 6.1, we show an example of how the optimal input mode for storage,
followed by retrieval, differs from the optimal storage mode. We used the adiabatic
kernel (5.76) to model the storage interaction, and we evaluated the expression (6.9)
for Ktotal numerically. This is easily done by discretizing the coordinates so that K
6.1 Collinear Retrieval
195
becomes a matrix. We use the same control pulse and detuning for both storage
and readout, Ω = Ωr and ∆ = ∆r , and then Ktotal is given by the product of
two copies the matrix K. Taking the SVD of the result yields a radically different
optimal input mode to that found from the SVD of K alone. The difference can be
understood by considering the shape, in space, of the spin wave Bout (z) produced by
the storage interaction. When optimizing storage alone, this is given by the optimal
output mode ψ1 (z) found from the SVD of the storage kernel K. The shape of this
mode generally takes the form of a decaying exponential, as shown in part (b) of
Figure 6.1. This shape is consistent with Beer’s law absorption: as the signal pulse
propagates through the ensemble it is increasingly likely to be absorbed, so that at
the exit face, at z = 1, there is very little probability for the signal pulse to excite
an atom. Therefore the spin wave decays in magnitude steadily from the input to
the exit face. But forward retrieval from a spin wave of this shape is problematic:
if the amplitude of the spin wave is concentrated close to the entrance face of the
ensemble, a retrieved excitation, that has been converted into a signal photon, must
propagate a large distance through the ensemble before reaching the exit face. There
is therefore a high probability that the retrieved photon will be re-absorbed, so that
it never emerges from the ensemble, and this greatly reduces the efficiency of the
read out process. For forward retrieval, it is much better that the bulk of the spin
wave is concentrated towards the exit face of the ensemble, so that re-absorption
losses are minimized. In fact, the optimal spin wave mode for retrieval is precisely
the space-reverse of the mode generated by the storage process; this can be derived
6.1 Collinear Retrieval
196
from the symmetry relating the storage and retrieval kernels. It is the fact that the
optimal shape of the spin wave for retrieval is improperly matched to the spin wave
mode generated by optimal storage that complicates the combined optimization of
storage followed by retrieval. A compromise between these two shapes must be
found. Such a compromise is shown in part (d) of Figure 6.1: the amplitude of
the spin wave at z = 1 is much larger than for the spin wave that results from only
optimizing storage. And there is a marked suppression of the spin wave amplitude at
z = 0, since even though the storage interaction naturally excites atoms close to the
entrance face, it is extremely deleterious to the retrieval efficiency to concentrate the
stored excitations there. If we could run the storage process backwards, so that the
retrieval process was precisely the time-reverse of the storage process, we should be
able to achieve a retrieval efficiency equal to the storage efficiency, ηretrieve = ηstorage .
It should then be possible to achieve a combined efficiency of storage followed be
2
. However, even after optimizing
retrieval of ηcombined = ηstorage × ηretrieval = ηstorage
2
the combined efficiency, it still falls far short of this optimum, ηcombined ηstorage
.
2
In the example shown in Figure 6.1, ηcombined = 28%, whereas ηstorage
= 80%. The
tension between Beer’s law during storage and re-absorption during retrieval makes
forward retrieval generally inefficient. In the next section we consider backward
retrieval, which performs better.
197
Input mode
1
Spin wave
6
(a)
(b)
4
0.5
2
0
6
(c)
(d)
4
0.5
2
0
−2
0
2
0
0.5
1
Phase
Intensity
(arb. units)
0
1
Phase
Intensity
(arb. units)
6.2 Backward Retrieval
0
Figure 6.1 Forward retrieval. We consider collinear storage, followed by forward retrieval, using a Gaussian control
√ pulse Ω(τ ) =
2
Ωmax e−τ , with W = 9, so that Ωmax = (2/π)1/4 W = 2.68. The
optical depth is d = 300, and the detuning is ∆ = 15, in normalized
units. (a) shows the input mode that optimizes just the storage efficiency, and (c) shows the input mode that optimizes the combined
efficiency of storage followed by retrieval. The shape of the control
intensity profile is indicated by the black dotted lines. The former
optimization gives a storage efficiency of ∼ 90%, whereas the latter
optimization gives a combined efficiency of only 28%. In parts (b) and
(d) we show the ‘intensity’ |Bout (z)|2 and spatial phase of the spin
wave generated in the ensemble after the storage process is complete,
for the storage and combined optimizations, respectively. The combined optimization produces a spin wave that reduces re-absorption
losses by shifting the ‘centre of mass’ of the spin wave away from the
entrance face at z = 0.
6.2
Backward Retrieval
The arguments given in the previous section suggest that retrieving the signal field
backwards should be much more efficient than forward retrieval. If the storage interaction produces a spin wave with its amplitude concentrated toward the entrance
face of the ensemble, re-absorption losses are minimized if retrieved signal photons
propagate backwards to re-emerge from the entrance face. This is indeed the case,
but issues of momentum conservation arise if the energy splitting between the ground
6.2 Backward Retrieval
198
and storage states |1i and |3i is non-zero. To see why, consider the diagram shown
in part (a) of Figure 6.2. Momentum conservation requires that the wavevectors
ks , kc and κ associated with the signal, control and spin wave, respectively, sum
to zero. When this is the case, the storage interaction is said to be phasematched,
since the spatial phases accumulated by the optical fields as they propagate through
the ensemble are ‘matched’ to the spatial phase of the spin wave. This means that
the slowly varying envelopes of these fields are strongly coupled to the spin wave,
as described by the equations of motion for the quantum memory (see (5.106) in
Chapter 5, for instance). As we will see below, when momentum is not conserved,
the interaction is not phasematched, and destructive interference greatly reduces the
strength of the coupling, and with it the memory efficiency. Fortunately, the storage
process is always phasematched, because there is no fixed dynamical relationship between the spatial phase of the spin wave and its energy. This is because the atoms
comprising the spin wave do not interact with another, so that there is no coupling
at all between the spatial shape of the spin wave and the frequency splitting between
the ground and storage states. In the storage process, a signal photon is absorbed,
and a control photon is emitted. The spin wave therefore acquires a wavevector that
‘takes up the slack’, given by the difference of the signal and control wavevectors,
κ = ks − kc =
ωs − ωc
z,
c
(6.10)
6.2 Backward Retrieval
199
where z is a unit vector pointing along the positive z axis. In the last equality we
used the optical dispersion relation for a plane wave,
k = ω/c,
(6.11)
where k is the magnitude of the wavevector and ω is the angular frequency of
the wave (not the integrated Rabi frequency!). This relation fixes the momentum
associated with the spin wave as given by the difference between the signal and
control field frequencies, which is in turn fixed by the energy splitting that separates
the ground and storage states, in order to satisfy two-photon resonance. Even though
there is no intrinsic connection between the spatial phase of the spin wave and its
energy, the kinematics of the scattering process that generates the spin wave does, in
fact, tie these two quantities together. The efficiency of the retrieval process depends
critically on whether it is possible to phasematch the retrieval interaction. When we
attempt to retrieve the spin wave excitations by sending the readout control pulse
in the backward direction, momentum conservation requires that the wavevector of
the retrieved signal field is formed from the sum of the spin wave and the read out
control wavevectors,
ksr = κ + kcr .
(6.12)
Substituting (6.10) into (6.12), we obtain
ksr − kcr = ks − kc .
(6.13)
6.2 Backward Retrieval
200
For collinear storage, we have ks = ks z and kc = kc z. For backward retrieval, we
have ksr = −ksr z and kcr = −kcr z. Inserting these expressions into (6.13), and using
the optical dispersion relation (6.11) along with the two-photon resonance condition,
we get to the condition
ωc − ωs
ωs − ωc
=
.
c
c
(6.14)
As is clear from part (b) of Figure 6.2, this condition can only be satisfied when
ωs = ωc , and κ = 0, which requires that the ground and storage states are degenerate
in energy. Experimentally, it is important to be able to distinguish the weak signal
field from the bright control pulse, and the ability to spectrally filter these two fields
should not be surrendered lightly. Therefore it is very useful to consider storage in
memories where ωs 6= ωc , and then the issue of momentum conservation becomes
important when considering backward retrieval.
(a)
(b)
Figure 6.2 Phasematching considerations for backward retrieval.
(a) The momenta of the control field and spin wave must sum to
that of the signal field, from which they are ‘scattered’. The storage
process is automatically phasematched, since the magnitude of κ is
initially a free parameter, that is determined by ks and kc during
storage. (b) The spin wave momentum is pointing in the ‘wrong
direction’, when backward retrieval is attempted: it is not possible to
simultaneously satisfy the two-photon resonance condition ksr − kcr =
(ωs − ωc )/c, and the phasematching condition ksr = κ + kcr .
In order to optimize the combined efficiency of storage, followed by backward
6.2 Backward Retrieval
201
retrieval, we need to find an expression for Ktotal that describes the entire memory
interaction in this case. The SVD of this kernel will then provide us with the optimal
input mode. The equations of motion for the retrieval process have the same form
as those describing storage, but they must describe propagation in the backward
direction, with the z coordinate is reversed. The kernel Ktotal is simply constructed
using the adiabatic solutions for storage and retrieval separately, provided we are
careful about how the boundary conditions are ‘stitched together’. Let us denote the
flipped z coordinate for the retrieval process by z r , so that z r = L − z, in ordinary
units, where z is the coordinate describing propagation during storage. The atomic
coherence at the start of the retrieval process is identical — as always assuming no
decoherence — to that generated by the storage process. We can write this as
r
σ13,in
(z r ) = σ13,out (z),
(6.15)
where the σ13 (x) denotes the Raman coherence associated with an atom located at
the longitudinal position x. The spin wave B, defined in (4.43), makes use of the
slowly varying operators σ
e13 , introduced in (4.22) (see Chapter 4), which incorporate
an exponential factor eiω13 τ . Recall that τ is in fact the retarded time, τ = t − z/c,
so there is a spatial phase built into the spin wave; this represents the momentum
imparted to the spin wave by the optical fields that create it during the storage
6.2 Backward Retrieval
202
process. Equating the boundary conditions, as in (6.15), gives the relation
r
r
Bin
(z r ) = Bout (z) × e−iω13 τout × eiω13 τin ,
r
= Bout (z) × e−iω13 (tin −tout ) × e−iω13 (z−z
= Bout (L − z r )e2iω13 z
r /c
r )/c
.
,
(6.16)
Here we used the definition τ r = tr − z r /c = tr − L/c + z/c, and in the last line
we dropped some unimportant constant phase factors. The spatial phase factor in
(6.16) represents the phase mismatch shown in Figure 6.2 (b). Note that it vanishes
if ω13 = 0. The phase mismatch causes oscillations of the spin wave in space that
can dramatically reduce the retrieval efficiency. This can be seen by considering the
form of the retrieval map
Arout (τ r )
Z
1
=
Z0 1
=
r
K r (τ, z r )Bin
(z r ) dz r
r
K r (τ r , z r )eiδkz Bout (1 − z r ) dz r .
(6.17)
0
Here we have switched to normalized units, and defined the dimensionless phase
mismatch
δk =
2ω13 L
.
c
(6.18)
Note that δk is a negative quantity if the storage state |3i is more energetic than the
ground state |1i. It is clear that, regardless of the form of the retrieval kernel K r , or
of the spin wave Bout , the integral can be made to vanish if δk is made large enough.
6.2 Backward Retrieval
203
This explains why a phase mismatch can be very detrimental to the efficiency of
backward retrieval. More precisely, if kmax represents the width in k-space of the
integrand K r × Bout , the value of the integral becomes strongly suppressed when
δk kmax . It is therefore possible to mitigate the effect of the phase mismatch,
to some extent, by localizing the spin wave strongly in space, so that kmax becomes
large. Below, we show explicitly how to optimize storage, followed by backward
retrieval in the presence of a phase mismatch; the result essentially generates a more
strongly localized spin wave, for exactly this reason.
To perform the optimization, we construct the kernel Ktotal using the solution
(6.4) for the retrieval process, making the replacement z −→ z r . The result is
Arout (τ r )
Z
∞
Ktotal (τ r , τ )Ain (τ ) dτ,
(6.19)
Ωr∗ (τ r )Ω(τ )
√
× Ktotal [ω r (τ r ), ω(τ )] ,
W rW
(6.20)
=
−∞
with
Ktotal (τ r , τ ) =
and where
r
Z
Ktotal (ω , ω) =
1
r
K r (ω r , z r )K(1 − z r , ω)eiδkz dz r .
(6.21)
0
Suppose that the readout control pulse is identical in shape and frequency to the
storage control pulse, with Ωr = Ω and ∆r = ∆. Suppose also that the storage state
is degenerate with the ground state, so δk = 0. Then consider the kernel Ktotal in
6.2 Backward Retrieval
204
(6.21) with its output time argument flipped around,
Z
r
1
K r (1 − ω r , z r )K(1 − z r , ω) dz r
Ktotal (1 − ω , ω) =
0
Z
=
1
K(1 − z r , ω r )K(1 − z r , ω) dz r ,
(6.22)
0
The second line follows from the symmetry K r (1 − x, y) = K(1 − y, x) that connects
the retrieval kernel (6.6) to the storage kernel (5.76). This shows that Ktotal is,
in the case under consideration, very similar in structure to the normally ordered
kernel KN that would be formed from K(1 − z r , ω). The only difference is that
the first instance of K in (6.22) is replaced by its complex conjugate in KN . If K
is real, then this complex conjugation is not important, and Ktotal = KN , which
means that the optimal input modes derived from the SVD of Ktotal are equal to
the optimal input modes derived from the SVD of K alone. This observation shows
that the optimal input mode for storage alone is close to the optimal input mode
for storage followed by backward retrieval, if identical storage and readout control
pulses are used, and if there is no phase mismatch. The two optimizations are the
same whenever K is real, and more generally whenever ψ1 (z), the output spin wave
mode generated by optimal storage, is real.
In Figure 6.3 below we show an example of how the optimal input mode for
storage alone can differ from the optimal mode for storage followed by backward
retrieval. Without a phase mismatch (parts (c) and (d)), the difference between the
two optimizations is rather small, though non-negligible, for the example shown.
6.2 Backward Retrieval
205
When a significant phase mismatch is introduced (parts (e) and (f)) the optimizations differ more markedly. It is a general feature of the optimization of storage
with backward retrieval, that as |δk| is increased, the shape of the optimal input
mode approaches that of the control pulse. The reason is as follows. Recall that
the dispersion experienced by the signal field as it propagates causes its group velocity to differ from that of the control (see the end of §5.4 in Chapter 5). If the
signal initially overlaps with the control, then the coupling between these two fields
is initially high, but as the signal walks off from the control, the coupling decays
away. The bulk of the storage interaction therefore occurs near the entrance face
of the ensemble at z = 0. As is clear from part (f) of Figure 6.3, the spin wave
generated by the optimal input mode is indeed concentrated more closely towards
the entrance face. This makes the spin wave more strongly localized, which reduces
the ‘wash-out’ caused by the phase mismatch, as discussed above, while at the same
time reducing re-absorption losses.
A discussion of similar optimizations can be found in the work of Gorshkov
et al. [133] . In this work time-reversal arguments are employed to show that for
2
optimized storage with backward retrieval, ηcombined = ηstorage
if ψ1 (z) is real. This
condition is always satisfied in the light-biased limit (see the end of §5.2.8 in Chapter
5), when the anti-normally ordered kernel KA becomes real, independent of whether
or not the adiabatic approximation is satisfied (see (5.27 in Chapter 5)). As shown
above in the adiabatic limit, when ψ1 is not real, implementing ‘true’ time reversal
requires that the phase of the spin wave is conjugated. Without a practical method
1
206
Input mode
Spin wave
10
(a)
6
(b)
4
5
0
1
0
0
6
6
(c)
−2
0
2
Intensity
(arb. units)
4
2
2
0
0
8
6
4
2
0
6
(f)
4
2
0
0.5
1
Phase
(e)
0.5
0
(d)
4
0.5
0
1
2
Phase
Intensity
(arb. units)
0.5
Phase
Intensity
(arb. units)
6.2 Backward Retrieval
0
Figure 6.3 Backward Retrieval. We consider collinear storage followed by backward retrieval, using a Gaussian control pulse Ω(τ ) =
/Tc )2
Ωmax e−(τp
, with Tc = 0.1 and W = 100, so that Ωmax =
(2/π)1/4 W/Tc = 28.25. The optical depth is d = 300, and the
detuning is ∆ = 150, in normalized units. Parts (a) and (b) show
the optimal input mode for storage alone, alongside the spin wave
mode ψ1 (z) generated in the medium. We use the adiabatic form
for the storage kernel; comparison with the prediction derived from
the numerically constructed kernel verifies that the adiabatic approximation is well satisfied. The optimal efficiency for storage alone is
ηstorage ∼ 96%. In parts (c) and (d) we show the optimal input mode
for storage followed by backward retrieval, and the generated spin
wave, with degenerate ground and storage states, δk = 0. The opti2
mized total efficiency is ∼ 88%, which is about 5% less than ηstorage
.
The most notable difference between these two optimizations is the
appearance of a ‘wiggle’ in the phase of the input mode for the combined optimization. But this occurs while the signal intensity is rather
low, so it is actually less important than the more subtle re-shaping
that occurs. In parts (e) and (f), we show the optimal input mode
for storage and backward retrieval, along with the generated spin
wave, with a phase mismatch δk = −5. Note that this optimization
results in a spin wave with its ‘centre of mass’ concentrated more
closely towards the entrance face of the ensemble at z = 0 than the
other optimizations. This reduces the effective length of the ensemble
— the length over which there is significant excitation — which diminishes the effect of the phase mismatch. The optimized combined
efficiency of storage and retrieval is around 75% in this case.
6.3 Phasematched Retrieval
207
2
for doing this, ηcombined < ηstorage
, and the optimal mode for storage alone differs
from the mode that optimizes the combination of storage and backward retrieval.
In this and the previous section we have shown that retrieving the stored excitations from a quantum memory efficiently is a non-trivial problem. Forward retrieval
is plagued by re-absorption losses — a modematching issue. Backward retrieval is
beset by momentum conservation problems — a phasematching issue. In the next
section we present a solution to both of these problems, which requires a departure
from the one dimensional treatment we have worked with thus far.
6.3
Phasematched Retrieval
The best memory efficiency possible is achieved by optimizing collinear storage,
followed by backward retrieval, with δk = 0. But to spectrally filter the signal field
from the control, we should have δk 6= 0. By introducing a small angle between the
signal and control beams, both at storage and at read out, we can maintain proper
phasematching, even when δk 1, which allows for efficient retrieval. This idea
was proposed by Karl Surmacz, and the following treatment is adapted from our
paper [152] ; the numerical simulations were performed by me. The principle of the
scheme is easily understood by considering the phasematching diagrams in Figure
6.4. Here we assume for simplicity that all the beams used, both at storage and at
read out, are confined to a common plane, so that we need only consider two space
dimensions. The spin wave momentum is determined by the signal and control field
wavevectors according to (6.10). If we fix the detuning, so that ∆r = ∆, then
6.3 Phasematched Retrieval
208
the magnitudes of ks and kc are unchanged at read out, and then the condition
(6.12) uniquely defines the direction of ksr that is phasematched. By symmetry, the
angle θ between the signal and control beams is the same at read in as at read out.
Provided we adhere to the geometry shown in Figure 6.4 (a), the retrieval process
remains correctly phasematched for all choices of θ. To maximize the efficiency of
the memory, we should reduce θ as far as is possible, so that
1. we approach a collinear geometry, which maximizes the overlap of the signal
and control pulses, and
2. the spin wave generated by the storage process overlaps well with the optimal
spin wave mode for retrieval.
An heuristic choice of θ that satisfies these desiderata is that shown in part (b)
of Figure 6.4, with ks cos θ = kc . The signal and control wavevectors are close to
parallel, but they are arranged so that the spin wave momentum κ is orthogonal
to kc . This way, when the control field direction is reversed for backward retrieval,
no phase mismatch is introduced and the signal field emerges at an angle θ with
respect to the read out control pulse, as shown in part (a) of Figure 6.5. This choice
for θ assumes that the storage state is more energetic that the ground state, so
that δk is negative, and ks > kc . We also consider the possibility that the storage
state lies energetically below the ground state. In many systems the ground state is
prepared artificially by optical pumping (see §10.12 in Chapter 10), and it is then
quite feasible to select the ground state so that ω1 > ω3 . In this case, δk > 0 and
6.3 Phasematched Retrieval
209
ks < kc . We then choose θ so that kc cos θ = ks ; the geometry is shown in part (b)
of Figure 6.5.
(b)
(a)
Figure 6.4 Non-collinear phasematching. (a) the general geometry
required to phasematch storage and retrieval, when the signal beam
makes an angle θ with the control. (b) a geometry that closely approaches the collinear one, while preserving correct phasematching.
In the next section we include dispersion, which modifies the length
of ks and ksr .
Storage
Retrieval
(a)
(b)
Figure 6.5 Efficient, phasematched memory for positive and negative phase mismatches. We show the beam directions of the control (green) and signal (blue) fields at storage and retrieval, with (a)
δk < 0, and (b) δk > 0. Off-resonance, the angles used depend on
both δk and on the material dispersion (see §6.3.1 below).
6.3 Phasematched Retrieval
6.3.1
210
Dispersion
So far we have used the dispersion relation (6.11), which applies to light propagation
in vacuo. The phase mismatch δk quantifies the spatial phase imparted to the
spin wave with collinear storage calculated using only this vacuum dispersion. But
even in the case of degenerate ground and storage states, the spin wave is not
always entirely real, as discussed in the previous section (see part (b) of Figure 6.3).
This spatial phase arises dynamically, and has its origin in the material dispersion
experienced by the signal field as it propagates through the ensemble. The control
field is unaffected, since it couples states |2i and |3i, both of which are unpopulated.
But the signal field couples the populated ground state |1i to state |2i, and therefore
it propagates subject to an augmented refractive index. This effect can be described
by incorporating a dispersive term into our definition for the magnitude of the signal
field wavevector,
ks −→ kd = ks − kdisp ,
(6.23)
where, working in ordinary units, we define
kdisp =
dγ∆
|κ|2 ∆
= 2
.
|Γ|2
(γ + ∆2 )L
(6.24)
This definition is motivated by inspection of the adiabatic storage kernel (5.77)
derived in Chapter 5, which includes an exponential factor with an exponent whose
imaginary part contains the spatial phase kdisp z. The form of kdisp is consistent with
the refractive index associated with an absorptive resonance at ∆ = 0. Note that on
6.3 Phasematched Retrieval
211
resonance, kdisp vanishes, and the signal field wavevector assumes its vacuum value.
At large detunings, kdisp ∝ 1/∆, so it is generally small in the Raman limit, but not
negligible.
Incorporating the dispersive phase into the signal wavevector allows us to use
the phasematching scheme outlined in the previous section to compensate both for
non-degeneracy of the ground and storage states and for material dispersion in the
ensemble. By eliminating even this latter dynamical phase, we essentially render
the spin wave real, so that the efficiency of backward retrieval (albeit at a small
angle to the z-axis) is equal to the storage efficiency, which is optimal. And with all
spatial phases removed in this way, the optimization of storage alone is the same as
the optimization of storage with backward retrieval, and so it is only necessary to
perform the former, simpler optimization.
6.3.2
Scheme
Including material dispersion, then, the phasematching scheme is summarized by
the following choice for θ,
−1
θ = cos
r;
r = max
kd kc
,
kc kd
,
(6.25)
where kc = ωc /c, and kd is given by (6.23). This formula holds for arbitrary values
of δk, and for all detunings.
Provided that θ is sufficiently small, we expect that many of the results pertaining
to the optimization of collinear storage remain approximately valid. Our general
6.3 Phasematched Retrieval
212
strategy is to use the one dimensional theory of Chapter 5 — in particular the
adiabatic storage kernel (5.77) — to find the optimal temporal input mode, and
then implement the above phasematching scheme by introducing the small angle θ
between the signal and control fields.
Note that this scheme is experimentally appealing for three reasons. First, it
allows for storage with non-degenerate ground and storage states, so the signal and
control fields can be spectrally filtered. Second, introducing an angle between the
signal and control beams makes it possible to spatially filter the signal from the
control, with a pinhole or an optical fibre tip, for example. Third, the effect of
unwanted Stokes scattering can be eliminated. This is because any spin wave excitations produced by Stokes scattering — the process shown in part (b) of Figure 4.3
in Chapter 4 — will not have the same momentum as those produced by absorption
of the signal field. At read out, we choose the direction of the control pulse so that
retrieval of the stored signal field is correctly phasematched. This same control pulse
could potentially drive anti-Stokes scattering, converting the unwanted excitations
into noise photons with the same frequency as the retrieved signal. But this process
will generally not be correctly phasematched, and is therefore greatly suppressed.
All these features of the scheme constitute a compelling case for the utility of a
non-collinear phasematched memory protocol.
Here we should remark that this type of phasematching scheme produces a spin
wave with a larger momentum than arises from collinear storage, since |κ| ≥ |δk|. If
the atoms in the ensemble are free to move — for instance if the storage medium is
6.4 Full Propagation Model
213
a warm atomic vapour — then any large spatial frequencies in the spin wave may be
washed out as atoms diffuse over the phase fronts of the spin wave. Therefore our
phasematching scheme may be more susceptible to decoherence, if D > λκ = 2π/|κ|,
where D is the distance over which the atoms diffuse during the storage time (see
§10.5 in Chapter 10). This problem does not affect solid state memories, which are
appealing for this and many other reasons.
To investigate the efficacy of our phasematching scheme, we should consider the
effect of walk-off between the signal and control fields as they propagate, since when
θ 6= 0 their paths inevitably diverge. In the next section we present the results of
numerical simulations that account both for walk-off and diffraction. These vindicate
the scheme (6.25), and they show, as might be expected, that the beams should be
loosely focussed, so as to maximize the overlap of the pulses.
6.4
Full Propagation Model
We model the propagation of the signal and control fields in three dimensions: two
spatial dimensions, and time. The signal field propagates along the z-axis, and the
control field is launched at an angle θ to the z-axis, in the (x, z)-plane, intersecting
the signal beam in the centre of the ensemble. We consider that both fields are
Gaussian beams — defined below — focussed at the centre of the ensemble, as
depicted in Figure 6.6. The equations of motion for the interaction are given in (4.52)
at the end of Chapter 4. For numerical convenience, we work in normalized units that
render all quantities dimensionless, and as close as possible to ∼ 1. We summarize
6.4 Full Propagation Model
214
the normalizations we employ below; they differ slightly from the normalized units
used in the preceeding analytic treatments.
1. The z coordinate is measured in units of L, and the magnitudes ks , kc , are
measured in units of 1/L. We define the z coordinate so that it runs from − 12
to 12 , so that the centre of the ensemble lies at z = 0. This definition is useful
for parameterizing the control field.
2. The x coordinate is measured in units ws , where ws is the transverse beamwaist of the signal field. That is, at z = 0, the transverse signal field amplitude
2
2
is a Gaussian e−(x/ws ) in ordinary units, and by definition e−x in normalized
units. To capture all relevant dynamics, x runs from −3 to 3.
3. The time coordinate τ , which as usual is the retarded time τ = t − z/c, is
measured in units of Tc , the duration of the control field. This is a departure
from the time normalization used previously, in which times were measured
in units of 1/γ. With this new normalization, the control field has roughly
unit duration, as does the signal field (since the optimal signal pulse is generally comparable in duration to the control). All relevant dynamics are then
captured by considering times from τ = −3 up to τ = 3, or thereabouts.
6.4 Full Propagation Model
215
The equations of motion, written using these normalized units, are given by
p
iα2 2
∂x + ∂z A = − dγP,
2ks
p
∂τ P = −ΓP + dγA − iΩB,
∂τ B = −iΩ∗ P.
(6.26)
As in all of our previous analyses, we have dropped the Langevin noise operators
associated with spontaneous emission and decoherence, since they have no effect
on the efficiency of the memory. The normalization factor α = L/ws is the aspect
ratio of the ensemble, which appears to correctly account for the difference in units
used for the x and z coordinates. Note that the amplitudes A, P and B are not
the averaged quantities defined in (5.1) at the start of Chapter 5. They are closer
to those quantities given in (4.7), (4.42) and (4.43) in Chapter 4, which depend
on the transverse spatial coordinates ρ, as well as τ and z. In fact, since we have
dropped any dependence on the y-coordinate, we have implicitly averaged over this
dimension. But this has no effect on the equations.
6.4.1
Diffraction
The transverse derivative ∂x2 in (6.26) describes the diffraction of the signal field
as it propagates. Its importance in the dynamics is set by the ratio of α2 to ks .
Although both of these are generally large quantities, their ratio can be small, or
large, depending on how tightly the signal is focussed: if ws is small, diffraction
6.4 Full Propagation Model
216
Figure 6.6 Focussed beams. The signal (blue) and control (green)
beams are cylindrically symmetric Gaussian beams focussed at the
centre of the ensemble, with a small angle between them.
can become significant. In fact, the signal field Rayleigh range, which quantifies the
length of the region over which the focussed signal beam is well-collimated, is given
in our normalized units by zs = ks /(2α2 ). As expected, the contribution from the
diffractive term becomes significant when the Rayleigh range falls below 1, which is
the length of the ensemble.
6.4.2
Control Field
Walk-off between the signal and control field appears through the spatial dependence of the control field Ω. To construct the correct expression, consider first the
representation of a pulse with a Gaussian transverse profile, and also a Gaussian
6.4 Full Propagation Model
217
temporal profile, propagating along the z-axis,
Ωmax
Ω(x, τ, z) =
exp
W (z/zc )
(
i
1
− 2
R(z/zc ) W (z/zc )
x
wc
2
− i tan−1
z
zc
)
2
×e−τ .
(6.27)
Here zc = kc wc2 /(2α2 ) is the Rayleigh range of the control field, with wc the beamwaist of the control, measured in units of ws . The structure of (6.27) is well known
in optics, since the Gaussian transverse profile arises naturally as the lowest order
mode supported by the confocal cavity in a laser. When a laser beam of this form
is directed through a lens, it retains its Gaussian profile, but its width narrows as it
approaches the focal plane, and then widens out aftwerwards. The functions W and
R parameterize the beam size and the radius of curvature of the field phase fronts,
and are given by
p
1 + z2,
1
R(z) = z 1 + 2 .
z
W (z) =
(6.28)
The focus lies at z = 0, at which point R −→ ∞ and W = 1. The term involving
tan−1 is known as the Gouy phase; its effect on the memory efficiency is negligible,
but we include it for completeness. Direct substitution shows that (6.27) indeed
satisfies the paraxial wave equation
2
wc 2
i
∂ + ∂z Ω = 0.
4zc x
(6.29)
6.4 Full Propagation Model
218
When the control propagates at an angle θ to the z-axis, the control amplitude is described by a similar expression to (6.27), except that we rotate the (x, z) coordinates
by the angle θ. We should be careful to apply this rotation with t held constant,
not τ , and we must take account of the different units used for z and x. The correct
transformation is found by making the replacements
x −→ x0 = cos(θ)x + α sin(θ)z,
z −→ z 0 = cos(θ)z −
and
τ −→ τ 0 = τ +
sin(θ)
x
α
z − z0
.
c
(6.30)
The factors of α ensure that the units of x and z are interconverted as they should
be, and the modification to τ represents the change in the apparent velocity of the
control pulse in the reference frame of the signal field.
6.4.3
Boundary Conditions
The boundary conditions must be specified in much the same way as in the previous one dimensional analyses. One additional feature is the dependence on the x
coordinate. The boundary conditions associated with this new degree of freedom
are simply that all the quantities A, P and B vanish as |x| −→ ∞. In practice, this
is achieved by fixing the value of the signal amplitude A to zero at x = ±3 (i.e.
at the edges of the region covered by the numerics). We use Chebyshev spectral
collocation to treat the spatial derivatives when solving the system (6.26), and we
6.4 Full Propagation Model
219
build this boundary condition directly into the differentiation matrices; the method
is explained in Appendix E. Fixing A at these boundaries also fixes P and B, since
the interaction is local: there can be no atomic excitation where there is no input
light. To model the storage process, we launch the signal field along the z-axis with
a Gaussian transverse profile at the entrance face of the ensemble,
("
Ain (x, τ ) = A z = − 21 , x, τ = exp
# )
i
1
−
x2 φ1 (τ ),
R − 12 /zs
W 2 − 21 /zs
(6.31)
where φ1 (τ ) is the optimal input mode for storage alone, found using the one dimensional theory of the previous Chapter — we use the adiabatic kernel (5.77), so that φ1
is given by (5.78). As described above, the phasematching scheme should eliminate
the spatial phase of the spin wave, so that the optimal mode for collinear storage
alone should be close to optimal for phasematched storage and retrieval. Note that
the absolute magnitude of the signal field is irrelevant, because the memory interaction is linear. This boundary condition is also incorporated into the differentiation
matrices we use to solve the dynamics numerically.
For the storage process, both the spin wave B and the polarization P are initially
zero,
Bin (x, z) = B(τ −→ −∞, x, z) = 0,
Pin (x, z) = P (τ −→ −∞, x, z) = 0. (6.32)
We use an RK2 method for the time-stepping, and these boundary conditions are
trivially implemented by zeroing the vectors representing B and P at the collocation
6.4 Full Propagation Model
220
points on the first iteration. See Appendix E for a description of the RK2 method,
and how it is used in combination with spectral collocation to arrive at a solution.
6.4.4
Read out
To model the read out process, we solve the equations of motion once more, this time
with no incident signal field, and a pre-existing Raman coherence determined by the
storage process. We are only able to model the build up of the signal field along a
specific direction, the z-axis, and so we must re-orient our coordinate system so that
the z-axis points along the direction in which the retrieval process is phasematched.
This is the direction of ksr in Figure (6.4). The angle through which we rotate our
coordinate system is therefore the angle between ks and ksr in Figure (6.4), which is
given by


kc sin(θ)





π
=
θ0 = −2 sin−1  q


kd2 + kc2 − 2kc kd cos(θ)


2θ − π
if kc > kd
.
(6.33)
if kd > kc
The first result holds for general values of θ, while the second result applies to the
case where θ is chosen according to the scheme (6.25).
At read out, the signal field is initially zero,
Arin (xr , τ r ) = 0,
(6.34)
6.4 Full Propagation Model
221
as is the polarization,
r
Pin
(xr , z r ) = 0.
(6.35)
Here xr , z r are the rotated coordinates describing the retrieval process,
xr = cos(θ0 )x + α sin(θ0 )z,
z r = cos(θ0 )z −
sin(θ0 )
x.
α
(6.36)
As in (6.30), the factors of α appear to interconvert the units of x and z.
The initial spin wave at read out is set by the spin wave generated by the storage
process, but we must be careful to include any spatial phase factors arising from
the above coordinate transformation. When the signal and control fields are not
collinear, the definition of the slowly varying coherence σ
e13 (see (4.22) in Chapter
4) must be modified accordingly,
σ
e13 (r) = σ13 (r)eiω13 t+i(ks −kc ).r .
(6.37)
Here the argument r indicates that the coherence refers to an atom at that position
in the ensemble. The boundary condition for B is found by equating the coherences
at the end of storage and the start of read out,
r
σ13,in
(r r ) = σ13,out (r).
(6.38)
The spin wave amplitudes are built from the slowly varying coherences. Dropping
6.4 Full Propagation Model
222
some unimportant temporal phase factors, we therefore obtain
r
r
r
Bin
(r r ) = Bout (r)ei[(ks −kc ).r
r −(k
s −kc ).r]
.
(6.39)
Writing this boundary condition out explicitly in terms of the retrieval coordinate
system gives the relation
r
Bin
(xr , z r ) = Bout (xr c0 − s0 αz r , z r c0 + s0 xr /α) ×
(6.40)
exp i (ks − kc c)(1 − c0 ) − kc ss0 z r + i (ks − kc c)s0 − kc s(1 + c0 ) xr /α ,
where c and s denote cos(θ) and sin(θ), respectively, and c0 , s0 are equal to cos(θ0 )
and sin(θ0 ). Note that in the absence of dispersion, with ks = kd , the phase factor
identically vanishes; it is removed by the phasematching scheme. When dispersion is
significant, the slowly varying envelope Bout itself contains a spatial phase variation.
The phasematching scheme is then designed so that the exponential factor in the
second line of (6.40) cancels this spatial phase, rendering B r as smooth as possible,
for efficient retrieval.
6.4.5
Efficiency
The efficiency of the memory is calculated in two stages. First, we simulate storage
and evaluate the storage efficiency, which is given by
R ∞ R 1/2
ηstorage =
2
−∞ −1/2 |Bout (x, z)| dz dx
R∞ R∞
.
2
−∞ −∞ |Ain (x, τ )| dτ dx
(6.41)
6.5 Results
223
We then simulate retrieval, and we calculate the retrieval efficiency,
R∞ R∞
−∞
ηretrieval = R−∞
∞ R 1/2
|Arout (xr , τ r )|2 dτ r dxr
r
r r 2
r
r
−∞ −1/2 |Bin (x , z )| dz dx
.
(6.42)
The total memory efficiency is then given by ηcombined = ηstorage × ηretrieval .
6.5
Results
In Figures 6.7 and 6.8, we present the results of simulations performed to examine the
effectiveness of our phasematching scheme. We simulated both resonant EIT storage,
and also off-resonant Raman storage. Each simulation was run twice, once with a
tightly focussed control field, with wc = 1, and once with a loosely focussed control,
with wc = 2. In the latter case the energy in the control pulse was quadrupled, so
that the intensity of the control was the same in both cases.
Figure 6.7 shows the angle θ at which the combined efficiency of storage and
retrieval is largest, over a range of values of the phase mismatch δk. The numerical
solutions generally bear out the analytic prediction (6.25) of our phasematching
scheme. But the agreement with our prediction is much better when the control is
loosely focussed. This is to be expected, since transverse walk-off is less important
when the control is wider than the signal, and since a wider control diffracts less, so
that its intensity is more homogeneous over the ensemble. For EIT, collinear storage
with θ = 0 is only optimal when δk = 0, since there is no dispersion on resonance.
But for the Raman memory, collinear storage is optimal when δk ∼ −2, when the
6.5 Results
224
(a)
−3
θ (radians)
8
x 10
EIT
6
4
2
0
−4
8
θ (radians)
−3
−2
−1
0
1
2
3
4
−2
−1
0
1
2
3
4
−3
(b)
x 10
6
Raman
4
2
0
−4
−3
Figure 6.7 Effectiveness of our phasematching scheme. We plot
the angle θ at which the combined efficiency of storage followed by
retrieval is optimized. The blue solid line shows the prediction of our
phasematching scheme (6.25). The black filled circles correspond to
tight control focussing, with wc = 1; the red open diamonds correspond to loose control focussing with wc = 2. We plot the results for
a typical EIT protocol in (a): we set γ = 1, d = 30 and ∆ = 0, in
units normalized as described in §6.4 above. In (b) we present equivalent results for a Raman protocol d = 300, γ = 0.1 and ∆ = 15.
The bandwidth of the control is 10 times larger, with respect to the
natural atomic linewidth γ, in this latter protocol. To make up for
the increased detuning in the Raman protocol, the optical depth is
also much larger. In both protocols, the control field amplitude was
given by Ωmax = 5.
storage state lies significantly above the ground state in energy. This is because the
material dispersion alters the signal field wavevector, kd 6= ks , so that even with
degenerate ground and storage states, the spin wave acquires a spatial phase.
In Figure 6.8 we plot the optimal memory efficiencies obtained from the numerical simulations alongside the analytic predictions for the best collinear efficiencies,
again over a range of values of δk. The efficiencies obtained using the phasematching
scheme (6.25) are generally very close to the best efficiencies achieved in the sim-
6.5 Results
225
efficiency
(a)
1
0.8
0.8
0.6
0.6
0.4
0.4
0.2
0
−4
(c)
efficiency
(b)
1
Narrow EIT
−2
0
0.2
2
0
−4
4
(d)
1
0.8
0.6
0.6
0.2
0
−4
0.4
Narrow Raman
−2
0
0.2
2
4
−2
0
2
4
0
2
4
1
0.8
0.4
Wide EIT
analytic
collinear backwards
collinear forwards
phasematched
0
−4
Wide Raman
−2
Figure 6.8 Comparing phasematched and collinear efficiencies.
Plots (a) and (b) contain the results for the EIT protocol, with tight
and loose focussing, respectively. Plots (c) and (d) show equivalent
results for the Raman protocol. The parameters used are the same
as those used in produciing Figure 6.7. The solid lines represent the
efficiencies obtained if the phasematching scheme (6.25) is used. The
black filled circles are the best efficiencies achievable. The green dotted lines show the optimal efficiencies predicted for collinear storage
with backward retrieval, calculated using the one dimensional adiabatic kernel (6.21). The dashed red lines show the efficiency achieved
by collinear storage with forward retrieval, where the input signal
pulses were shaped using the adiabatic kernel (6.9), where this improved the efficiency.
ulations. These efficiencies are greatest when θ = 0, since at these points walk-off
between the signal and control is eliminated. As δk is increased, so does the angle θ
required for optimal efficiency, and walk-off therefore reduces the memory efficiency.
But the efficiency falls only slowly with increasing δk. With loose focussing it is
possible to exceed the optimal collinear efficiency when δk is large enough, using
either EIT or Raman storage. This demonstrates what we sought to show with our
simulations, that even including the effects of diffraction and walk-off, non-collinear
storage and retrieval, with proper phasematching, is preferable over collinear stor-
6.5 Results
226
age, with either forward or backward retrieval. The lower efficiencies observed for
tight control focussing confirms that diffraction and walk-off can be detrimental, but
provided sufficient laser energy is available, it is possible to reduce their effects by
simply widening the control beam waist.
The red dashed lines in Figure 6.8 are the result of simulating collinear storage
with forward retrieval, using the full numerical propagation model. For the EIT
protocol, we used the one dimensional adiabatic kernel (6.9) to predict the optimal
input profiles, and these input pulse shapes performed well — better than any other
pulse profiles we tried. This at least confirms the utility of the one dimensional
theory in this context. However, for the Raman protocol, this optimization did not
produce the best pulse profiles. In fact, optimal pulse profiles for backward retrieval
worked better, even though the excitation was retrieved in the forward direction.
So these input profiles were used to produce the red dashed lines in parts (c) and
(d). The reason for the failure of the forward retrieval optimization is that the
Raman protocol is more sensitive to the control intensity than is EIT. The control
intensity is highest in the centre of the ensemble, where it is focussed; its intensity
falls off towards the exit face of the ensemble. But as noted in §6.1.1, optimizing for
forward retrieval tends to shift the bulk of the spin wave towards the exit face. When
diffraction is introduced in the full model, this optimization is no longer beneficial,
since it shifts more of the spin wave to where the control intensity is lower. This
explains why the pulses optimized for backward retrieval performed better in the
simulations of forward retrieval for the Raman protocol. If the control beam waist
6.5 Results
227
is increased above 2.4, diffraction effects become small enough that the optimization
for forward retrieval does indeed produce better results than the backward retrieval
optimization, for both the EIT and Raman protocols.
Comparison of Figures 6.7 and 6.8 reveals that the efficiency falls off sharply in
the region kd > kc , to the left of the point at which θ = 0. That is, the EIT efficiency
falls sharply for δk < 0, and the Raman efficiency falls quickly as δk is reduced below
around −2. The efficiency remains high even for large angles when kd < kc . This
is easily explained by considering the geometry in part (b) of Figure 6.4. When
kd < kc , the retrieved signal field propagates exactly backwards with respect to the
direction of the input signal beam, since θ0 = π (see (6.33)). Therefore the region
over which the retrieved signal experiences gain overlaps precisely with the region
in which the spin wave is deposited during the storage process. We could say that
the ‘retrieval overlap’ is high, and the efficiency is correspondingly large. On the
other hand, when kc < kd , θ0 6= π, and the retrieved signal field propagates at
an angle to the direction defined by the input signal field. The retrieval overlap
suffers as θ is increased, and this explains the sharp drop in memory efficiency in
this regime: although the memory is correctly phasematched, the atomic excitations
are not distributed favourably for efficient retrieval.
The foregoing discussion suggests that, as a rule, positive phase mismatches are
preferable to negative ones, although dispersion complicates things slightly. That is
to say, for efficient phasematched retrieval, it is generally better to arrange for the
ground state |1i to have a higher energy than the storage state |3i. Although this
6.6 Angular Multiplexing
228
may be rather non-standard, when the ground state manifold is prepared by optical
pumping, it is not problematic.
We have succeeded in showing that proper phasematching can indeed boost the
efficiency of retrieval from a quantum memory. In the next section we describe how
the angular selectivity of phasematching offers the possibility of storing multiple
modes in a single ensemble.
6.6
Angular Multiplexing
There is a sense in which an ensemble memory represents a degree of ‘overkill’ if
just a single photon is to be stored. The number of atoms is many times larger
than the number of input photons, so there is an enormous redundancy built into
the memory. A more efficient use of resources would aim to store multiple input
pulses; in this way a single physical memory allows for highly parallel information
processing. Several schemes that make use of the idea of multimode memories have
already been proposed [17,38,153,154] , in connection with both quantum computing
and quantum repeaters. A general difficulty with multimode storage in an ensemble
is the possibility that the control field used for storing one signal pulse, may also
retrieve another signal pulse that was previously stored. Phasematching provides a
way to suppress this kind of unwanted coupling.
We consider storing multiple pulses in a single ensemble by changing the angle θ
between the signal and control for each input pulse. After a pulse has been stored,
momentum conservation picks out a single, unique direction, in which the retrieval
6.6 Angular Multiplexing
229
process is phasematched, as shown in part (a) of Figure 6.4. As long as the control
field used to store a subsequent pulse propagates in a different direction to this, the
previously stored pulse will not be retrieved. By successively scanning the angle
of either the signal or control field, it is possible to store many pulses in a single
ensemble. And it is also possible to selectively retrieve just one of the stored pulses,
by aligning a retrieval control pulse with the appropriate direction for the chosen
mode.
The directional selectivity afforded by phasematching is demonstrated using our
numerical model in part (a) of Figure 6.9. Here we simulated EIT storage, with θ
chosen using the phasematching scheme (6.25). Upon retrieval, however, we rotated
away from the optimal angle by θd . In the case δk = 0, we have θ = 0 and the spin
wave momentum vanishes, so that all retrieval directions are phasematched. The
efficiency is therefore only limited by walk-off, so it remains high over a large range
of deviation angles θd . But when we set δk = 1, phasematching selects a unique
retrieval direction, and the efficiency falls off quickly as θd is increased. This shows
how it is possible to ‘switch off’ retrieval from a particular spin wave by tuning the
angle of the control.
6.6.1
Optimizing the carrier frequencies
Angular multiplexing requires that we implement storage with a range of angles θ
between the signal and control fields. Now, for a given pair of frequencies ωc , ωs for
the control and signal fields, there is only one angle that satisfies the phasematching
6.6 Angular Multiplexing
(b)
efficiency
(a) 1
0.8
0.6
0.4
0.2
0
0
230
0.02 0.04 0.06 0.08 0.1
θd (radians)
Figure 6.9 Angular multiplexing. In (a) we plot the dependence of
the memory efficiency on the deviation angle θd , for EIT storage with
δk = 0 (dotted line) and δk = 1 (solid line). We used a wide control
beam, with wc = 9; all other parameters are the same as those used
for the EIT protocol in Figures 6.7 and 6.8. The efficiency is only
limited by walk-off when δk = 0, but when δk = 1 phasematching
restricts the range of angles over which efficient retrieval is possible.
In part (b) we illustrate the principle of multimode storage by angular
multiplexing. The number of pulses that can be stored is given by
N = ∆θ/δθ, where ∆θ is the largest angle, permitted by walk-off,
for which efficient retrieval is possible, and where δθ is the smallest
angle separating two modes, such that storage of one mode does not
result in the accidental retrieval of another previously stored mode.
Inspecting part (a), we see that ∆θ ∼ 0.04, and δθ ∼ 0.02, so N ∼ 2.
scheme (6.25). As our simulations showed, this angle provides the optimal memory
efficiency. To retain this optimum as θ is scanned, we should change the signal and
control frequencies accordingly. For example, in the case that ωs < ωc (i.e. δk > 0),
the phasematching scheme imposes the relationship ωs = ωc cos(θ) + ckdisp between
θ, ωs and ωc . In the Raman protocol, it is possible to perform this kind of fine
tuning on the frequencies, since the efficiency changes slowly with the detuning. Of
course, EIT is defined by the condition that ∆ = 0, and so for this protocol there is
no freedom to tune the frequencies in this way.
6.6 Angular Multiplexing
6.6.2
231
Capacity
How many modes can be stored in such a multiplexed memory? One constraint is
that it must be possible to resolve one retrieved signal field from another. Therefore
the minimum angle δθ by which the direction of phasematched retrieval must change
between two modes must be greater than the angular divergence of the retrieved
signal field. This angular divergence is set by Fraunhofer diffraction, which yields
the condition δθ & λs /ws — the angle between consecutive modes must exceed the
ratio of the wavelength and beam waist of the signal field.
Equally important is the requirement that there is no ‘cross-talk’ between neighbouring modes — that δθ is large enough that retrieval of an undesired mode is
well-suppressed by its momentum mismatch. An estimate of the magnitude of the
momentum mismatch is given by δθks (see Figure 6.10). This momentum mismatch
is directed approximately perpendicular to the z-axis, so we can approximate the
phase accumulated by the mismatch as δθks ws (working in ordinary units). For good
suppression of unwanted retrieval, this phase should exceed 2π, roughly, and from
this we again derive the condition δθ & λs /ws . Therefore the multimode capacity of
a multiplexed memory is bounded by the number of diffraction limited modes that
can be addressed. For the example in Figure 6.9 (a) this condition implies δθ ∼ 0.01,
which appears to be broadly correct.
The number of modes that can be stored is given by N = ∆θ/δθ, where ∆θ is the
largest angle at which efficient retrieval is possible, as limited by walk-off between
the signal and control fields. We approximate ∆θ as the angle subtended by the
6.6 Angular Multiplexing
232
Figure 6.10 Minimum momentum mismatch. Two signal fields
with momenta k1 and ks differ in their directions by a small angle
δθ. When retrieval for one field is phasematched, retrieval for the
other is not, with the momentum difference shown in red, which
points essentially across the ensemble. It’s magnitude is roughly δθks
(ignoring any dispersion that may slightly alter the lengths of k1 , ks
away from ks ). The total phase accumulated through propagation
across the ensemble, of width ws , is then roughly δθks ws . When
this phase is larger than 2π, phasematching is sufficiently selective to
effectively suppress any accidental retrieval of stored excitations.
control waist across the length of the cell, ∆θ ∼ wc /L (in ordinary units). The
multimode capacity is then given by N ∼ wc ws /λs L. This is essentially equal to
the geometric mean of the Fresnel numbers Fs , Fc associated with the signal and
control fields,
N∼
p
Fs Fc ,
where
Fs,c =
2
ws,c
,
Lλs,c
(6.43)
where we used the fact that the control wavelength λc is generally rather close to
the signal wavelength. The Fresnel number of a beam quantifies the number of
diffraction limited modes it supports, so this result has the interpretation that the
number of modes that may be stored in a multiplexed memory is given by the average
(in the sense of the geometric mean) number of modes supported by the optical fields
used. For the example used in Figure 6.9 (a), the formula (6.43) predicts N ∼ 3,
and this agrees reasonably well with what might be estimated ‘by eye’ (see caption
of Figure 6.10).
6.6 Angular Multiplexing
233
A high capacity multiplexed memory requires wide beams, with Fs,c 1. In
addition, our numerics showed that wc should be at least 2 (better 3) times larger
than the signal beam waist. The feasibility of this kind of memory depends on the
availability of energetic lasers that can maintain a high intensity, even when very
loosely focussed. But laser technology has advanced sufficiently that this is certainly
possible for beams with a width on the order of centimeters, giving multimode
capacities on the order of 100.
In the next chapter we continue to investigate multimode storage, both in the
Raman and EIT protocols covered so far, and in a number of alternative memory
protocols.
Chapter 7
Multimode Storage
At the end of the last Chapter we described a way to store multiple pulses in a single
ensemble memory. We considered scanning the propagation direction of either the
control or the input signal beams: each independently stored pulse represented a
separate spatial mode, characterized by the direction of its wavevector. However, in
many practical situations, it is desirable to fix the alignment of all components in an
optical system. It may therefore be more useful to consider using a different degree
of freedom to define a basis of modes. As discussed at the start of Appendix C, we
might also consider polarization or spectral modes. And in fact since polarization is
only two dimensional, any such consideration of multimode storage must focus on
spectral modes. Equivalently, we might talk about temporal modes, since time and
frequency are Fourier conjugates; they are two interchangeable representations of
the same underlying space. This space can be thought of as the space of all possible
input profiles for the signal field. The method used in the previous Chapters to
7.1 Multimode Capacity from the SVD
235
optimize the storage efficiency of a quantum memory is immediately useful in this
connection. The SVD of the storage kernel provides a ‘natural’ basis for the space
of signal profiles: the input modes {φk }. In this Chapter we apply the SVD to the
study of multimode storage: we show how to optimize and quantify the multimode
capacity of the Raman, EIT, CRIB and AFC memory protocols. The following
treatment expands upon the account published in Physical Review Letters [155] .
7.1
Multimode Capacity from the SVD
Suppose that we would like to store a ‘message’ in a quantum memory. We must
select an alphabet in which to write the message. If we encode the message in an
optical signal, each letter in the alphabet is represented by a mode of the optical
field. The multimode capacity of the quantum memory is the size of the largest
alphabet we can use for our message compatible with efficient storage. In general,
this capacity will depend on the particular alphabet chosen: some modes of the
optical field cannot be stored at all. To take an absurd example, we certainly could
not encode any letters as X-rays! The multimode capacity also depends on the
brightness of the signals used to encode the message. For instance, when two similar
modes are bright, we can distinguish them and encode two letters, whereas when
they are dim we can encode only a single letter (see Figure 7.1).
In a quantum communication protocol, security requires that we encode each
letter using just a single photon (see §1.5 in Chapter 1). This means that there can
be no overlap whatsoever between the modes comprising the alphabet. That is, we
7.1 Multimode Capacity from the SVD
A
236
B
Bright
Dim
Figure 7.1 Bright overlapping modes are distinct. Two “letters”
A and B are encoded with similar optical pulses. When they are
bright, the letters can be distinguished; not so when their intensity
is reduced.
must find an orthonormal basis of modes. The multimode capacity of a quantum
memory in this context is therefore the number of orthonormal modes that can
be stored. It remains true that this capacity depends on the choice of basis. For
example, it may not be the same for time bin modes as it is for frequency modes
(see §C.1 in Appendix C). What, then, is the optimal basis? Which encoding
maximizes the capacity? It is the basis formed by the input modes {φk } determined
from the SVD of the storage kernel K. That is, for any number N , storage of the
first N input modes φ1 , . . . , φN is more efficient than the storage of any other set
of N orthonormal modes. Each input mode φk is stored with efficiency ηk = λ2k . In
the optimal situation of phasematched backward retrieval without dispersion, the
retrieval efficiency is also equal to ηk (see the discussion towards the end of §6.2 in
6). The total memory efficiency for the k th mode is then ηk2 = λ4k . The multimode
capacity is given by the width of the ‘distribution’ of efficiencies (see Figure 7.2).
Below we consider two ways of quantifying this width.
7.1 Multimode Capacity from the SVD
237
Figure 7.2 Visualizing the multimode capacity. The total memory
efficiency — the efficiency of storage followed by retrieval — for the
k th mode is ηk2 = λ4k . These efficiencies define a decaying distribution, and the multimode capacity is related to the width, or ‘variance’
of this distribution. The Schmidt number quantifies this width, but
takes no account of the absolute efficiencies. If we introduce a threshold efficiency ηthreshold , we can define an operational measure N for
the multimode capacity by considering the number of modes with
memory efficiency greater than the threshold.
7.1.1
Schmidt Number
The Schmidt number S is defined as follows [156,157] ,
1
S = P∞
,
ek4
k=1 η
where
η2
ηek2 = P∞k
2.
j=1 ηk
(7.1)
The ηek2 are defined so that they sum to unity, turning them into effective probabilities. In fact, if we consider that these probabilities are the eigenvalues of a quantum
mechanical density matrix, describing a statistically mixed state, then S is the inverse of the purity of this state [158] . From the definition (7.1) it’s clear that S = 1
if there is just one non-zero singular value. In general, if there are N equal singular
values, with all others vanishing, then S = N , so the Schmidt number does indeed
count the number of modes that may be stored with equal efficiency. When the
singular values are not equal, but instead decay in magnitude smoothly as the index
7.1 Multimode Capacity from the SVD
238
k increases, S still provides a sensible measure of the number of coupled modes. The
normalization of the η’s makes S independent of the overall efficiency of the memory:
it measures how many modes are involved in the memory interaction with comparable strength, independent of what that strength is. This makes S an independent
figure of merit for characterizing the performance of a quantum memory. One might
choose to optimize the efficiency (i.e. the optimal efficiency) of a quantum memory,
or the Schmidt number, or both. In most situations, however, it is more useful to
quantify the ability of a memory to store multiple modes efficiently, so that a metric
that combines these two measures is desirable. We introduce such a metric in the
next section. But before abandoning the Schmidt number entirely, we observe that
it is possible to develop an intuition for the value of S — at least whether or not
it is large — by looking at the form of the Green’s function describing the storage
interaction (see Figure 7.3). The reason is that the contribution of a single mode to
the Green’s function necessarily takes the form of a product of two functions; each
with contours orthogonal to the other’s: one function represents the input mode;
the other the corresponding output mode. Now, the Schmidt number is independent
of the basis of modes we consider, because the singular values are. We are therefore
always free to choose a basis of modes comprised of localized pulses. Each mode
is orthogonal to the others if the pulses do not overlap. The product of two such
modes produces a contribution to the Green’s function in the form of a rectangular
‘blob’, with sides parallel to the coordinate axes, and rounded edges if the pulses
have smooth edges. Counting the number of such blobs required to make up the
7.1 Multimode Capacity from the SVD
239
Green’s function provides an estimate of the number of modes required to construct
it. Any feature of the Green’s function that traces out a curve is ‘multimode’, as is
any feature that runs at an angle to the coordinate axes. The Schmidt number is
sometimes useful because it is amenable to this kind of dead reckoning, as we will
see in §7.3.2 below.
(a) Single mode
(b) Multimode
Figure 7.3 The appearance of a multimode Green’s function. Inspection of the contours of the Green’s function describing a memory
interaction offers insight into the degree to which it is multimode, as
quantified by the Schmidt number S. By reconstructing the approximate shape of these contours using non-overlapping rectangles, one
can estimate the number of modes of which it is composed. In (a)
we show an archetypal single mode Green’s function, which looks like
an isolated ‘hill’; a single elliptical contour is drawn. The contour is
roughly approximated by just a single rectangle, indicating that only
one mode contributes to its structure. In (b) we illustrate how curved
or angled contours admit a decomposition involving more modes.
7.1.2
Threshold multimode capacity
To be more quantitative, we introduce a threshold efficiency ηthreshold that allows
us to delineate those modes that are stored and retrieved with acceptable efficiency
from those that are not. We then define the multimode capacity as the largest
number of modes N for which the mean memory efficiency, averaged over all N
7.1 Multimode Capacity from the SVD
240
modes, exceeds ηthreshold . The average memory efficiency of the first k input modes
is given by
Pk
Λk =
2
j=1 ηj
k
.
(7.2)
Then the multimode capacity can be written as
N=
∞
X
k=1
βk ,
where
βk =





1
if Λk > ηthreshold




0
otherwise
.
(7.3)
A more conservative estimate could replace the average efficiency (7.2) with the
minimum efficiency ηk2 . The results presented below are not qualitatively altered by
making this alternative definition, but as N becomes large, this definition represents
an increasingly pessimistic estimate of the performance of a memory for storing
random ‘words’ drawn from the alphabet defined by the {φk }. The capacity defined
in (7.3) is a convenient measure, with a clear interpretation.
In the next section, we investigate the scaling of N for various memory protocols.
We set ηthreshold = 70%; this value makes the form of the scaling apparent for
parameters that are tractable numerically, and indeed experimentally. We return
to the one dimensional model used in Chapter 5, since we are able to write down
explicit expressions for the storage kernels in this case. The simulations presented in
the previous Chapter confirm that the one dimensional approximation works well,
provided that the control field is more loosely focussed than the signal field (see §6.5
in Chapter 7).
7.2 Multimode scaling for EIT and Raman memories
7.2
241
Multimode scaling for EIT and Raman memories
In Chapter 5 we showed that the best possible storage efficiency for a Λ-type quantum memory is found from the SVD of the anti-normally ordered kernel (5.27) [133] .
We repeat it here for convenience,
√ d
0
KA z, z 0 = e−d(z+z )/2 I0 d zz 0 .
2
(7.4)
Here we are using normalized units for the z coordinates. The kernel (7.4) is valid
for both EIT and Raman protocols, in the light-biased limit where the control pulse
is sufficiently energetic (see §5.2.8 in Chapter 5). The eigenvalues of this optimal
kernel are equal to the largest possible values of the storage efficiencies ηk = λ2k ,
so we can calculate the multimode capacity N by diagonalizing KA . This can be
done numerically without difficulty, and we present the results of such a numerical
diagonalization in part (a) of Figure 7.4 below. In part (b) we plot the capacity
predicted by the Raman kernel (5.95) introduced in §5.3 of Chapter 5. In both cases
the capacity rises only slowly with the optical depth d. To explain these results, we
briefly return to the treatment given in §5.2.3 in Chapter 5. There we derived an
approximate expression (5.21) for the storage efficiencies, valid in the limit d 1:
ηk = e−4παk /d ,
(7.5)
7.2 Multimode scaling for EIT and Raman memories
242
√
where αk is the k th zero of the function J0 (2 2πx). The zeros of J0 are distributed
approximately linearly along the real line (see §15.4 in Chapter XV in G.N. Watson’s
famous treatise [159] ). Therefore the αk increase, roughly, with the square of the index
k. Setting αk ∼ qk 2 for some constant q, we find that the index of the mode whose
total efficiency ηk2 falls below ηthreshold is given approximately by
r
k∼
ηthreshold √
× d.
8πq
(7.6)
The multimode capacity N should scale in the same way, so that we expect
N∼
√
d,
(7.7)
for large optical depths, regardless of the threshold ηthreshold . Inspection of Figure
7.4, which was generated by setting ηthreshold = 70%, indeed reveals a scaling of
N≈
√
d/3.
7.2.1
A spectral perspective
The kernel (7.4) is derived by considering the optimal efficiency for absorption into
the excited state |2i, and the square root scaling of N with d can be understood by
considering the bandwidth of the absorption line associated with an ensemble of twolevel atoms. If each atom has a Lorentzian lineshape with natural linewidth 2γ, the
absorption profile of the entire ensemble is given by exponentiating the single-atom
7.2 Multimode scaling for EIT and Raman memories
243
profile,
2dγ 2
.
F (ω) = exp − 2
γ + ω2
(7.8)
The full width at half maximum (FWHM) of F is given by
r
∆ω = 2
2dγ 2
− γ 2 ≈ 2γ
ln 2
r
2d
,
ln 2
(7.9)
where the approximation holds for d 1. An estimate of the multimode capacity is
provided by counting the number of spectral modes with FWHM 2γ that ‘fit’ inside
this absorption profile, which procedure yields
N∼
∆ω √
∼ d.
2γ
(7.10)
We can therefore identify the origin of the square root scaling of N with d as the
scaling of the bandwidth of the absorption of an ensemble with its optical thickness.
This scaling is rather poor: in order to store two modes in such a memory, we must
quadruple the optical depth required for storing a single mode. If this optical depth
were divided up into four separate ensemble memories, we could store four modes
— one mode in each memory — and so the multimode scaling for EIT and Raman
memories is decidedly sub-optimal. The reason is that when the optical depth is
increased, new atoms are added ‘behind’ the old: they absorb at the same frequencies
as the other atoms, and so they improve the absorption at these frequencies, but
they do not provide much coupling at other frequencies — other modes. In the next
7.2 Multimode scaling for EIT and Raman memories
244
section we will show that adding an inhomogeneous broadening to the ensemble
can improve the multimode scaling, by redistributing optical depth over a range of
frequencies. This turns the square root scaling with d into a linear one, consistent
with what one might expect to achieve from operating separate memories in parallel.
And this is essentially the way such broadened protocols work, albeit within a single
physical ensemble.
(a)
(b)
Optimal kernel
Raman kernel
8
6
4
2
0
0
100
200
300
400
500 0
20
40
60
80
100
Figure 7.4 Multimode scaling for Raman and EIT memories. (a):
the multimode capacity found by numerical diagonalization of the
optimal kernel KA in (7.4). This represents the best possible scaling
for both EIT and Raman memories; the capacity scales only with the
square root of the optical depth. (b): we also show the multimode
capacity found by diagonalizing the Raman kernel (5.95), which is
parameterized by the Raman memory coupling C (see §5.3 in Chapter
5). This kernel is valid in the far off-resonant limit — though if the
control pulse intensity is increased sufficiently, the optimal kernel
KA should be used instead. We plot the Raman capacity against C 2 ,
since this quantity is proportional to d. The same square root scaling
with ensemble density is evident. Equivalently, if d is held constant,
the Raman capacity scales with the square root of the control pulse
energy. We used ηthreshold = 70% in both plots.
7.3 CRIB
7.3
245
CRIB
The CRIB memory protocol is introduced in §2.3.3 in Chapter 2. Storage is achieved
by direct absorption into the excited state |2i, which is artificially broadened by
application of an external field. Once storage is complete, the excitation is ‘shelved’
by application of a short, bright control pulse, which transfers the excitation to
the storage state |3i. To implement retrieval, another control pulse transfers the
excitation back to |2i, and the inhomogeneous broadening is ‘flipped’, so that the
atomic dipoles re-phase and the signal field is re-emitted.
The same considerations regarding retrieval discussed in the previous Chapter
apply to this protocol, just as they do to Raman and EIT memories: retrieval in
the forward direction is inefficient due to re-absorption losses [160] , while backward
retrieval is vulnerable to phasematching problems. However, in this protocol, the
control pulse is applied after the signal has been absorbed, so it is possible to
distinguish the two fields temporally. Spectral and spatial filtering is therefore less
important, and so it is feasible to use an ensemble where the states |1i and |3i are
degenerate. This allows for efficient collinear storage, followed by phasematched
backward retrieval. Since there is no dispersion on resonance, the stored excitation
has no spatial phase, and the retrieval efficiency is equal to the storage efficiency,
2
with ηcombined = ηstorage
. In the following, we restrict our attention to this situation,
which is optimal.
7.3 CRIB
7.3.1
246
lCRIB
We first consider lCRIB, in which the broadening is applied longitudinally [84,161,162] .
The resonant frequency of the |1i ↔ |2i transition varies linearly along the z-axis.
Since the control field is only applied after the signal field has been resonantly
absorbed, the equations of motion for lCRIB are given by the system (5.106), with
Ω = 0, and with the spatial variation of the detuning included.
√
∂z A(z, τ ) = − dP (z, τ ),
∂τ P (z, τ ) = −Γ(z)P (z, τ ) +
where Γ(z) = 1 − i∆(z), with ∆(z) = ∆0 z −
1
2
√
dA(z, τ ),
(7.11)
. Here we have returned to the
normalized units of Chapter 5, with all frequencies measured in units of γ, and with
z running from 0, at the entrance face of the ensemble, up to 1, at the exit face. The
width of the applied spectral broadening is ∆0 . As usual, we solve these equations
by applying a unilateral Fourier transform over the z-coordinate. Using the formula
(D.27) from Appendix D, we obtain the transformed system
√
e − √1 Ain = − dPe,
−ik A
2π
√
e
∂τ Pe = − 1 − ∆0 ∂k + i ∆20 Pe + dA,
(7.12)
7.3 CRIB
247
where Ain is the temporal profile of the incident signal field. Solving the first equation
e and substituting the result into the second equation yields
for A
(∂τ − ∆0 ∂k ) Pe = − 1 +
i ∆20
+
i kd
√
d
Pe + i √
Ain .
2πk
(7.13)
Now if we temporarily define the composite variable s = k + ∆0 τ , we can replace
the derivatives on the left hand side with a single time derivative,
√
∂τ Pe = −f (s, τ )Pe + i √
d
Ain ,
2π(s − ∆0 τ )
(7.14)
where the partial derivative ∂τ is taken with s held constant, and where we have
defined the function
d
.
f (s, τ ) = 1 + i ∆20 + i s−∆
0τ
(7.15)
Integrating (7.14) gives the solution
Peout (s) = Pein (s)e−
RT
−∞
f (s,τ ) dτ
√ Z T
RT
d
0
0
+ i√
Ain (τ )e− τ f (s,τ ) dτ dτ,
2π −∞
(7.16)
where Pout (s) = P (s, T ) is the atomic excitation at the moment τ = T that the
short control pulse is applied, marking the end of the storage interaction. We do not
explicitly model the atomic dynamics induced by the control; we simply assume that
it is sufficiently intense to transfer — effectively instantaneously — all the atomic
excitations into the storage state |3i, so that Pout −→ Bout . We set Pin = 0, since no
atoms are excited at the start of the storage process. We then perform the integral
7.3 CRIB
248
in the exponent of the integrand in the second term, and convert back from s to k,
to arrive at the following expression for the lCRIB storage map in k-space,
Z
∞
e
K(k,
T − τ )Ain (τ ) dτ,
eout (k) =
B
(7.17)
−∞
e defined by
with the kernel K
√
d
e
K(k, τ ) = i √ e−(1+i∆0 /2)τ × k id/∆0 (k + ∆0 τ )−id/∆0 −1
2π
(7.18)
e = 0 when τ < 0. The multimode scaling of lCRIB is determined
for τ ≥ 0, with K
by the singular values of this kernel. Unfortunatey we cannot extract these singular
values directly because there is a singularity at k = −∆0 τ . A number of alternatives
are open to us. First, we can return to the equations of motion, and construct the
Green’s function in (z, τ )-space directly by numerical integration (see §5.4 in Chapter
5 and §E.5 in Appendix E). This kernel is not singular, and is therefore amenable to
a numerical SVD. We’ll refer to this as simply the ‘numerical method’. Second, we
can remove the singularity in (7.18) by applying a small regularization to k. That is,
we replace k by k − i, where is some small real number. This shifts the singularity
off the real axis, making the kernel well-behaved. But it also changes the singular
values, since it is not a unitary transformation. We fix this with the following
procedure. We apply a numerical inverse Fourier transform from k-space back to zspace. This can be performed very efficiently with a fast Fourier transform (FFT), an
implementation of which is standard in Matlab. The Fourier transform is unitary, so
7.3 CRIB
249
it does not affect the singular values. Next, we multiply the result by the exponential
factor e−z . By the shift theorem (see §D.3.4 in Appendix D), this compensates for
the regularization, so that the result is equal to the Fourier transform of (7.18)
without the regularization applied. Because the Fourier transform is unitary, we
can extract the singular values of (7.18) by taking a numerical SVD of the kernel
generated by this procedure. We will refer to this as the ‘Fourier method’.
We use both of these methods below. But to gain some insight into the form of
the multimode scaling of lCRIB, we now introduce a third approach. The approach is
only valid for very large broadenings, but it clarifies the scaling behaviour exhibited
by the numerical techniques just described.
7.3.2
Simplified Kernel
We perform a series of unitary transformations that simplify the form of the kernel
(7.18), while leaving the singular values unchanged. The first of these transformations is to trop the exponential factor iei∆0 τ /2 , since it is just a phase rotation. Next
we drop the factor k id/∆0 . This is, again, a pure phase rotation, which is well defined
as long as ∆0 6= 0 and k 6= 0. The second of these conditions arises because there is a
logarithmic singularity in the phase of k id/∆0 at k = 0. The effect of this singularity
is small when d/∆0 is small, so in the following we assume a large broadening, with
∆0 d. The resulting kernel is
√
d
e
K(k,
τ ) = √ e−τ (k + ∆0 τ )−id/∆0 −1 .
2π
(7.19)
7.3 CRIB
250
We now take the inverse Fourier transform, from k-space back to z-space. I have to
confess to not knowing how to perform this transform, but Mathematica provides
an answer! Combining this with the shift theorem gives
√
K(z, τ ) = α (d/∆0 )
de−τ z id/∆0 ei∆0 zτ ,
(7.20)
where α is given by
1
α(x) = − e−πx/2 sinh(πx)Γ(ix),
π
(7.21)
with Γ denoting the Euler Gamma function (not the complex detuning!). Finally,
we note that the factor z id/∆0 is another pure phase rotation, well-defined if ∆0 6= 0
and z 6= 0. We can drop it without affecting the singular values, again in the limit
∆0 d, which gives the simple kernel,
√
K(z, τ ) = α(d/∆0 ) de(i∆0 z−1)τ .
(7.22)
Now we form the anti-normally ordered kernel KA , by integrating the product of
two of these kernels from τ = 0 to τ = ∞ (see §3.3.1 in Chapter 3),
KA (z, z 0 ) =
d
1
|α(d/∆0 )|2 ×
.
∆0
2/∆0 − i(z − z 0 )
(7.23)
This kernel has the simple structure we have been seeking. It takes the form of a
Lorentzian peak centred on the line z = z 0 , with a width set by ∆0 , and a height
determined by the ratio d/∆0 . We can draw two conclusions from the form of this
7.3 CRIB
251
kernel.
First, in the limit of large broadening ∆0 d, the Schmidt number becomes
independent of the optical depth, being determined only by ∆0 . This follows from
the fact that the functional form of (7.23) does not depend on d, which only affects its
overall magnitude. The Schmidt number does not depend on this overall magnitude,
because of the normalization of the η’s in (7.1), and so S is independent of d.
The contours of (7.23) form a strip along the diagonal line z = z 0 . Application
of the estimation technique described in Figure 7.3 suggests that the multimode
capacity is proportional to the ratio of the length to the width of the strip, so that
S ∝ (2/∆0 )−1 ∼ ∆0 (see Figure 7.5). Numerics confirm this to be the case: the
Schmidt number of an lCRIB memory rises linearly with the applied broadening
(see Figure 7.6).
The second conclusion we can draw from the structure of (7.23) is that the
threshold multimode capacity N rises linearly with the optical depth d — a significant improvement over the square root scaling derived for EIT and Raman memories
in §7.2. To see why, consider a situation where the optimal memory efficiency η12
exceeds the threshold efficiency ηthreshold by a reasonable margin, so that N ≈ S
(see Figure 7.7). If we double the applied broadening, the Schmidt number doubles.
But the ratio d/∆0 is then halved, so that the overall efficiency falls below ηthreshold ,
and N ≈ 0. In fact, the function α(x) ≈ 1 for x 1, so that halving d/∆0 approximately halves the ηk , and divides the total memory efficiencies ηk2 by four. To
bring the overall efficiency back to its previous value, above the threshold, we must
7.3 CRIB
252
(a) Many modes
(b) Few modes
1
1
0.5
0.5
00
0.5
1
00
0.5
1
Figure 7.5 Scaling of Schmidt number with broadening. In (a)
and (b) we plot |KA | for large and small broadenings of ∆0 = 100
and ∆0 = 10, respectively, with d = 30 in both cases. Below these
plots we illustrate the mode-counting procedure described in Figure
7.3, which provides a way to understand why the Schmidt number
increases linearly with the applied broadening (see Figure 7.6).
double the optical depth. The multimode capacity N is then doubled, because we
have both doubled S and maintained the correspondence N ≈ S by keeping the
overall efficiency above ηthreshold . This argument is quite general: we can increase
the number of modes contributing to the storage interaction by increasing the applied broadening, but the coupling is ‘shared’ between all these modes, so we must
increase the optical depth at the same time in order to maintain efficient storage
over all the modes. If our aim is to maximize N , there is an optimal value for the
ratio d/∆0 , which depends on ηthreshold . If the ratio is too large, we can afford to
increase the broadening and introduce more modes without bringing the efficiency
below the threshold. Converseley, if the ratio is too small, we should sacrifice some
7.3 CRIB
253
(a)
(b)
1
30
20
0.5
10
0
0
100
200
300
0
0
100
200
300
Figure 7.6 Comparison of the predictions of the kernels (7.23) and
(7.18). (a): the optimal total memory efficiency η12 predicted by the
simplified kernel (7.23) (green dotted line) is plotted alongside the
efficiency predicted by the kernel (7.18), using the ‘Fourier method’
(blue solid line), as a function of the applied broadening. The optical
depth is set at d = 30; the simplified kernel compares well with the
Fourier method for ∆0 & 3d. (b): the Schmidt number predicted by
the two kernels. The agreement between the two, even at the boundaries of the regime of validity of the simplified kernel, is excellent.
The linear scaling of S with ∆0 expected from the form of (7.23) is
clear.
modes and reduce the broadening to boost the efficiency above the threshold. This
analysis shows that the multimode capacity N of an lCRIB memory scales linearly
with d, provided that the broadening ∆0 is increased linearly with d at the same
time.
The preceding discussion applies in the limit ∆0 d. We confirm the persistence
of linear multimode scaling outside this regime using the direct numerical method
described earlier, in which we construct the Green’s function by integrating the
equations of motion. The result is plotted in Figure 7.8 at the end of §7.3.3 below,
alongside the results for EIT and Raman shown earlier, and the results for tCRIB,
which we now turn to.
7.3 CRIB
254
(a)
(b)
(c)
Figure 7.7 Understanding the linear multimode scaling of lCRIB.
In the limit of large broadening, the Schmidt number of the lCRIB
storage interaction depends linearly on the width of the applied spectral broadening, but the overall storage efficiency is determined by
the ratio d/∆0 . To double the multimode capacity N , given a fixed
threshold efficiency ηthreshold , we need to double the Schmidt number,
while keeping the overall efficiency the same. The process is shown
in parts (a) to (c). We begin with η12 > ηthreshold , so that S ≈ N . In
(b) we double the spectral broadening, which doubles S, but reduces
the efficiency (in fact, by roughly a quarter). To return to the same
efficiency as in (a), in part (c) we double the optical depth. This
has no effect on S, but it returns the ratio d/∆0 to the same value
as in part (a). By doubling the optical depth, we have suceeded in
doubling the multimode capacity. These arguments explain why the
threshold capacity N depends linearly on d for lCRIB.
7.3.3
tCRIB
In a tCRIB memory, the direction of the broadening is perpendicular to the z-axis.
That is, the resonant frequency of the atoms varies across the ensemble (see part
(a) of Figure 2.10 in Chapter 2). The theoretical treatment requires that we divide
the ensemble into frequency classes, where all the atoms in one frequency class have
the same detuning from the signal field carrier frequency. Thorough treatments are
given by Gorshkov et al. [163] , and also by Sangouard et al. [160] , and we adapt their
techniques to our purpose, which is to derive the Green’s function for the tCRIB
memory interaction.
Suppose that the applied broadening produces an inhomogeneous line with spec-
7.3 CRIB
255
tral absorption profile p(∆). This means that a proportion p(∆)d∆ of the the atoms
have their resonant frequencies shifted by ∆ away from their nominal frequency. The
profile is normalized, so that
Z
p(∆) d∆ = 1.
(7.24)
The total optical depth d is divided amongst all the frequency classes, so that the
optical depth contributed by atoms detuned by ∆ is d(∆)d∆ = dp(∆)d∆. The
equations of motion for the storage process are found by a straightforward generalization of the system (7.11) to the case of multiple frequency classes (and with no
variation of ∆ with z in this case, of course),
Z p
∂z A = −
d(∆)P (∆) d∆,
∂τ P (∆) = −Γ(∆)P (∆) +
p
d(∆)A.
(7.25)
Here we have emphasized the functional dependence of the complex detuning Γ(∆) =
1 − i∆, and we have defined P (∆) as the slowly varying polarization associated with
the frequency class of atoms detuned by ∆. We quickly encounter difficulties if we
attempt to solve this system of equations with our usual trick of Fourier transforming
over z. Instead, we get to the solution by applying a unilateral Fourier transform
over time, from τ to ω. Note that here ω is the frequency conjugate to the retarded
time τ ; it certainly has no relation to the integrated Rabi frequency used in Chapter
7.3 CRIB
256
5! The transformed equations are given by
e = −
∂z A
Pin (∆)
−iω Pe(∆) − √
2π
Z p
d(∆)Pe(∆) d∆,
= −Γ(∆)Pe(∆) +
p
e
d(∆)A.
(7.26)
Solving the second equation, and substituting the result into the first yields the
following equation for the signal field,
r
e z) = −
[∂z + df (ω)] A(ω,
d
2π
Z p
p(∆)Pin (z; ∆)
d∆,
Γ(∆) − iω
(7.27)
where we have defined the lineshape function
Z
f (ω) =
p(∆)
d∆,
1 − i(∆ + ω)
(7.28)
which is essentially the convolution of the Lorentzian spectral response of the atoms
with the inhomogeneous profile p(∆). Recall that in our normalized units the natural
atomic linewidth is defined to be equal to 1. We can immediately solve for the signal
field,
r
e z) = e−df (ω)z A
ein (ω) −
A(ω,
d
2π
Z p
Z
p(∆)
0
z
0
e−df (ω)(z−z )
Pin (z 0 ; ∆) 0
dz d∆.
Γ(∆) − iω
(7.29)
ein is the spectrum of the incident signal field. Now that we are in possession of
Here A
a solution for the signal, it is a matter of algebra to construct the Green’s function.
7.3 CRIB
257
But it is not enough to find the storage map alone. We can write down an expression
for the atomic excitation at the end of the interaction, but since it is distributed over
all frequency classes, it is not clear what we should optimize for efficient storage.
There is no single spin wave whose norm represents the storage efficiency. Therefore,
we construct the kernel Ktotal that relates the input signal to the retrieved signal
field (see §3.4 in Chapter 3). The modes found from the SVD of this kernel have a
clear interpretation as those modes that are eventually retrieved from the memory.
To find this kernel, we first solve for the polarization Pout at the end of the storage
process, setting Pin = 0. This requires taking an inverse Fourier transform from ω
back to τ = T , where T marks the end of the storage interaction,
1
Pout (z, ∆) = √
2π
Z
∞
e−iωT Pe(ω, z; ∆) dω,
(7.30)
−∞
where, using the second line of (7.26),
Pe(ω, z; ∆) =
=
p
e z)
d(∆)A(ω,
Γ(∆) − iω
p
d(∆)e−df (ω)z
ein (ω),
×A
Γ(∆) − iω
(7.31)
where in the second line we used (7.29). This completes the description of the storage
process. At time τ = T , the control pulse shelves all the excited atoms. To describe
retrieval of the excitations, we again use (7.29). As usual we use a superscript r to
identify quantities associated with the retrieval process. There is no input field for
er = 0. The retrieved signal field, at the exit face of the ensemble with
retrieval, so A
in
7.3 CRIB
258
z r = 1, is given by
r
erout (ω r ) = −
A
d
2π
1
Z p
Z
r
r
p (∆ )
e−df
r (ω r )(1−z r )
0
r (∆r , z r )
Pin
dz r d∆r .
Γ(∆r ) − iω r
(7.32)
We ‘stitch’ together the storage and retrieval interactions by making the identification
r
Pin
(∆r , z r ) = Pout (∆, z).
(7.33)
This says that the atomic polarization is the same at the end of the storage process
as it is at the start of the retrieval process. But in between storage and retrieval, the
inhomogeneous profile is flipped, so that red-detuned atoms become blue-detuned
and vice-versa. This is crucial for reversing the atomic dynamics so that they re-emit
the signal field. We model this flipping of the detunings by setting
∆ = −∆r ,
(7.34)
so that frequency classes with positive and negative detunings swap places. For
backward retrieval with no phase mismatch — the optimal situation — we also
swap the z-coordinate:
z = 1 − zr.
(7.35)
Note that in this case the frequency ω r is conjugate to the retarded time τ r , which is
the time coordinate in a frame moving at the speed of light backwards with respect to
the initial z-axis used for the storage process. We now combine the relations (7.35)
7.3 CRIB
259
and (7.34) with (7.33), and substitute this into (7.32), using (7.30) and (7.31). The
result is
er (ω r ) =
A
out
Z
∞
e total (ω r , ω)A
ein (ω) dω,
K
(7.36)
−∞
where the Green’s function for the memory interaction is given by
r
−iωT
e total (ω , ω) = −e
K
Z
p
Z 1
pr (∆r )p(−∆r )
d
r
r
r
r
e−d(1−z )[f (ω )+f (ω)] dz r .
d∆
×
r
r
r
[Γ(∆ ) − iω ][Γ(−∆ ) − iω]
2π 0
(7.37)
Now we assume that the inhomogeneous profile is the same for both the storage
and retrieval processes, and symmetric about the unbroadened resonance. then
pr (∆r ) = p(∆r ) = p(−∆r ), and the lineshape is also unchanged, f r = f . Performing
the integrals over ∆r and z r , we find
r
−d[f (ω )+f (ω)] − 1
e total (ω r , ω) = 1 e
K
,
2π 2 + i(ω r + ω)
(7.38)
where we have dropped the exponential factor of e−iωT , since it represents only
an unimportant phase rotation. Here we note that a nearly identical treatment
was first given by Gorshkov et al. [163] . The crucial difference for us is that we
used a unilateral Fourier transform, rather than a Laplace transform, to solve the
ein and A
er have a natural
equations. This means that the transformed amplitudes A
out
interpretation as the spectra of the input and retrieved signal fields. The norm of
the signal spectrum is the same as the norm of the temporal profile, by Parseval’s
theorem (or, alternatively, by energy conservation), and so the SVD of the kernel
7.3 CRIB
260
Ktotal in (7.38) tells us about the storage efficiency of the memory, and indeed, its
multimode capacity.
Now that we have found an explicit form for the Green’s function describing
tCRIB, the multimode capacity can be found by taking its SVD: the singular values
of Ktotal are equal to the ηk . In Figure 7.8 we plot the resulting prediction for the
multimode scaling of tCRIB. We used a rectangular broadening profile with total
width ∆0 ,
p(∆) =





1/∆0
if |∆| ≤ ∆0 /2,




0
otherwise.
(7.39)
And we optimized N with respect to the width ∆0 , using a threshold ηthreshold =
70%. It is clear that the scaling of N with d is linear. Furthermore, it is the same
as the scaling of lCRIB, whose multimode capacity is also plotted. This suggests
that the multimode capacity of both CRIB protocols is identical. The scaling of
unbroadened EIT and Raman protocols, as derived from (7.4), is shown for comparison. Clearly CRIB dramatically outperforms equivalent unbroadened protocols:
given the same optical depth — the same total number of atoms — more modes of
the optical field can be stored by adding a controlled broadening. We found that for
both tCRIB and lCRIB, the multimode capacity scaled roughly as N ∼ d/25, and
the optimal width for the spectral broadening scaled as ∆opt
0 ∼ 9d/5. This confirms
the validity of the arguments given at the end of §7.3.2 for lCRIB.
Just as the poor multimode scaling of unbroadened protocols can be explained
by considering the absorption bandwidth of a homogeneous ensemble (see §7.2.1), so
7.3 CRIB
261
we can understand the improved scaling of CRIB by considering the inhomogeneous
linewidth. Consider the storage of a spectral mode with bandwidth 2γ. An optical
depth of order 10 is required to efficiently absorb the incident light. To store N
such spectral modes ‘side by side’ in frequency, we should have a total optical depth
d ∼ 10N , spread over a spectral width ∆0 ∼ 2γN . That is, the multimode capacity
scales linearly with d, and so does the broadening ∆0 , which is precisely what we have
found to be the case for CRIB. And it is clear why there is an optimal broadening:
if ∆0 is too large, there is insufficient optical depth over the bandwidth of a mode
for efficient absorption.
In the next section, we consider a modification to the Raman protocol which
takes advantage of spectral broadening to improve its multimode scaling.
20
10
0
0
100
200
300
400
500
Figure 7.8 Multimode scaling for CRIB memories. The multimode capacity N is shown as a function of the total optical depth d
for tCRIB (green dashed line), lCRIB (red dotted line) and, for comparison, unbroadened EIT and Raman protocols (solid blue line).
For the tCRIB calculation, a numerical SVD was applied directly
to the kernel (7.38). For lCRIB, we constructed the storage kernel
by solving the equations of motion (7.11) numerically — we found
this to be the most reliable method, and of course no approximations
are required. Nonetheless some numerical error is apparent at large
broadenings, since the numerical problem becomes increasingly stiff
as ∆0 increases. For both the CRIB protocols, we optimized N over
∆0 , and as expected, the optimal broadening width was found to
scale linearly with d. We set ηthreshold = 70% for all calculations.
7.4 Broadened Raman
7.4
262
Broadened Raman
Is it possible to improve the multimode capacity of the EIT and Raman protocols by
incorporating a spectral broadening? Here we consider a simple modification to the
Raman protocol, in which a longitudinal broadening is applied to the storage state.
That is, the energy of the state |3i varies linearly along the z-axis, covering a range
of frequencies ∆0 . The treatment is very similar to that of lCRIB given earlier,
and we will see that it is indeed possible to recover the linear multimode scaling
characteristic of CRIB. As usual, we start with the equations of motion describing
one dimensional propagation. Adapting the system (5.106) to the present scenario,
with an added broadening, we have
√
∂z A(z, τ ) = − dP (z, τ ),
∂τ P (z, τ ) = −ΓP (z, τ ) +
∂τ B(z, τ ) = i∆0 z −
1
2
√
dA(z, τ ) − iΩ(τ )B(z, τ ),
B(z, τ ) − iΩ∗ (τ )P (z, τ ).
(7.40)
The first term on the right hand side of the Heisenberg equation for B just describes
a position-dependent energy shift of the storage state. We make no attempt to solve
these equations exactly. Rather, we proceed directly to study the adiabatic limit,
in which any driven dynamics are much slower than the timescale 1/∆ set by the
detuning. In this limit, we can eliminate the polarization P , by setting the left hand
side of the second equation to zero (see §5.3.3 in Chapter 5, and also the papers by
Gorshkov et al. [133] ). Solving the resulting algebraic equation for P , and substituting
7.4 Broadened Raman
263
the result into the other two equations yields the system
√
d
Ω(τ ) d
∂z +
A(z, τ ) = i
B(z, τ ),
Γ
Γ
√
|Ω(τ )|2
Ω∗ (τ ) d
1
∂τ +
− i∆0 z − 2 B(z, τ ) = −i
A(z, τ ).
Γ
Γ
(7.41)
As we did for lCRIB, we now apply a unilateral Fourier transform over z. The
adiabatic equations of motion in k-space are then
e τ) =
A(k,
√Γ Ain (τ )
2π
√
e τ)
+ iΩ(τ ) dB(k,
d − ikΓ
√
2
∗
∆0 |Ω(τ )| e
Ω (τ ) d e
∂τ − ∆0 ∂k + i
+
B(k, τ ) = −i
A(k, τ ).
2
Γ
Γ
,
(7.42)
Again we encounter the combined derivatives ∂τ − ∆0 ∂k , which we deal with by
transforming from the coordinates (k, τ ) to (s, τ ), where s = k+∆0 τ . The derivatives
can then be replaced with ∂τ , where now s is held constant. Substituting the first
equation of (7.42) into the second, and integrating, we arrive at the solution
−
eout (s) = B
ein (s)e
B
R∞
−∞ f (s,τ ) dτ
r
−i
d
2π
Z
∞
−∞
R∞
Ω∗ (τ )
0
0
e− τ f (s,τ ) dτ Ain (τ ) dτ,
d − iΓ(s − ∆0 τ )
(7.43)
where the function f is given by
∆0 |Ω(τ )|2
d
+
1−
.
f (s, τ ) = i
2
Γ
d − iΓ(s − ∆0 τ )
(7.44)
7.4 Broadened Raman
264
ein = 0. At the end of the storage interaction,
To model the storage process, we set B
at time τout −→ ∞, the coordinate s is given by s = k + ∆0 τout . That is, k is related
eout , so it does not affect
to s by a constant offset. This does not affect the norm of B
the efficiency of the memory. In the following we therefore make the replacement
s −→ k, since k has the clearer physical meaning: it is the spatial frequency of the
spin wave. We also drop the first term i∆0 /2 from f , since it represents only a phase
rotation, which also does not affect the memory efficiency. The storage map for a
broadened Raman memory can then be written as
Z
∞
eout (k) =
B
e
K(k,
τ )Ain (τ ) dτ,
(7.45)
−∞
where the k-space storage kernel is given by
r
e
K(k,
τ ) = −i
R
1 ∞
d ∗
0 2
0
0
Ω (τ )g(k, τ )e− Γ τ |Ω(τ )| [1−dg(k,τ )] dτ ,
2π
(7.46)
with g defined by
g(k, τ ) =
1
.
d − iΓ(k − ∆0 τ )
(7.47)
Note that the broadening introduces a timescale into the dynamics that is not set
by the control field, so we cannot conveniently write this kernel in terms of the
integrated Rabi frequency used in Chapter 5 (see (5.46)). But if we set ∆0 = 0,
(7.46) does indeed reduce to the standard adiabatic storage kernel (5.74) for a Λ-type
memory.
7.4 Broadened Raman
265
We can extract the singular values from (7.46) with a numerical SVD. This can be
a little problematic, since for large detunings the kernel becomes nearly singular. But
it is easy to resolve this issue by introducing a regularization, Fourier transforming,
and then compensating — this ‘Fourier method’ is described at the end of §7.3.1
above, where we applied it to the k-space storage kernel for lCRIB, which is also
singular. In Figure 7.9 we show how the square root scaling of the unbroadened
protocol is transformed into linear scaling upon application of a broadening. To
see how this scaling arises from the structure of the storage kernel, we repeat the
arguments used for lCRIB which allow us to simplify the kernel. We consider the
special case of control field with a rectangular profile, since this allows us to perform
e
the integral in the exponent of K,
Ω(τ ) =





Ωmax
for 0 ≤ τ ≤ T ,




0
otherwise.
(7.48)
The kernel evaluates to
r
e
K(k,
τ ) = −i
i dΩ22max
Γ ∆0
d
g(k,
τ
)
2
Ωmax g(k, τ )e−Ωmax T /Γ
.
2π
g(k, T )
(7.49)
In the Raman limit ∆ 1, we can write Γ ≈ −i∆, to obtain
2
2
C∆ −iΩ2max T /∆
i C
1−i C
e
K(k,
τ ) = −i √
e
× [g(k, T )] ∆0 T × [g(k, τ )] ∆0 T ,
2πT
(7.50)
7.4 Broadened Raman
266
where C is the Raman coupling (this is defined in (5.94) in §5.3 of Chapter 5). In
this far-detuned limit, the exponential factor is a pure phase rotation that we are
free to remove, and so is the factor involving g(k, T ), since g is purely real and it is
here raised to a purely imaginary power. As in our treatment of lCRIB, there is a
logarithmic singularity at g(k, T ) = 0, but its effect becomes negligible in the limit
∆0 C 2 /T of large broadening. Dropping these terms, and some other spurious
phase factors, we find
C
e
K(k,
τ) = √
2πT
d
− k − ∆0 τ
∆
i
C2
∆0 T
−1
.
(7.51)
Taking the inverse Fourier transform, with the help of Mathematica, and the shift
theorem, and dropping a further phase factor, we find the z-space kernel
K(z, τ ) = α C 2 /∆0 T
p
C 2 /T e−i∆0 zτ ,
(7.52)
where the function α is defined in (7.21) in §7.3.2 above. This kernel is nearly
identical in form to (7.22). Instead of being damped in time by the factor e−τ due
to the spontaneous lifetime of the excited state, the kernel is instead truncated at
τ = T , when the control field is switched off. But the functional form is the same:
an exponential parameterized only by the spectral broadening ∆0 . The magnitude
of the kernel is also the same as for (7.22), except the optical depth d has been
swapped for the quantity C 2 /T . By analogy with the facts we know for lCRIB, we
therefore conclude the following. First, in the limit of large broadening, the Schmidt
7.4 Broadened Raman
267
number of the Raman memory depends only on ∆0 . Second, the multimode capacity
N , given some threshold efficiency, scales linearly with C 2 /T , provided that as this
quantity is increased, the applied broadening ∆0 is also increased proportionately.
Note finally that C 2 ∝ d, so the multimode capacity scales linearly with the optical
depth, as it does for CRIB.
These assertions are confirmed by the results of a numerical SVD performed on
the kernel (7.46) using the ‘Fourier method’ described above. We used a Gaussian
control pulse, but the shape of the control makes no difference to the multimode
scaling in the adiabatic limit. This is to be expected, since in this regime the dynamics adiabatically follow the control, and so its temporal profile only affects the
shapes of the input modes, not the efficiency of the memory. For the parameters
shown, we found N ∼ d/300, with an optimal broadening width of ∆0 ∼ d/77.
We should remark that the multimode capacity is much smaller than that found
for CRIB, or even than is predicted by the optimal kernel KA in (7.4). This is
because the memory operates far off resonance, so that a much higher optical depth
is required to achieve a strong interaction. Of course the other advantages of Raman storage — for instance, broadband capability, tunability, and insensitivity to
unwanted inhomogeneous broadening — are retained in the present scheme. Since
investigating this protocol, a demonstration has been implemented by Hétet et al.
in Canberra [164] , although only a single optical mode was stored and retrieved: this
is quite difficult enough to start with! An analysis of this protocol was also recently
conducted by Moiseev and Tittel [165] .
7.5 AFC
268
10
5
0
0
1000
2000
3000
Figure 7.9 The multimode scaling of a broadened Raman protocol. The blue solid line shows the multimode capacity calculated
from the kernel (7.46) using the ‘Fourier method’, optimized over the
width ∆0 of the applied broadening. The green dashed line shows
the square-root scaling obtained if ∆0√is set equal to zero. We used a
2
Gaussian√control pulse, with Ω(τ ) = 10de−(τ /0.1) , and a detuning
of ∆ = 90d. These parameters maintain the adiabatic condition
that ∆ Ωmax , so that we remain in the Raman limit as the coupling is increased. We used a threshold efficiency of 70% for both
calculations.
The last memory protocol we consider is the AFC memory protocol proposed by
Afzelius et al. in Geneva [91] , which is introduced in §2.3.4 of Chapter 2.
7.5
AFC
In the AFC protocol, an ensemble with a naturally broad inhomogeneous absorption
line is prepared by optical pumping, producing an atomic frequency comb. That is,
atoms are removed (or pumped into a ‘shelf’ state) that have resonant frequencies
lying between the ‘teeth’ of a spectral comb, so that the ensemble only absorbs light
at the evenly spaced frequencies of the comb teeth. The great advantage of this
approach, is that the broad spectral bandwidth covered by the ensemble absorption arises naturally. Adding more teeth to the comb to increase the absorption
7.5 AFC
269
bandwidth only requires that fewer atoms are removed, the density or size of the
ensemble need not be increased. This should be contrasted with CRIB, in which
broadening the absorption bandwidth requires an increase in the total optical depth,
if the same level of absorption is to be maintained. As a result, the multimode capacity of AFC does not depend on the density of the ensemble, making it by far the
most ‘multimode’ protocol yet proposed.
We model AFC using precisely the same approach as we used for tCRIB. We
assume that we have succeeded in preparing an ensemble so that it has an inhomogeneous absorption profile that takes the form of a series of M equally spaced
resonances, covering a total spectral width ∆0 , each with optical depth d,
p(∆) =
M
X
δ(∆ − δj ),
where
δj = ∆0
j=1
Note that we have defined p so that
R
1
j−1
−
.
M −1 2
(7.53)
p(∆) d∆ = M . The total optical depth associ-
ated with the entire frequency comb is then dtotal = M d. We reserve the designation
d for the optical depth associated with a single tooth of the frequency comb, since
this is set by the density of the ensemble. This definition allows direct comparison
of AFC with the other protocols studied in this Chapter, where d quantifies the
physical resources — density, length — required to build the memory.
The lineshape function for AFC is a sum of Lorentzian lines,
f (ω) =
M
X
j=1
1
1 − i(δj + ω)
(7.54)
7.5 AFC
270
(again, recall that the ‘1’ in the denominator represents the natural atomic linewidth
in our normalized units). We construct the Green’s function for the AFC memory
in the same way as we did for tCRIB. The only difference is that the inhomogeneous
profile is not flipped around for retrieval: the discrete structure of the comb means
that the phase of the atomic dipoles undergoes periodic revivals, without requiring
any external meddling. We still consider phasematched retrieval in the backward
direction, however, since this is the optimal situation. The equivalent expression to
(7.37) for AFC is therefore
e total (ω r , ω) = −e−iωT
K
Z
d
p(∆)
d∆×
r
[Γ(∆) − iω ][Γ(∆) − iω]
2π
Z
1
e−d(1−z
r )[f (ω r )+f (ω)]
dz r ,
0
(7.55)
where we used ∆r = ∆, pr = p and f r = f . Performing the integrals, and dropping
the unnecessary phase factor, we find the Green’s function for AFC to be
r
r
−d[f (ω )+f (ω)] − 1
e total (ω r , ω) = 1 f (ω ) − f (ω) × e
K
.
2π
ωr − ω
f (ω r ) + f (ω)
(7.56)
The first quotient on the right hand side exhibits a removable singularity at ω r = ω,
but this can be dealt with using L’Hôpital’s rule:
−2df (ω) − 1
e total (ω, ω) = 1 f 0 (ω) e
,
K
2π
2f (ω)
where
f 0 (ω) =
M
X
j=1
iω
.
[1 − i(δj + ω)]2
(7.57)
In Figure 7.10 we show the multimode scaling for AFC derived from the SVD of this
kernel. In part (a) we show the multimode capacity N as function of the tooth optical
7.5 AFC
271
depth d, for various numbers of comb teeth. For each number of teeth M , the square
root scaling characteristic of an unbroadened memory is apparent. But it is possible
to increase the multimode capacity arbitrarily, just by adding more teeth to the
comb. In principle, the multimode capacity of this memory protocol is infinite! Of
course, there is a limitation in practice. First, the number of modes stored can never
exceed the number of atoms in the ensemble. A more important restriction however
comes from the spectral width of the initial inhomogeneous line. To achieve efficient
retrieval, the teeth comprising the comb must be ‘well-separated’, so that they do
not overlap in frequency. If they overlap, the re-phasing of the atomic dipoles will
not be complete, and there will be only partial retrieval of the signal field. Because
the lineshape function f is a sum of Lorentzians, which are functions that do not
have compact support, a rather large frequency separation between the comb teeth
is desirable. Therefore, as more teeth are added to the comb, the total width ∆0 of
the comb must be increased, in order to accomodate the increased number of teeth
with the same separation between them. Eventually, the width of the comb will
approach the width of the initial inhomogeneous absorption profile from which the
comb was prepared. More teeth cannot then be added without compromising the
efficiency of the retrieval process. We found that the memory efficiency begins to
suffer seriously if the finesse F falls below around 30, with F defined by
F =
∆0
.
2(M − 1)
(7.58)
7.5 AFC
272
The finesse is just the ratio of the tooth separation δj+1 − δj and the natural atomic
linewidth 2 (the FWHM in normalized units). For our threshold of ηthreshold = 70%,
a finesse of greater than 100 was required.
In part (b) of Figure 7.10 we show the multimode scaling of AFC as a function
of the total optical depth dtotal , alongside the scaling for CRIB and unbroadened
protocols. For these protocols, dtotal = d; the plots are simply reproduced from
Figure 7.8. This plot shows that, as we might expect, the multimode capacity of
AFC does indeed scale linearly with dtotal . That is, N remains proportional to the
total number of atoms available for the protocol. The advantage of AFC lies in the
use of a ‘natural resource’ of atoms, so that increasing the number of atoms used in
the protocol does not require a higher ensemble density. Note however, that even
as a function of dtotal , the AFC protocol still outperforms CRIB, with a scaling of
N ∼ 2d/25, and an optimal comb width of ∆opt
0 ∼ 5d.
This concludes our investigation of multimode storage. The techniques used are
quite general, and are readily applicable to new quantum memory protocols as they
are invented. Experimental implementations of multimode storage will likely by
challenging, but the technical advantages for applications of memory to quantum
communication makes the research a worthwhile endeavour.
Practically, it is hard to imagine how one might encode or detect photons in
the optimal input modes φk (τ ). These modes have non-trivial shapes, which do
not necessarily overlap well with the ‘natural’ modes of photon detectors. Generally these detectors, be they avalanche photodiodes (APDs), photomultiplier tubes
7.5 AFC
273
(PMTs) or superconducting bolometers (SSPDs), are engineered to have a broad,
flat spectral response. This means they couple to a temporal mode that looks like
a short, sharp, spike. The natural modes for single photon detection are therefore
time-bin modes, where information is encoded in the time of arrival of a photon.
These time-bin modes are not usually the same as the φk , so time-bin encoding is
sub-optimal, and the multimode capacity predicted by the SVD cannot be reached.
But the discrepancy between the optimal capacity N and the achieved capacity
depends on the overlap between the subspace of signal profiles spanned by the timebin modes with the subspace spanned by the first N optimal modes. As N becomes
large, this overlap grows, and the capacity for time-bin modes rapidly approaches
the optimal capacity. Without attempting a detailed calculation, the plausibility of
this claim can be appreciated by comparing parts (a) and (b) in Figure 7.5 in §7.3.2.
Here, we estimated the Schmidt number by decomposing the Green’s function using non-ovelapping rectangular pulses as spatial modes. If we consider a Green’s
function defined in the temporal domain, such pulses are precisely time-bin modes.
And it is clear that in part (a), the multimode Green’s function is more faithfully
reconstructed than its few-mode counterpart in part (b). This is quite general. As
a Green’s function becomes multimode, it admits an increasingly fine-grained decomposition in terms of time-bin modes. This shows that the time-bin basis quickly
becomes ‘just as good’ as the φk . Therefore the multimode capacity for time-bin
modes fast approaches that calculated from the SVD. A similar argument applies
if a frequency-bin encoding is used, or, for that matter, if any other orthonormal
7.5 AFC
274
basis is chosen for the signal alphabet. The multimode capacities calculated in this
chapter can accordingly be characterized as a tight upper bound on the performance
of the protocol concerned.
In the next chapter, we wrap up our analysis of quantum memory by addressing
the optimization of storage when the signal field is given, and cannot be ‘shaped’.
7.5 AFC
275
(a)
(b)
40
150
20
30
100
14
20
8
50
10
2
0
0
0
100
200
300
400
500
0
100
200
300
400
500
Figure 7.10 The multimode scaling of the AFC memory protocol.
(a): We show the multimode capacity of AFC for various numbers of
comb teeth — indicated by the numbers in the plot — as a function
of the optical depth d associated with a single comb tooth. For each
number of comb teeth M , the capacity scales with the square root of
d, just like an unbroadened protocol. But adding more teeth allows
to increase N arbitrarily, so the multimode capacity is not limited
by the optical depth. (b): Here we show the multimode capacity
of AFC (green dashed line) as a function of the total optical depth
dtotal = M d associated with the entire frequency comb. This shows
that N scales linearly with the total number of atoms involved in
the protocol, just as it does for the CRIB protocols. For comparison,
the capacities of tCRIB, lCRIB and EIT/Raman protocols are also
shown (red dotted line, black dot-dashed line and blue solid line,
respectively). For these latter protocols dtotal = d, so these capacities
are just taken from Figure 7.8. It is interesting that the capacity of
AFC is larger than that of CRIB, even when evaluated as a function
of dtotal . Of course, d is the more relevant physical resource for AFC,
since this sets the ensemble density. We optimized the capacities
plotted for AFC in both (a) and (b) over the spectral width ∆0 of the
frequency comb. The threshold efficiency used was ηthreshold = 70%,
as usual.
Chapter 8
Optimizing the Control
The SVD has proved an extremely useful tool for the analysis of quantum storage.
Given a set of parameters, including the ensemble geometry and density, as well
as the detuning and the control pulse profile, it is possible to construct a Green’s
function that contains all the information we might want to know about the memory
interaction. In particular, taking its SVD provides us with the optimal input mode,
so that we can achieve efficient storage by shaping the signal field. But suppose that
we are simply given a signal field, and asked to store it. This is probably the more
likely situation in practical applications, where we have control over our memory,
but not over the source that generates the signal field to be stored. In this case we
are not able to shape the signal so that it matches the optimal mode φ1 . How do we
achieve efficient storage in this situation? We must try to deform the optimal mode
φ1 into the shape of our signal! This can be done by shaping the control field in
order to sculpt the Green’s function; in this chapter we will explore how the correct
8.1 Adiabatic shaping
277
shape for the control may be found.
8.1
Adiabatic shaping
The adiabatic limit is precisely the limit in which the natural atomic response is
sufficiently fast that the atoms can ‘follow’ the driving fields. In this limit, the atomic
response function is completely determined by the control field profile. Therefore
the adiabatic limit is the regime we should work in, if our aim is to be able to affect
the Green’s function by shaping the control. Fortunately, another consequence of
adiabatic following is that the Green’s function may be completely parameterized by
the integrated Rabi frequency ω, defined in (5.46) in §5.2.6 of Chapter 5. Later in
Chapter 5 we derived an analytic solution for the adiabatic storage kernel of a Λ-type
ensemble memory (5.77). It is extremely convenient that the optimal input mode
found from the SVD of this kernel applies universally for all control profiles (within
the adiabatic limit, of course). Given a control profile, it is easy to find the temporal
shape of the optimal input mode, simply by converting back from the ω to τ . The
conversion is given by (5.78). The optimization problem in this case is then simple.
We must adjust the control field until the shape of the optimal temporal mode is the
same as the shape of the signal field we intend to store. Changing the shape of the
control pulse has no effect on the singular values, since the storage kernel depends
only the control pulse energy through W . Therefore, when the shaping is complete,
the same storage efficiency is achieved, as would have been if we had shaped the
signal field and left the control fixed.
8.1 Adiabatic shaping
278
In Figure 8.1 we show some examples of this kind of optimization. The control
field is parameterized by a vector Ω of 2N real numbers, giving the real and imaginary parts of Ω(τ ) at a set {τj } of N discrete points in time. We choose the points τj
to lie on a Chebyshev grid, since this makes polynomial interpolation of the control
field to other times numerically stable (see §E.1.2 in Appendix E). Starting with
the initial guess Ω(τj ) = Ain (τj ), we optimize Ω using a simplex search algorithm —
‘fminsearch’ in Matlab — to minimize the norm of the difference Ain (τ ) − φ1 (τ ).
At each iteration the optimal mode φ1 (τ ) is determined from Ω by interpolating to
find the envelope Ω(τ ), and then using (5.78). Convergence does not take longer
than 1 minute on a 3 GHz machine. After each optimization, we use the optimized
control profile along with the numerical method described in §5.4 of Chapter 5 to
construct the storage kernel without the adiabatic approximation. The resulting optimal mode is then shown for comparison with the signal profile to be stored. This
provides a direct way to examine the accuracy of the optimizations. The method
used relies on the validity of the adiabatic kernel (5.77), and it is discernible that the
optimization performs better in the Raman limit than for EIT, since the adiabatic
approximation is less robust on resonance (this is discussed in §5.2.9 of Chapter 5).
As with so much of this thesis, an excellent account of closely related work can
be found in the papers of Gorshkov et al. [133] , who studied this problem in the lightbiased limit of large control pulse energy, where the optimal kernel (5.27) is valid (see
§5.2.8 in Chapter 5). They did not use the SVD, or the integrated Rabi frequency
ω, so their approach is a little more convoluted, but nonetheless it is accurate and
8.2 Non-adiabatic shaping
279
ingenious.
8.2
Non-adiabatic shaping
The opposite to the adiabatic limit is the limit of a very short, broadband control
pulse. If the pulse is sufficiently short, there is no time for the coherence induced
on the |2i ↔ |3i transition to couple to the optical polarization P connecting states
|1i and |2i. In this limit, the shape of the control field makes no difference to the
form of Green’s function describing absorption of the signal. Instead, the control
simply induces Rabi oscillations between |2i and |3i that transfer population from
the excited state to the storage state. The efficiency of this population transfer
is given by sin2 (θ/2), where θ = 2
R
Ω(τ ) dτ is the pulse area (note the difference
between this dimensionless quantity and the integrated Rabi frequency ω, which is
relevant in the adiabatic limit). Perfect transfer is achieved by an infinitely short
π-pulse: this kind of instantaneous map is assumed in the CRIB and AFC protocols
treated in the last Chapter, and in the derivation of the optimal kernel (5.27) in
Chapter 5. If we use such a control pulse, the optimal signal input mode is fixed
by the homogeneous and inhomogeneous lineshape of the ensemble. If the signal
field mode does not coincide with this optimal mode, there is nothing we can do to
optimize the storage efficiency.
But there is a large middle-ground between the adiabatic regime and the extreme
case of a π-pulse control. In this middle-ground, the analytic solution for the storage
kernel (5.77) is of limited use, but the shape of the control field still has an effect, at
8.2 Non-adiabatic shaping
280
least to some degree, on the form of the storage Green’s function. Since the analytic
solution breaks down, we must resort to numerical solutions of the equations of
motion to optimize the control. The method we use is the most direct method one
could imagine. Given a vector Ω parameterizing the control, we construct the full
control profile Ω(τ ) by interpolation, and then we integrate the equations of motion
numerically, and we extract the storage efficiency η. We then use a simplex search
algorithm to optimize the elements of Ω so as to maximize η — using fminsearch
in Matlab, we minimize −η. This approach works well, but relies on the ability to
repeatedly solve the equations of motion quickly and accurately. Fortunately, this
is possible using the method described in Appendix E, where we use Chebyshev
spectral collocation for the spatial propagation, and RK2 for the time-stepping.
To improve the convergence of the optimization, it helps to start with an initial
guess that is already close to optimal. We can make use of the SVD in this connection. We start with an initial guess Ω, determined by setting Ω(τ ) = Ain (τ ).
This is motivated by the fact that φ1 (τ ) = Ω(τ ) in the adiabatic limit, when the
coupling is small (i.e. when d or C is small), so it is a reasonable opening gambit.
Next we use the numerical method described in §5.2.9 of Chapter 5 to construct the
resulting Green’s function, and we take its SVD to extract the optimal input mode
φ1 (τ ). In general, since the coupling is not small, φ1 (τ ) 6= Ain (τ ) — this is why the
optimization is necessary. By fitting a cubic spline interpolant to φ1 , and also to
(k)
Ain , we are able to build a sequence of n functions {φ1 } that represent a smooth
(1)
(n)
deformation of this optimal mode into Ain . That is, φ1 = φ1 , and φ1
= Ain , with
8.2 Non-adiabatic shaping
281
the intervening functions lying ‘between’ these two. This is easily done in Matlab by
(k)
building the φ1
using splines with coefficients found by interpolating between the
coefficients for φ1 and Ain . Now, Ω already describes the optimal control for storing
(1)
the mode φ1 , by construction. We now use the optimization method described
(2)
in the previous paragraph to optimize Ω for storing the mode φ1 . The hope is
(1)
that this is sufficiently ‘close’ to φ1
that the optimization will converge quickly.
We then repeat the optimization, this time using the newly optimized vector Ω as
(3)
our initial guess, and using the mode φ1
as the target input mode. The pattern
is now clear. We iterate these steps, each time switching the target input mode
(k)
from φ1
(k+1)
to φ1
, and using the previous result for Ω as an initial guess. As we
procede, the target modes approach Ain , and Ω approaches the vector describing
the optimal control for storing Ain . The rationale behind this approach is that the
control remains near-optimal at all times. We begin with an optimal control for the
‘wrong’ target, and by degrees we deform this target into Ain , all the time ‘bringing
along’ the control. We found that this method is helpful in the most non-adiabatic
situations, where the initial guess Ω(τ ) = Ain (τ ) is particularly poor. When the
dynamics are more adiabatic, optimizing for Ain directly in a single step is often
sufficient.
In Figure 8.2 we show some examples of this type of optimization. Clearly the
numerical optimization works well where the analytic optimization described in the
previous section fails. The method is not particularly time consuming on a modern
computer, although it is of course slower than the analytic optimization. It is quite
8.2 Non-adiabatic shaping
282
general though, and it provides an easy way to optimize the storage efficiency for a
given input, and also to check the final proximity of the optimal input mode to the
desired shape.
Gorshkov et al. have also studied the optimization of the control field outside the
adiabatic limit, in the fourth of their series of papers on the subject [166] . They use
an interesting approach involving gradient ascent, in which they derive an explicit
formula for the incremental change in the control profile that will improve the storage
efficiency. Numerical solution of the equations of motion provides them with this
incremental change, and by iterating they are able to generate the optimal control
for storage. This is a robust and efficient optimization that produces very similar
results to the method we describe here. They comment that they are not able to
verify the optimality of their results, and this is true for our optimizations too. But
in both cases, the SVD of the numerically constructed Green’s function does make it
possible to check the resemblance between the desired input mode, and the optimal
mode resulting from the control. When these two modes match up, as they do in
the cases shown in Figure 8.2, it is clear that the optimization has converged. When
these modes do not match, of course, it is not obvious whether the result is a global
or simply a local optimum.
This completes our theoretical treatment of optimal storage and retrieval from
ensemble memories. In the next Chapter we present a derivation of the equations
of motion for the memory interaction in diamond, since this is one of the media we
have used in our experiments, and the theory of Chapter 4 is not directly applicable.
8.2 Non-adiabatic shaping
283
Finally, in Chapter 10 we review the status of our experiments, before concluding
the thesis in Chapter 11.
1
Broadband Raman
(a)
Broadband EIT
8
6
(b)
6
0.5
4
4
2
2
0
0
−0.2
0
0
−0.2
0.2
1
0
0.2
Non-adiabatic
Narrowband EIT
8
(c)
(d)
6
6
0.5
0
−2
0
2
4
4
2
2
0
Phase
Intensity (arb. units)
284
Phase
Intensity (arb. units)
8.2 Non-adiabatic shaping
0
−2
0
2
Figure 8.1 Adiabatic control shaping. Given an input signal profile, as well as the available control pulse energy W , optical depth d
and detuning ∆ we construct the adiabatic kernel (5.77), and extract
the optimal input mode, as a function of the integrated Rabi frequency ω. We then optimize the control profile Ω(τ ) until the temporal profile of the optimal mode matches the given signal. In parts (a)
2
to (d), the signal is assumed to be a Gaussian, Ain (τ ) = e−[(τ −τs )/Ts ] .
Its intensity profile |Ain (τ )|2 is shown as a dashed black line. The
timing of the signal pulse τs is arbitrary, and in each case we choose
it so as to aid the convergence of the optimization. The temporal
intensity profile |Ω(τ )|2 of the optimized control is represented by
the blue solid line, with its temporal phase shown by the red solid
line (referred to the axes on the right-hand side). We used N = 21
Chebyshev points to parameterize the control in all cases. The green
dotted lines show the temporal intensity profiles of the optimal input
mode determined from the optimized control using numerical integration to construct the Green’s function. These should coincide with
the signal mode, if the optimization has succeeded. Part (a) shows
the result for an optimization with W = 2.5, ∆ = 150 and d = 300,
and a signal duration of Ts = 0.1, all in normalized units. This describes off-resonant Raman storage. The adiabatic approximation is
well satisfied, and the optimization performs well. In part (b) we
show the result for the same optimization when the detuning is set
to zero, which describes broadband EIT. Here the adiabatic approximation is not well satisfied, and the optimization performs rather
poorly. In part (c) we optimize the control for a narrowband signal
with Ts = 1 (note the difference in time scale on the horizontal axis).
The optimization performs much better. Finally in part (d) we reduce the optical depth to d = 10, and we increase the control energy
to W = 49. The detuning is set to 0, and the signal duration remains
Ts = 1. Despite the narrow signal bandwidth, the adiabatic approximation that worked well for part (c) is now ‘broken’, because of the
large Rabi frequency, and the optimization fails entirely. The storage
efficiencies achieved in parts (a) to (d) were 94%, 74%, 94% and 54%,
respectively. For comparison, the efficiencies that would have been
achieved if the signal field were exactly equal to φ1 (τ ) were 97%, 99%,
96% and 82%.
8.2 Non-adiabatic shaping
285
Broadband EIT
10
(a)
6
(b)
4
0.5
5
2
0
−0.2
0
0.2
0
−0.2
1
0
0.2
0
Non-adiabatic
Narrowband EIT
6
(c)
(d)
10
4
0.5
5
2
0
−2
0
2
0
−2
0
2
Phase
Intensity (arb. units)
Phase
Intensity (arb. units)
Broadband Raman
1
0
Figure 8.2 Non-adiabatic control shaping. Parts (a) to (d) show
the results of the direct numerical optimization described above,
where n = 10 steps were used in deforming the initial target Ain (τ ) =
2
(1)
φ1 (τ ) into the final target Ain (τ ) = e−[(τ −τs )/Ts ] . The optimizations each ran in around 2 minutes on a 3 GHz machine. The parameters are identical to those used in parts (a) to (d) of Figure 8.1.
The numerical optimization copes well with non-adiabatic dynamics, and in all cases comparison of the target signal mode with the
optimized input mode shows that the optimizations have met with
some success. In part (b), the broad bandwidth of the signal makes
the adiabatic approximation poor, and it is noticeable that the optimized control profile features a small oscillation which is not present
in the control in part (c). The adiabatic approximation fails entirely
in part (d), and here the control involves a large oscillation, with its
energy distributed into two ‘pulses’. These oscillations are typical
of non-adiabatic shaping, since the ‘ringing’ of the atomic dynamics must be compensated by the control to produce a smooth input
mode. The storage efficiencies achieved in parts (a) to (d) were 97%,
99.5%, 97% and 81%, respectively. For comparison, the efficiencies
that would have been achieved if the signal field were exactly equal
to φ1 (τ ) were 97%, 99.6%, 97% and 82%.
Chapter 9
Diamond
In this chapter we build a theoretical description of quantum memory in diamond.
The end result is a set of equations with precisely the same form as (5.107) describing
Raman storage in a vapour, so that our analysis of optimal storage and retrieval
applies in diamond just as it does to atomic vapours (see §5.3.3 in Chapter 5). But
some justification of this claim is required, and so we present a derivation below.
9.1
Diamond Scheme
Diamond is a singular material, both physically and aesthetically. It is the hardest
and most transparent mineral, with the highest thermal conductivity of any material,
and a very large refractive index (around 2.4). These properties arise in part from
its extremely simple structure. It is comprised entirely of carbon atoms; each is
connected to four others by strong covalent bonds. The bonds are all equivalent,
and this symmetry produces a tetrahedral arrangement of atoms that is exceedingly
9.1 Diamond Scheme
287
robust. The diamond structure can be visualized by convolving a basis of two carbon
atoms with a face-centred cubic bravais lattice [167] , as shown in Figure 9.1.
Figure 9.1 The crystal structure of diamond. The diamond lattice
is FCC (face-centred cubic); here we show just a single unit cell,
outlined for clarity with thin ‘rods’. Each basis is shown as a pair
of atoms connected by a thick ‘tube’, one atom is coloured red; the
other grey. The grey atom of each basis is located at the sites of the
FCC lattice. Each atom is connected by four covalent bonds (not
shown) to its neighbours, forming a tetrahedral pattern.
If the basis is deformed, the bonds produce a large restoring force, and so in
response to an impulse the atoms of the basis can undergo harmonic oscillations
relative to one another. The frequency of these ‘internal’ vibrations is rather high,
because the interatomic bonds are very ‘stiff’. It is these high frequency vibrational
modes we seek to excite when implementing a quantum memory in diamond.
Eventually any relative motion within a basis couples to collective motion of
the basis with respect to its neighbouring bases, and the energy is dissipated as
9.2 Quantization
288
waves of motion of the lattice sites with respect to one another: sound waves. This
process limits the lifetime of a diamond quantum memory, and in fact the lifetime
— on the order of picoseconds — is much too short to make diamond a useful
medium for quantum storage. Nonetheless, it is possible to store very broadband
pulses in diamond, because the oscillation frequency is so high, and a solid-state,
room-temperature, broadband quantum memory is interesting in its own right.
9.2
Quantization
Just as the electromagnetic field can be quantized, revealing photons as the constituents of light, so the harmonic oscillations of a crystal can be quantized. The
quanta of crystal vibrations are known as phonons. Our aim in the present context
is to describe the coherent mapping of a single photon to a single phonon in the
diamond crystal.
The crystal vibrations may be quantized by imposing periodic boundary conditions on the atomic displacements within a cubic crystal of side length L. To
illustrate this procedure we consider a one dimensional wave b(z) = eikz describing
the atomic displacement at position z. Periodic boundary conditions require that
b(0) = b(L), so that we must have kL = 2πm for some integer m. Therefore the
momenta associated with crystal vibrations are quantized, with k = 2πm/L. The
quanta are phonons.
The smallest non-zero wavevector allowed is found by setting m = 1, whence we
obtain the mode separation δk = 2π/L. There is also an upper limit to the allowed
9.2 Quantization
289
wavevectors. The wavelength of any vibration cannot be meaningfully defined if it
falls below 2a, where a is the lattice constant — the distance separating neighbouring
atoms. The wave is only ‘sampled’ at the atomic positions, and the spatial frequency
of the waves cannot exceed the sampling rate. The maximum wavevector, set by
this coarseness of the crystal structure, is ∆k = π/a. As shown in Figure 9.2 below,
any wave with a larger wavevector k is physically indistinguishable from a wave with
momentum k − 2∆k, and so we only consider phonons with wavevectors lying within
the range [−∆k, ∆k]. This region in k-space is known as the first Brillouin zone,
or just the Brillouin zone. That all physically distinct phonon modes are contained
within the Brillouin zone can be confirmed by counting them: the number of modes
in the Brillouin zone is 2∆k/δk = L/a = N , where N is the number of atoms in
the crystal. So the number of phonons is equal to the number of atoms. This must
be true in one dimension, since each atom has one vibrational degree of freedom.
The generalization to three dimensional vibrations turns the Brillouin zone into a
three dimensional volume in k-space, often with a non-trivial shape. But this is not
important for us. As in most of the rest of this thesis, we will use a one dimensional
model to describe the quantum memory.
The lattice constant in diamond is around 3.6 Å, so the Brillouin zone boundary
lies at ∆k ∼ 1010 m−1 . By comparison, the wavevector associated with visible light
at around 500 nm is about 107 m−1 . Any momentum imparted to the crystal by
interaction with optical fields is therefore very small, on the scale set by the crystal
lattice. The excitations produced by a quantum memory may safely be considered to
9.3 Acoustic and Optical Phonons
290
lie at the zone centre, with δk ∼ 0. This, of course, greatly simplifies our theoretical
displacement
description.
Figure 9.2 Phonon aliasing. A crystal vibration with a wavelength
smaller than 2a is physically equivalent to one with a longer wavelength. In the example shown, the blue solid line represents the profile
of a short wavelength vibration, and the dashed blue line shows the
profile of the equivalent longer wavelength vibration. Note that the
atomic positions, indicated by the red circles, are identical for both
waves. The black circles show the equilibrium positions of the atoms,
equally spaced by the lattice constant a.
9.3
Acoustic and Optical Phonons
Phonons come in two varieties, as we have already hinted. Acoustic phonons are
the quanta of sound waves in a crystal. They represent compression and rarefaction within the crystal lattice, and the energy associated with this lattice distortion
clearly vanishes as its wavelength becomes large, since in the limit of infinite wavelength there is no distortion, and therefore no restoring force. As mentioned above,
optical wavelengths are already much larger than a unit cell, so the energies of optically accessible acoustic phonons are very small. A typical dispersion relation for
acoustic phonons is shown in part (a) of Figure 9.3. The energy of acoustic phonons
rises linearly with their wavevector k for k ∆k. Scattering from low-energy zone-
9.3 Acoustic and Optical Phonons
291
centre acoustic phonons is conventionally known as Brillouin scattering, but it is not
of interest to us here. We will focus instead on Raman scattering, which in crystals
refers to the scattering of light from so-called optical phonons. These represent the
second variety of phonon; they are the quanta associated with the high-frequency
internal vibrations of the crystal bases. The bases oscillate essentially independently
of their neighbours, provided the wavelength of the oscillation is not so small that
their neighbours are ‘pulling’ on them. In the limit of infinite wavelength, the optical phonon energy does not vanish, but is set by the frequency associated with
the normal modes describing the natural internal oscillations of the basis. A typical dispersion relation for optical phonons is shown in part (a) of Figure 9.3. The
non-vanishing energy of optical phonons at the zone-centre distinguishes them from
acoustic phonons. The Raman interaction in diamond couples an incident signal
field to these zone-centre optical phonons, and it is these phonons that play the
role of the storage state in a diamond quantum memory. Broadly speaking, these
phonons are like the metastable state |3i used for storage in the atomic systems
discussed previously.
9.3.1
Decay
In many crystals, the basis is composed of unlike atoms or ions, and so the optical
phonons are associated with an electric dipole moment. This means they couple
strongly to the electromagnetic field — that is why they are given the designation
‘optical’. However in diamond, which is homopolar, with both atoms in its basis
9.3 Acoustic and Optical Phonons
(a)
292
(b)
Optical branch
Acoustic branch
Figure 9.3 Phonon dispersion. (a): typical dispersion curves for
acoustic and optical phonons. The former have negligible energy near
the zone-centre, which is the region to which we have access with
optical fields. The latter have large energies, and a flat dispersion
relation close to the zone-centre. It is the zone-centre optical phonons
that we use to store an incident photon. (b): The decay of optical
phonons is dominated by the Klemens channel, in which anharmonic
coupling allows a zone-centre optical phonon (black dot) to decay
into a pair of acoustic phonons with large, opposite momenta (red
dots). We only show the dispersion in the first Brillouin zone. The
dotted lines on the left and right hand sides indicate the boundaries
of this zone, which are identified, meaning that any we could wrap
the plots around a cylinder and stitch these two lines together: they
represent physically equivalent momenta.
identical, there is no separation of charge associated with internal oscillations. The
optical phonons in diamond are therefore not, in fact, optically active. That is, they
do not directly radiate or absorb electromagnetic radiation. This is advantageous for
quantum memory, since the optical phonons in diamond are accordingly longer-lived
than in many other materials. As touched upon above, the dominant decoherence
process for these phonons is the decay into acoustic phonons via anharmonic couplings: the covalent bonds do not behave like perfect springs, and their deviation
from Hooke’s law allows the optical and acoustic phonons to exchange energy. To
conserve momentum the acoustic phonons are produced in pairs with approximately
opposite momenta, as illustrated in part (b) of Figure 9.3. This process is known
9.3 Acoustic and Optical Phonons
293
as the Klemens channel [168,169] . The anharmonicities that give rise to the Klemens
channel are largely geometrical in origin, and the lifetime of optical phonons is only
weakly affected by temperature [170] . The purity and quality of the crystal also contribute, but all diamonds have an optical phonon lifetime1 τp . 10 ps.
9.3.2
Energy
There are three optical phonons in diamond, corresponding to basis oscillations in
three orthogonal directions. The three phonons are degenerate at the zone centre,
because diamond is symmetric with respect to the interchange of these three directions. With a one dimensional interaction we will excite just a single phonon mode;
the degeneracy of the modes means we need not worry about which mode this is.
The optical phonon energy at zone centre is Ep = 0.17 eV, which corresponds to a
wavenumber of νp = 1332 cm−1 , or an angular frequency of ωp = 2.5 × 1014 s−1 .
This would correspond to an infra-red wavelength of λp = 7.5 µm. The large phonon
energy in diamond is advantageous for the following reasons:
First, the energy scale kB T associated with room temperature (T ∼ 300 K) is
around 1/40 eV, which is much smaller than the phonon energy. Therefore there
are very few thermally excited optical phonons at room temperature: using the
Boltzmann formula pthermal = e−Ep /kB T we predict a population of around 1.7×10−3
thermal phonons per mode. Therefore demonstrating quantum memory at room
1
These phonon lifetimes were studied in our research group by Felix Waldermann, and later
by K.C. Lee and Ben Sussman, using a technique they named TCUPS [171] . In these experiments,
a pair of delayed pump pulses directed through a diamond crystal produces a corresponding pair
of Stokes pulses. The phase coherence, as measured from the visibility of spectral interference,
between the two Stokes pulses directly measures the coherence of the optical phonons.
9.4 Raman interaction
294
temperature in diamond is feasible.
Second, the bandwidth of the stored signal field cannot exceed the phonon frequency, since ωp sets the frequency difference between the signal and control fields,
and these should not overlap spectrally. Since the phonon frequency is so large, the
signal bandwidth can be large, meaning that a short pulse can be stored. Taking
τs ∼ 1/ωp as a rough estimate of the shortest signal pulse duration that can be
stored, we find τs ∼ 1000τp . If the ratio of the shortest storable pulse duration
to the maximum storage time τs /τp is taken as a figure of merit for a memory, a
diamond quantum memory is actually rather impressive!
9.4
9.4.1
Raman interaction
Excitons
The Raman interaction in diamond involves an intermediate state, just as it does in
the atomic case considered in earlier chapters. The optical fields are detuned from
resonance with this state, but the interaction nonetheless requires strong coupling
to this state to mediate the storage of the signal field. The relevant intermediate
state in diamond is an exciton. To understand what this is, recall that the electronic
orbitals in an extended crystal arrange themselves into disjoint bands. Electronic
band structure arises from Bragg scattering of electrons from the periodic potential
associated with the regular lattice of atomic nuclei in the crystal. As the De Broglie
wavelength of an electron approaches 2a, the reflected and transmitted components
9.4 Raman interaction
295
of the electronic wavefunctions interfere destructively, leading to the appearance of
forbidden energy bands, containing no allowed electronic states, at the edges of the
Brillouin zone. This is illustrated in Figure 9.4.
(a)
Free
(b)
Periodic
(c)
Bands
Figure 9.4 Band structure. (a): the dispersion relation of a free
electron is parabolic, since its kinetic energy E = p2 /2m is quadratic
in k. (b): electrons in a periodic lattice must have a periodic dispersion relation. To a first approximation, this is found by simply
adding ‘copies’ of the free electron dispersion relation at intervals of
2∆k = 2π/a in k-space. (c): scattering of the electrons from the periodic potential produced by the atomic cores results in anti-crossings
in the dispersion relations, leaving energy gaps, shown shaded in gray.
Here we have removed any electronic states lying outside the first
Brillouin zone (red shaded area), since they are not physically meaningful.
An exciton is produced when an electron in the lower energy band, the valence
band, is promoted into the upper energy band, the conduction band, leaving behind
a hole. The hole in the valence band simply represents the absence of an electron,
but it is convenient to think of it as a particle in its own right, with positive charge,
negative energy and negative mass. An exciton is the combined system of an electron
in the conduction band and a hole in the valence band (see Figure 9.5). In fact, since
these two particles have opposite charge, they attract one another, and it is possible
for them to form a bound system, similar to a hydrogen atom, or the positronium
9.4 Raman interaction
296
of particle physics. The binding energy is rather small, however, and here we will
treat the electron and hole as if they were free particles2 .
Conduction band
Valence band
Figure 9.5 An exciton. A photon (not shown) promotes an electron (blue dot) from the valence band into the conduction band,
leaving behind a positively charged hole (red dot). Note that the
curvature of the valence band for the hole has been reversed, since
the hole has negative energy. In this picture, the electron and hole
have approximately equal and opposite momenta, so that the total
momentum of the exciton is small, consistent with the small momentum of the photon.
9.4.2
Deformation Potential
An incident photon can produce an exciton if it has sufficient energy to breach
the band gap. This process provides the coupling between the optical fields and
the diamond. To excite an optical phonon, there should be some coupling between
excitons and phonons. In polar crystals, there are direct couplings between the
dipole fields of excitons and phonons, in the form of the Fröhlich and piezoelectric
interactions [173–176] . In diamond, these long-range electric couplings are absent, but
2
A precise characterization of the Raman cross-section in diamond does require an account of
bound excitons [172] , but our aim is simply to study the feasibility of using this interaction for
quantum storage, so we sacrifice rigour for simplicity
9.4 Raman interaction
297
there remains a short range coupling known as the deformation potential. The origin
of this interaction can be understood as follows. The diamond structure takes the
form of two interpenetrating FCC lattices, offset from one another. A zone centre
optical phonon — with infinite wavelength — may then be interpreted as the rigid
displacement of one sub-lattice with respect to the other. At any instant, the crystal
structure is accordingly deformed, much as if it were subject to an external strain,
and the electronic band structure, which depends on electron scattering from the
crystal potential, is altered. Therefore electronic energy levels are coupled to crystal
vibrations. More specifically, excitons are coupled to optical phonons. Optical
phonons are not energetic enough to create or destroy an exciton outright, but
an exciton with some momentum k can scatter from the deformation potential to
produce an exciton with a momentum k 0 , and an optical phonon with momentum κ.
Phasematching of this process over the length of the crystal ensures that momentum
is conserved, with κ = k − k 0 .
We now have to hand all the ingredients necessary to unpick the Raman interaction in diamond, which is shown in Figure 9.6,
1. A signal photon produces a virtual exciton — ‘virtual’ because the energy
of the signal photon is smaller than the band gap. This is analogous to the
virtual, or dressed state to which the signal couples in atomic systems, when
it is detuned from the excited state |2i.
2. The virtual exciton scatters from the deformation potential to produce an
exciton with a different momentum, and also a zone-centre optical phonon.
9.4 Raman interaction
298
3. The remaining virtual exciton recombines — the electron decays back into the
valence band, filling the hole — and emits a control photon.
From this description, it’s clear that the Raman interaction in diamond is third
order, rather than second order as it is in the atomic systems considered in earlier
Chapters. In principle this makes it weaker than in atomic systems, but the extremely high density of electrons in the solid state more than makes up for the extra
perturbative order, and the Raman cross section in diamond is, in fact, extremely
large.
(a)
(b)
Figure 9.6 The Raman interaction in diamond. (a): An energy
level diagram for the Raman quantum memory interaction. The relevant states are written in the form |n, mi, where n is the number
of excitons involved, and m is the number of optical phonons. A signal photon (blue wavy arrow) produces a virtual exciton (indicated
by the dotted line, detuned from the real exciton state). The deformation potential interaction then produces another virtual exciton,
and an optical phonon (orange arrow). Finally, the virtual exciton
recombines, emitting a control photon (green arrow) and leaving a
single optical phonon behind. (b): A Feynman diagram for the same
process. Here the dotted lines indicate the world lines of the virtual
electron and hole comprising the intermediate exciton. The optical
phonon is indicated by the orange spiral.
9.5 Propagation in Diamond
9.5
299
Propagation in Diamond
It is advantageous to describe both the optical fields and the excitations of the
diamond crystal in the Heisenberg picture, in order to treat propagation. This
was done for the atomic case in Chapter 4 by studying the dynamics of the flip
operators σjk describing the atomic evolution. The analysis was greatly simplified
by considering just three discrete atomic states. In diamond however the electronic
states form a quasi-continuum in each energy band. It is therefore not immediately
obvious how the approach of Chapter 4 carries over to the present case.
In addition to this issue, there is the problem of how to describe the local dynamics of the crystal excitations. To examine the spatial distribution of these excitations,
we would like to obtain equations for the phonon or exciton amplitudes at some position z within the crystal. In the atomic case the interaction was entirely local,
since each atom scattered light at a point, and independently of all other atoms.
The local dynamics was therefore determined by the Hamiltonian of a single atom,
and propagation was treated by summing these local contributions. The situation
in diamond is conceptually different. First, phonons are global excitations of the
crystal. Second, the electrons in a crystal are not localized around the atomic cores;
rather they form a quasi-free Fermi gas distributed over the volume of the crystal.
The above problems are essentially cosmetic, as we’ll see below. Our general
strategy is as follows. To work in the Heisenberg picture, we write the crystal
9.5 Propagation in Diamond
300
Hamiltonian H in second-quantized form,
H=
X
Hαβ |αihβ|,
(9.1)
α,β
where in the sum, both α and β run over a complete set of states, and where the
Hαβ = hα|H|βi are the matrix elements of the Hamiltonian connecting these states.
The dynamics are then determined from the Heisenberg equations of motion for the
flip operators |αihβ|. This approach broadly mirrors that used in Chapter 4 (see
§4.3).
To extract the spatial variation of the crystal excitations we seek to express the
crystal Hamiltonian in the form
1
H=
L
Z
L
H(z) dz,
(9.2)
0
where H is an effective local Hamiltonian. It turns out that this representation of
the Hamiltonian emerges naturally from the periodic structure of the crystal lattice.
9.5.1
Hamiltonian
The Hamiltonian for a diamond quantum memory is comprised of three parts,
H = HER + HEL + H0 .
(9.3)
9.5 Propagation in Diamond
301
The first two contributions represent the interaction of the electrons in the diamond
with the radiation field and with the lattice respectively. The last part accounts for
the energy of the excitations in the diamond. We neglect the Hamiltonian HL of
the free radiation field, as we did in §4.4 in Chapter 4, since it plays no part in the
equations of motion.
9.5.2
Electron-radiation interaction
The Hamiltonian HER is simply the A.p interaction introduced in §C.4 in Appendix
C. This form of the light-matter interaction is more appropriate than the E.d electric
dipole Hamiltonian, because the electrons in diamond are not localized around the
atomic cores. Instead, they are spread over the entire volume of the crystal, in so
called Bloch waves3 .
Signal and control fields
We divide the vector potential into two parts; the
weak signal field and the strong classical control,
A = As + Ac .
(9.4)
Ac (t, z) = vc Ac (t, z)eiωc (t−nz/c) + c.c.,
(9.5)
The control field is written as
3
The vector potential is not actually a physical field, and strictly we should apply the PZW
transformation (C.20) to the Bloch states, in order that our treatment is gauge invariant [177,178] .
However, this transformation simply shifts the electron momentum (c.f. (C.23)), and has no effect
on the transition matrix elements. In any case, in the Heisenberg picture we are free to choose a
gauge such that A(t = 0) = 0, whence the PZW transformation becomes trivial [177] .
9.5 Propagation in Diamond
302
where we have included the factor n = 2.417 in the exponent, which is the refractive
index of diamond. We treat the signal field quantum mechanically, so it is written
in second-quantized notation as
Z
As (z) = vs
g(ω)
a(ω)e−iωnz/c dω + h.c.,
ω
(9.6)
where we have used the one dimensional formula (C.10) along with (C.13) from
§C.4 in Appendix C. The mode amplitude g(ω) =
p
~ω/4π0 nAc includes the
refractive index. Just as we did in (4.6) in Chapter 4, we anticipate the compact
spectral support of the signal pulse about its carrier frequency ωs by pulling the
mode amplitude g(ω)/ω out of the integral and defining a slowly varying envelope
operator A,
As (t, z) =
Here gs =
√
vs gs
A(t, z)eiωs (t−nz/c) + h.c..
ωs
(9.7)
2πg(ωs ). We have also introduced the time dependence of the operators
arising in the Heisenberg picture (see Appendix B).
Electron wavefunctions
It is sufficient to consider the interaction of the optical
fields with just a single active electron. This is because the complicated many-body
physics governing the behaviour of all the electrons in the crystal can be swept under
the rug of Fermi-Dirac statistics: the Pauli-exclusion principle prevents more than
one electron from occupying each state, and since all valence band states in the
crystal are initially occupied, any electron-electron scattering in this band is ‘frozen
9.5 Propagation in Diamond
303
out’ because no electron can change its state without occupying a previously filled
orbital. The upshot of this is that we may consider p to be the momentum operator
for a single electron.
The wavefunction ψk,n (r) of an electron with wavevector k in the nth energy band
is given by the product of a spatial phase factor with a periodic Bloch function,
ψk,n (r) = eikz uk,n (r).
(9.8)
The Bloch functions have the same translational symmetry as the crystal lattice,
uk,n (z + a) = uk,n (z), which is a consequence of Floquet’s theorem. Note that such
states are not exactly eigenstates of the momentum operator; this is why the A.p
interaction can induce electronic transitions.
Matrix elements
We are interested in the matrix element hα|HER |βi connecting
two quantum states. Neglect for the moment the state of the signal mode. This matrix element is then given by the spatial overlap of the electronic orbitals describing
the initial and final states, with the operator A.p inserted between the two orbitals:
e
hα|HER |βi = −
m
Z
ψα∗ (r)A(z).pψβ (r) d3 r.
(9.9)
crystal
Here the indices α, β are standing in for the wavevectors and band indices of the
initial and final orbitals. Now, the coordinate representation of the momentum
operator is p = −i∇. Applying this to ψβ (r), and using (9.8), we can write the
9.5 Propagation in Diamond
304
matrix element as
e
hα|HER |βi = −
m
Z
ei(kβ −kα )z A(z). [u∗α (r)(p + kβ )uβ (r)] d3 r.
(9.10)
crystal
Since all the momenta are very close to the zone-centre, the spatial variation of
the exponential factors in the integrand is very slow, and this is also true for any
variation of the optical field A(z) (which contains only slowly varying envelopes
or similarly long-wavelength exponential factors). On the other hand, any rapid
spatial variation of the Bloch functions uα,β is periodic, repeated in every unit cell.
We therefore factorize the integral into two parts, as follows,
e X i(kβ −kα )zj
hα|HER |βi = −
A(zj ).
e
m
Z
j
u∗α (r)(p
3
+ kβ )uβ (r) d r , (9.11)
unit cell
where zj is the position of the j th unit cell. We define the matrix element pαβ as N
times the overlap integral inside the square brackets, where N = AL/a3 is the total
number of unit cells in the crystal. Taking the continuum limit for the sum over the
zj , we can write
e
hα|HER |βi = −
pαβ .
mL
Z
L
A(z)ei(kβ −kα )z dz.
(9.12)
0
We have now succeeded in separating out the local from the bulk dynamics. It only
remains for us to introduce the flip operators |αihβ| in a convenient form.
9.5 Propagation in Diamond
Excitons
305
The momentum operator has negative parity, just as the atomic dipole
operator djk does (see §4.3.1 of Chapter 4). Therefore pαα = 0, so there is no
coupling of any state to itself. As mentioned above, if all electrons are in the valence
band, an electron has ‘nowhere to go’, since all the valence band states are occupied.
The only possibility is to promote an electron into the empty conduction band,
creating an exciton. This process, and its time-reverse — exciton recombination
— are the only important scattering processes involved in HER . Once an exciton
has been created, it is of course possible for either the conduction band electron or
one of the valence electrons to undergo scattering (this latter process is the same as
scattering of the hole) within their respective bands via the A.p interaction. But
the energies involved are much smaller than the photon energies in the signal or
control fields, so these processes do not conserve energy, and may be neglected. The
Hamiltonian can therefore be written as

HER (t) = −
e
mL
Z
0
L
A(t, z).

X
ν,k


pνk,0 s†νk eikz + h.c. dz,

(9.13)
where the subscript 0 on p denotes the crystal ground state, and where s†νk creates an
exciton with momentum k and energy ων . In the notation of (9.12), k = kβ −kα . The
Hermitian conjugate component destroys an exciton, representing recombination.
The energy of an exciton does not depend on its wavevector k, because the electron
and hole comprising the exciton might have large but opposite momenta, giving
a small total momentum, but a large total energy. Therefore the exciton state
9.5 Propagation in Diamond
306
is independently parameterized by the two quantities ν and k. Excitons, being
composed of pairs of fermions (a hole is a quasiparticle obeying fermionic statistics),
are bosons. The annihilation operators sνk therefore satisfy the same commutation
relation as photon mode annihilation operators (see Appendix C),
h
i
sνk , s†µk0 = δν,µ δk,k0 .
(9.14)
A merciful simplification is achieved by neglecting any dependence of the matrix
elements pνk,0 on ν or k, since the dependence of the Bloch functions describing the
electron and hole on wavevector is rather weak. We thus write pνk,0 = ip, where p
is the constant magnitude of the matrix element, and where the factor of i appears
because the momentum operator is purely imaginary. With these simplifications,
the Hamiltonian can now be written in the form of (9.2), with the effective local
Hamiltonian given by


X †
ie
sνk eikz − h.c. .
HER (t, z) = − p.A(t, z) 
m
(9.15)
ν,k
As a final step, we can perform the sum over momenta in (9.15) explicitly. We define
the local exciton operator
1 X
Sν (z) = √
sνk e−ikz ,
L k
(9.16)
9.5 Propagation in Diamond
307
which satisfies the commutation relation
h
i
Sν (z), Sµ† (z 0 ) = δν,µ δ(z − z 0 ).
(9.17)
The local Hamiltonian then takes the form
"
#
√
X
ie L
†
p.A(t, z)
HER (t, z) = −
Sν (z) − h.c. .
m
ν
9.5.3
(9.18)
Electron-lattice interaction
The Hamiltonian HEL for the electron lattice interaction is just given by
HEL = V |with phonon − V |no phonon ,
(9.19)
where V is the potential experienced by an electron, generated by all the atomic
cores and other electrons. Let the displacement between the sub-lattices of diamond
caused by a zone-centre optical phonon by given by u. If this displacement is small,
a Taylor expansion to first order gives
HEL =
∂V
.u.
∂u
(9.20)
When we quantize the lattice vibrations, the leading factor becomes the deformation
potential matrix elements, or just the ‘deformation potentials’ [174,179,180] , and the
second factor becomes the operator for the optical phonon amplitude. For vibrations
9.5 Propagation in Diamond
308
along the direction vp with wavevector κ the phonon amplitude operator can be
written [180,181]
uκ = gκ vp eiκz (b†κ + b−κ ).
(9.21)
Here b†κ creates an optical phonon with momentum κ. Phonons are bosons, so that
we have (see (C.8) in Appendix C)
[bκ , b†κ0 ] = δκ,κ0 .
(9.22)
The phonon mode amplitude gκ , which has the dimensions of length (it is the lattice
displacement due to a single phonon), is given by
r
gκ =
~
,
N M ωκ
(9.23)
where M is the mass of a carbon atom and ωκ is the phonon frequency.
For reasons discussed in the previous section, the only states where electronic
scattering can occur are the exciton states, with an electron in the conduction band
and hole in the valence band. The action of the deformation potential is therefore to
destroy an exciton with energy and momentum (ν, k), and to produce a new exciton
with modified parameters (ν 0 , k 0 ). Summing over all possibilities gives the expression
HEL =
X
κ,µ,ν,k,k0
Z
0
L
1
0
vp .Dνµkk0 κ s†µk0 sνk ei(k −k)z × gκ eiκz (b†κ + b−κ )
aL
dz. (9.24)
The factor of 1/a appears to give the deformation potentials D the dimensions of
9.5 Propagation in Diamond
309
energy. The integral over space, along with the factor of 1/L, arises in precisely
the same way as it did in (9.12) above. A dramatic simplification of (9.24) is possible, since the deformation potentials D depend only very weakly on the phonon
and exciton momenta so close to the zone centre. By the same token, the phonon
frequency ωκ = ωp is independent of κ close to the zone centre (see Figure 9.3).
Dropping these dependencies, we can perform the summations over k, k 0 and κ to
obtain the following expression for the effective local Hamiltonian,
HEL (z) =
h
i
L3/2 g X
Dµν Sµ† (z)Sν (z) B † (z) + B(z) ,
a µ,ν
(9.25)
where Dµν = Dνµ = vp .Dµν is the real magnitude of the deformation potential
connecting excitons with energies ωµ and ων , and where we have defined the local
phonon operator
1 X
B(z) = √
bκ e−iκz ,
L κ
(9.26)
i
B(z), B † (z 0 ) = δ(z − z 0 ).
(9.27)
which has the commutator
h
9.5.4
Crystal energy
The energy H0 of the excited crystal is simply found by counting the number of
excitons and phonons. Using the number operators for these particles (see (C.9) in
9.6 Heisenberg equations
310
Appendix C), we find
H0 =
X
ων s†νk sνk +
ν,k
X
ωκ b†κ bκ .
(9.28)
κ
Or, in terms of the operators S and B, the local energy takes the form
H0 X
=
ων Sν† (z)Sν (z) + ωp B † (z)B(z).
L
ν
9.6
(9.29)
Heisenberg equations
Now that we have constructed the Hamiltonians describing the Raman interaction,
we can write down the Heisenberg equations governing time evolution of the operators Sν and B. The dynamics of A are derived in the next section using Maxwell’s
equations, as was done for the atomic case in Chapter 4.
Commutation of Sν and B with H, using the relations (9.17) and (9.26), yields
the equations
√
e
Lg X
Dµν Sµ (B † + B),
∂t Sν = iων Sν + √ p.A + i
a
m L
µ
√
D Lg X
∂t B = iωp B + i
Dµν Sµ† Sν .
a
µ,ν
(9.30)
9.6 Heisenberg equations
9.6.1
311
Adiabatic perturbative solution
e in a rotating frame, so that
Let us define local operators Seν and B
Seν
= Sν e−iων (t−nz/c) ,
e = Be−iωp (t−nz/c) .
B
(9.31)
For notational convenience, let us also define
B(t, z) = B † (t, z) + B(t, z)
e † (t, z)e−iωp (t−nz/c) + B(t,
e z)eiωp (t−nz/c) .
= B
(9.32)
e are then given
The equations of motion for the slowly varying operators Seν and B
by
√
e
Lg X
−iων (t−nz/c)
√ p.Ae
=
Dµν Seµ eiωµν (t−nz/c) B,
+i
a
m L
µ
√
X
e = i Lg e−iωp (t−nz/c)
∂t B
Dµν Seµ† Seν e−iωµν (t−nz/c) ,
a
µ,ν
∂t Seν
(9.33)
where ωµν = ωµ − ων . The spatial phase factors are included for convenience when
considering the exponentials within A.
e the local phonon amplitude, in terms
Our aim is to obtain an equation for B,
of the signal and control fields. We achieve this by eliminating the intermediate
excitons Seν adiabatically. The procedure is related to that used in §5.3.3 in Chapter
9.6 Heisenberg equations
312
5. We start by formally integrating the equation for Seν in (9.33),
Seν (t) =
Z t
e
0
√ p.
A(t0 )e−iων (t −nz/c) dt0
m L
0
√
Z t
X
Lg
0
+i
Dµν
Seµ (t0 )B(t0 )eiωµν (t −nz/c) dt0 .
a
0
µ
(9.34)
e = 0) = 0, since there are no excitons in the crystal initially.
We have set S(t
Unfortunately (9.34) does not provide a direct solution for Seν , because it is coupled
to all the other excitons through the summation on the right hand side. We settle
instead for a perturbative solution. Substituting the first term on the right hand
side of (9.34) into the second term yields the approximate result
Seν (t) =
Z t
e
0
√ p.
A(t0 )e−iων (t −nz/c) dt0
(9.35)
m L
0
#
√
Z t"
Z t0
Lg X
e
0
00 −iωµ (t00 −nz/c)
00
√ p.
+i
Dµν
A(t )e
dt B(t0 )eiωµν (t −nz/c) dt0 ,
a
m L
0
0
µ
This solution is correct to first order in the deformation potential. To perform the
integrals in (9.35), we use the fact that the time variation of the exponential factors
is much faster than the temporal dynamics of the optical fields, and also faster than
the dynamics of the crystal excitations produced by these fields. This adiabatic
approximation requires that the detunings ∆ν = ων − ωs of the excitons from the
signal frequency are all much larger than the bandwidths of the signal or control
fields. In diamond, the bandgap is in the ultraviolet, so this condition is always very
well satisfied if optical frequencies are used. We proceed by pulling all slowly varying
9.6 Heisenberg equations
313
amplitudes out of the integrals and integrating only the exponentials. The resulting
e
expression for Seν contains 12 terms. Inserting this into the equation (9.33) for B
provides us with the dynamical description we have been looking for, although now
there are 144 terms! Fortunately, of the first order terms in the product Seµ† Seν , only
very few contribute significantly to the memory dynamics: most terms are oscillating
at high frequencies, so that they average to zero. After some legwork, we obtain
e = iKΩ∗ A,
∂t B
(9.36)
where we have defined the control field Rabi frequency
Ω(t, z) =
evc .pAc (t, z)
,
m~
(9.37)
and where the coupling constant K, with the dimensions of (length)−1/2 × (time)1/2 ,
is given by
K=
Dµν
Dµν
g
ep.vs gs
1 X
×
×√
+
. (9.38)
~a
~ωs m
L µν (ωµ + ωc )(ων + ωs ) (ωµ − ωc )(ων − ωs )
Let us suppose that we are sufficiently close to resonance that ων + ωs ων − ωs .
This need not be the case, because the bandgap in diamond is very large, but it
is a convenient simplification to neglect the ‘counter-rotating’ terms with summed
frequencies in their denominators (c.f. §4.6 in Chapter 4). The coupling now takes
9.7 Signal propagation
314
the approximate form
K=
Dµν
g
ep.vs gs
1 X
×
×√
,
~a
~ωs m
L µ,ν (∆µ + ωp )∆ν
(9.39)
where we have used the Raman resonance condition ωs = ωc + ωp . The double
appearance of the detuning in the denominator is characteristic of third-order scattering. Equation (9.36) is very similar in form to the equation for the spin wave
(5.107) derived in Chapter 5. Thus encouraged, we proceed in the next section to
derive the equation describing the propagation of the signal field.
9.7
Signal propagation
The electric field of the signal is the solution to Maxwell’s wave equation, with the
polarization in the diamond acting as a driving term (see (4.34) in Chapter 4). The
polarization is the dipole moment per unit volume. That is, er =
R
P dV , where
r is the one-electron position operator. We can relate the matrix elements of r to
those of the momentum operator p, as follows (see (F.6) in §F.2 of Appendix F),
pαβ = −imωαβ rαβ ,
(9.40)
9.7 Signal propagation
315
where ωαβ is the frequency splitting between the energy eigenstates |αi, |βi. Using
this relation, a second-quantized form for the polarization operator can be found:
P
pνk,0 † ikz p0,νk
sνk e −
sνk e−ikz
ων
ων
ν,k
ep X 1 e† −iων (t−nz/c)
√
= −
Sν e
+ h.c. .
mA L ν ων
=
ei X
mAL
(9.41)
Making the SVE approximation (see (4.39) in Chapter 4), we find the propagation
equation
∂z +
n µ0 ωs2 ∗ e
∂t A = −
v .Ps ,
c
2gs ks s
(9.42)
where Pes is the component of the polarization oscillating at the signal frequency ωs .
Substituting the solution (9.35) into (9.41), performing the integrals with the help
of the adiabatic approximation, and retaining only those terms with the appropriate
time dependence, we arrive at the final equation for the signal,
∂z +
n
e ∗ ΩB,
e
∂t + iχ A = iK
c
(9.43)
e is very similar to K,
where K
X
ωs Dµν
e = g × ep.vs gs × √1
K
,
~a
~ωs m
L µ,ν ων (∆µ + ωp )∆ν
(9.44)
9.7 Signal propagation
316
and where χ represents a spatial phase picked up by the signal due to the crystal
dispersion,
evs .pgs 2 1 X ωs
×
χ=
.
~mωs L ν ων ∆ν
(9.45)
e is probably spurious, and may arise from the
The discrepancy between K and K
secular approximation we use in eliminating terms oscillating at the ‘wrong’ frequencies [177,178] . In any case, the difference between the two expressions is small close
to resonance, and never more than an order of magnitude. In what follows, we set
e −→ K.
K
The coupling constants are admittedly rather dense combinations of various material and optical parameters. In the next section we will try to extract a prediction
for the coupling strength of a diamond quantum memory. But for the moment,
what is important is the form of the equations (9.36) and (9.43). They are essentially identical to the Raman memory equations of §5.3.3 in Chapter 5. We
can make the similarity explicit with a number of coordinate transformations and
re-normalizations. First, we introduce the retarded time τ = t−nz/c (note the presence of the refractive index n). When the equations are written in terms of τ and
z, the derivative ∂z + nc ∂t becomes simply ∂z , while the time derivative ∂t becomes
∂τ . Next, we remove the dispersive factor χ with a phase rotation, by making the
transformation
A −→ A = Aeiχz ,
e −→ B = Be
e iχz .
B
(9.46)
9.7 Signal propagation
317
To remove the dependence on the control field profile, we introduce the normalized
integrated Rabi frequency ω = ω(τ ) — defined in (5.46) in Chapter 5 — and the
dimensionless transformed variables
√ A(z, τ )
α(z, ω) = i W
,
Ω(τ )
√
β(z, ω) =
LB(z, τ ).
(9.47)
Here we have also re-scaled the longitudinal coordinate by L so that z runs from
e with K (as mentioned above),
0 up to 1. As a final simplification, we replace K
and we assume that K is real. The equations of motion for the diamond quantum
memory then take the form
∂z α = −Cβ,
∂ω β = Cα,
(9.48)
where the dimensionless Raman memory coupling is
C=
√
LW K.
(9.49)
This tells us that a diamond memory will behave in precisely the same way as a
Raman memory based on an atomic vapour (see (5.109) in §5.3.3 of Chapter 5). We
know that efficient storage is possible if the Raman memory coupling C is of the
order of unity (we should have C & 2). And all of the results pertaining to optimal
shaping carry over. So given the coupling C, we can directly find the optimal input
mode φ1 for the signal, using (5.97).
9.8 Coupling
9.8
318
Coupling
Here we show that efficient storage is possible in a small sample of diamond, so that
diamond quantum memory need not be a luxury enjoyed only by the super rich.
To estimate the memory coupling C, we must evaluate the summations in (9.39).
To do this we first write Dµν = Dδµ,ν , which holds approximately, since the wavefunctions of distinct excitons overlap poorly [172] . Next we assume that both conduction and valence bands are parabolic around the zone centre (as shown in Figure
9.5), so that we may write the energy of an exciton as [182]
ων −→ ω(k) = ω0 +
~k 2
,
2m?
(9.50)
where ~k is the momentum of the electron relative to the hole, and where m? is
an effective reduced mass for the exciton that describes the local band curvature.
Parameterizing the exciton energies in this way, the summations in K may be rewritten approximately as an integral over a sphere in k-space,
X
µ,ν
Dµν
(∆µ + ωp )∆ν
−→
≈
D
(9.51)
(∆ν + ωp )∆ν
ν
Z kmax
AL
4πk 2
D
dk,
(2π)3 0
(ω0 − ωs + ωp + ~k 2 /2m? ) (ω0 − ωs + ~k 2 /2m? )
X
where kmax is an arbitrary, suitably large cut-off. The integral can be performed
9.8 Coupling
319
with a partial fraction expansion and a trigonometric substitution,
Z
κ
0
b−3/2
a−c
k2
dk =
(a + bk 2 )(c + bk 2 )
b−3/2
(a − c)
=
#
Z √bκ
a
c
dx −
dx
(9.52)
a + x2
c + x2
0
0
p
√
p
i
h√
a tan−1
b/aκ − c tan−1
b/cκ .
"Z
√
bκ
Choosing kmax sufficiently large, we get to the result [182]
X
µν
Dµν
ALD
≈
(∆µ + ωp )∆ν
4πωp
2m?
~
3/2 √ p
∆ + ωp − ∆ ,
(9.53)
where ∆ = ω0 − ωs is the detuning of the signal field from the conduction band
minimum.
We can express W in terms of the energy Ec in the control pulse as follows,
Z
∞
W =
−∞
|Ω(τ )|2 dτ =
2πα|vc .p|2
× Ec ,
~m2 ωc2 An
(9.54)
where here α = e2 /4π0 ~c = 1/137 is the fine structure constant.
We estimate the momentum matrix elements as vc .p ≈ vs .p = ~ × 2π/a. This
is justified by noting that the band gap at the zone centre is produced by Bragg
scattering that mixes electrons with k = 0 and k = 2∆k = 2π/a, and it is this latter
component that is responsible for interband transitions.
Other parameters are as follows. The bandgap4 in diamond is ~ω0 = 13 eV, and
4
This energy corresponds to the direct gap in diamond (at the zone centre, or ‘Γ-point’). The
conduction band minima occur elsewhere in the Brillouin zone (~ω0 ∼ 5 eV at the ‘L-point’), but
transitions to these states are mediated by phonons. They are therefore suppressed somewhat.
Although they may not be insignificant, we neglect these processes here.
9.8 Coupling
320
the deformation potential has a value of around D ≈ 7~ω0 ≈ 90 eV. As mentioned
before, the phonons have wavenumber νp = 1332 cm−1 . The lattice constant in
diamond is a = 3.6 Å. The refractive index is n = 2.417, and the mass of a carbon
atom is M = 12 a.m.u. (indeed, by definition). For simplicity we assume a reduced
exciton mass m? = m. We then consider a crystal of length L = 1 mm, illuminated
by beams with waists 100 µm in diameter. If the control pulse is taken from a
modelocked pulse train with an 80 MHz repetition rate and an average power of
10 mW, it has an energy of 0.13 nJ. Using a signal field with central wavelength
λs = 800 nm, we estimate a memory coupling of
C = 2.4.
(9.55)
The optimal storage efficiency is therefore around 99.9% (see Figure 5.5 in §5.3 of
Chapter 5). This demonstrates that efficient storage in a small sample of diamond
is extremely feasible. Of course, the estimate we have made is very crude, but
since neither the laser energy nor the sample size required are close to any practical
limitations, a downward revision of C by a factor as large as 100 could still be
accommodated. In fact (9.55) is something of an underestimate, since we have
neglected the counter-rotating terms in K. When the signal is so far detuned from
the conduction band edge, these terms still contribute significantly.
9.9 Selection Rules
9.9
321
Selection Rules
The large Stokes shift (i.e. the large phonon energy) in diamond makes it easy
to distinguish the signal and control fields spectrally. For instance, if the signal
field wavelength is 800 nm, the control wavelength is around 894 nm. In addition,
however, it turns out that the crystal symmetry requires the signal and control fields
to be orthogonally polarized. The Raman interaction that couples the input and
output fields is constrained to be proportional to an irreducible representation of
the symmetry group associated with the optical phonons. The zone-centre phonons
are always described by a subgroup of the crystal point group — the group of
reflections and rotations that leaves the crystal unchanged. In the case of diamond,
the crystal point group is the cubic group m3m (sometimes written Oh ), and the
optical phonons transform as the subgroup Γ+
5 (sometimes written T2g ). The optical
fields are 3 dimensional vectors, and so the Raman interaction must be proportional
[180]
to the 3 × 3 irreducible representation of the group Γ+
5 , which takes the form




+
Γ5 (x) = 




1







+

1  , Γ5 (y) = 






1

1 





 , Γ+ (z) =  1
5







1
where zero elements have been left blank for clarity, and we have assumed that the
z-axis is aligned with the [001] direction (that is, parallel to a vertical edge of the
cubic unit cell; see Figure 9.1). The Raman interaction requires that vc = Γ+
5 (z)vs ,
from which it is clear that vc should be perpendicular to vs . This polarization



,



9.10 Noise
322
selection rule adds to the experimental attractiveness of a diamond memory.
9.10
Noise
The foregoing analysis of storage in diamond implicitly ignored the possibility of
unwanted couplings. For instance, the intense control pulse can stimulate strong
Stokes scattering, creating optical phonons and Stokes photons in pairs. This is the
same problem as that shown for the atomic case in part (b) of Figure 4.3 in §4.7 of
Chapter 4. Here it is exacerbated because the detuning from resonance is typically
so large that both processes — storage of the signal and Stokes scattering — occur
with roughly equal probabilities. In §6.3.2 of Chapter 6, a solution to this problem
is described that involves introducing a small angle between the signal and control
beams, and this solution certainly carries over to the diamond memory. Another
interesting possibility is that of modifying the optical dispersion so that the unwanted
Stokes light cannot propagate. It would be possible to do this by building a Bragg
grating — alternating layers of diamond and air — with a spatial frequency equal
to that of the unwanted Stokes light. Interference within the Bragg structure would
then suppress any Stokes generation. But we won’t consider this further; a proofof-principle demonstration of broadband storage in room temperature diamond is
challenging enough, without engaging in micro-fabrication [183] .
In the next chapter, we review the experimental progress made in our group
towards the goal of demonstrating a Raman quantum memory.
Chapter 10
Experiments
Although most of this thesis is theoretical, I had initially intended to build a working
quantum memory. I have not succeeded as yet, but the ‘Memories’ subgroup is
continuing its efforts in this direction. In this chapter we discuss some of the ongoing
experimental work; its goals and future prospects.
10.1
Systems
The experimental programme divides into three projects.
1. Diamond,
2. Quantum dots,
3. Atomic vapour.
My experimental research has been focussed on the last of these; the theoretical
analysis in the preceeding chapter constitutes my contribution to the first. We will
10.1 Systems
324
not describe the quantum dot project here; suffice it to say that quantum dots may
be thought of as artificial atoms, so that a sample containing many dots behaves
like an atomic ensemble, in which Raman storage may be implemented.
Light storage in atomic vapour is becoming standard in quantum optics. In almost all cases, resonant EIT is used, and narrowband diode lasers provide the signal
and control fields [74,150,151] . Our research group has some considerable experience
with ultrafast lasers, however, and it was decided that a broadband Raman memory
would be interesting. Atomic vapour is an ideal system for demonstrating such a
memory; indeed this is the system that is considered when deriving the memory
equations in Chapter 4.
A common feature of all the experiments is the requirement of strong Raman
coupling between a laser field and a material sample. The easiest way to verify the
existence of strong coupling is to observe strong Stokes scattering. This is therefore
the first step in all our experiments — the general strategy is shown schematically
in Figure 10.1. Unfortunately, we have not yet been able to achieve this first prerequisite in the atomic vapour experiments. This is rather depressing, but we are
persevering, since there are many improvements to be made. In the rest of this chapter, we will introduce the atomic species used in our experiments with atomic vapour.
We then discuss the theory of Stokes scattering, and we estimate the strength of the
coupling we expect to achieve. Finally, we describe the experimental techniques we
have developed in our attempts to see strong Stokes scattering. We finish with a
discussion of the planned realization of a Raman quantum memory.
10.2 Thallium
Laser
325
Raman pump
Stokes
Vapour cell
Filter
Figure 10.1 Observing Stokes scattering as a first step. Strong
Raman coupling between a bright laser and the atomic ensemble is
required for a Raman quantum memory. The ability to produce,
and detect, strong, stimulated Stokes scattering is a sine qua non for
implementing the memory.
10.2
Thallium
The first incarnation of the atomic vapour experiment used thallium (Tl) as the
storage medium. Thallium is an extremely toxic poor metal, with a history of use
in rat poison, and homicide generally. However the atomic structure of thallium
exhibits a well-defined Λ-system, with a large Stokes shift (see Figure 10.2). For
this reason, attempts were made to build a Raman quantum memory with thallium
vapour, provided by means of a heated glass cell containing solid thallium. After
around a year of unsuccessful attempts to observe strong Raman scattering from
thallium vapour, it was realized that the vapour pressure of thallium is too low for
a strong Raman interaction to be engineered (see the discussion in §10.9.1, and part
(a) of Figure F.1 in Appendix F).
10.3 Cesium
326
7S1/ 2
378 nm
F=1
F=0
6P3/ 2
F=2
F=1
1283 nm
6P1/ 2
F=1
F=0
Figure 10.2 Thallium atomic structure. The three levels of the Λsystem are discernible, marked by the thickened line segments. The
hyperfine structure (indicated by the fine branches) is ignored, being
negligible in the ground P -wave manifolds. The large Stokes splitting
between the J = 1/2, 3/2 states in the P -state manifold makes it ideal
for broadband storage. Unfortunately Thallium has a low vapour
pressure, making an efficient thallium quantum memory impractical.
10.3
Cesium
To increase the Raman coupling, an atomic species with a much higher vapour
pressure was required. Our current experiment uses cesium1 (Cs), which is many
orders of magnitude denser than thallium at room temperature (see part (b) of
Figure F.1 in Appendix F). Cesium is a soft, gold-coloured alkali metal that is,
thankfully, non-toxic, although it reacts explosively on contact with water, even the
water vapour in air! Our cesium is sealed in an evacuated glass cell, along with a
small amount of neon (10 Ne), which acts as a buffer gas (see §§10.4 and 10.5 below).
There is only one stable isotope of cesium, namely
133 Cs.
Needless to say we refer
only to this isotope in what follows; our sample is naturally isotopically pure. The
1
Cesium is the American spelling; the British spelling Caesium retains some of the latinate
flavour if its etymology. The word derives from the latin for ‘sky’, because the 7P ↔ 6S1/2 doublet
lines are a brilliant blue colour. However American spell-checkers and journal styles have worn me
down, and now the British spelling seems odd.
10.3 Cesium
327
Λ-system is implemented in the so-called cesium D2 line at 852 nm. This is the
second of the strong ‘doublet lines’ (the D1 line is at 894 nm) characteristic of alkali
metals — the same doublet in Sodium illuminates the night-time activities of most
of this planet. The relevant atomic structure is shown in Figure 10.3.
852 nm
9.2 GHz
Figure 10.3 Cesium atomic structure. The F = 3 and F = 4
hyperfine levels in the 6S1/2 ground state — the famous ‘clock states’
— provide the ground and metastable states for the Λ-type quantum
memory. The 6P3/2 manifold collectively provides the excited state.
The upper 6P3/2 state is split by the hyperfine interaction: The nuclear spin
of I = 7/2 combines with the total electronic spin of J = 3/2 to produce four
hyperfine levels with total angular momentum quantum numbers F = 2, 3, 4 and 5.
The interaction is weak however, and the splitting between each state is around 200
MHz. For the purposes of the memory, we therefore treat the excited state manifold
as a single state, which plays the role of |2i in the Λ-system.
The hyperfine interaction is much stronger in the 6S1/2 electronic ground state.
The reason is that the ground state is an S-wave, meaning that it has no orbital
angular momentum. It therefore has no azimuthal phase, and so it remains well-
10.4 Cell
328
defined at the origin without vanishing — a wavefunction with such a phase is multivalued at the origin unless it is zero, so higher orbitals must disappear at the origin.
In the ground state, then, the electron penetrates into the cesium nucleus. There is
therefore a strong magnetic dipole coupling between the nuclear and electronic spins,
known as the Fermi contact interaction, which acts to separate the two hyperfine
states with F = 3 and F = 4 by an enormous 9.2 GHz. These two states are
sometimes known as the clock states because coherent oscillations between them are
used in cesium atomic clocks. In fact the hyperfine splitting is now defined to be
exactly 9, 192, 631, 770 Hz, with the duration of the second being a derived quantity.
The clock states form the two lower states in the quantum memory Λ-system.
10.4
Cell
We contain the cesium vapour in a glass cell (the TT-CS-75-V-Q-CW from Triad
Technology in Colorado, USA). The cell is 10 cm in length, and is made of glass that
is resistant to heating up to temperatures of 500 ◦ C. The cell windows are 25 mm
across, and are made of optically polished quartz, to which an anti-reflection coating
has been applied which reduces reflection losses for light at the D2 wavelength of
852 nm down to around 0.1%. The cell is evacuated, and then a small sample of
solid cesium is introduced. Finally, the cell is backfilled up to a pressure of 20 torr
(∼ 2700 Pa) with neon gas. The reason for introducing this buffer gas is explained in
§10.5. The cell is then hermetically sealed, by pinching shut the glass tube through
which the cell contents are delivered.
10.4 Cell
10.4.1
329
Temperature control
The cell is wrapped in heating tape: a thin weave of high resistance wires surrounded
by thermal insulation. A thermocouple connected to an electronic temperature
controller allows one to set and maintain the cell temperature as desired. The glass
protuberance remaining after the cell is sealed provides a convenient ‘cold finger’
— an unheated region protruding from the cell where the cesium preferentially
condenses. As long as the cell windows remain hotter than this cold finger, cesium
does not collect on the cell windows, and the cell remains transparent to our laser
beams.
10.4.2
Magnetic shielding
The hyperfine levels in cesium are not pure quantum states. A hyperfine state with
total angular momentum quantum number F is (2F +1)-degenerate, being comprised
of Zeeman sublevels with quantum numbers m = −F, −F + 1, . . . , F − 1, F . These
quantum numbers represent the projection of the atomic angular momentum along
some axis, known as the quantization axis (see §F.4 in Appendix F). In the presence
of a magnetic field, the atomic angular momenta precess around the direction of the
field, and it is natural to define the quantization axis as aligned with this direction.
With this definition, the quantum numbers m remain good quantum numbers, but
the degeneracy of the Zeeman sublevels is lifted — this is known as the Zeeman
shift. Optical transitions between the Zeeman sublevels occur subject to selection
rules, which determine whether or not a transition conserves angular momentum,
10.4 Cell
330
based on the polarization of the incident light. With a judicious choice of polarized
lasers, one can prepare the ensemble in just one Zeeman sublevel. Such a ‘spin
polarized’ ensemble is used by Julsgaard et al. when implementing their continuousvariables memory in cesium [122] (see §2.4 of Chapter 2). This type of ensemble
state is exquisitely sensitive to external magnetic fields, and it is common to build
a magnetic shield of so-called µ-metal (an alloy with a high magnetic permeability
that deflects magnetic field lines) around the vapour cell.
In our system, the spectral bandwidth of the laser pulses comprising the signal
and control fields is much larger than any feasible Zeeman shift one could produce
(see §10.6). Therefore the Zeeman sublevels cannot be resolved in the quantum
memory interaction, and so we neglect the Zeeman substructure of the hyperfine
states. There is no need to spin-polarize the ensemble, and no need for magnetic
shielding.
In §F.4 of Appendix F, we show that orthogonal circular polarizations are not
coupled by the Raman interaction. This is a further reason why it is not useful to
polarize the ensemble: it might have been possible to use the Zeeman selection rules
to our advantage (as is explained in the Appendix), but this result obviates this
possibility.
An unpolarized ensemble is in a mixed quantum state. It can be thought of as
several independent sub-ensembles, each with a different spin polarization. Each subensemble interacts coherently with the optical fields, however, and the theoretical
description of the memory interaction given in Chapter 4, 5 remains valid for each
10.5 Buffer gas
331
sub-ensemble.
The lack of any magnetic shielding means that there may be stray magnetic
fields that introduce a distribution of frequencies into the evolution of the Raman
coherence through the Zeeman shift. This may cause the spin wave to dephase, so
that the coherence is lost. However, as long as the magnetic fields remain constant
over the memory lifetime, the spin wave will re-phase, because the number of Zeeman
components is finite (this periodic re-phasing of a discrete set of oscillators is the
principle behind retrieval from the AFC quantum memory [91] ; see §2.3.4 in Chapter
2). The periodic beating between the Zeeman sublevels restricts the times at which
efficient retrieval is possible to those times where the spin wave is ‘in phase’, but the
efficiency of the memory is not adversely affected.
The above discussion justifies our decision not to build a magnetic shield for our
cesium cell, and indeed not to attempt to polarize the vapour. However, since we
have not yet been able to observe strong Stokes scattering, we cannot be sure that
a magnetic shield would not help. We are investigating the construction of such a
shield.
10.5
Buffer gas
The 20 torr of neon buffer gas is added in order to extend the time that the cesium
atoms spend in the interaction region. The cesium atoms are deflected by collisions
with the neon atoms, so that their motion is diffusive rather than ballistic. The
mean time between collisions is given by 1/γp , as can be verified by computing
10.5 Buffer gas
hτ i =
Rτ
0
332
τ ps (τ ) dτ using the ‘survival’ distribution (F.15) in Appendix F. If we
follow the trajectory of a cesium atom, it will trace out a random walk with an
average step length l = vth /γp , where vth =
p
2kB T /M is the average thermal
velocity of the atom. After N such steps, the root-mean-square displacement of the
atom from its starting point is D =
√
N l. The beam diameter is roughly
√
A, so an
atom escapes the beam after a time tescape , where
D(tescape ) ∼
⇒ tescape ∼
√
A,
AM γp
.
2kB T
(10.1)
A typical diffusion time for a beam with diameter 100 µm is around 10 µs at room
temperature, which compares with around 0.5 µs in the absence of a buffer gas.
During the course of this Brownian motion, the atom undergoes around 105 collisions. As described in Appendix F, these collisions randomize the phase of the
optical polarization, causing pressure broadening. But the hyperfine coherence —
the spin wave — is unaffected. The ground state hyperfine levels represent different
orientations of the cesium nuclear spin, but neon is spinless, so there is no magnetic
dipole interaction between the atomic nuclei in a collision. The Raman coherence is
therefore maintained, despite the frequent collisions of cesium atoms with the buffer
gas. The natural lifetime of the hyperfine coherence is several thousand years, so
the memory lifetime is set by tescape .
As well as slowing the escape of the cesium atoms, collisions with the buffer gas
10.6 Control pulse
333
change their velocities, so that the cesium atoms diffuse spectrally: as the cesium
atoms’ velocities change, so do their Doppler shifted resonant frequencies. Before
an atom leaves the interaction region, its resonance will ‘wander’ across the entire
Doppler profile (see §10.10 below). This means that the atoms can be optically
pumped (see §10.12 below) using a narrowband laser tuned into resonance with
stationary atoms, since all atoms will eventually drift into resonance with the pump
laser. Without a buffer gas, it is necessary to dither the pump laser frequency in
order to address all the atoms, which is inconvenient (although certainly possible).
10.6
Control pulse
The control pulse, or alternatively the Raman pump light, is sourced from a bespoke
Ti:Sapphire laser oscillator: a Spectra-Physics Tsunami. The oscillator is actively
modelocked using an acousto-optic modulator installed in the laser cavity, and the
laser produces a pulse train with a repetition rate of 80 MHz, which is set by the
cavity round-trip time.
10.6.1
Pulse duration
The bandwidth of the control pulse cannot exceed the 9.2 GHz Stokes splitting between the two lower hyperfine states, if a quantum memory is to be implemented.
However the gain bandwidth of Ti:Sapphire is very wide (several hundred nanometers!), so some care was taken to limit the bandwidth of the laser. A Gires-Tournois
Interferometer (GTI) is installed inside the laser cavity. This is essentially a Fabry-
10.6 Control pulse
334
Perot etalon, with the rear of the two plates having a reflectivity of 100%. All
incident energy is therefore reflected, but the spectral phase inside the cavity undergoes periodic jumps, with the free spectral range determined by the plate separation.
Any pulse in the laser cavity with a bandwidth spanning one of these phase jumps
experience catastrophic dispersion, which greatly reduces the efficiency with which
the pulse — being heavily distorted — can extract energy from the Ti:Sapphire
crystal. The bandwidth of the laser is therefore limited to bandwidths smaller than
the free spectral range of the GTI. Use of a GTI with an unusually large plate separation produces a modelocked pulse train with a spectral bandwidth or around 1.5
GHz (∼ 3.6 pm at 852 nm).
x
Mirror
Beamsplitter
Mirror
Retroreflector
intensity (arb. units)
Photodiode
1000
500
0
150
200
250
x (mm)
300
Figure 10.4 First order autocorrelation. The set up is essentially
what is known as a Fourier transform spectrometer. An incident
pulse is split into two components, and one is delayed with respect to
the other. The arrangement with the retroreflector and facing mirror
makes the movable arm insensitive to misalignment as the retroflector
is translated over a large distance.
To characterize the pulse train, measurements of the first and second order correlation functions of the output light were made. A Michelson interferometer was
10.6 Control pulse
335
built, so that interference of the optical field with a delayed copy of itself could
be observed. The first order field autocorrelation, shown in Figure 10.4, produces
a fringe pattern whose envelope is equal to the Fourier transform of the spectral
intensity I(ω) of the laser output (see (F.16) in Appendix F for an example of this
relationship). The spectral intensity is found to be approximately Gaussian, with
a FWHM bandwidth of 1.5 GHz, as mentioned above. This is consistent with a
Fourier-transform-limited pulse duration of 300 ps, but it may be that distortions
to the spectral phase of the pulse train ‘smear out’ the pulses, producing longer durations. To investigate this possibility, we performed a second order interferometric
autocorrelation. The experimental set up is shown in Figure 10.5. Two delayed
copies of an incident pulse are focussed into a non-linear crystal — a small piece of
β-Barium Borate (BBO). Blue light scattered in the forward direction results from
the frequency upconversion of one pulse with its delayed counterpart. The envelope
of the resulting fringe pattern can be related to the pulse duration. Although the
shape of the autocorrelation envelope is not uniquely related to the pulse envelope
(it is not possible to ‘invert’ the autocorrelation to retrieve the pulse profile), it is
possible to infer the presence of spectral phase distortions by inspection of the lower
portion of the interferogram. In the presence of ‘chirp’ (a drift in the carrier frequency of the pulse as a function of time), interference between the leading edge of
one pulse, and the trailing edge of the other, is suppressed, since the carrier frequencies are no longer commensurate. This causes a distinctive narrowing of the lower
portion of the interferogram that is not present in our measurement. On the basis of
10.6 Control pulse
336
this result, we conclude that the pulses produced by our laser source are close to being transform-limited, with approximately Gaussian spectral and temporal profiles.
The FWHM duration of each pulse is around 300 ps, and the spectral bandwidth is
approximately 1.5 GHz. On the one hand, this is much narrower than the Stokes
splitting, as required. On the other hand, the pulses are much more broadband than
have been used to date in quantum memory experiments (pulse durations of 100’s
of microseconds are the shortest that are employed for EIT [74] ).
Filter
BBO
Lens
x
Mirror
Beamsplitter
Mirror
Retroreflector
intensity (arb. units)
Photodiode
600
400
200
0
150
200
250
x (mm)
300
Figure 10.5 Second order interferometric autocorrelation. A nonlinear crystal is inserted, and the frequency upconversion of one pulse
with its delayed counterpart is detected. The inset shows the measured envelope of the interferogram. Again, the red lines are Gaussian fits. These data, along with the data in Figure 10.4, are the only
experimental data in this thesis that were taken by me!
10.6.2
Tuning
The plate separation of the GTI can be temperature tuned using a knob on top of the
laser, and this also provides a convenient way to smoothly tune the laser frequency
over small frequency shifts, of order 10 or 20 GHz (it appears that the spectral phase
10.7 Pulse picker
337
introduced by the GTI introduces differential gain over its free spectral range, which
allows small changes to the FSR to sweep the laser frequency). Larger frequency
shifts can be ‘dialled in’ using a birefringent filter (this is actually an intra-cavity
Lyot filter — see §10.13.2 below).
10.6.3
Shaping
Although a great deal of the theoretical work in this thesis deals with the optimization of quantum storage by appropriate shaping of the control or signal pulse profiles,
it is not feasible to shape the pulses used in this experiment. They are too short
to be shaped electronically, and two narrowband to be shaped spectrally! However,
since we are still struggling to demonstrate storage in the first place, we are content
to walk before attempting a steeplechase. The excellent work of Novikova et al. [74]
on resonant shaped storage is a vindication of the theory of Gorshkov et al. [133] , and
by extension the theory in this thesis, since they are so closely related (at least at
optical depths less than 50 or so [151] ).
10.7
Pulse picker
The time between consecutive pulses in the output of our laser is 12.5 ns. This
is much shorter than the lifetime of the Raman coherence (see §10.5 above), so if
the pulse train from the laser is sent ‘as is’ into the cell, the ensemble does not
recover after each Raman interaction. If one’s aim is to generate strong Stokes
scattering, this may be beneficial. The Stokes process can be stimulated both the
10.8 Stokes scattering
338
presence of Stokes photons, and by the presence of spin wave excitations, so the
Raman coherence produced by a previous pulse may stimulate the Raman interaction
in subsequent pulses. However it is not possible to investigate the efficiency of a
quantum memory if different realizations of the memory interaction are coupled. To
demonstrate the quantum memory, it is necessary to reduce the repetition rate of the
laser so that each implementation of the memory is independent of all others. This
is done with a ‘pulse picker’, which is a fast optical switch based on the electro-optic
effect, known as a Pockels cell. The device is synchronized with the laser output,
and the optical switch is set to transmit every 80, 000th pulse, reducing the laser
repetition rate to 1 kHz. The ensemble now has 1 ms to ‘reset’ between pulses,
which is plenty of time.
In the next section we introduce the theory of Stokes scattering.
10.8
Stokes scattering
The requirements for observing strong Stokes scattering are very similar to the
requirements of a Raman quantum memory. A medium must be prepared in the
ground state of a Λ-system. There should be strong Raman coupling between an
incident laser pulse and the excitations of the medium. And it should be possible
to observe scattering at the Stokes-shifted wavelength by filtering out the strong
Raman pump.
A description of Stokes scattering runs along very similar lines to that of a Raman
quantum memory: the signal and control fields simply trade places, becoming pump
10.8 Stokes scattering
339
and Stokes fields (see part (a) of Figure 1.6, or part (b) of Figure 4.3). However,
instead of absorption, the Stokes field experiences gain, with energy being transferred
from the pump field into the Stokes field, while at the same time excitations are
generated in the Raman medium.
We are interested in transient Stokes scattering, which refers to the regime in
which the duration of the Raman pump pulse is much shorter than the lifetime of
the excitations in the medium. This is consistent with the requirement that the
same medium should be useful as a quantum memory, which requires that the spin
wave excitations far outlive the optical pulses.
The equations describing transient Stokes scattering in the adiabatic limit can
be written in the following simple form,
∂z α = Cβ † ,
∂ω β = Cα† .
(10.2)
Here all the notation has the same meaning as in (5.109) from §5.3.3 in Chapter 5,
except that α is a dimensionless annihilation operator for the Stokes mode. Recall
that ω is the integrated Rabi frequency (defined in (5.46) in §5.2.6 of Chapter 5),
where now the relevant Rabi frequency is that of the Raman pump pulse. The coupling constant C is the same as the Raman memory coupling, so the efficiency of
both Stokes scattering and quantum memory are characterized by the same number.
The equations (10.2) describe a squeezing, or Bogoliubov transformation as opposed
to a beamsplitter interaction: optical and material excitations are produced in cor-
10.8 Stokes scattering
340
related pairs. This interaction is the basis of the DLCZ quantum repeater protocol,
since the Stokes mode becomes entangled with the ensemble (see §1.6.4 in Chapter
1). The equations can be solved in precisely the same manner as the system (5.109).
The solution for the Stokes field is
r
h √
i
C2
0 α (ω 0 ) dω 0
αout (ω) = α0 (ω) +
I
2C
ω
−
ω
1
0
ω − ω0
0
Z 1 h p
i
I0 2C (1 − z)ω β0 (z) dz.
+C
Z
ω
(10.3)
0
Here the I’s denote modified Bessel functions, which describe exponential gain. We
are interested in the intensity of spontaneously initiated Stokes scattering, when
there are initially no spin wave excitations and no photons in the Stokes mode.
The average number of Stokes photons produced is given by the normally ordered
product
Z
hNout i =
0
1
†
hαout
(ω)αout (ω)i dω.
(10.4)
On substitution of (10.3) into (10.4), a number of terms vanish. The cross terms
involving the products hα0† (ω)β0† (z)i and hβ0 (z)α0 (ω)i are both zero, since in the
first case the inner product between singly-excited and vacuum states is taken, and
in the second case the vacuum state is annihilated. The same is true of the term
involving hα0† (ω)α0 (ω)i. The only non-vanishing term involves the anti-normally
ordered combination hβ0 (z)β0† (z)i. To evaluate this term, we observe that β is a
bosonic annihilation operator which satisfies the same commutation relation as B
10.8 Stokes scattering
341
(see (4.45) in Chapter 4, or (9.27) in Chapter 9),
h
i
β(z), β † (z 0 ) = δ(z − z 0 ).
(10.5)
We therefore have that hβ0 (z)β0† (z 0 )i = hδ(z −z 0 )+β0† (z 0 )β0 (z)i = δ(z −z 0 ). Inserting
this result into the expression for hNout i, we obtain2
hNout i = 2C 2 I02 (2C) − I12 (2C) − CI0 (2C)I1 (2C).
(10.6)
Figure 10.6 shows the average number of scattered Stokes photons as a function of
the coupling C. C = 1 corresponds to hNout i ∼ 1, and this may be thought of as
marking the onset of the stimulated scattering regime, when previously scattered
Stokes photons stimulate further scattering, so that the scattering efficiency begins
to grow exponentially with C. For efficient Raman storage, a quantum memory
requires C ∼ 1 also, so the possibility of producing stimulated Raman scattering is
a necessary condition for implementing a Raman quantum memory in any system.
In the next section we explain how to calculate a prediction for the coupling
constant C, when an atomic vapour is used as the memory medium (a calculation
of C for the case of a diamond quantum memory was undertaken at the end of the
previous Chapter).
2
This result appears as equation (38) in the seminal paper on Stokes scattering by Michael
Raymer and Jan Mostowski [139] , who specialized to the case of a square pump pulse.
10.9 Coupling
342
10
10
10
10
10
5
0
−5
0
2
4
6
8
Figure 10.6 Stokes scattering efficiency. As the coupling C increases, the number of Stokes photons scattered rises, growing exponentially for C & 1. The model neglects depletion of the energy in
the pump pulse: of course the number of Stokes photons scattered
cannot exceed the number of photons in the pump pulse, or indeed
the number of atoms in the ensemble.
10.9
Coupling
Throughout this thesis we have made reference to the quantities d, Ω, C etc...
Fortunately these quantities are not difficult to calculate in the context of an atomic
vapour.
10.9.1
Optical depth
Combining the definition (5.12) in §5.2.3 of Chapter 5 with (4.50) in §4.11 of Chapter
4, the optical depth can be written as
d=
|d∗12 .vs |2 ωs nL
.
2γ0 ~c
(10.7)
The spontaneous emission rate 2γ can be expressed in terms of the dipole moment,
using Fermi’s golden rule to derive the stimulated emission rate, and then Einstein’s
10.9 Coupling
343
relations to connect this to the spontaneous rate [107] . The result is
2γ =
3 |d |2
ω21
12
.
3π0 ~c3
(10.8)
Here we have neglected any factors arising from degeneracy of the states involved.
The factor of 1/3 represents an average over all spatial direcitons. Substituting
(10.8) into (10.7), and assuming that the signal polarization vs is aligned with the
dipole moment d∗12 , gives the result
d =
3
× nλ2 L
4π
∼ nλ2 L,
(10.9)
where λ = 2πc/ω21 is the wavelength associated with the |1i ↔ |2i transition, and
where we have made the approximation ωs ≈ ω21 (any detuning is much smaller than
an optical frequency). (10.9) is consistent with the notion that the scattering crosssection of an atomic transition is roughly λ2 . The approximation in the second line of
(10.9) is not generally accurate, but it is extremely useful as an ‘order of magnitude’
estimate for the optical depth. The number density n can be found from the vapour
pressure, given the temperature of the atomic ensemble (see §F.1 in Appendix F).
For an optical transition with λ ∼ 1 µm and a typical ensemble length of L ∼ 1 cm,
we have d ∼ n[m−3 ] × 10−14 . As a rule of thumb, this allows one to easily estimate
the atomic number density required for an efficient memory (with d & 100).
In general, a more accurate value for the optical depth should be calculated by
10.9 Coupling
344
using empirical values for the rate γ (recall that 2γ is the spontaneous emission rate)
and for the dipole moment d12 . Sometimes data tables list the values of oscillator
strengths associated with atomic transitions. These are dimensionless numbers that
quantify the dominance of a transition over other possibilities within the atom. The
connection between the oscillator strength and the dipole moment for a transition
is derived in §F.2 of Appendix F. The optical depth can be expressed in terms of
the oscillator strength f12 for the |1i ↔ |2i transition as follows,
d=
π~α
m
×
f12 nL
,
γ
(10.10)
where m is the electron mass and α = 1/137 is the fine structure constant. Figure
10.7 below shows the variation of the optical depth of our cesium cell, with a length
of 10 cm, as a function of temperature.
10.9.2
Rabi frequency
The Rabi frequency is given by
Ω=
d23 .vc Ec
,
~
(10.11)
where Ec is the electric field amplitude of the control associated with its positive
frequency component. The instantaneous intensity of the control field is given by
1
Ic = 0 c × |2Ec |2 .
2
(10.12)
10.9 Coupling
345
8
Optical depth
10
6
10
4
10
2
10
300
350
400
450
Temperature, K
500
Figure 10.7 Cesium optical depth. This is the optical depth associated with the complete 6P3/2 manifold. The optical depth is the
same whether we consider transitions from the F = 3 ground state
level, or the F = 4 level. The dipole moment for the cesium D2
line is around 10−29 Cm, according to the data provided by Daniel
Steck [184] (this figure has been divided by a factor of 3, because we
consider linearly polarized incident fields). The length of our cell is
10 cm. The number density is found from the vapour pressure curve
in Figure F.1 in Appendix F.
An estimate of the energy in the control pulse is then given by Ec = Ic ATc . Neglecting any complex phase in Ec , we can express the peak Rabi frequency of the control
in terms of this energy,
Ωmax
d23 .vc
×
≈
~
s
Ec
.
20 cATc
(10.13)
This can, of course, be expressed in terms of the oscillator strength for the |2i ↔ |3i
transition using the conversion formula given in Appendix F. If the control pulse
is taken from a pulse train — for example the output of a modelocked laser — we
can write Ec = P/R, where P is the average power in the beam, and where R is the
pulse repetition rate. The pulse duration Tc needs to be measured (see §10.6 above).
10.9 Coupling
10.9.3
346
Raman memory coupling
The Raman memory coupling is defined in ordinary units as C =
√
dγW /∆. An
approximate expression, assuming a top-hat shape for the control pulse, is
2
C ≈ Tc dγ ×
Ωmax
∆
2
.
(10.14)
If (10.13) is used in this expression, the factors of Tc cancel, and the assumption of a
top-hat profile can be dropped. Since adiabatic evolution requires that ∆ Ωmax ,
the above form of the Raman coupling makes it easy to see why an efficient, adiabatic
Raman memory should have Tc dγ 1 [133] . That is to say, the bandwidth δc ∼ 1/Tc
of the pulses stored (the optimal signal bandwidth is close to the control bandwidth)
should be small compared with dγ.
When estimating the magnitude of the Raman coupling, the following form can
be useful3 ,
C2 ≈
~
mA∆
2
× (πα)2 × (f12 f23 ) × Na Nc ,
(10.15)
where Na is the number of atoms in the ensemble that are illuminated by the optical
fields, and where Nc is the number of photons in the control pulse. Putting f12 ≈
f23 ≈ 1 provides an upper limit to the Raman coupling that is achievable with a
given number of atoms and control photons.
As an example, an optical depth of d ∼ 3.8 × 104 is predicted for our cesium cell,
of length 10 cm, when heated to 90 ◦ C. Setting the control field beam waist to 330
3
This form of the coupling appears in our Rapid Communication on Raman storage [77] , but
there is a typo! The form given here is correct.
10.9 Coupling
347
µm gives a Rayleigh range (see §6.4 of Chapter 6) of 5 cm, so that the beam remains
collimated over the full length of the cell. Setting the average power from our laser
oscillator to 600 mW gives a pulse energy of 7.5 nJ. One such pulse, focussed as
just described and detuned by ∆ = 20 GHz from the D2 line is sufficient to produce
a Raman memory coupling of C ∼ 2.4, corresponding to a storage efficiency of
∼ 99.9%. The memory balance is then R ∼ 0.5, so the memory is well within
the Raman regime (i.e. the interaction is balanced; see §5.3.1 in Chapter 5). The
detuning is around 50 times larger than the Doppler linewidth (see §10.10 below),
so the inhomogeneous broadening may be neglected. Adiabaticity is guaranteed
because the detuning is by far the largest frequency involved in the interaction,
∆/γ ∼ 104 , ∆Tc ∼ 40, ∆/Ωmax ∼ 6.
10.9.4
Focussing
Substitution of (10.10) and (10.13) into (10.14) reveals that the Raman coupling
depends on the geometry of the ensemble in the following way,
C2 ∝
L
.
A
(10.16)
This is because the optical depth only depends on the length, and the squared Rabi
frequency depends only on the inverse of the beam area. The ensemble is addressed
by collimated laser beams, which in the ideal case (when the laser is working well)
have a Gaussian transverse profile. The Rayleigh range zR of a focussed Gaussian
10.9 Coupling
348
beam of wavelength λ is related to its cross-sectional area at the focus by diffraction,
zR =
A
.
λ
(10.17)
The beam remains well collimated over a region of length zR either side of the focus,
but after this is quickly diverges, and the intensity drops rapidly. Therefore we may
consider that the length of the ensemble over which the memory interaction is reasonably strong is limited by the Rayleigh range, L ∼ zR . Making this identification
in (10.16), and using the relation (10.17), we find
C2 ∝
1
.
λ
(10.18)
That is, the geometrical dependence of C drops out. The Raman coupling is independent of how the beams are focussed. Loosely focussed beams are better described
by the one-dimensional theory, and indeed numerical simulations show that a loosely
focussed control beam improves the memory efficiency (see 6.5 of Chapter 6). But
if the beams are too loosely focussed, their Rayleigh range will extend beyond the
cell length, and the coupling will be limited by the dimensions of the cell. Therefore
the optimal situation obtains when the Rayleigh range of the beams is matched to
the cell length, zR = L/2.
10.10 Line shape
10.10
349
Line shape
Figure 10.8 shows an absorption spectrum for the D2 line, measured using a laser
with a spectral bandwidth of 1.5 GHz, with the cell heated to 70 ◦ C. A number
of factors contribute to the shape of the D2 absorption lines in a warm cesium
vapour. The absorption linewidth of each transition in the D2 line is around 250
MHz at room temperature for a single atom. But the large optical depth widens
the absorption lines of the ensemble (see §7.2.1 in Chapter 7, and §10.11 below).
The dominant line broadening mechanism is Doppler broadening, with pressure
broadening contributing at the level of ∼ 10 MHz. These mechanisms are explained
in §F.3 of Appendix F.
The Doppler width is calculated using (F.12) by setting M = 133 a.m.u. — the
mass of a cesium atom — and ω0 = 2πc/λ0 , with λ0 = 852 nm the cesium D2 line
wavelength.
The pressure-broadened linewidth is calculated from (F.18), but the parameters
used are those of the buffer gas. The cesium cell contains neon, which was back-filled
up to a nominal pressure of pbuffer = 20 torr. The buffer gas reduces the mean free
path of the cesium atoms, so that they stay in the interaction region — the volume
illuminated by the light beams — for a longer time than they otherwise would (see
§10.5 below). The number density of neon is much greater than the number density
of cesium, at all reasonable temperatures, so a cesium atom is much more likely
to collide with a neon atom than with another cesium atom. The relevant number
density in estimating the pressure-broadened linewidth is therefore n = pbuffer /kB T0 ,
10.10 Line shape
350
1000
Signal, mV
800
Absorption spectrum
No cell
600
400
200
0
−10
0
10
Detuning, GHz
20
Figure 10.8 Cesium D2 absorption spectrum. This plot appears
thanks to Klaus Reim, a D.Phil student currently dividing his time
between the cesium and quantum dot projects. A weak laser is
scanned across the cesium D2 line at 852 nm. The blue line shows
the signal from a photodiode placed after the cesium cell. The red
dotted line indicates the signal detected if the cell is removed. The
two dips correspond to the two ground state hyperfine levels: on the
left is the F = 3 state; the F = 4 state is on the right. The 9.2 GHz
splitting between these states is evident. The hyperfine structure of
the upper 6P3/2 state is not resolved. The laser used has a Gaussian
spectrum with a FWHM bandwidth of ∼ 1.5 GHz (see §10.6), and
this contributes to the wide absorption linewidth. Doppler broadening, along with the ‘smearing’ effect of the hyperfine splitting in
the upper state, and the large optical depth (d ∼ 104 at 70 ◦ C), accounts for the remainder of the linewidth, with pressure broadening
contributing negligibly.
where T0 ∼ 300 K is the temperature at which the cell was filled. The cell is sealed,
so the number density of the buffer gas is fixed. The relevant collision velocity is
the thermal velocity of Neon atoms, calculated using M = 10 a.m.u.. A reasonable
figure for the collision cross-section σ is found by choosing datom = (300 + 50)/2 pm,
which represents an average of the atomic diameters of a cesium and neon atom.
Figure 10.9 shows the variation in the Doppler and pressure-broadened linewidths
with temperature for our cesium cell. The temperature dependence is very weak,
10.11 Effective depth
351
since both linewidths scale linearly with the thermal velocity of the atoms in the
vapour cell, which in turn scales with
√
T . It’s clear that Doppler broadening dom-
inates over pressure broadening. The linewidth is similar to the hyperfine splitting
in the 6P3/2 excited state manifold, so these hyperfine states are not resolved in our
sample. Fortunately it is not important, either for optical pumping (see §10.12 below), or for the quantum memory interaction, to distinguish the hyperfine structure
of the excited state. Provided that Raman storage is implemented with a detuning
that is large compared to both the Doppler linewidth and the hyperfine splitting,
a theoretical description that ignores these complications remains appropriate. In
fact, Gorshkov et al. have shown that even resonant storage is unaffected by Doppler
broadening provided the ensemble is sufficiently optically thick [163] . However, characterizing the optical depth of the ensemble is not entirely straightforward, in the
presence of line-broadening.
10.11
Effective depth
If a weak probe beam is tuned into resonance with one of the D2 line transitions,
the measured attenuation is much less than would be predicted on the basis of the
optical depth estimated using (10.9). The reason for this is the redistribution of
optical depth over a wide spectral range, because of Doppler broadening. Suppose
one measures the transmitted power Pout , and the input power Pin . The effective
optical depth is defined by
Pout = Pin × e−2deff .
(10.19)
10.11 Effective depth
352
9
10
Linewidth, Hz
Doppler
8
10
Pressure
7
10
6
10
300
350
400
450
Temperature, K
500
Figure 10.9 Absorption linewidth. The pressure (green) and
Doppler (blue) linewidths γ/2π for our cesium sample, calculated
using the formulae (F.12) and (F.18) derived in Appendix F. The
red dotted line shows the half-width at half-maximum of the resulting Voigt profile, which is the line profile resulting from the convolution of the pressure-broadened Lorentzian and Doppler-broadened
Gaussian lineshapes. Doppler broadening dominates however, and
the lineshape is essentially Gaussian.
The spectrum of the transmitted beam can be found from (7.29) in Chapter 7,
setting z = 1 and Pin = 0. In the presence of pressure-broadening, the polarization
decay rate is modified,
γ −→ γ 0 = γ + γp /2,
(10.20)
and the pressure broadened optical depth dp = d×γ/γ 0 should be used. The effective
optical depth is then given by
deff
1
= ln
2
R
R
Iin (ω) dω
Iin (ω)e−2dp <[f (ω)] dω
,
(10.21)
where Iin (ω) is the spectral intensity profile of the probe beam, and where the
lineshape function f (ω) is defined in (7.28) in Chapter 7, the only difference being
10.12 Optical pumping
353
that all frequencies are normalized by γ 0 instead of γ.
Using a beam with an average power of 1 µW, taken from the output of a
modelocked Ti:Sapphire oscillator with a spectral bandwidth of 1.5 GHz (see §10.6),
we measure deff ∼ 3 at a cell temperature of 90 ◦ C. If we invert the formula (10.21),
using γ = 16.4 MHz (the spontaneous lifetime of the cesium D2 line is τ = 1/2γ = 30
ns), γ 0 ≈ 70 MHz and γd ≈ 260 MHz, we infer a ‘real’ optical depth of d ≈ 3,500.
This is consistent with the prediction d = 3,700, found from (10.7) using the number
density plotted in part (b) of Figure F.1.
10.12
Optical pumping
Even though the hyperfine splitting is extremely large, it represents a very small energy gap at room temperature, with Esplitting ≈ kB T /680. Therefore the populations
of both the clock states are equal. From the perspective of quantum storage, this
means there is a large thermal background of incoherent material excitations that
would swamp any stored signal. But the situation is worse than this: the memory
interaction is totally destroyed if the populations of the two lower states are equal.
As signal photons are absorbed, they are also produced by Stokes scattering from
the thermal population in the storage state. The gain from thermally seeded Stokes
scattering exactly balances the quantum memory absorption, and the memory is
rendered useless. This effect is demonstrated in Figure 10.10.
For an efficient quantum memory, therefore, one of the lower two states must be
emptied. This is done in the laboratory by by optical pumping.
10.12 Optical pumping
(b)
Intensity
(arb. units)
(a)
354
Figure 10.10 Equal populations destroy quantum memory. (a):
Absorption of a Gaussian signal pulse with the storage state empty.
The memory efficiency is 73% (b): With equal populations in both
ground and storage states, the absorption vanishes. The memory efficiency here is ∼ 0.01%. These plots are produced by numerically
integrating the full system of Maxwell-Bloch equations, without making the linear approximation introduced in §4.5 of Chapter 4. There
are 8 equations to solve, one for each optical field (i.e. the signal
and the control pulses), one for the population of each state in the Λsystem, and three equations describing the coherences between pairs
of states. The numerical methods used are described in Appendix
E. We consider a Raman quantum memory, with ∆ = 150, d = 300,
and W = 122.5, and a Gaussian control pulse with duration Tc = 0.1
(working in normalized units with frequencies expressed in terms of
γ, the natural linewidth of the excited state — pressure and Doppler
broadening are not modelled). The Raman memory coupling is then
C = 1.28, and the balance is R = 0.64 (see §5.3.1 in Chapter 5).
The signal field profile is identical to the control profile (no optimization is applied), but its amplitude is 1000 times smaller than that
of the control (the relative amplitudes of the signal and control are
important when the equations are not linearized).
Suppose we wish to use the F = 4 state as the ‘ground state’ for the memory
— such a choice allows for efficient phasematched backward retrieval, as discussed
in §6.3 of Chapter 6. We then need to pump all the atoms into the F = 4 state,
leaving the F = 3 state completely empty. We accomplish this by tuning a laser
into resonance with the |F = 3i ↔ |2i transition, where |2i stands for any of the
states in the 6P3/2 manifold. The laser field excites atoms into |2i, which then decay
10.12 Optical pumping
355
by spontaneous emission back into both the F = 3 and F = 4 ground states (the
branching ratio is roughly equal). The laser field then re-excites the atoms, and
the process repeats (see Figure 10.11). Consider the effect of consecutive cycles of
excitation and decay. With each cycle, only around 50% of the initial population
of the F = 3 state ends up back in F = 3. The rest ends up in the F = 4 state.
Since there is no laser field exciting atoms out of this state, population builds up in
F = 4, while the F = 3 state is eventually emptied entirely. This is the principle
behind optical pumping.
Figure 10.11 Optical pumping. A CW diode laser is tuned into
resonance with the |F = 3i ↔ |2i transition. Spontaneous emission
redistributes the excited population more-or-less equally between the
two ground hyperfine levels. After several cycles, population builds
up in the un-pumped F = 4 state, while the F = 3 state is emptied.
10.12.1
Pumping efficiency
To characterize the optical pumping efficiency a simple measurement was made (this
was done by Virginia Lorenz and Klaus Reim), based on an experiment performed
by Jean-Louis Picqué in 1974 [185] . The set-up is illustrated in Figure 10.12. We use
an external cavity diode laser to provide the pumping light. This is a single-mode
10.13 Filtering
356
diode laser with a diffraction grating placed in front of it in the Littrow orientation:
the first order of diffraction is directed back into the laser. By fine-tuning the
grating angle, it is possible to select those frequencies that experience the greatest
optical gain (other frequencies are misaligned by the grating and are lost). The
diode laser frequency can therefore by stabilized and tuned using the grating. The
specular reflection from the grating provides the laser output, which is a continuous
wave (CW) beam with a linewidth of around 100 MHz. Using the measurement
scheme shown in Figure 10.12, we detect the fluorescence signal from a weak probe
beam. The suppression of this signal with the pump beam set at 30 mW over the
signal with the pump blocked indicates that we achieve a pumping efficiency of
around 95%, with the cell temperature set at T = 50 ◦ C. At higher temperatures,
the optical pumping efficiency drops as the energy in the pump is absorbed by the
larger number of atoms. We are currently installing an ECDL with an output power
of 100 mW, which should improve our pumping efficiency for higher temperatures.
10.13
Filtering
The greatest experimental challenge posed by the implementation of a Raman quantum memory is the ability to filter out the very weak — even single photon — signal
field from the strong classical control pulse. The problem is particularly difficult in
cesium, because the Stokes shift of 9.2 GHz corresponds to a very small spectral
shift. At 852 nm — the D2 resonance wavelength — the signal and control frequencies differ by around 20 pm. The required filter contrast is of order 108 or better,
10.13 Filtering
357
Photodiode
Lens
Lock-in
Pump
Cs cell
Heaters
Beamsplitter
Probe
Chopper
Laser diode
Grating
Figure 10.12 Verifying efficient optical pumping. The beam from
an external cavity diode laser (ECDL) tuned into resonance with
the |F = 3i ↔ |2i transition is split into two parts at an asymmetric
beamsplitter (just a microscope slide). 96% of the diode beam is used
to pump the cesium atoms in the heated cell. 4% of the beam is sent
through an optical chopper, which applies a 1 kHz modulation, before
being directed through the cell at a small angle to the pump beam.
Fluorescence is collected by a lens and focussed onto a photodiode. A
lock-in amplifier is used to isolate the component of the fluorescence
due to the modulated probe beam. This signal quantifies the optical
pumping efficiency. The inset shows the reduction in the fluorescence
signal as the pump power is increased. A pumping efficiency of 95%
is achieved. These data appear thanks to Virginia Lorenz and Klaus
Reim.
so that no single filter provides sufficient rejection of the pump: several filters must
be used in tandem. In this section we give details of the various filtering techniques
we have at our disposal.
10.13 Filtering
10.13.1
358
Polarization filtering
The Raman interaction couples orthogonal linear polarizations: a horizontally polarized control pulse will store a vertically polarized signal pulse, and vice versa,
provided the detuning is sufficiently large. And any Stokes scattering induced by a
linearly polarized pump pulse is polarized orthogonally to the pump. These facts are
derived in §F.4 in Appendix F. Therefore the pump can be effectively rejected after
the cell using a polarizer aligned orthogonally to the pump polarization. We use
Glan-Laser polarizers from Foctek, which use total internal reflection at the boundary of two calcite crystals with perpendicular optic axes to preferentially transmit
one linear polarization. The extinction ratio of these polarizers is of order 10−6 .
However, stress-induced birefringence of the cell windows can distort the pump polarization, causing leakage of the pump light through the polarizer. The effective
polarization extinction is therefore of order 10−4 .
10.13.2
Lyot filter
A Lyot filter is a polarization-based spectral filter, popular among astronomers because of its wide working numerical aperture. In general a Lyot filter consists of
a number of concatenated filter stages, with each stage improving the finesse of
the filter (that is, reducing the spectral width of the pass-band, while increasing
the spectral width of the region over which frequencies are blocked). We built the
simplest possible type of Lyot filter, which involves a single stage. The filter then
has a sinusoidal transmission as a function of frequency. Figure 10.13 shows the
10.13 Filtering
359
structure of the filter. It is essentially a Mach-Zender interferometer for polarization: an incident pulse is split into fast and slow polarization components inside a
birefringent retarder. The polarizations are then combined on a polarizer, where
they interfere to produce spectral fringes. The free spectral range (i.e. the period
of the fringes in frequency) depends inversely on the length and birefringence of the
material used for the retarder. In order to be useful as a spectral filter for Stokes
scattering on the cesium D2 line, a free spectral range of ∆f = 2 × 9.2 GHz is
required, so that transmission of the Stokes light is accompanied by rejection of the
Raman pump light (see Figure 10.14). Calcite has the largest birefringence available, with |no − ne | = 0.17, but 10 cm of calcite are still required to produce the
required free spectral range! We use three pieces of calcite, each around 3 cm long,
mounted in series, along with a pair of Foctek Glan-laser polarizers (see previous
section). The filter contrast is limited to 99%, however, because of phase distortions
arising from the surface roughness of the calcite faces. The Lyot filter can be tuned
in frequency over one free spectral range by tilting one of the calcite crystals slightly,
which alters the optical path length.
10.13.3
Etalons
Since the Lyot filter does not have sufficient contrast for filtering the Stokes light, we
ordered some custom Fabry-Perot etalons. These are fixed air-gap etalons — pairs
of optically flat glass plates that form a planar cavity with a discrete transmission
spectrum. The free spectral range of the etalons depends on the separation L of
10.13 Filtering
360
(a)
(b)
optic axis
Fringes
B1
P2
P1
Retarder
B2
Figure 10.13 Lyot filter. (a): The filter consists of a piece of
birefringent material of length L — a retarder — placed between
polarizers P1 and P2. The optic axis of the retarder is aligned at
45◦ to P1, so that an incident pulse of wavelength λ (purple) is split
into ordinary and extraordinary polarization components (red and
blue pulses). One component is delayed with respect to the other,
because the refractive indices no , ne associated with each component
are different. The phase retardance is given by φ = 2π|no − ne |L/λ.
When the components recombine at P2, they interfere, producing
sinusoidal spectral fringes, with a free spectral range ∆f = c/L|no −
ne |. (b): Analogy with a Mach-Zender interferometer, in which two
delayed pulses are mixed on a beamsplitter, producing fringes.
the plates as ∆f = c/2L, since 2L is the distance an optical wave must traverse
to make a round-trip of the cavity. Requiring a free spectral range of 18.4 GHz,
as in the case of the Lyot filter, sets the etalon plate separation to be 8.2 mm. As
mentioned in §10.6 below, the laser bandwidth is around 1.5 GHz, so the width of
the pass band should not be smaller than this, in order to transmit the full Stokes
spectrum (which follows the Raman pump spectrum). The maximum finesse is then
F = 18.4/1.5 ∼ 12. The finesse of a Fabry-Perot etalon is fixed by the reflectivity
R of the plates comprising the cavity, according to the formula
√
π R
.
F=
1−R
(10.22)
10.13 Filtering
361
Pump
Stokes
9.2 GHz
Figure 10.14 Stokes filtering. A Lyot filter with a free spectral
range of 18.4 GHz can be used to suppress the Raman pump, while
transmitting the weaker Stokes field. The finesse F = ∆f /δf , where
δf is the FWHM of the pass band, is only 2, so this filter is far from
ideal.
The above considerations therefore fix the plate reflectivity to be 78%. This limits
the out-of-band extinction to around 95%, so one of these etalons is not quite as
good a filter as the Lyot filter. However, they are considerably more convenient,
being very easy to align, and much smaller! Six custom etalons with these specifications, and a large clear aperture of 15 mm, were ordered from CVI Melles Griot.
Initially it was intended to concatenate several etalons together, in series, so as to
combine their extinction. However, when the etalons are placed one after another,
their behaviour becomes complicated by interference effects arising from reflections
between the etalons. This is mitigated somewhat because the etalons we use have
anti-reflection coatings on their outside faces, which are also deliberately wedged.
Nonetheless it can be problematic to use the etalons consecutively.
10.13 Filtering
10.13.4
362
Spectrometer
Most recently, a grating spectrometer was built to filter the Stokes light (this was
done by Virginia Lorenz and Klaus Reim). The resolution required is R = λ/δλ >
852/0.02 ≈ 106 . The resolution of a grating spectrometer is limited by the number
of lines ruled on the grating, R ≈ N . The spacing between consecutive lines cannot
fall below half the optical wavelength, or the first order of diffraction does not exist.
For light at λ = 852 nm, this limits the maximum groove frequency to less than
2300 lines mm−1 . Using this groove frequency, a grating roughly 40 cm across is
required. Such a large grating is not available, but a spectrometer with a grating
around 6 cm across has been built, using a large off-axis parabolic mirror to focus
the diffracted light onto a photon-counting CCD camera. Although the resolution
of this spectrometer is sub-optimal, the ability to visualize the spectrum of the light
downstream from the cesium cell makes it an extremely useful apparatus.
All the above techniques have been used in an attempt to observe strong Stokes
scattering from the cesium cell. So far we have not been successful, but our combined
filter contrast is still improving.
10.13.5
Spatial filtering
The utility of spatial filtering for a quantum memory is discussed in §6.3.2 of Chapter
6. In a quantum memory, where both signal and control beams are directed by the
experimenter, it is possible to introduce a small angle between the beams so that
the strong control pulse can be blocked after the cell. When looking for stimulated
10.13 Filtering
363
Stokes scattering however, it is necessary to look along the axis of the Raman pump,
since the Raman gain is restricted to this direction. Therefore it is not generally
possible to spatially filter Stokes scattering, although attempting to detect scattered
light at very small angles from the pump direction may be sensible. A pinhole, or
even an optical fibre — fibre in-coupling is very spatially selective — can be used;
we are investigating these possibilities.
It is possible, however, to look for Stokes scattering in the backward direction.
The duration of our pulses is sufficiently long that they extend, in space, over the
length of the cell: cTc ∼ L. This means that a Stokes photon emitted near the exit
face of the cell in the backward direction, while illuminated by the leading edge of
the pump pulse, can propagate backwards, nearly all the way to the entrance face
of the cell, while all the time the cell remains illuminated by the pump, so that the
photon experiences Raman gain all along the cell’s length (see part (a) of Figure
10.15). The experimental set-up shown in part (b) of Figure 10.15 has been used
to look for this type of Stokes scatter. Strong stimulated fluorescence has been
observed, and the signal is extremely clear, because there is no need to filter out
the control pulse: it is propagating in the opposite direction! However the search
is still on for Stokes scattering, the signature being that the Stokes signal should
tune in frequency as the pump frequency is tuned, rather than remaining at the D2
resonance, as it does currently.
Similar arguments would suggest that a Raman quantum memory in which the
control and signal pulses are counter-propagating might be efficient, as long as the
10.14 Signal pulse
364
(a)
(b)
Figure 10.15 Backward Stokes scattering. (a): If the spatial extent of the Raman pump pulse is comparable to the cell length,
a backward scattered Stokes photon experiences Raman gain as it
propagates backward. (b): A polarizing beamsplitter can be used
to deliver the control pulse, and separate the orthogonally polarized
stimulated Stokes light that is scattered backwards. This essentially
eliminates the background from the Raman pump. We are currently
looking for Stokes light using this method.
pulses are sufficiently long. The propagation theory of Chapters 4 and 5 is not applicable in this situation, but the numerical model presented in Chapter 6 can be
used to show that the memory efficiency indeed remains high in this case. Such
an arrangement removes the demanding filtering requirements, and is therefore an
appealing possibility. A classical optical memory based on off-resonant Brillouin
scattering using counter-propagating pulses in an optical fibre has in fact been implemented [186] , so there is some precedent for this approach.
10.14
Signal pulse
Before implementing a single-photon quantum memory, a proof-of-principle demonstration using a weak coherent pulse for the signal field is planned. A single-photon
10.14 Signal pulse
365
source for the signal field is not, therefore, of immediate concern. However it remains
challenging to generate the signal field. Currently we have only one laser oscillator.
It is therefore necessary to generate the signal pulse from the control. The idea is to
sample a small portion of the control field, perhaps using a beamsplitter, and then
to apply a frequency modulation to shift the carrier frequency of the sampled pulse
by 9.2 GHz, so that this frequency shifted pulse can act as the signal.
Raman modulator One way to achieve this frequency modulation is to use the
cesium atoms themselves, by inducing strong Stokes scattering of a control pulse.
The Stokes sideband can then be sent into a second cesium cell along with the remaining control, where storage can be implemented. However, the Stokes scattering
process is somewhat aleatoric: large fluctuations in the intensity of the Stokes light
may make it difficult to assess the reliability of the memory.
EOM A second possibility is to use an Electro-Optic Modulator (EOM) to apply
the frequency shear. This is a device containing an electro-optically active crystal,
whose refractive index can be altered by the application of an external voltage.
Subjecting the crystal to a sinusoidally varying potential with frequency Ω will
imprint the waveform onto the phase of an optical wave passing through the crystal,
E(t) −→ E(t) × eiφ sin(Ωt) ,
(10.23)
10.15 Planned experiment
366
where φ quantifies the amplitude of the phase modulation. Fourier transforming
(10.23) reveals the presence of sidebands, separated by the modulation frequency,
e
E(ω)
=
∞
X
e + kΩ),
sgn(k)|k| J|k| (φ) E(ω
(10.24)
k=−∞
where the sideband amplitudes are given by Bessel functions. With the choice
φ ∼ 0.2, the energy in the first sideband represents around 2% of the total transmitted energy, while all higher order sidebands may be neglected. Use of such a
modulator allows one to reproducibly generate a weak signal pulse with the correct
frequency shift from the control pulse. A modulator with the ability to operate in
the microwave X-band at 9.2 GHz has been ordered from New Focus.
10.15
Planned experiment
The current experimental set-up is in a state of flux, we are continually improving
our filtering contrast and optical pumping efficiency, in the hope of detecting the
strong stimulated Stokes signal that indicates there is sufficient Raman coupling to
implement a Raman memory. Plans for the final implementation of the memory are
tentative, being contingent on the eventual success of these early stages. However in
the spirit of optimism we present in Figure 10.16 below a schematic of the possible
layout of an experimental demonstration of a cesium Raman memory. The optical
pumping beams are not shown, for clarity, although of course efficient optical pumping is critical. The off-axis geometry for phasematched retrieval described in Chapter
10.15 Planned experiment
367
6 is used, and we assume that the atoms have been prepared by optical pumping in
the upper F = 4 state for this purpose. The angle between the control and signal
beams is around 2◦ , and we assume that the control is more loosely focussed than
the signal (the signal focus can be tightened by expanding the signal beam before it
enters the confocal system, but we have not shown the beam expander). Since we
seek only to demonstrate the feasibility of the memory, a long memory lifetime is
not important, and so the delay between the storage and retrieval control pulses is
adjusted by a mechanical delay stage. It is only feasible to move such a stage over a
few feet, which corresponds to just a few nanoseconds of variability in the memory
storage time, but this is sufficient for our purposes. The quarter wave plate in the
control beam following the cell rotates the control polarization through 90◦ (since
the control beam traverses it twice, the combined effect being that of a half-wave
plate). The rotated control then retrieves the signal field into the orthogonal polarization mode to the polarization mode of the incident signal field. This allows the
retrieved signal to be re-directed to a detector using a polarizing beamsplitter. The
efficiency of the memory can be quantified by comparing the energy in the retrieved
signal field to that of the incident signal. The incident signal pulse is generated from
the control using an EOM (see §10.14 above), and an etalon removes the fundamental (the unmodulated light transmitted through the EOM). A Pockels cell is used
to reduce the repetition rate of the Ti:Sapphire laser, as described in §10.7.
We look forward to overcoming our present difficulties and assembling the above
apparatus, or a variation thereupon, in the near future.
10.15 Planned experiment
368
Delay stage
Ti:Sapphire
Pulse
Picker
L1
Block
E1
Cs cell
L2
QWP
EOM
E2
Retrieved signal
Figure 10.16 A possible design for demonstration of a cesium
quantum memory. The pulse train from a Ti:Sapphire oscillator
passes through a pulse picker to reduce the pulse repetition rate.
Consider a single pulse. A beamsplitter redirects a portion of the
pulse into an EOM, and the first sideband, shifted by 9.2 GHz is isolated from the output using a Fabry-Perot etalon E1 (see §10.13.3).
This is the input signal field. The remainder of the initial pulse is used
as the control field. It is directed through the cesium cell at a small
angle to the signal pulse using a confocal arrangement (lenses L1 and
L2): the signal is (hopefully) stored in the cell. The transmitted control field is sent through a variable delay line, and its polarization is
rotated through 90◦ , before being sent back through the cell. The
stored signal field is retrieved with the orthogonal polarization to the
incident signal, and is sent by a polarizing beamsplitter through an
etalon E2 (to remove any residual control) to a detector. Note that
we have not shown the optical pumping beams, which are critical.
This concludes the thesis. In the next chapter we summarize the results of the
present research.
Chapter 11
Summary
This thesis has been concerned with the problem of storing light. I can recall feeling
some puzzlement, lying in bed on a school night, at how completely my room darkened when the light was switched off. Why did you have to keep pouring more and
more light into a room? Well, light is an ephemeral beast. But the possibility of
a material that remembers its illumination — not with the feeble pallor of glow-inthe-dark paint, but with the unmistakable vigour of a laser pulse — is remarkable. I
am not the first to study such media, and the current research was undertaken in the
aftermath of the successes of light stopping by EIT. The main contributions of this
thesis are theoretical: a fairly general framework for the analysis and optimization
of light storage has been developed. The framework provides a unified description
of both EIT and Raman storage, and generalizes to tCRIB, lCRIB, AFC and broadened Raman protocols. Further applications of the formalism are expected. The
use of the SVD has been crucial to the success of the theoretical programme. Many
370
of the results in the thesis are simple adaptations of well-known facts from linearalgebra to the particular case under study. Another important ingredient of the
thesis is numerical simulation. The propagation of optical fields through an atomic
ensemble is always described by a set of coupled linear partial differential equations,
and these are particularly easy to solve on a modern computer.
The ‘take-home’ results are as follows.
1. Any quantum memory is a linear system, with storage and retrieval interactions described by Green’s functions, which are essentially large matrices.
2. The SVD of the Green’s functions provides a complete characterization of the
memory. It allows one to immediately identify the optimal input mode.
3. The singular values of the Green’s function are invariant under unitary transformations. This fact can be very useful in analyzing the memory interaction.
One can work in the Fourier domain, or indeed in an entirely unfamiliar coordinate system.
4. The efficiency of a memory is limited by its optical depth [133] .
5. Explicit expressions for the Green’s functions describing storage in a Λ-type
atomic ensemble are provided. The Rosen-Zener kernel holds whenever the
control has a hyperbolic secant temporal profile. The adiabatic kernel holds for
all detunings and control pulse shapes that satisfy the adiabatic approximation.
The Raman kernel holds for adiabatic interactions that are far-detuned and
‘balanced’.
371
6. In the adiabatic limit, a coordinate transformation links the results for different
control profiles. Therefore the SVD only needs to be computed once for a
given control pulse energy. Changes to the control profile simple change the
coordinate transformation.
7. A Raman memory may be characterized as a multimode beamsplitter interaction between optical and material modes. A single set of modes describes
both the transmitted fields and the stored excitations. A single number C
characterizes the efficiency of a Raman memory.
8. The Green’s function can always be constructed numerically, so that the optimal input modes and memory efficiencies can be found at any detuning,
regardless of whether or not the interaction is adiabatic.
9. Retrieval of the stored excitations can be problematic. Forward retrieval suffers from re-absorption losses. Backward retrieval is not phasematched if the
ground and storage states are non-degenerate. Numerical simulations verify that an off-axis geometry allows for efficient backward retrieval with nondegenerate states. Loose focussing of the control field is desirable.
10. The multimode capacity of a quantum memory can be evaluated by considering the SVD of the Green’s function. The multimode scaling of EIT, Raman, tCRIB, lCRIB, AFC, and a broadened Raman protocol is studied. Unbroadened ensembles have poor multimode scaling. Adding an inhomogeneous
broadening improves the scaling. The AFC protocol has the best multimode
11.1 Future work
372
scaling of all the protocols studied.
11. If one is not able to shape the signal pulse, optimal storage can still be achieved
by instead shaping the control. But to solve the optimization problem, the
equations of motion for the memory must be solved numerically. This can
be done rather quickly, however. A simple optimization algorithm works well
for finding the optimal control profiles. The SVD allows one to verify the
optimality of the numerical solutions. This optimality suffers as the interaction
becomes less adiabatic.
12. A Raman memory in bulk diamond, based on the excitation of optical phonons,
is feasible. It is shown that the equations of motion describing the Raman
interaction in diamond have precisely the same form as the Raman equations
describing storage in an atomic vapour. The form of the coupling constant C
is derived.
13. Attempts to implement Raman storage in cesium vapour have been made, but
it is proving difficult even to generate and detect Stokes scattering. I am not
a good experimentalist!
11.1
Future work
There is a great deal of experimental work to do. It may be that there is a good
reason why our attempts to build a Raman memory have been unsuccessful: we
should either make the memory work, or find this reason, in the coming year.
11.1 Future work
373
An intriguing theoretical challenge is how to make use of the stored excitations
once they are in place. Is it possible to perform computationally interesting operations on the spin wave in an atomic ensemble? Can operations be designed that
allow different stored modes in a multimode memory to interact? Work on this front
has begun in the literature [17,187,188] , but this is likely to be a rich seam.
If you have survived this far, I am very grateful for your attention! I hope that
some of the results in this thesis are useful to other workers in the field, even if only
as a warning of what not to try.
Appendix A
Linear algebra
Physics is generally concerned with change: the evolution of a system over time, or
the response of a system to an external agent. The easiest, and most uninspiring
situation to analyze, is when there is no change and no response. Linear algebra
is concerned with the much more interesting situation arising when the response
depends linearly on some parameter. On the one hand, this is almost always an
approximation that only holds for small changes in a parameter, and small responses.
So linear algebra rarely provides an exact description. On the other hand, the
linear approximation can be successfully applied to nearly every physical system!
Linear algebra is therefore useful in almost every branch of physics, as well as in
mathematics and science generally. In our case, the linear response of a quantum
memory is certainly an approximation, valid when the signal field does not contain
too many photons.
Here we summarize various concepts that are needed to properly understand the
A.1 Vectors
375
singular value decomposition as it pertains to the optimization of quantum storage.
There is a significant overlap with the formalism of quantum mechanics, so we
will also take this opportunity to review some aspects of that formalism. A clear
and comprehensive introduction can be found in Nielsen and Chuang’s quantum
information bible [158] .
A.1
Vectors
A vector is essentially a list of numbers. It also helps to keep in mind the image of
a vector as an arrow (see Figure A.1). This analogy cannot always be made with
rigour, but it provides a convenient visualization. The numbers comprising a vector
are the components of the arrow along the coordinate axes. When writing down
these components, implicit reference is therefore always made to some coordinate
system. It’s clear that we could rotate the coordinate axes — altering the vector’s
components — without changing the arrow in Figure (A.1), and in this sense, a
vector transcends its components. Nonetheless, it will be useful to write down the
vector components for concreteness.
We will use two equivalent sets of notation for a vector labelled ‘v’. Either v, or
|vi. The first symbol, in bold face, is in general use. The second — a ket — is an
example of ‘Dirac notation’, used only in the context of quantum mechanics. Dirac
notation is at times very convenient, and it will help to be able to use these two
types of notation interchangeably.
The number of components of a vector is called the dimension of the vector. If
A.1 Vectors
376
Figure A.1 A vector. On the left is a representation in ‘component
form’. On the right the same mathematical object is drawn as an
arrow. The direction and length of the arrow are determined by its
components α, β. Implicitly, a coordinate system (thinner arrows) is
used to define the components.
there are n components, the vector is said to be n-dimensional. We will adopt the
convention that v is a column vector;






v = |vi = 






v1 


v2 

.
.. 

. 


vn
(A.1)
Vectors can be added together, provided they have the same dimension,

 v 1 + w1


 v +w
 2
2
v + w = |vi + |wi = 

..

.



v n + wn






.





(A.2)
A.1 Vectors
377
And a vector can be multiplied by a number, say α, like this


 αv1 




 αv 

2 

αv = α|vi = 
 . .
 . 
 . 




αvn
(A.3)
Using these operations, vectors can be combined together to make new vectors, and
it is convenient to think of all these possible vectors as inhabiting a ‘space’, known
as an n-dimensional vector space. Our own universe, with three spatial dimensions,
can be thought of as a 3-space.
A.1.1
Adjoint vectors
The components of a vector do not have to be real numbers. In quantum mechanics,
and many other applications, they are generally complex numbers. A useful concept
in this case is the Hermitian adjoint v † of a vector v. This is simply another vector,
this time a row vector, with each component equal to the complex conjugate of the
corresponding component of v,
v † = hv| =
v1∗
v2∗
...
vn∗
.
(A.4)
The symbol hv| is known, rather unfortunately, as a ‘bra’, for reasons that will
become clear. These row vectors (bras) can be added or multiplied by numbers in
the same way as column vectors (kets), and so they form their own vector space,
A.1 Vectors
378
sometimes known as the adjoint space.
Every vector v has a corresponding Hermitian adjoint v † ; every ket has its
corresponding bra. And Hermitian conjugation is involutive: The Hermitian adjoint
of v † is v again.
A.1.2
Inner product
Vectors can be multiplied together in ways as various as mathematicians are inventive. The inner product — sometimes scalar product — is defined as the sum of the
component-wise products of a bra and a ket with the same dimension,
v † w = hv|wi = v1∗ w1 + v2∗ w2 + . . . + vn∗ wn .
(A.5)
This type of product between two vectors is not another vector; it’s just a number.
For some reason it’s rather satisfying to take an inner product, and Paul Dirac’s
notation anticipates something of this satisfaction. When a bra hv| encounters a ket
|wi they merge to become a ‘braket’ hv|wi, and so fulfill their destiny.
It’s quite common to speak of taking the inner product of two kets, |vi and
|wi. In this case it is understood that one of the kets has to be replaced by its
corresponding bra before using (A.5). Note that hv|wi = hw|vi∗ , so one should be
consistent about which of the two kets is replaced.
A complex vector space with an inner product defined as in (A.5) is known as
a Hilbert space. Quantum mechanics is a theory about vectors in Hilbert space; as
such it is extremely simple. It is the reconciliation of this mathematical structure
A.1 Vectors
379
with what we know about the real world that makes the theory so difficult to pin
down.
A.1.3
Norm
Having defined the inner product, we can now define the norm of a vector. This is
defined as the square root of the inner product of a vector with itself,
√
v = ||v|| = || |vi || =
v†v =
p
hv|vi.
(A.6)
This is a positive, real quantity that grows with the size of the components of v. In
fact, substituting in the definition (A.5) and applying Pythagoras’ theorem shows
that the norm is simply the length of the arrow representing the vector v. Some
further geometrical manipulations reveal that the inner product is related to the
angle θ between the arrows representing two vectors, as follows (see Figure A.7),
v † w = hv|wi = vw cos θ.
(A.7)
An immediate consequence of this is that the inner product hv|wi vanishes when
θ = π/2, that is, when v is perpendicular to w. Often, the word orthogonal is
used instead of perpendicular, especially in the case that the vectors involved are
complex, or when they have dimension greater than 3, since then the notion of an
angle is less transparent.
A.1 Vectors
380
Figure A.2
A.1.4
The inner product of two vectors.
Bases
We have already mentioned in passing the concept of a coordinate system, with
respect to which the components of a vector are defined. We drew the coordinate
axes in Figure A.1 as black arrows. The axes themselves are therefore described by
a pair of vectors, one pointing along the x-axis; the other along the y-axis. Let’s
call them |xi and |yi. If we fix the length (the norm) of these vectors as 1, then we
can write any vector v directly in terms of |xi and |yi,


 vx 
 = vx |xi + vy |yi.
|vi = 


vy
(A.8)
The set of two vectors {|xi, |yi} is a basis from which we can construct any other
2-dimensional vector. In fact, it’s clear that any two vectors, as long as they point
in different directions, can serve as a basis. The nice feature of the set {|xi, |yi} is
that these two vectors are orthogonal to each other, hx|yi = 0. This is particularly
convenient, because the components of |vi can be found directly by taking inner
products, vx = hx|vi, vy = hy|vi. A basis of this kind is almost always preferable to
non-orthogonal bases. Such a basis, with mutually orthogonal basis vectors of unit
A.2 Matrices
381
norm, is called an orthonormal basis. When we speak of a coordinate system, or
coordinate axes, we are implicitly making reference to an orthonormal basis.
A.2
Matrices
A matrix is essentially an array of numbers, laid out on a rectangular grid, as follows:


 M11 M12 . . . M1n


 M
M22 . . . M2n

21
M =
 .
..
..
..
 .
.
.
.
 .


Mm1 Mm2 . . . Mmn











(A.9)
The numbers Mij comprising a matrix are known as its elements. The dimension
of a matrix is specified by two numbers, the number of rows, and the number of
columns in the matrix. In the example (A.9) M has dimension m × n. Just as with
vectors, matrices are greater than the sum of their parts: the actual values of the
elements of a matrix are not important. To visualize a matrix, one should imagine
a process, in which a vector is transformed into another vector (see Figure A.3).
For this reason, matrices are sometimes referred to as maps, since they map one
vector onto another. The term operator is also used, since a matrix can be viewed
as an operation — rotation, or reflection, say — applied to a vector. The way this
operation is performed mathematically is via matrix multiplication, written like this,
w = M v,
or
|wi = M |vi.
(A.10)
A.2 Matrices
382
This multiplication is evaluated by combining the elements of M and the components
of v to form w, in the following way. Define a set of column vectors

 M1j


 M

2j
mj = |mj i = 
 .
 .
 .


Mmj






,





(A.11)
so that each column of the matrix M is given by one of these vectors,
 



M =



 


 


 


 

 m   m  ...
 1   2 

 


 


 

 

 

 
 m  .
 n  

 

 
(A.12)
The vector w is then given by a weighted sum of the mj , with coefficients equal to
the components of v,
w = v1 m1 + v2 m2 + . . . + vn mn .
(A.13)
From (A.13) it is clear that the number of columns of M must be the same as the
dimension of v for this multiplication to be possible. Thus the number of columns
of M sets the dimension of the vectors upon which M can act. Similarly the number
of rows of M sets the dimension of the vectors mj , and therefore the dimension of
the output vector w. So an m × n matrix is an operator that acts on n dimensional
A.2 Matrices
383
vectors to produce m dimensional ones.
Figure A.3 A matrix acting on a vector. Here M maps the initial
vector v (black) onto the final vector w (red), via a rotation and a
‘dilation’ (length increase). The values of the matrix elements Mij
depend on the components of v and w, which in turn depend on the
direction of the coordinate axes (thin arrows). Rotating the coordinate axes would change the Mij , but the transformation represented
by M would be the same. In this sense, a matrix is more fundamental
than its elements.
Matrices with the same dimensions can be added together; M + N is just the
matrix whose elements are given by the sum of the corresponding elements of M
and N . And of course they can be multiplied by numbers. αM is a matrix whose
elements are αMij . Incidentally, these properties mean that the space of all matrices
is actually also a vector space. But this will not be important for us.
Two matrices can be multiplied together to produce a new matrix. In the product
M N = Q, each of the column vectors qj of Q are formed from the corresponding
column vector nj of N , by combining the column vectors mj of M in a weighted
sum like (A.13), with coefficients given by the components of nj . So M acts on each
column of N to produce the columns of Q.
To multiply a row vector (or a bra) by a matrix, we simply treat the row vector
A.2 Matrices
384
as a 1 × n matrix, and apply the above rule. We therefore obtain
hv|M =
hv|m1 i hv|m2 i . . . hv|mn i
.
(A.14)
Matrices also have Hermitian adjoints. The Hermitian adjoint M † of M is given
by swapping the rows and columns of M , and taking the complex conjugate of all
its elements,


∗
M11



 M∗

12
†
M =
 .
 .
 .


M1n∗
∗
M21
...
∗
M22
∗
. . . Mm2
..
.
..
.
∗
Mm1
..
.
∗
∗
M2n
. . . Mmn





.





(A.15)
Using this definition, it’s easy to check that Q† = N † M † (note the reversed order of
M and N ), and that (M |vi)† = hv|M † . These facts are useful when manipulating
expressions involving several matrices.
A.2.1
Outer product
The outer product of two vectors |vi and |wi is a matrix, written as


v1 w1∗
v1 w2∗
···



 v w∗ v w∗ · · ·
 2 1
2 2
†
|vihw| = vw = 
 .
..
..
 .
.
.
 .


vn w1∗ vn w2∗ · · ·
∗
v 1 wm
∗
v 2 wm
..
.
∗
v n wm





.





(A.16)
A.2 Matrices
385
Each column of this matrix is just |vi, multiplied by the corresponding element of
hw|, so that its structure can be visualized as that of a row vector hw| with column
vectors |vi ‘hanging’ from it. The Dirac notation is very satisfying in this context,
since the result of applying the operator |vihw| to a third vector |xi is written like
this,
|vihw||xi = |vihw|xi = hw|xi|vi.
(A.17)
That is, the matrix product of |vihw| with |xi is just the same as the inner product
of hw| and |xi, multiplied by the vector |vi. If |xi = |wi, the result is w2 |vi.
As |xi deviates away from |wi, the inner product hw|xi gets smaller and smaller,
until it vanishes, when |xi is orthogonal to |wi. A natural interpretation for the
operator (A.16) is therefore as a kind of ‘switch’ that maps an input from |wi to |vi.
Operators of this kind are sometimes known as flip operators, or transition operators,
in quantum mechanics. Breaking down larger operators into flip operators can often
provide valuable insights.
A.2.2
Tensor product
A further generalization of the outer product is the tensor product. The tensor
product is used to combine vector spaces together to produce a new, larger space.
Suppose we have an n-dimensional vector space V, and also an m-dimensional space
W. The tensor product V ⊗ W of these two spaces would be the space of all vectors
with dimension nm. Vectors v and w from the smaller spaces can be combined
together via the tensor product to produce a vector v⊗w inhabiting the larger space.
A.2 Matrices
386
And similarly matrices M and N acting on the smaller spaces can be combined
together to produce an operator M ⊗ N , that acts on vectors in the tensor product
space V ⊗ W. The result of applying the combined operator to the combined vector
is the same as the result of applying the operators to the vectors separately, and
then taking the tensor product:
(M ⊗ N )(v ⊗ w) = (M v) ⊗ (N w).
(A.18)
A common example arising in quantum mechanics is the tensor product of a pair
of 2-dimensional vectors, representing the state of a pair of qubits (a pair of electron spins perhaps). Suppose one qubit is in the state labelled |0i, and the other
is in the state |1i. The combined state |ψi of both is found by taking the tensor product of these two vectors, |ψi = |0i ⊗ |1i. Sometimes the more compact
notation |01i is employed, where the meaning should be clear from the context.
But other 4-dimensional vectors, which cannot be represented as tensor products
of 2-dimensional vectors, can exist in the 4-dimensional tensor product space. For
√
instance, the vector |ψi = (|01i + |10i)/ 2 is a valid state in quantum mechanics (see (1.7) in Section 1.6.4 of Chapter 1). It cannot be written in the form
|state 1i ⊗ |state 2i, but it is a 4-dimensional vector, produced by adding together
two vectors that can be written in this form. Vectors of this kind, that exist in the
tensor product space, but cannot be written as a tensor product of vectors from
the component spaces, are known as non-separable. In quantum mechanics, they
A.2 Matrices
387
represent states that are entangled.
The tensor product of two matrices M (with dimension m × n) and N (with
dimension p × q) is found by ‘attaching’ a copy of M to each element of N , as shown
below,



 N11



 M 
 ...
11 







Np1



M ⊗N =





 N11



 .

 Mm1 
 ..





Np1

···
..
.
···
N1q 

.. 
. 



Npq
..
.
···
..
..
.
···
 N11 · · ·

 .
..
M1n 
.
 ..


Np1 · · ·
.

···
 

N1q 

.. 
. 
 ···


Npq




Mmn 



N1q  
 

.. 


.  
 
 


Npq



..
.
.

 

N11 · · · N1q  
 


..
.. 
..

.
.
.  
 

 
Np1 · · · Npq
(A.19)
The procedure for vectors is identical; the vectors are just treated as matrices with a
single column (or row, in the case of bras). A bit of head scratching will verify that
this definition, when combined with standard matrix multiplication (A.13), satisfies
the requirement (A.18). An important property of the tensor product is as follows.
If {|ii} is an orthonormal basis for one space, and {|ji} is an orthonormal basis for
a second space, the set of tensor product vectors {|ii ⊗ |ji} is an orthonormal basis
for their tensor product space.
A.3 Eigenvalues
A.3
388
Eigenvalues
Consider a matrix R representing a reflection about the x-axis, as shown in Figure
A.4. A vector |1i lying along the x-axis is not changed by the action of this matrix.
That is, it is its own reflection. So we have R|1i = |1i. A second vector |2i lying
along the y-axis is flipped around by R. Its reflection points in the opposite direction
to itself, so R|2i = −|2i. Other vectors are altered in more complicated ways when
they are reflected, so that the vector resulting from the application of R is not related
to the original vector by a simple numerical factor (1 or −1 in the two cases above).
The vectors |1i and |2i are examples of vectors for which the action of R is the same
as multiplication by a number. These ‘special’ vectors are known as eigenvectors of
R. In general, any matrix M has a set of eigenvectors {|ii}, such that
M |ii = λi |ii.
(A.20)
Here the number λi is the eigenvalue corresponding to the eigenvector |ii. For the
example given above, we had λ1 = 1 and λ2 = −1. The eigenvectors and eigenvalues
contain all the information required to reconstruct the transformation implemented
by M ; they represent the essence of a transformation, and as such they are of
paramount importance in linear analysis, and central to quantum mechanics.
A.3 Eigenvalues
389
Figure A.4 Eigenvectors and eigenvalues. The matrix R represents
reflection in the x-axis (horizontal axis). The eigenvectors of R are
those vectors pointing along, or perpendicular to the x-axis, since
the application of R to one of these vectors produces the same vector
again, multiplied by some number.
A.3.1
Commutators
In general, matrix multiplication is not commutative. That is, M N 6= N M ; the order
in which matrices are multiplied is important. This makes sense when matrices are
viewed as representing transformations of vectors (see Figure A.5).
Often it is useful to examine the commutator of two matrices, defined by
[M, N ] = M N − N M.
(A.21)
If M and N were just numbers, their commutator would always vanish, but for
matrices often it does not. In quantum mechanics, the commutator of different
physical quantities may be non-zero, and this non-vanishing of the commutator
can be viewed as the source of a great many of the strange features of quantum
mechanics.
A.3 Eigenvalues
390
3
2
2
1
1
3
Figure A.5 Non-commuting operations. Here M represents a reflection around the y-axis, while N is an anti-clockwise rotation
through 90 degrees. The red arrows are numbered in order, with
1 the initial vector, 2 the result after the application of one of the
transformations, and 3 the result after both transformations have
been applied. On the left, M is applied first, and then N . On the
right, N is applied first, followed by M . The results are different
because the matrices M and N do not commute. The notation can
be counter-intuitive: the product M N represents the application of
N first, with M applied afterwards.
If the two matrices N , M do commute, then they have common eigenvectors. To
see this, suppose that |ui is an eigenvector of N , with eigenvalue λ. If we take the
product M N |ui (that is, we apply N to |ui first, and then M ), the result is simply
λM |ui. On the other hand, if [M, N ] = 0, we can swap the order of M and N , to
get N M |ui. That is,
N (M |ui) = λ(M |ui).
(A.22)
Therefore the vector M |ui is also an eigenvector of N , with the same eigenvalue
λ. M |ui must be parallel to |ui, so that M |ui = µ|ui. That is, |ui is also an
eigenvector of M , with some new eigenvalue µ. This fact is intimately connected
with the epistemology of quantum mechanics.
A.4 Types of matrices
A.4
391
Types of matrices
There are some types of matrix that are particularly important, both for the calculations in this thesis, and for quantum mechanics generally.
A.4.1
The identity matrix
The identity matrix, often denoted by I, is the matrix equivalent of the number 1.
It is the matrix that results in no change when it is multiplied by another matrix —
it represents the operation ‘doing nothing’. That is, IM = M I = M . And of course
the identity does not change a vector either, I|vi = |vi, hv|I = hv|. The identity
matrix is a square matrix (i.e. dimension m×m), with ones along its main diagonal,
and zeros everywhere else (the zero elements are left blank below to avoid clutter),

 1




I=






1
..
.
1





.





(A.23)
Sometimes care should be taken to ensure that the correct dimension m of I is used,
so that the multiplication is possible. Usually this is quite clear from the context,
but the symbol Im can be used when the size of I needs to be specified.
A.4 Types of matrices
A.4.2
392
Inverse matrix
The inverse M −1 of a matrix M is the matrix that ‘undoes’ the action of M . It is
the matrix equivalent of a reciprocal. The inverse satisfies the relations M −1 M =
M M −1 = I. It is clear that taking the inverse of a matrix is also involutive, since the
inverse of an inverse is just the original matrix, (M −1 )−1 = M . If M is rectangular,
with m < n, then M describes a map from a larger space into a smaller space,
so that some information is inevitably lost, in the sense that there are different
vectors in the input space that are mapped to the same vector in the output space.
Therefore M cannot have an inverse — it is impossible to ‘undo’ this type of map. It
is generally the case that only square matrices, with m = n, have a matrix inverse.
It is possible to define a pseudo-inverse, that represents the closest approximation
of a true inverse, for any matrix (even rectangular ones), but we will not make use
of the pseudo-inverse [189–192] . Calculating the inverse of a matrix can be rather
involved, and although an algorithm for inverting 3 × 3 dimensional matrices is
taught to students in school, matrix inversion is rarely performed explicitly. Lloyd N.
Trefethen is a prominent numerical analyst who teaches a course on computational
linear algebra at Oxford University. His reaction to a suggestion that students should
consult W. H. Press’s famous book on numerical techniques was
The only way to annoy a numerical analyst more than by inverting
a matrix, is to use Numerical Recipes.
A.4 Types of matrices
393
The formula for the inverse of a 2 × 2 matrix M is simple however,
−1

 a b 

M −1 = 


c d

=

 d −b 
1

.

ad − bc 
−c a
(A.24)
The quantity ad−bc is known as the determinant of the matrix — sometimes denoted
by vertical bars, |M | — since it determines whether or not the inverse of M exists:
if |M | = 0, the formula for the inverse ‘blows up’. In this case, the matrix does not
have in inverse, which implies that there exists some vector |vi such that M |vi = 0.
Clearly it is not possible to invert this expression. This provides a convenient way to
find the eigenvalues of a matrix. If we want to find |ui such that M |ui = λ|ui, then
we must have that (M − λI)|ui = 0, and therefore we require that |M − λI| = 0.
A.4.3
Hermitian matrices
A Hermitian matrix is equal to its Hermitian adjoint, H = H † . It is the matrix
equivalent of a real number, and in fact its eigenvalues are all real numbers. To see
this, consider the quantity k = hi|H|ji. On the one hand, using the definition of H,
along with (A.20), we have
k = (H † |ii)† |ji = (H|ii)† |ji = (λi |ii)† |ji = λ∗i hi|ji.
On the other hand, we have
k = hi|(H|ji) = λj hi|ji.
A.4 Types of matrices
394
Taking the difference of these, we get k − k = (λ∗i − λj )hi|ji = 0. If we set i = j,
we must have that λ∗j − λj = 0, since hj|ji > 0 is the square of the norm of |ji.
Therefore λj = λ∗j , that is, the eigenvalues of H are real numbers. At the same
time, if we set i 6= j, we must have that hi|ji = 0, which means that different
eigenvectors of H are all orthogonal to one another. Note that we are free to scale
the eigenvectors |ii so that they have length 1. If we do this, the set of eigenvectors
{|ii} of a Hermitian matrix is an orthonormal basis. The eigenvectors define a
‘natural’ coordinate system for the space of vectors upon which H acts. And it’s
very practical to work with this coordinate system, since the effect of H on each
basis vector reduces to multiplication by the corresponding eigenvalue,
H|vi = H (v1 |1i + v2 |2i + . . . + vn |ni) = v1 λ1 |1i + v2 λ2 |2i + . . . + vn λn |ni. (A.25)
A.4.4
Diagonal matrices
A diagonal matrix D is a matrix with zeros everywhere except along its main diagonal,


 D11




D=










.





D22
..
.
Dmm
(A.26)
Clearly diagonal matrices must always be square, with n = m. The identity matrix
is a diagonal matrix with Djj = 1. Diagonal matrices are very easy to work with.
A.4 Types of matrices
395
For example, the square of a diagonal matrix D2 = DD is another diagonal matrix
2 . The inverse D −1 of a diagonal matrix is just another
with its elements equal to Djj
diagonal matrix with all its elements equal to 1/Djj . Any two diagonal matrices
commute with one another, [D1 , D2 ] = 0, and the eigenvalues of a diagonal matrix
are just equal to its elements, λj = Djj , with its eigenvectors being the basis vectors
of the coordinate system with respect to which the matrix elements are defined.
This last property is important. Diagonal matrices are wonderfully simple to
manipulate, but surely it is very unlikely that any interesting matrices are diagonal.
The point is that all matrices are diagonal matrices (or more correctly, most square
matrices), as long as you write them down with reference to the correct coordinate
system! This coordinate system is the one defined by the eigenvectors of the matrix,
and when written down using this basis, the elements of the matrix are just its
eigenvalues.
A brief inspection of (A.25) reveals that in fact, when written down with reference
to the coordinate system defined by the eigenvectors |ii, H is actually diagonal, with
elements H11 = λ1 , H22 = λ2 , etc...
For this reason, the process of finding the eigenvalues and eigenvectors of a matrix
is sometimes referred to as diagonalization, since this calculation is simply what is
required to convert a matrix into a diagonal one. The eigenvalue decomposition can
be written as
M = W DW −1 ,
(A.27)
where D is a diagonal matrix containing the eigenvalues of M , and where the eigen-
A.4 Types of matrices
396
vectors of M comprise the columns of the matrix W . Very efficient algorithms exist
for finding this decomposition; the results in this thesis rely heavily on the speed
and precision of the LAPACK routines implemented in MATLAB.
A.4.5
Unitary matrices
A unitary matrix U is a matrix whose inverse is equal to its Hermitian adjoint,
U −1 = U † . A unitary matrix represents a rotation in space, so that |wi = U |vi is a
vector pointing in a different direction to |vi, but with the same norm — the same
length. To see why, consider the norm of |wi, w2 = hw|wi = hv|U † U |vi. But since
U † = U −1 , we have that U † U = I, so w2 = hv|I|vi = v 2 . That is, unitary matrices
preserve the norm of vectors upon which they act.
Figure A.6 A unitary transformation. U represents a rotation
from an initial (black) into a new (red) coordinate system. The
columns of U are unit vectors comprising an orthonormal basis for
the new coordinate system.
A rotation can be thought of as a transformation from one orthonormal coordinate system to another, as shown in Figure A.6. Associated with this new coordinate
system is an orthonormal basis {|ii}, and these vectors are actually the columns of
U , ui = |ui i = |ii. To see this, consider the product K = U † U . From the definition
A.4 Types of matrices
397
of the Hermitian adjoint, the rows of U † are the bras hui |,





†
U =





hu1 |
..
.
hum |




.




(A.28)
Applying the matrix multiplication described in (A.13), we find that each element
of the product matrix K is given by an inner product, Kij = hui |uj i. But since U
is unitary, K = I, the identity, so that we must have hui |uj i = δij , where δij is the
kronecker delta symbol (δij = 1 if i = j, and 0 otherwise). Therefore the column
vectors |ui i form an orthonormal basis. If U is applied to a vector |vi pointing
along the x-axis, with components v1 = 1, vj6=1 = 0, the result is |u1 i. In the same
way, each coordinate axis is mapped by U to a new axis |ui i, and so U represents a
rotation into a new orthonormal coordinate system defined by its columns.
Incidentally, the above arguments serve to demonstrate that the inner product
hu|vi of two vectors |ui and |vi is always independent of the coordinate system used
for writing out the components of |ui and |vi. Changing the coordinate system
is done by applying a rotation |ui → U |ui, |vi → U |vi, and the inner product is
then hu|U † U |vi = hu|vi. Changing coordinates makes no difference. This is to be
expected of course, since (A.7) makes no reference to any coordinates.
It is worth noting that U † is unitary, if U is. Therefore the columns of U † also
form an orthonormal basis, and so the rows of U form an orthonormal basis.
Note also that the product of two unitary matrices U , V is also unitary: U V (U V )† =
A.4 Types of matrices
398
U V V † U † = U IU † = I. Two rotations composed together can always be thought of
as a single rotation.
Unitary matrices play a central role in quantum mechanics, and we will encounter
them in the optimization of quantum memories.
Appendix B
Quantum mechanics
In this Appendix we give a brief review of the structure of quantum mechanics. This
is intended as a pedagogical precursor to Appendix C, on quantum optics. We will
make use of the concepts developed in Appendix A.
Quantum Mechanics was developed in the early twentieth century, primarily as a
theory of atomic physics. In the days before Google, interdisciplinary communication
was more difficult, and in fact Werner Heisenberg re-invented matrices in order to
formulate his version of quantum theory [193] . The incarnation we present here uses
the notation introduced by Paul Dirac [194] , and we follow broadly the excellent
account given by Nielsen and Chuang [158] .
B.1 Postulates
B.1
Postulates
B.1.1
State vector
400
In quantum mechanics, the state of a system is described by a ket |ψi. The simplest
vector is a 2-dimensional one, and this describes the simplest type of quantum system
— a qubit. More complicated systems are described by higher dimensional vectors.
B.1.2
Observables
Quantities, like energy, momentum or position — any observable that might be measured — are represented by matrices that act on the state vector. These matrices are
always Hermitian, and this guarantees that their eigenvalues are real numbers (see
Section A.4.3 in Appendix A). In addition, the eigenvectors of Hermitian matrices
form an orthonormal basis: they define a coordinate system.
B.1.3
Measurements
Quantum mechanics provides the following recipe for making predictions about measurements. The observable being measured is assigned to a Hermitian operator H.
Making this assignment correctly is left up to the skill and imagination of the physicist. This operator is diagonalized, yielding its eigenvalues {λi } and eigenvectors
{|ii}. The eigenvalues are real numbers, and each one represents a possible numerical outcome of the measurement: the number you might see on an oscilloscope
screen, for example. Each eigenvalue λi is associated with an eigenvector |ii, and
these eigenvectors define a coordinate system. The state vector of the system |ψi is
B.1 Postulates
401
written with reference to these coordinates, known as the measurement basis,
|ψi = ψ1 |1i + ψ2 |2i + . . . + ψm |mi.
(B.1)
The probability pi that the measurement yields the result λi is then given by |ψi |2 ,
the squared magnitude of the ith component of |ψi in the measurement basis. A
more compact way to write this is
pi = |hψ|ii|2 .
(B.2)
This is known as the Born rule, after Max Born who proposed it in 1926 [195] . Immediately after the measurement has been completed, the state of the system changes,
essentially instantaneously, according to the measurement result, |ψi → |ii. This is
known as the collapse postulate.
The average value of the measurement result is often useful. This is sometimes
called the expectation value of the quantity H, since it is the number one would
expect when repeating the measurement many times. The expectation value is
given by hHi =
P
i pi λi ,
and a bit of thought shows that this is equal to hψ|H|ψi.
The fact that hHi = hψ|H|ψi is another convenience of Dirac notation.
B.1.4
Dynamics
It must always be the case that the probabilities pi sum to unity,
P
i pi
= 1. This just
codifies the assertion that we must always get some result from a measurement, even
B.1 Postulates
402
if the result is ‘no signal’. Using the Born rule, this means that
P
2
i |ψi |
= hψ|ψi = 1.
That is, the norm of a state vector in quantum mechanics is always exactly equal
to 1. The norm can never be altered by any dynamical process, which immediately
fixes all dynamics in quantum theory to be unitary. In other words, given some
initial state |ψ0 i, and a final state |ψi, we must have
|ψi = U |ψ0 i,
(B.3)
where U is a unitary operator that advances the system from the initial to the final
state. Differentiating (B.3) with respect to the time t, we obtain the equation of
motion
∂t |ψi = U̇ |ψ0 i
= U̇ U † |ψi,
(B.4)
where the overdot indicates the time derivative of U . Now, the requirement that
the norm of |ψi does not change can be expressed by the condition ∂t (hψ|ψi) = 0.
Substituting in (B.3) gives
hψ0 |U̇ † U + U † U̇ |ψ0 i = 0,
(B.5)
from which we derive the condition that the operator (U̇ U † ) is skew-Hermitian,
meaning that it changes sign under Hermitian conjugation. Any skew-Hermitian
B.2 The Heisenberg Picture
403
operator can be represented as the product of the imaginary unit i with a Hermitian
operator H, and making this replacement in (B.4) gives us the Schrödinger equation
∂t |ψi = iH|ψi.
(B.6)
The operator H is known as the Hamiltonian. Schrödinger’s great insight was to
identify H as the operator associated with the energy of the system. In (B.6) it is the
energy, represented by H, that sets the rate of change of the state vector. Systems
with high energy evolve quickly, with fast oscillations, while low energy systems are
more sluggish.
B.2
The Heisenberg Picture
The above discussion was based on the so-called Schrödinger picture, in which the
quantum state |ψi evolves in time. It is possible to formulate quantum mechanics
differently, and sometimes it is easier to solve a problem by using this different
formulation. The results are identical, regardless of how the calculations are done. In
Heisenberg’s formulation, the quantum state |ψ0 i of a system at some initial time is
fixed. It does not change with time. Instead, the operators acting on the state vector
evolve in time. As an example, consider a Hermitian operator A associated with
some quantity that we might want to measure. Here we use the symbol A instead
of H; we reserve the symbol H for the Hamiltonian from now on. In the Heisenberg
picture the operator A depends on the time at which we make the measurement.
B.2 The Heisenberg Picture
404
In order for this formulation to work, we must have that the expectation value
hAi predicted by either formalism is the same. Denoting the fixed operator in the
Schrödinger picture with a subscript S, we must have
hψ0 |A|ψ0 i = hψ|AS |ψi,
⇒
hψ0 |A|ψ0 i = hψ0 |U † AS U |ψ0 i,
A = U † AS U.
∴
(B.7)
That is, the time evolution of an operator in the Heisenberg picture is found by sandwiching the Schrödinger operator between two copies of U , the same operator that
generates the time evolution of the state in the Schrödinger picture. Differentiating
(B.7) with respect to time, we find
∂t A = U̇ † AS U + U † AS U̇
= U̇ † U A + AU † U̇ .
(B.8)
Note also that [U † , U̇ ] = 0, since
U † U̇ = U † U̇ U U † = U † U U̇ U † = U̇ U † ,
(B.9)
where we used the fact that [U, U̇ ] = 0 (a little thought shows that an operator must
always commute with its derivative; see Section A.3.1 in Appendix A). Therefore
(B.8) can be re-written in terms of the Hamiltonian, to produce the Heisenberg
B.2 The Heisenberg Picture
405
equation
∂t A = i[A, H].
(B.10)
This is the fundamental equation of motion in the Heisenberg picture; it plays the
same role as the Schrodinger equation does in the Schrodinger picture — generating
time evolution.
B.2.1
The Heisenberg interaction picture
Often we are interested in analysing the behaviour of a system when it is subjected
to a weak external field. Of specific relevance in this thesis is the case of an atom
illuminated by a laser: the internal electric fields generated by the charges within
the atom are much stronger than the electric fields within the laser beam, so the
laser acts as a weak external perturbation, on top of the much stronger interactions
binding the atom together. In such cases, it is convenient to separate out the strong
and weak contributions to the energy of a system. Suppose that we can divide
the Hamiltonian into two parts, H = H0 + Hint , where H0 dominates, and Hint
represents a comparatively small interaction. The large contribution H0 will make
the operator A change very quickly (as can be seen from the form of (B.10), where a
large energy produces rapid oscillations in time). This rapid oscillation can obscure
any interesting effects arising from the interaction Hamiltonian Hint . To extract
e in the following way,
these interesting effects, we define a slowly varying operator A
e = U0 AU † .
A
0
(B.11)
B.2 The Heisenberg Picture
406
Here U0 is the time evolution operator associated with the Hamiltonian H0 . That
is, U0 satisfies U̇0 U0† = iH0 . Differentiating (B.11) with respect to time, and using
the Heisenberg equation (B.10), we find
e = U̇0 AU † + U0 (∂t A)U † + U0 AU̇ † ,
∂t A
0
0
0
e + U0 (i[A, H0 + Hint ]) U † − iAH
e 0,
= iH0 A
0
e H0 ] + i[A,
e H0 ] + i U0 AHint U † − U0 Hint AU † .
= −i[A,
0
0
(B.12)
Conveniently, the first two terms cancel. The last term can be re-written in a
e = U0 Hint U † , whence we
compact form, if we define a modified Hamiltonian H
0
obtain the interaction picture equation of motion
e = i[A,
e H].
e
∂t A
(B.13)
Appendix C
Quantum optics
Quantum optics is the study of the quantum features of light. The theory requires a
treatment of ensembles of identical photons, which are easily created and destroyed
in their interaction with atoms. Therefore the techniques of quantum field theory
must be employed, in order to deal with the creation and destruction of identical
particles. In this Appendix we briefly review the quantum mechanical description
of the electric field associated with a propagating light beam, before describing the
form of the interaction between light and matter.
C.1
Modes
Classically, light is a transverse electromagnetic wave. Apart from its amplitude,
it has three degrees of freedom that must be specified to uniquely determine its
properties. These are (i) its polarization, (ii) its frequency and (iii) its propagation
direction.
C.1 Modes
408
The polarization is the direction along which the electric field oscillates; it is
a vector in a plane perpendicular to the propagation direction. It is easy to see
that the space of polarizations is simply a 2-dimensional vector space. In fact,
due to the possibility of phase delays between different polarization directions, it is
actually a complex vector space — a Hilbert space (see Section A.1.2 in Appendix
A). Nonetheless it is a 2-dimensional vector space.
The same is true for the other degrees of freedom. That is, the space of frequencies is a vector space. It is a space with an uncountably infinite number of
dimensions, since the different possible frequencies are infinitely closely spaced, but
it is no different in character to the space of polarizations. And similarly for the
propagation direction: there are an infinite number of infinitely closely spaced propagation directions, and the set of all of these forms a vector space.
Already, in talking of these vector spaces, we have made implicit reference to a
basis for each of them. We talk of two perpendicular directions for polarization. Or
different directions of propagation. These are labels that we use to keep track of
dimensions in a vector space, and they are intuitive and natural. But any basis is as
good as any other. For example, instead of talking about different frequencies, we
could talk about different arrival times. Or we could think of left and right circular
polarizations as the polarization basis. It is useful in quantum optics to be flexible
about the basis we use to describe an optical field. A common concept is therefore
that of the optical mode.
A mode is a member of an orthonormal basis for one of the vector spaces associ-
C.1 Modes
409
ated with a light field. So, horizontal polarization is a polarization mode, since it is
one of a pair of perpendicular polarizations that form a basis for the space of possible light polarizations. The other, orthogonal mode, is vertical polarization. And a
single frequency ω labels a spectral mode. It is orthogonal to another frequency ω 0 ,
because two plane-waves with these frequencies have a vanishing inner product,
Z
0
eiωτ e−iω τ dτ = 0,
(C.1)
when ω 6= ω 0 . Equivalently, we could label different temporal modes t and t0 . These
are orthogonal because two delta-functions with these timings also have a vanishing
inner product,
Z
δ(τ − t)δ(τ − t0 ) dτ = 0,
(C.2)
when t 6= t0 .
An optical mode is a member of a basis for the full space of all possible optical
fields. This space of all possible fields is just the tensor product of the vector spaces
associated with each degree of freedom. And a basis for the full space is found
by taking the tensor product of the bases used for each degree of freedom (see
Section A.2.2 in Appendix A). That is, an optical mode is the tensor product of a
polarization mode, a spectral mode and a spatial mode.
Once a basis of modes is settled upon, it is possible to introduce the concept of
a photon. A photon is an excitation of an optical mode. Sometimes it is useful to
remember that photons are only defined with respect to a basis of modes. Although
C.2 Quantum states of light
410
photons are often contrasted with waves as an embodiment of the particulate nature
of light, they do not have to be localized, like tiny bullets. The ‘shape’ of a photon
is the shape of the mode of which it is an excitation.
C.2
C.2.1
Quantum states of light
Fock states
Suppose we consider a plane wave optical mode. That is, a mode with a linear polarization (horizontal, say), a single frequency ω, and a single propagation direction
k, where k is the wavevector of the mode. An excitation of this mode has a fixed
energy, given by the Planck formula E = ~ω, so a single photon in this mode is an
eigenstate of the Hamiltonian for the field. Similarly, if we excite two photons in
this mode, we have a state with twice the energy, E = 2~ω. This is also an energy
eigenstate, but with a different eigenvalue. It follows that these two states must
be orthogonal. And by extension, each photon number state is orthogonal to every
other photon number state. If we use the notation |ni to denote the state with n
photons, we must have
hn|mi = δnm .
(C.3)
Changing the basis of optical modes from plane waves to some other basis cannot
change this orthogonality, since the inner product is invariant under unitary transformations. Therefore (C.3) holds generally, for different photon number states of
an arbitrary optical mode.
C.2 Quantum states of light
411
The orthonormal basis of photon number states |ni, associated with excitations
of some given optical mode, is known as the Fock basis for that mode. The photon
number states are sometimes known as Fock states, and the space for which they
form a basis is accordingly Fock space. The Fock space represents the final degree of
freedom associated with an electromagnetic quantum state: the amplitude. That is
to say, the more photons in a mode, the more intense the field. Thus the quantum
state of an electromagnetic field is fully specified by the tensor product of 4 vector
spaces: the polarization, spectral and spatial modes (collectively specifying an optical mode), and finally the Fock space (specifying the photon number: the energy in
the field; its brightness).
C.2.2
Creation and Annihilation operators
A marked difference between optical fields and material systems is the impermanence of photons. Generally the atoms and electrons in a quantum memory are
considered to be indestructible. They are not created or destroyed by their interactions. But photons can be absorbed and re-emitted. So we must describe processes
that change one Fock state into another — processes that change the number of
photons excited into a given mode. This description is accomplished by introducing
a creation operator a† , that adds a single photon to an optical mode. The similarity of the symbol ‘†’ for Hermitian conjugation to a ‘+’ sign serves as a useful
mnemonic for this operator’s function. The effect of applying a† to an empty optical
mode |0i, containing no photons, is to produce the state |1i, with a single photon.
C.2 Quantum states of light
412
Further applications of a† add extra photons, with contributions from all possible
permutations of arranging these photons (see Figure C.1). These contributions must
be included, since photons are bosons, meaning that their state must be unchanged
by swapping any pair of photons. The Fock states created by the action of a† are
not correctly normalized, so that a numerical factor, accounting for the number of
permutations, must be included,
(a† )n |0i =
√
n!|ni.
(C.4)
Another way to write this is
a† |ni =
√
n + 1|n + 1i.
(C.5)
Taking the norm of (C.5), we have
hn|aa† |ni = (n + 1)hn + 1|n + 1i = n + 1,
⇒ aa† |ni = (n + 1)|ni.
(C.6)
That is, the Hermitian conjugate a = (a† )† is an annihilation operator that removes
a photon. And from (C.6) we see that
a|ni =
√
n|n − 1i.
(C.7)
C.2 Quantum states of light
413
Note that a ‘kills’ the empty vacuum state, a|0i = 0, which is fortunate, since there
cannot be fewer than zero photons in a mode! It is often useful, when manipulating
Figure C.1 Symmetrized photons. n applications of the photon
creation operator a† to the vacuum state |0i produces a symmetrized
n photon state, with contributions from all n! permutations of the n
photons. Swapping any two photons leaves the state unchanged, as
required by Bose statistics.
expression involving these operators, to be able to reverse their ordering. This is
done using their commutator which, applying (C.5) and (C.7), is given by
hn|[a, a† ]|mi = 0,
hn|[a, a† ]|ni = 1,
⇒ [a, a† ] = 1.
(C.8)
The commutator (C.8) expresses what is known as the canonical commutation relation; commutators of this form are common to creation and annihilation operators
for all bosonic fields.
Another useful operator is the number operator N = a† a, a Hermitian operator
that satisfies the eigenvalue equation
N |ni = n|ni,
so that N counts the number of photons excited into a particular mode.
(C.9)
C.3 The electric field
C.3
414
The electric field
Electric fields are associated with separated charges, while magnetic fields are associated with moving charges. Electrons move rather slowly in most ordinary forms
of matter, and accordingly their interaction with light is dominated by its electric
component. In this thesis, we treat light fields as if they were purely electric waves,
an approximation that is very well satisfied provided that light intensities are not
sufficient to produce a relativistic electron plasma.
The electric field associated with a beam of light, as might be generated by a
laser, can be expressed in terms of the annihilation operators a(ω) associated with
plane waves propagating along the beam [107] ,
Z
E(z) = iv
g(ω)a(ω)e−iωz/c dω + h.c.,
(C.10)
where z is the longitudinal position along the beam, v is a unit polarization vector in the plane perpendicular to the beam and g(ω) =
p
~ω/4π0 Ac is the mode
amplitude. Here A is the cross-sectional area of the beam, 0 is the permittivity
of free space and c is the speed of light. Note that in principle the electric field is
an observable quantity, that we could measure (although at optical frequencies it is
not generally possible to directly measure the electric field, at radio frequencies it
certainly is feasible). And so E is a Hermitian operator, as expected.
The annihilation operators a(ω), labelled by the frequency ω of the mode upon
C.4 Matter-Light Interaction
415
which they act, satisfy the commutation relation
[a(ω), a† (ω 0 )] = δ(ω − ω 0 ).
(C.11)
This expresses the fact that operators for different frequency modes do not ‘see’
eachother, so they commute, while when ω = ω 0 , the canonical relation (C.8) is
satisfied. The delta function is the appropriate generalization for the case when the
modes are labelled by a continuous parameter, such as ω.
C.4
Matter-Light Interaction
Generally light interacts with matter through electrons. In most quantum memory
protocols these are the optically active outer electrons bound to some atoms. We
will also consider scattering in a diamond crystal, and here the electrons are more
appropriately described as free, or quasi -free particles. The Hamiltonian describing
the interactions are slightly different in these two cases; here we briefly review their
origin, and the relationship between them.
C.4.1
The A.p Interaction
The interaction of an electron with the electromagnetic field is found by incorporating the appropriate electromagnetic term, associated with so-called U (1) gauge
symmetry, into the Lagrangian density. The effect of this term is to modify the
C.4 Matter-Light Interaction
416
momentum p of the electron,
p −→ p − eA,
(C.12)
where e is the electronic charge, and where A is the magnetic vector potential.
The potential A is not actually an observable field. The electric and magnetic
fields are related to its derivatives, but the absolute value of A is arbitrary to some
extent. Different choices for the functional form of A — known as different gauges —
produce different Hamiltonians, with differing degrees of calculational convenience;
the physical predictions of the theory are unchanged of course. A standard choice of
gauge in quantum optics is the Coulomb gauge, which requires that A is divergence
free, ∇.A = 0. With this choice, the physical electric and magnetic fields are given,
respectively, by
E = −∂t A,
B = ∇ × A.
(C.13)
Using (C.13) we can express the potential A in the form (C.10), with the replacement
ig(ω) −→ g(ω)/ω.
The Hamiltonian for an electron, with mass m, in an electromagnetic field is
found by substituting the ‘canonical momentum’, given by (C.12), into the Hamiltonian for a ‘bare’ electron,
(p − eA)2 A
H=
+ 0
2m
2
Z
E 2 + c2 B 2 dz.
(C.14)
This is known as the minimal coupling Hamiltonian. The first term is the kinetic
C.4 Matter-Light Interaction
417
energy of the electron, with the transformation (C.12) included. The second term,
in square brackets, represents the ‘free field’ energy: this is the energy of the electromagnetic fields, in the absence of the electron. The integral extends over all space,
or at least, over the entire region occupied by the fields. The contribution from the
magnetic field B is very small, but the contribution from the electric field E is more
significant. Inserting (C.10) for E shows that the free field energy takes the form
R
(N + 12 )~ω dω. The term involving the number operator N simply expresses the
Planck formula E = ~ω, so that the energy in the field increases with the number of
photons excited. The term proportional to
1
2
is known as the zero-point energy: the
energy of the vacuum. It is rather unfortunate that this energy is infinite (since it is
integrated over all frequencies), but it is possible to work around these technicalities
with some mathematical sleight-of-hand, known as renormalization [196] . In any case
we will not be concerned with the zero point energy.
Multiplying out the first term in (C.14), we obtain a term of the form p2 /2m,
which just describes the ‘bare’ kinetic energy of the electron, without the field.
There is a term A2 /2m, which describes the field acting back on itself — this type
of non-linear back action is generally negligibly small. And there is a term of the
form −eA.p/m. This describes the coupling of electronic momenta to the vector
potential. In situations where electrons are spread over an extended region, such as
in a crystal, this interaction dominates the atom-light coupling.
C.4 Matter-Light Interaction
C.4.2
418
The E.d Interaction
When electrons are well-localized, such as when bound into atoms, a more convenient form of the interaction Hamiltonian can be derived. This is accomplished
formally by means of a unitary transformation due to Power, Zienau and Woolley
(PZW) [197,198] . In general it is desirable to eliminate explicit reference to the vector
potential A in the Hamiltonian, since then the equations are manifestly gauge invariant — it is quite clear that there can be no-dependence on the choice of gauge. The
PZW transformation removes A from the Hamiltonian, and introduces interactions
between the physical field E and the moments of the atomic charge distribution. To
see how this is done, we will need two results. The first is the equal-time commutator
of E and A, the amplitudes of the electric field and the vector potential,
[A(z), E(z 0 )] = −i~
e
δ(z − z 0 ),
A0
(C.15)
This is easily derived from (C.10) and (C.13) using (C.11). It is well known that in
quantum mechanics momentum and position generally satisfy the relation [x, p] = i~,
and indeed the form of (C.15) when z = z 0 reflects the fact that in the Coulomb
gauge the field E is actually the ‘momentum’ that is conjugate to the ‘coordinate’
A in the electromagnetic Hamiltonian.
The second result we need is that
eC De−C = D + [C, D],
(C.16)
C.4 Matter-Light Interaction
419
whenever [C, D] is just a number (i.e. not another operator). Here the exponential
of an operator is defined according to the series
eC =
∞
X
Cn
n=1
n!
.
(C.17)
The result (C.16) is straightforward to derive. Consider the product C n D. Using
the commutator, we can ‘pull’ the operator D through C, in the following way,
C n D = C n−1 (CD) = C n−1 (DC + [C, D]) = (C n−1 D)C + [C, D]C n−1 .
(C.18)
Repeating this procedure recursively, we obtain
C n D = DC n + n[C, D]C n−1 .
(C.19)
Re-writing the left hand side of (C.16) using the series (C.17), and applying (C.19),
we arrive at (C.16).
With these preliminaries, we can introduce the PZW transformation. Suppose
that the action of the light is to make an optically active electron oscillate; it remains bound to an atom, but it is ‘wiggled’ by the field. This is certainly what we
expect would happen classically. The atomic polarization, distinct from the optical polarization, is a useful concept in this situation. It is the ‘dipole moment per
unit volume’, where the dipole moment is the product of the electronic charge and
displacement. Suppose that the electron, with charge −e, is displaced a distance x
C.4 Matter-Light Interaction
420
along the polarization direction v of the incident light field. The dipole moment is
d = −exv, and the atomic polarization is P (z) = dδ(z)/A, where the delta function
describes a single dipole placed at the position z = 0. To express the interaction energy associated with the atomic polarization, we introduce a unitary transformation
of the Hamiltonian,
†
H → U HU ,
with
iA
U = exp
~
Z
P (z ).A(z ) dz .
0
0
0
(C.20)
This transformation simply changes the coordinate system with respect to which the
quantum states |ψi of the atom-light system are defined. Essentially it is nothing
more than a cosmetic change, but it has a marked effect on the form of the Hamiltonian. Applying the transformation to the free-field part of (C.14), we have U E 2 U †
= (U EU † ).(U EU † ), with
Z
iA
U EU † = vU EU † = v E(z) +
P (z 0 )[A(z 0 ), E(z)] dz 0
~
1
= E(z) − P (z).
0
(C.21)
That is, the PZW transformation adds a component proportional to the electron displacement into the electric field. Using (C.21), the free-field Hamiltonian, neglecting
the small contribution from the magnetic field B, becomes
A
0
2
Z
2
E dz −→
=
2
Z A
1
0
E − P dz
2
0
Z A
1 2
2
0
E + 2 P dz 0 − d.E(z = 0).
2
0
(C.22)
C.4 Matter-Light Interaction
421
The term proportional to E 2 represents the ‘bare’ free-field energy, with no electron
present. The P 2 term represents an unimportant ‘self-interaction’ of the electron.
But the last term, proportional to E.d, represents the interaction of the physical
electric field, at the position of the electron, with the electronic dipole moment. It
is known as the electric dipole interaction Hamiltonian, and it serves as the basis
for the analysis of all the atomic quantum memory protocols in this thesis.
Finally, we note that the electron momentum acquires a component proportional
to the vector potential under the PZW transformation,
iA
p −→ p +
~
=
Z
A(z 0 )[P (z 0 ), p] dz 0
p + eA(z = 0),
(C.23)
where this time the commutator of the electronic momentum and position was used,
[x, px ] = i~. In the approximation that the wavelength of the light is much longer
than the spatial extent of the atom — a limit valid for all interactions at optical frequencies — we can set A ≈ A(z = 0), and then the vector potential A is completely
eliminated from the Hamiltonian.
Thus when electrons are tightly bound into atoms, the only significant interaction
with optical fields occurs through the electric dipole interaction.
C.5 Dissipation and Fluctuation
C.5
422
Dissipation and Fluctuation
In this section we address the issue of loss in quantum systems. Specifically, we seek
a theoretical description of the decoherence in a quantum memory: the constituent
atoms may emit photons into random directions, or collide with one another, and
these processes partially destroy the quantum information stored in the memory. In
Chapter 4 we use the Heisenberg picture to describe the propagation of light through
a quantum memory, and so we should account for losses using the Heisenberg picture.
In the following we use a simple model to show how the equations of motion for a
quantum system are modified by the presence of losses. Fortunately it is well known
that the results are not significantly altered by refining the model.
Our model consists of a single bosonic mode, our system, coupled to a bath of
other bosons — a reservoir. We use bosons because their commutation relations
are simple, but this model applies rather well to an ensemble quantum memory.
As shown in §4.11 in Chapter (4), both the atomic polarization and the spin wave
appearing in the equations of motion of an ensemble memory have approximately
bosonic commutation relations. In fact, in diamond, the optical phonons that constitute the spin wave really are bosons (see Chapter 9). The reservoir of modes to
which the system is coupled could be the electromagnetic field, and indeed this is
an excellent description of spontaneous emission losses affecting the atomic polarization.
The Hamiltonian contains the free field energy of the bath, and also terms that
represent processes in which a photon in the system is destroyed, while a photon in
C.5 Dissipation and Fluctuation
423
the reservoir is created, or vice versa. Working in the Heisenberg interaction picture,
so that the free-field energy of the system (and zero point energy of the reservoir)
is removed, we have
Z
e =
H
ωb† (ω)b(ω) dω + κ
Z h
i
a† b(ω) + b† (ω)a dω,
(C.24)
where b(ω) destroys a photon with frequency ω in the reservoir, and where a destroys
a photon in the system. The equal-time commutators of these operators are given
by
[a, a† ] = 1,
and
[b(ω), b† (ω 0 )] = δ(ω − ω 0 ).
(C.25)
Using the Heisenberg equation (B.13), we derive the following equations of motion
for the annihilation operators,
Z
∂t a = −iκ
b(ω) dω,
(C.26)
∂t b(ω) = −iωb(ω) − iκa.
(C.27)
Integrating (C.27) and substituting the result into (C.26), we obtain
Z ∂t a(t) = −iκ
−iωt
e
Z
b0 (ω) − iκ
t
e
−iω(t−t0 )
0
a(t ) dt
0
dω,
(C.28)
0
where the operators b0 (ω) = b(ω, t = 0) represent the initial state of the reservoir.
Performing the frequency integral in the second term produces a delta-function (see
C.5 Dissipation and Fluctuation
424
(D.9) in Appendix D), which selects out the time t = t0 , and then (C.28) becomes
∂t a(t) = −γa(t) + F (t),
(C.29)
where
Z
F (t) = −iκ
b0 (ω)e−iωt dω
(C.30)
is known as a Langevin noise operator, and where γ = πκ2 is an exponential decay rate. Equations of this form were first used to describe the classical Brownian
motion of colloidal particles buffeted by the molecules of a warm fluid. There, the
term involving F represents the fluctuating force arising from collisions with the
randomly moving molecules, and a similar interpretation is helpful in the present
case. The operator F mixes a component of ‘white noise’ into the dynamics of a,
which otherwise would simply decay exponentially with a rate γ. That the noise is
white, with a flat power spectrum, can be seen from examining its temporal correlation functions. Suppose that initially the reservoir contains some small thermal
excitations, so that there are n̄ photons, on average, in a mode with frequency ω.
Then we have
hb†0 (ω)b0 (ω 0 )i = n̄δ(ω − ω 0 ).
(C.31)
C.5 Dissipation and Fluctuation
425
The normally ordered correlation function of the noise operator is then given by
†
0
Z Z
0 0
n̄δ(ω − ω 0 )ei(ωt−ω t ) dω dω 0 ,
Z
1
0
= 2πκ2 n̄ ×
eiω(t−t ) dω
2π
hF (t)F (t )i = κ
2
= 2γ n̄δ(t − t0 ),
(C.32)
(C.33)
and the anti-normally ordered correlation function is similarly given by
hF (t)F † (t0 )i = 2γ(n̄ + 1)δ(t − t0 ),
(C.34)
where we used the commutator in (C.25). The noise is therefore delta-correlated,
meaning that it is completely random from moment to moment. There is no correlation with earlier times, and the noise changes ‘infinitely quickly’; a signature of
unlimited-bandwidth white noise. The infinite bandwidth is a consequence of the
fact that the reservoir coupling κ was assumed to be frequency independent. This
assumption is known as a Markov approximation, since it means that the dynamics
of a does not explicitly depend on its past values. As is clear from the form of
(C.29), the Markov approximation gives rise to exponential decay of a: the ubiquity
of exponential decay in quantum systems serves to confirm the robustness of this
approximation as a model for a wide variety of dissipative processes.
The noise F is just sufficient to preserve the expectation value of the equal-time
commutator in (C.25), so that a remains a bona fide bosonic operator. This can be
C.5 Dissipation and Fluctuation
426
seen by inserting the formal solution to (C.29),
−γt
a(t) = a(0)e
Z
t
+
0
e−γ(t−t ) F (t0 ) dt0 ,
(C.35)
0
into the commutator (C.25),
Z tZ t
0
00
e−γ(2t−t −t ) h[F (t0 ), F † (t00 )]i dt0 dt00
h[a(t), a† (t)]i = h[a(0), a(0)† ]ie−2γt +
0
0
Z t
0
e2γt dt0
= e−2γt 1 + 2γ
0
= 1.
(C.36)
This close connection between fluctuations, represented by F , and damping, represented by γ, is a manifestation of the fluctuation-dissipation theorem, discovered
first by Einstein.
Finally, we note that the solution (C.35) allows us to solve for the time evolution
of the number operator for the system, giving the result
hN i = ha† (t)a(t)i = e−2γt ha† (0)a(0)i + 1 − e−2γt n̄.
(C.37)
This shows that in the infinite future, with t −→ ∞, the system relaxes into thermal equilibrium with the reservoir, with N −→ n̄. And this thermal equilibrium
condition can often be taken as the initial state of the system, if we want to analyze
processes which drive it out of equilibrium.
Appendix D
Sundry Analytical Techniques
In this Appendix we define the Dirac delta function and the unilateral and bilateral
Fourier transforms used in this thesis. We discuss some of their properties, and
finally we demonstrate some results relating to the Fourier transform of certain
Bessel functions. First though, we introduce a useful technique that expedites these
calculations.
D.1
Contour Integration
Contour integration is a powerful method for evaluating integrals that would otherwise be difficult, or impossible, to perform. To see how it works, we consider the
integral
Z
Ix =
b
f (x) dx
a
(D.1)
D.1 Contour Integration
428
of some smooth function f (x). The anti-derivative F (x) of f is the function that
satisfies ∂x F = f , and (D.1) can be simply expressed in terms of F as Ix = F (b) −
F (a). That is, the integral of f between two points is given by the change in
‘height’ of its anti-derivative between these same two points (see Figure D.1 (a)).
Now suppose that we introduce the possibility of complex arguments for f , so that
we consider all the values of f (z), where z = x + iy is an arbitrary complex number.
We can now think of f as a two dimensional surface, lying above the (x, y)-plane:
as we vary the real and imaginary coordinates x and y, the value of f traces out a
characteristic landscape. We could define an integral
Z
B
Iz =
f (z) dz
(D.2)
A
from some initial point A = xA + iyA to a final point B = xB + iyB . But now that
we are working in a two dimensional plane, the endpoints aren’t enough to specify
the integral completely. We need to know the path that we should take to get from
A to B. On the other hand, the anti-derivative F (z) also describes some kind of
surface in the (x, y)-plane, as shown in part (b) of Figure D.1, and the integral is
again given by Iz = F (B) − F (A). So in fact, the integral does not depend on the
path taken, it only depends on its endpoints. The proof of this fact is known as the
Cauchy integral theorem. Suppose that we integrate along a closed loop, so that the
endpoint B is equal to the starting point A. Then F (B) = F (A) and the integral
vanishes. This is interesting, since we could build the first integral (D.1) into the
D.1 Contour Integration
429
complex integral (D.2), and possibly use this result to draw some conclusions. For
instance, suppose we start our integral at A = (a, 0) in the (x, y)-plane, integrate
along the real axis (i.e. the line y = 0) to the point (b, 0), and then loop back in a
semi-circle to finish where we started at B = A. The straight part of the loop, along
the real line, is just equal to Ix . The whole integral must vanish, because we finish
where we started. Therefore if we can find an expression for the curved part of the
integral — call it Ic — then we must have Ix = −Ic , in order that the two parts of
the integral cancel out. In this way, interesting results about real integrals can be
derived by considering their behaviour in the complex plane.
(a)
(b)
Figure D.1 Contour integrals. (a) a function f and its antiderivative F . (b) The anti-derivative is extended into the complex
plane, and we consider integrating around a loop that includes the
original integral along the real line.
D.1.1
Cauchy’s Integral Formula
So far we have only considered smooth functions. When f contains a discontinuity,
the Cauchy integral theorem does not apply. Suppose that we want to integrate the
function f (z)/(z − z0 ), which ‘blows up’ when z = z0 . We consider integrating along
D.1 Contour Integration
430
a circular loop around the singularity at z0 . We can parameterize the integral using
polar coordinates z = z0 + reiθ , with dz = ireiθ dθ. We are free to make the radius
r as small as we like, since as above the details of the path taken don’t matter; it
only matters that we encircle the singularity. We find
I
f (z)
dz = lim
r→0
z − z0
where the symbol
H
Z
0
2π
f (z0 + reiθ )
× ireiθ dθ = 2πi × f (z0 ),
reiθ
(D.3)
indicates integration around a loop. The singularity is like a
witness to our passage around the loop, so that we do not quite get back to where
we started on completing our roundtrip. The contribution 2πif (z0 ) to the integral is
known as the residue associated with the singularity at z = z0 . In general, wherever
a function blows up, it leaves a residue that contributes to an integral that encloses it.
Differentiating the formula (D.3) with respect to z0 tells us how to handle ‘stronger’
singularities of the form 1/(z − z0 )n ,
I
D.1.2
∂zn−1
f (z0 )
f (z)
0
dz
=
2πi
×
.
n
(z − z0 )
(n − 1)!
(D.4)
Typical example
We now give an example of how to use these formulas to evaluate an integral. The
method is typical. Consider the integral
Z
∞
Ix =
−∞
eix
dx.
x
(D.5)
D.1 Contour Integration
431
This is equivalent to the contour integral
I
Iz = lim
R→∞
eiz
dz,
z
(D.6)
where the integration path is a closed semicircle with radius R, such that the straight
portion runs along the real line, as shown in Figure D.2. To see that Iz = Ix , note
that the function eiz is damped with a positive imaginary argument, ei(x+iy) =
eix e−y . Therefore any contribution to Iz from the curved portion of the path, all of
which has y > 0, vanishes in the limit R −→ ∞. Note that if the integrand were
conjugated, so that it contained the exponential e−ix , we would have to choose to
close the contour in the lower half of the plane, in order that the curved part of the
integral should vanish.
Now the integral Iz is not zero, because the integrand has a singularity at z = 0.
It’s a little inconvenient that this lies exactly on the integration path. We deal with
this by moving the singularity inside the integration path by a small amount , in
order to apply (D.3); we then take the limit −→ 0 when we are done:
I
Ix = Iz = lim lim
→0 R→∞
eiz
dz
z − i
= lim 2πi × e−
(D.7)
= 2πi.
(D.8)
→0
This technique for performing integrals is indispensable in Fourier analysis, as discussed below. And of course it is quite elegant!
D.2 The Dirac Delta Function
432
Figure D.2 Upper closure. The integration path is comprised of a
straight portion running along the real axis, and a semicircle in the
upper half of the complex plane. In the limit of infinite radius R, the
straight portion becomes equal to the real integral Ix , and the curved
portion vanishes.
D.2
The Dirac Delta Function
We use the Dirac delta function δ(x) a great deal in this thesis. Where it is not
introduced ‘by hand’, prompted by physical arguments, it is most often encountered
in the form of an integral over plane waves,
δ(x) =
1
2π
Z
∞
eikx dk.
(D.9)
−∞
This integral is not really well defined, but its properties become apparent from the
following limit,
Z T
1
δ(x) = lim
eikx dk
T →∞ 2π −T
T sinc(xT )
= lim
,
T →∞
π
(D.10)
where sinc(θ) = sin(θ)/θ. For each value of T , this describes a narrow peak of width
∼ 1/T and height T , centred at x = 0. As T → ∞, the peak becomes infinitely tall
D.2 The Dirac Delta Function
433
and narrow. But its integral remains finite. To see this, consider the integral
∞
−i
Ix =
sinc(x) dx =
2
−∞
Z
∞
Z
−∞
eix
dx −
x
Z
∞
−∞
e−ix
dx .
x
(D.11)
The first integral on the right hand side is given by (D.8). Applying the same
technique to the second integral, we must use an integration path that is closed
in the lower half of the complex plane, as explained in §D.1.2. When we shift the
singularity using z → z − i, it moves into the upper half of the plane, so the
contour used for the second integral does not enclose the singularity, and therefore
the integral vanishes. Therefore we find that
Ix =
−i
{2πi − 0} = π.
2
(D.12)
Integrating the delta function then yields
Z
∞
1
δ(x) dx = lim
T
→∞
π
−∞
Z
∞
T sinc(xT ) dx = 1.
(D.13)
−∞
The Dirac delta function is an ideal ‘spike’, with unit area, and so it has the extremely
useful property that
Z
∞
f (x)δ(x − x0 ) dx = f (x0 ).
−∞
(D.14)
D.3 Fourier Transforms
D.3
D.3.1
434
Fourier Transforms
Bilateral Transform
The most familiar Fourier transform is the bilateral transform fe(k) of a function
f (z),
1
fe(k) = Fz {f (z)} (k) = √
2π
Z
∞
f (z)eikz dz.
(D.15)
−∞
The name ‘bilateral’ refers to the lower limit of integration, which is −∞ in this
case, so that the whole function f (z), including its values for negative z, is involved
in forming the Fourier transform fe.
D.3.2
Unitarity
√
The factor of 1/ 2π makes the transform unitary, meaning that the norm of the
the transform fe is the same as the norm of f ,
Z
∞
−∞
Z ∞Z ∞Z ∞
1
0
f (z)f ∗ (z 0 )eik(z−z ) dk dz dz 0
2π
Z ∞ −∞
Z ∞ −∞ −∞
0
=
δ(z − z 0 )eik(z−z ) f (z)f ∗ (z 0 ) dz dz 0
−∞
Z−∞
∞
=
|f (z)|2 dz,
|fe(k)|2 dk =
−∞
where the plane wave expansion (D.9), and the property (D.14) were used.
(D.16)
D.3 Fourier Transforms
D.3.3
435
Inverse
The inverse Fourier transform takes essentially the same form, except that the integral kernel is conjugated,
f (z) =
Fk−1
Z ∞
n
o
1
e
fe(k)e−ikz dk.
f (k) (z) = √
2π −∞
(D.17)
This is easily demonstrated by substituting (D.15) into (D.17) and using (D.9).
D.3.4
Shift
From the definition (D.15) we can derive a simple formula for the Fourier transform
of a shifted function f (z + a), where a is some constant,
Fz {f (z + a)} (k) =
=
Z ∞
1
√
f (z + a)eikz dz
2π −∞
Z ∞
1
√
f (x)eik(x−a) dx
2π −∞
= e−ika fe(k),
(D.18)
where we changed integration variables from z to x = z + a in the second line. This
shows that a shift in the coordinate z corresponds to a phase rotation in the Fourier
domain. Identical arguments show that for the inverse transform
n
o
Fk−1 fe(k + a) (z) = eiza f (z).
(D.19)
D.3 Fourier Transforms
D.3.5
436
Convolution
The bilateral convolution of two functions f and g is defined by the integral
1
f ∗ g(z) = √
2π
Z
∞
1
f (x)g(z − x) dx = √
2π
−∞
Z
∞
f (z − x)g(x) dx.
(D.20)
−∞
The Fourier transform of a convolution is given by the product of the Fourier transforms of the convolved functions. That is,
Fz {f ∗ g(z)} (k) =
=
Z ∞Z ∞
1
f (x)g(z − x)eikz dz dx
2π −∞ −∞
Z ∞Z ∞
1
f (x)eikx g(y)eiky dxdy
2π −∞ −∞
= fe(k)e
g (k),
(D.21)
where we used a change of variables y = z −x with dy = dz. This shows also that the
inverse Fourier transform of a product of two functions is given by the convolution
of their individual inverse transforms.
D.3.6
Transform of a Derivative
The nth derivative ∂zn f (z) of a function can be expressed using the inverse (D.17) as
∂zn f (z)
=
1
√
2π
Z
∞
−∞
fe(k)∂zn e−ikz dk
n
o
= Fk−1 (−ik)n fe(k) (z),
(D.22)
D.4 Unilateral Transform
437
and therefore we must have that
Fz {∂zn f (z)} (k) = (−ik)n fe(k).
(D.23)
That is, the Fourier transform converts differentiation into ordinary multiplication;
clearly this is a boon when tackling differential equations.
D.4
Unilateral Transform
The unilateral Fourier transform is identical to its bilateral counterpart, except that
it involves an integral over positive coordinates only,
1
fe(k) = Fz {f (z)} (k) = √
2π
Z
∞
f (z)eikz dz.
(D.24)
0
We have used the same notation to denote both types of transform: we will be
careful to distinguish between them when the difference is important. This type of
transform is appropriate for functions f (z) which vanish, or are undefined, when
z < 0. As an example, the function f might describe the temporal response to
an interaction at ‘time’ z = 0, in which case causality requires that f vanishes for
z < 0. For these types of functions, the unitarity property (D.16) still holds, and
the inverse transform is given precisely by (D.17). That is, the inverse is bilateral.
D.4 Unilateral Transform
D.4.1
438
Shift
The shift theorem (D.18) is not directly applicable to the unilateral transform, because of the semi-infinite integral domain. But since the inverse transform is bilateral, the shift theorem (D.19) does apply to the inverse transform.
D.4.2
Convolution
The unilateral convolution, or causal convolution of two functions f and g is defined
similarly to the bilateral convolution (D.20), except that the integral is limited by
causal consistency,
1
f ∗ g(z) = √
2π
Z
0
z
1
f (x)g(z − x) dx = √
2π
Z
z
f (z − x)g(x) dx.
(D.25)
0
To calculate the transform of this convolution, we need to take care with the limits
of integration, as shown in Figure D.3. We then obtain
Fz {f ∗ g(z)} (k) =
=
=
Z ∞Z z
1
f (x)g(z − x)eikx dxdz
2π 0
0
Z ∞Z ∞
1
f (x)g(z − x)eikz dzdx
2π 0
Z ∞ Zx∞
1
f (x)eikx g(y)eiky dydx
2π 0
0
= fe(k)e
g (k).
(D.26)
Just as for the bilateral case above, the unilateral Fourier transform of a causal
convolution is given by the product of the transforms of the convolved functions.
D.4 Unilateral Transform
439
(a)
(b)
Figure
R ∞ R z D.3 Integration limits. (a) Schematic of the integral
dx dz. (b) Schematic of the same integral with the order of in0
0
R∞R∞
tegration reversed, 0 x dz dx. Each arrow represents an instance
of the inner integral at a fixed value of the outer variable.
D.4.3
Transform of a Derivative
The real utility of the unilateral transform is in the treatment of derivatives. Consider the unilateral Fourier transform of ∂z f . Integrating parts,
1
√
2π
Z
∞
e
0
ikz
Z ∞
i∞
1 h ikz
1
∂z f (z) dz = √
e f (z)
f (z)ikeikz dz
−√
0
2π
2π 0
1
= −ik fe(k) − √ f (0).
(D.27)
2π
This is the same result as (D.23) for the bilateral transform, except for the presence
of the boundary condition f (0). Note that we assumed f (z −→ ∞) = 0: all the
functions we will consider satisfy this property, referred to alternately as boundedness
or integrability. Physically, this simply says that the effects of any interaction fall
away to nothing as we move away from their source.
D.5 Bessel Functions
D.4.4
440
Laplace Transform
The unilateral Fourier transform occupies a middle ground between the bilateral
Fourier transform on the one hand, and the well-known Laplace transform on the
other. The Laplace transform Lz {f (z)} (s) is generated from the unilateral Fourier
transform by making the replacement ik → s. The Laplace transform is more common since even unbounded functions generally possess a Laplace transform, due
to the exponential damping provided by the integral kernel. However this damping destroys the unitarity of the transform. For this reason, the unilateral Fourier
transform better suits our purposes in this thesis, since the efficiency of a quantum
memory is unchanged by working with this type of transform.
D.5
Bessel Functions
In this section we define the ordinary and modified Bessel functions that represent
the propagators for quantum memories operated adiabatically (see Chapter 5). The
nth order ordinary Bessel function of the first kind — denoted by Jn — is defined
by the infinite series
∞
X
(−1)m z 2m+n
Jn (2z) =
,
m!(m + n)!
(D.28)
m=0
where n is a non-negative integer. The corresponding modified Bessel funtion In is
defined identically, except that the factor of (−1)m is missing from each term of the
sum. Conversion between ordinary and modified Bessel functions is reminiscent of
D.5 Bessel Functions
441
the relation between trigonometric and hyperbolic ratios,
Jn (iz) = in In (z).
(D.29)
Indeed the J functions have the appearance of decaying cosines, while the I functions
all grow exponentially with increasing z. Just as sin(x) and cos(x) take the values
0 and 1, respectively, at the point x = 0, so it is also true that J0 (0) = I0 (0) = 1,
while Jn>0 (0) = In>0 (0) = 0.
D.5.1
Orthogonality
As can be easily shown using the series (D.28), the Bessel functions Jn (z) satisfy
the differential equation
z∂z (z∂z ) Jn = (n2 − z 2 )Jn .
(D.30)
Consider the ith zero ai of Jn , which satisfies Jn (ai ) = 0. Defining s = z/ai and
X(s) = Jn (ai s), we can re-write (D.30) in the ‘normalized’ form
s∂s (s∂s ) X = (n2 − a2i s2 )X,
(D.31)
with X(1) = 0. We can repeat this procedure for another zero of Jn , say aj , giving
s∂s (s∂s ) Y = (n2 − a2j s2 )Y,
(D.32)
D.5 Bessel Functions
442
where now s = z/aj and Y (s) = Jn (aj s), so that Y (1) = 0. We now multiply (D.32)
by X and (D.31) by Y, subtract the two resulting equations and divide through by
s, to obtain
(a2i − a2j )sXY = ∂s [Xs∂s Y − Y s∂s X] .
Integrating this from s = 0 to s = 1, and using the fact that X(1) = Y (1) = 0, the
right hand side vanishes. Dividing through by a2i − a2j and converting back into the
Jn notation, we find the orthogonality condition
1
Z
sJn (ai s)Jn (aj s) ds = 0,
when
i 6= j.
(D.33)
0
Alternatively, writing z = s2 , we have
Z
1
√
√
Jn (ai z)Jn (aj z) dz = 0,
when
i 6= j.
(D.34)
0
D.5.2
Memory Propagator
In Chapter 5 we are faced with taking the inverse Fourier transform of the function
e−ia/k
fe(k) = n+1 ,
k
(D.35)
where a is some constant and n = 0, 1, 2.... This is done with the help of contour integration, as described at the beginning of this Appendix. Using the series
D.5 Bessel Functions
443
expansion of the exponential, we have
1
f (z) = √
2π
"
∞
1
−∞
k n+1
Z
#
∞
X
1 −ia m −ikz
e
dk.
m!
k
(D.36)
m=0
We must therefore perform the integral
∞
e−ikz
−∞
k n+m+1
Z
Ik =
dk.
(D.37)
We can equate this to the complex integral
I
Iz = lim lim
→0 R→∞
e−ikz
dk,
(k + i)n+m+1
(D.38)
where the integration contour is closed in the lower half of the complex plane, as
shown in Figure D.4, and where the regularization shifts the singularity into the
interior of the contour. Using Cauchy’s integral formula (D.4), we find
(
Iz = lim
→0
"
∂ n+m e−ikz
−2πi × k
(n + m)!
= −2πi ×
(−iz)n+m
(n + m)!
#
)
k=−i
.
(D.39)
Substituting this result into the series (D.36), and comparing with (D.28), we obtain
the result
z n/2
√
√ f (z) = (−i)n+1 2πΘ(z)
Jn 2 az .
a
(D.40)
D.5 Bessel Functions
444
The function Θ(z) is known as the Heaviside step function, defined so that Θ(z) = 0
when z < 0, and Θ(z) = 1 for z > 0. We include it in the above expression to
represent its causal nature. That f (z) = 0 for negative z can be seen by considering
the integration contour used to evaluate the Fourier transform. When z becomes
negative, the integration contour must be closed in the upper half of the complex
plane, so that it no longer encloses the singularity, which therefore no longer leaves
a residue, and the integral vanishes.
Figure D.4 Lower closure. Here the integrand is damped in the
lower half of the complex plane, so we close the integration contour
in this region.
We will also need the inverse transform of (D.35) when n = −1,
n
o
f (z) = Fk−1 e−ia/k (z).
(D.41)
Again, inserting the series expansion for the exponential, we have
Z
∞
1 X (−ia)m ∞ e−ikz
f (z) = √
dk.
m
2π m=0 m!
−∞ k
(D.42)
The first term, with m = 0, involves an integral with no singularity. In fact, it is an
D.5 Bessel Functions
445
integral over plane waves — a delta function. We separate off this first term, and
re-index the remaining terms,
1
f (z) = √
2π
Z
∞
1 X (−ia)m+1 ∞ e−ikz
e−ikz dk + √
dk.
2π m=0 (m + 1)! −∞ k m+1
−∞
Z
∞
(D.43)
Then we perform the singular integrals using contour integration, as before, to obtain
√
∞
ia X (−ia)m (−iz)m
2πδ(z) − √
Θ(z)
2π m=0 (m + 1)!m!
r
√
√ a
2π δ(z) − Θ(z)
J1 2 az .
=
z
f (z) =
(D.44)
We make use of this result in Chapter 5, where it relates transmitted and incident
signal fields in a Raman quantum memory.
D.5.3
Optimal Eigenvalue Kernel
In Chapter (5) we claim that it is possible to derive a certain kernel KA from a
storage kernel K by direct integration — the result is used in the work of Gorshkov
et al. The Fourier transforms presented above provide one way to confirm it. The
two results we need are as follows. First, we need the unilateral Fourier transform
of an integral over the product of two storage kernels:
Z
A = Fz,z 0
0
∞
√
√
0
e−ax × J0 (2 bxz)e−uz × J0 (2 cxz 0 )e−vz dx (k, k 0 ).
(D.45)
D.5 Bessel Functions
446
Using the result (D.40) along with the shift theorem (D.19) we obtain
A =
=
Z ∞
0
1
e−ibx/(k+iu) eicx/(k +iv)
−ax
e
×
×
dx
2π 0
k + iu
k 0 + iv
1
1
×
.
0
2π akk + i(va + c)k + i(ua + b)k 0 − uva − bv − cu
(D.46)
The second result we need is the unilateral Fourier transform of an anti-normally
ordered kernel:
o
n
p
0
B = Fz,z 0 e−α(z+z ) J0 (2 βzz 0 ) (k, k 0 ).
(D.47)
Transforming over z first, and then z 0 , we get
(
)
0
i
e−iβz /(k+iα)
−αz 0
B = √ Fz 0 e
×
(k 0 )
k + iα
2π
Z ∞ −[α+iβ/(k+iα)−ik0 ]z 0
i
e
=
dz 0
2π 0
k + iα
1
1
= −
×
.
2π kk 0 + iα(k + k 0 ) − β − α2
(D.48)
(D.49)
Comparing these two results, and grinding through some tedious algebra, it is not
hard to demonstrate the claimed equality.
Appendix E
Numerics
In this Appendix we review the numerical methods used in this thesis. Our aim
is to solve a system of coupled linear partial differential equations (PDEs) in space
and time. For pedagogical purposes, we will consider the simple example presented
in §3.5 in Chapter 3: A resonant signal pulse with amplitude A propagates through
an ensemble of classical atoms, whose response to the optical field is described by
the average displacement B of the atomic electrons from their equilibrium positions.
The equations of motion are given by
∂τ B = −iαA,
and
β
∂z A = −i B.
α
(E.1)
(The second equation follows from the first line of (3.34)). Since we are only concerned with the method of numerical solution, we set α = β = 1, and we normalize
the τ and z coordinates so that they are dimensionless, with z running from 0 to 1,
448
and τ running from 0 to 10. The boundary conditions are Ain = A(z = 0, τ ) and
Bin = B(z, τ = 0): the profile of the incident signal pulse, and the initial atomic
excitation, respectively. Generally Ain will take the form of a pulse, reminiscent of a
Gaussian, and Bin = 0, if the atoms are prepared in their ground states. To find the
solution numerically, we discretize the space and time coordinates on a finite grid,
z −→ zj ,
τ −→ τk ,
(E.2)
and we represent the continuous functions A(z, τ ), B(z, τ ) as matrices A, B whose
elements approximate those functions sampled at the grid points,
Ajk = A(zj , τk ),
Bjk = B(zj , τk ).
(E.3)
We solve the system (E.1) via the method of lines [199] . The technique is so named
because we integrate the spatial derivative ∂z ‘in one go’, whereas we integrate the
temporal derivative ∂τ incrementally, stepping forwards in time iteratively. The solution, starting as a series of points in space at τ = 0, evolves gradually forwards
in time, and each of the points in space traces out a line, over time, that describes
the solution at that position in space. It is a feature of the implementation of the
technique that the temporal discretization is much finer than the spatial one, and
so the numerics are conducive to this mental picture, which is illustrated schematically in Figure E.1. The spatial derivative is performed using a spectral method,
which uses the value of the function at all the spatial points in order to compute
E.1 Spectral Collocation
449
Figure E.1 The method of lines. The spatial coordinate is discretized on a coarse grid. The solution at each spatial point traces
out a line as the temporal integration proceeds stepwise on a fine
grid.
the derivative. For smooth functions, this is extremely accurate, even when very
few spatial points are used. Before describing the method used for the temporal
derivative, we introduce this spectral method more fully.
E.1
Spectral Collocation
An excellent introduction to this field of numerical analysis is provided in Nick
Trefethen’s book Spectral Methods in Matlab [200] — which is hosted on the internet.
Other treatments include the books by Gottlieb and Orzag [201] , and by Boyd [202] .
The easiest spectral method to understand is that involving a Fourier transform. As
shown in §D.3.6 in Appendix D, Fourier transformation of a function f from z to
k converts differentiation by z into multiplication by −ik. It is also clear that the
operation of taking a Fourier transform, or indeed its inverse, is a linear one. Suppose
that we discretize the coordinates z and k, so that the function f is approximated by
E.1 Spectral Collocation
450
a vector f , with components fj = f (zj ). In this context the points zj at which f (z)
is sampled are known as collocation points. The linearity of the Fourier transform
then allows us to represent it as a matrix acting on f . The composition of the
operations of taking the Fourier transform, multiplying by −ik, and then taking the
inverse Fourier transform, can therefore also be represented as a single matrix, D.
An approximation fz to the derivative of f at the collocation points is then given
by
fz = Df .
(E.4)
The matrix D is dense, meaning that there are generally very few elements of D
that are zero, so all of the elements of f are involved in determining fz . As a result,
spectral methods can be very accurate, even if very few collocation points are used.
This accuracy comes at the price of requiring a dense matrix multiplication, which is
computationally more expensive than a more ‘local’ method — such as the method
we employ for the temporal integration (see below). As it happens there is a more
efficient way to implement a discrete Fourier transform (DFT), known as the fast
Fourier transform (FFT), which was developed by Cooley and Tukey in 1965 [203] .
The FFT does not use explicit matrix multiplication, and it is generally employed
for Fourier spectral methods. We do not use a Fourier spectral method in this
thesis, however. One reason for this is that a DFT works by fitting a finite number
of complex exponentials — which are periodic — to the vector f . Since the basis
functions are periodic, the vector f must be periodic. Of course any function f (z)
defined on a finite domain [0, 1] can be made periodic by ‘gluing’ copies of the
E.1 Spectral Collocation
451
function onto either end of the domain, as shown in Figure E.2. But if f (0) 6= f (1),
the resulting ‘extended’ function will contain discontinuous jumps, where the copies
were glued together. These discontinuities represent a very rapid change in f , so
that the spectrum fe(k) contains very high spatial frequencies. Given that f , and
the approximated Fourier transform fe, are comprised of only a finite number of
collocation points, the accuracy of the DFT is therefore severely compromised by
these discontinuities. Fourier spectral methods are not well suited to dealing with
functions that are not ‘really’ periodic, for this reason. Since we are modeling the
spatial distribution of atomic excitations, which need have no intrinsic periodicity,
it behooves us to use a method that is not subject to this limitation. Such a method
is provided by polynomial differentiation matrices.
(a)
(b)
Figure E.2 Periodic extension. (a): the function f satisfies f (0) =
f (1), so that its periodic extension is smooth. The Fourier spectral
derivative of this function is accurate. (b): now f (0) 6= f (1), and the
periodic extension of f contains sharp discontinuities, which erode
the accuracy of a Fourier spectral derivative.
E.1 Spectral Collocation
E.1.1
452
Polynomial Differentiation Matrices
To generalize spectral differentiation to the case of non-periodic functions, we should
use non-periodic basis functions. Instead of fitting the vector f with complex exponentials, as in the Fourier case, we fit f with an algebraic polynomial p(z). We can
then easily differentiate this to get p0 (z), and the vector approximating the derivative of f is found by evaluating p0 at the collocation points. The basis functions
from which we construct p are specified by requiring that p evaluates to f at the
collocation points. We write p in terms of a set of basis functions {pj }, and the
elements of f , as
p(z) =
X
fk pk (z).
(E.5)
k
Fixing p(zj ) = fj imposes the condition
pk (zj ) = δjk .
(E.6)
The interpolant p(z) is a polynomial of degree N − 1, if there are N collocation
points, and therefore the basis functions pk are also polynomials of degree N − 1.
The condition (E.6) means that each of the N collocation points, except for zk , are
roots of pk (z), and therefore
pk (z) =
1 Y
(z − zi ),
ak
i6=k
with
ak =
Y
i6=k
(zk − zi ).
(E.7)
E.1 Spectral Collocation
453
Here ak is just the normalization required so that pk (zk ) = 1. Differentiating the
relation (E.5), and setting z = zj gives
p0 (zj ) =
X
p0k (zj )fk .
(E.8)
k
Identifying the elements of fz with p0 (zj ), and using the definition (A.13) in Appendix A for matrix multiplication, we see the elements of the differentiation matrix
D are given by
Djk = p0k (zj ).
(E.9)
Taking the logarithm of (E.7) and differentiating, we find
p0k (z) = pk (z)
X
i6=k
1
.
z − zk
(E.10)
The elements of D are then given by the general formulae
Djk =



P


 i6=j
1
zj −zi



a

 ak (zjj−zk )
E.1.2
if j = k,
(E.11)
otherwise.
Chebyshev points
We are now able to construct a differentiation matrix that works on an arbitrary
set of collocation points {zj }. A grid of equally spaced points is the natural choice,
but this is in fact disastrously unstable: the interpolant p develops pathological
oscillations near the edges of the domain, close to z = 0 and z = 1. This is known
E.1 Spectral Collocation
454
as the Runge phenomenon, and it is dealt with by using a set of collocation points
that cluster together near the domain edges. To develop an intuition for why this
might be, observe that the magnitude of one of the basis functions can be written
in the form
|pk (z)| =
eV (z)
,
|ak |
where
V (z) =
X
ln |z − zi |.
(E.12)
i6=k
Notice that V (z) has the same functional form as the electrostatic potential due to
infinite lines of charge intersecting the z-axis at the collocation points zi . Therefore the size of the basis functions is exponentially sensitive to the potential energy
associated with the ‘charge distribution’ describing the collocation points. The interpolant p is well-behaved when this energy is constant across the domain [0, 1].
If we allowed the collocation points to move along the z-axis according to the mutual repulsion between them represented by V , they would arrange themselves in
a minimum-energy configuration which renders the potential flat, and the interpolant would be stable. In this configuration, the points are clustered together at
the domain boundaries, and this explains why choosing a set of collocation points
according to this configuration avoids the Runge phenomenon.
A stable set of collocation points is provided by the Chebyshev points, which are
the projections onto the domain of equally spaced points along a semicircle joining
the domain boundaries (see Figure E.3):
zj =
1
2
n
h
io
1 − cos π(j−1)
.
N −1
(E.13)
E.2 Time-stepping
455
By substituting the points (E.13) into the formula (E.11), the appropriate Cheby-
(a)
(b)
collocation points
Figure E.3 Chebyshev Points. (a) Chebyshev collocation points
are clustered towards the boundary of the domain. They are the
downward projections of equally spaced points along a semicircle of
radius 12 centred at z = 12 . (b) The method of lines with Chebyshev
spectral collocation.
shev differentiation matrix can be calculated. We use Matlab for all of our numerical computations, and the differentiation matrices are conveniently generated by
a simple script called cheb.m, which is available online, and can be found in Nick
Trefethen’s book [200] , along with many detailed examples of its application.
Having introduced the method we use for the spatial derivatives, we now discuss
the temporal derivatives, before describing how these two techniques are combined
to solve the system (E.1).
E.2
Time-stepping
Suppose that we would like to solve the differential equation
f 0 (τ ) = g [f (τ ), τ ] ,
with
f (0) = f0 ,
(E.14)
E.2 Time-stepping
456
where the prime now denotes differentiation with respect to τ . We assume that
the boundary condition f0 and the function g are known. Note that g may depend
on f . For instance, in our example, B 0 = −iA, but A depends on B. Now, the
simplest numerical approach is to make a finite difference approximation to the
time derivative. In terms of the discretized functions and coordinates, we have
fk+1 − fk
δτ
= g (fk , τk )
⇒ fk+1 = fk + δτ g (fk , τk ) ,
(E.15)
where δτ = τk+1 − τk is the time step — assumed to be independent of k. The
second line of (E.15) is a recursion relation, relating the future value of f to the
present values of f and g. Starting with the boundary condition at τ = 0, we
can use this relation to step forward in time, gradually building up the solution
for f . This method is known as a first order Euler method, since errors of order
δτ accumulate in the numerical solution. In our numerics, we use a second-order
Runge-Kutta (RK2) method, which is only slightly more complicated, but accurate
up to errors of order δτ 2 . The Runge-Kutta method reduces the errors by using
values of the known function g at intermediate points, between τk and τk+1 . The
recursion formula therefore requires that we discretize our functions on two grids: a
primary grid, with times τk , which we denote with the same notation as used above,
and a secondary, intermediate grid, with times τ k = τk + 32 δτ . The second order
E.3 Boundary Conditions
457
Runge-Kutta method we employ for time stepping is then written as
fk+1 = fk +
δτ g (fk , τk ) + 3g f k , τ k ,
4
(E.16)
where the value of f on the secondary grid is approximated with the first order Euler
formula (E.15), so that
f k = fk + 23 δτ g (fk , τk ) .
E.3
(E.17)
Boundary Conditions
It is clear how to implement the boundary condition Bin on B using the above
time-stepping algorithm. This boundary condition simply tells us the initial values
to use, on the first time step. However it is not so clear, from our discussion of
spectral methods, how to implement the boundary condition Ain on A, which holds
at z = 0 at all times. This is done by incorporating the boundary condition into
the dynamical equation for A. First consider the discretized version of the equation
∂z A = −iB, which is
DA = −iB,
(E.18)
where D is a Chebyshev differentiation martrix, which acts on each column of A to
produce the corresponding column of −iB. There is no mention of the boundary
condition so far. The discretized form of the boundary condition on A can be written
E.3 Boundary Conditions
458
as
1 × [A11 A12 A13 · · · A1N ] = [Ain (τ1 ) Ain (τ2 ) Ain (τ3 ) · · · Ain (τN )] .
(E.19)
That is, 1 multiplied by the first row of A — which corresponds to the values of A
at z = z1 = 0 — is equal to the discretized boundary condition. This equation can
be ‘built in’ to the dynamical equation (E.18) simply by replacing the first row of
D with the first row of the identity matrix I, and by replacing the first row of −iB
with the discretized boundary condition. The equation becomes
Db A = −iB b ,
(E.20)
where the superscript b — for ‘boundary’ — indicates the modifications
h
b
b
b
b
D11
D12
D13
· · · D1N
i
= [1 0 0 · · · 0] ,
h
i
= i × [Ain (τ1 ) Ain (τ2 ) Ain (τ3 ) · · · Ain (τN )] .(E.21)
b
b
b
b
B11
B12
B13
· · · B1N
and
All the other elements of Db and B b are the same as those of D and B.
Now that we have fixed the boundary condition for A, we can solve for A in
terms of B. Formally, we can invert the modified differentiation matrix Db to get
A = −iDb−1 B b .
(E.22)
E.4 Constructing the Solutions
459
Of course, we do not know all the values of B until we have performed the timestepping, and this requires knowledge of A. In order to get to the solutions for A and
B, we must build up B column by column, using the time-stepping, and solving for
each column of A in turn using the above procedure. We are now ready to describe
how this works.
E.4
Constructing the Solutions
Let us denote the columns of A by ak , and similarly the columns of B by bk :
 



A=



 


 


 


 

 a   a  ...
 1   2 

 


 


 



 a
 M












,



 



B=



 


 


 


 

 b   b  ...
 1   2 

 


 


 



 b
 M









(E.23)
where M is the number of temporal discretization points. The algorithm proceeds
as follows. First, we use the boundary condition Bin to set the values of b1 ,


 B11


 B
 21

 .
 .
 .


BN 1
  Bin (z1 )
 
 
  B (z )
  in 2
=
 
..
 
.
 
 
 
Bin (zN )







.





(E.24)
Then, we solve for a1 using the formula
ak = −iDb−1 bbk .
(E.25)



,



E.4 Constructing the Solutions
460
Here bbk is the k th column of B b . That is to say, bbk = bk , except that its first element
is replaced by the signal field boundary condition, (bbk )1 = iAin (τk ). Now that we
have both b1 and a1 , we implement the first stage of the RK2 iteration, which is to
approximate the first time step on the intermediate grid, b1 , using (E.17),
bk = bk − i 23 δτ ak .
(E.26)
Before we can implement the second part of RK2, we must approximate a1 . This is
b
done by using (E.25) again, this time replacing bbk with bk . That is,
b
ak = −iDb−1 bk .
(E.27)
This time, the modified vector contains the signal boundary condition evaluated
b
at τ = τ k . That is, bk
= iAin τk + 23 δτ . Finally, the second part of RK2 is
1
implemented, which provides us with b2 . The formula is
bk+1 = bk − i
δτ
(ak + 3ak ) .
4
(E.28)
We have now succeeded in constructing b2 from our knowledge of b1 , and along the
way we have found a1 . Iterating this procedure, we can construct b3 and a2 , and
then b4 and a3 , and so on. In this way we proceed forward in time until we reach
the end of our time domain, at τ = τM = 10. This completes the numerical solution.
Here we comment that a dramatic increase in computational speed is achieved
E.4 Constructing the Solutions
461
if Gaussian elimination is used instead of matrix inversion in (E.25) and (E.27).
Gaussian elimination is an efficient method for solving the matrix equation Ax =
y to obtain x, without explicitly calculating the inverse A−1 . In Matlab, this is
implemented by the ‘backslash’ operator. For example, the Matlab code for (E.25)
might look like
A(:,k)=-i*inv(Db)*Bb(:,k),
where the ‘inv’ function calls a matrix inversion routine. A much more efficient
way to perform the same calculation would be coded as follows,
A(:,k)=-i*( Db\Bb(:,k) ).
This latter method becomes crucial to the feasibility of our chosen spectral
method when modeling complicated dynamics.
In order to produce an informative plot, we can use polynomial interpolation in
order to replace the coarse Chebyshev grid of spatial collocation points with a finer,
equally spaced grid. This utilizes the full accuracy of the spectral method, since the
values at the collocation points are implicitly derived from the global interpolant p(z)
at each time step. In practice, it is often faster, and quite acceptably accurate, to
use a piecewise spline interpolant, which glues together multiple cubic polynomials
in a continuous way, to join the collocation points. Routines for implementing this
are standard in Matlab. In Figure E.4 we plot example solutions for A and B. Even
using only 5 spatial collocation points, the solutions are accurate to ∼ 1%.
E.5 Numerical Construction of a Green’s Function
462
Figure E.4 Example solutions. We plot the the squared moduli
|A|2 and |B|2 of the solutions for A and B, found using the method
of lines, as a function of z and τ . We used N = 5 spatial collocation
points, on a Chebyshev grid, and M = 20 equally spaced time steps.
The Matlab code runs in 0.02 seconds on a 3 GHz machine. The
2
signal boundary condition is a Gaussian pulse, Ain (τ ) = e−(τ −2.5) ,
and we assume there are no atomic excitations initially: Bin (z) = 0.
The black lines indicate the time evolution of the spatial collocation
points. The smooth surfaces are generated by spline interpolation in
between these points. The red lines indicate the boundary conditions.
The interpolated solutions are accurate to ∼ 1%, even with only 5
collocation points - this rapid convergence is a remarkable feature of
spectral collocation.
The behaviour is as might be expected: the signal field is absorbed as it propagates through the ensemble, with the transmitted field having undergone significant
temporal distortion due to the spectral hole burnt by the atomic absorption line.
The atomic excitation grows in time, assuming a roughly exponential shape in space,
at the end of the interaction, consistent with Beer’s law.
E.5
Numerical Construction of a Green’s Function
Now that we are able to solve the system of coupled PDEs describing a quantum
memory, for given boundary conditions, we can find the Green’s function — the
storage kernel — for the interaction. As described in §5.4 in Chapter 5, this is done
E.5 Numerical Construction of a Green’s Function
463
by solving with delta function boundary conditions for the signal field, successively
varying the timing of the delta function to construct an approximation to the kernel.
As usual for storage, we set Bin (z) = 0. Of course, it is not possible to implement
a true delta function numerically. Instead, the delta function Ain (τ ) = δ(τ − τk ) is
represented as
[A11 A12 · · · A1k · · · A1N ] = [0 0 · · · 1 · · · 0] .
(E.29)
That is, along the line z = 0, the signal field is identically vanishing at all times,
except at τ = τk , where it takes the value 1. Since the equations are linear, the
absolute magnitude of Ain is not important, it only matters that the incident signal
field is non-zero at just the single time step τ = τk . Due to the causal nature of the
interaction, it is not necessary to numerically integrate from τ = 0 up to τ = τk .
Since the signal field is zero in this region, and since there is no atomic excitation,
the dynamics are trivial in this region. Therefore we only integrate from τ = τk up
to τ = τM .
Let bkM denote the vector representing the atomic excitation at the end of the
interaction at τ = τM , produced by a delta function incident signal at τ = τk . The
numerical approximation to the Green’s function is the matrix whose k th column is
E.5 Numerical Construction of a Green’s Function
464
bkM ,
 



K=






 b1
 M


 


 










 ...






 bM
 M












 b2
 M





.



(E.30)
Constructing the Green’s function is a more demanding computation than simply
solving an instance of the equations with some particular boundary condition. First,
the system of PDEs must be solved multiple times, and second, the sharp nature of
the delta function boundary condition generally requires a smaller time step than
is required for a smooth boundary condition. Below, in Figure E.5, we plot the
numerically constructed Green’s function for our example system, alongside the
analytic result, expressed as a Bessel function, which is derived in (3.38) in §3.5
of Chapter 3. With a sufficiently fine grid, the numerical and analytic results are
indistinguishable.
numerical
analytic
Figure E.5 A numerically constructed Green’s function. We used
N = 30 spatial collocation points, and M = 4000 time steps. The
computation takes around 10 minutes on a 3 GHz machine; the agreement with the analytic result is excellent.
E.6 Spectral Methods for Two Dimensions
E.6
465
Spectral Methods for Two Dimensions
In Chapter 6 we describe simulations carried out in three dimensions — two space
dimensions, and time. Needless to say, these simulations are much more time consuming than those involving only one spatial dimension. The method we use is
essentially identical to that described in §E.4 above, except that now we use spectral collocation to deal with both spatial dimensions: we still use RK2 for the time
stepping. In this section, we explain how to extend the spectral method to two
dimensions. For concreteness, we will consider solving the simple example system
(E.1) using only spectral collocation. The method is rather memory-intensive, and
it quickly becomes unwieldy as the problem becomes complex, but it has the appeal
of being very direct: we encode all of the dynamics into a single matrix equation,
from which we extract the solutions in a single step. To see how this works, consider
the following representation of the dynamical equations (E.1),



 ∂z i   A 


 = 0.



i ∂τ
B
(E.31)
If we could invert the matrix of derivatives, we might solve for A and B. But
this matrix has no unique inverse: boundary conditions are needed. We proceed by
discretizing the coordinates, and approximating (E.31) numerically using Chebyshev
differentiation matrices to represent the differential operators. We then incorporate
the boundary conditions into the resulting matrix equation, much as we did in §E.3
previously. Finally, we extract the solutions for A and B using Matlab’s backslash
E.6 Spectral Methods for Two Dimensions
466
operator.
In order to represent (E.31) numerically, we must convert A and B into column
vectors. First, we discretize both the z and τ coordinates on a Chebyshev grid. We
use N spatial collocation points and M temporal collocation points. The functions
A and B are then represented as N × M matrices, whose columns describe spatial
variation, and whose rows describe evolution in time. Let us denote the columns of
A and B by {ak } and {bk }, as before (see (E.23)). We vectorize the matrices A and
B by concatenating adjacent columns into a single column vector, of total length
NM,

 a1


 a
 2
A = vec(A) = 
 .
 .
 .


aM






,






 b1


 b
 2
B = vec(B) = 
 .
 .
 .


bM






.





(E.32)
We now generate the differentiation matrices that will simulate the action of ∂z and
∂τ on A and B. Suppose that Dz is the differentiation matrix designed for the
spatial Chebyshev grid defined on [0, 1] (Dz is therefore exactly the same as the
matrix D discussed in §E.4). This acts on a column ak of A to produce the required
derivative. The matrix Dz , which acts on A as a partial space derivative, is therefore
E.6 Spectral Methods for Two Dimensions
467
given by


 Dz



Dz

Dz = 





..
.
Dz





 = I ⊗ Dz ,





(E.33)
where on the right hand side we have used the tensor product notation (see §A.2.2 in
Appendix A). This expresses the fact that Dz acts on the spatial part of the vector
A (within each column of A), whereas the temporal part (across columns of A) is
unaffected. In a similar vein, we assemble the temporal partial derivative ∂τ by first
generating a Chebyshev differentiation matrix Dτ for the temporal Chebyshev grid.
Since the τ coordinate runs from 0 up to 10, the collocation points are given by
τk = 10 ×
1
2
n
h
io
1 − cos π(k−1)
.
M −1
Because of the factor of 10 appearing here, we have Dτ =
(E.34)
1
10 D,
where D is an
M × M Chebyshev differentiation matrix generated for the domain [0, 1], like the
one used for the spatial derivative previously. The operator Dτ representing a partial
temporal derivative on B is then given by
Dτ = Dτ ⊗ I.
(E.35)
This operator acts only on the rows B, leaving its columns unaffected. We are now
E.6 Spectral Methods for Two Dimensions
468
able to write down our numerical approximation to (E.31), which is



 Dz i   A 


 = 0.



i Dτ
B
(E.36)
We can express this compactly as
LX = 0,
(E.37)
in an obvious notation. In order to solve for X — i.e. for A and B — we must
insert the boundary conditions for the signal field and atomic excitations. As in
§E.3 above, the general approach is to incorporate the equation
1 × ( X at the boundary ) = ( appropriate boundary condition ) .
(E.38)
For instance, to fix the signal field boundary condition Ain (τ ), we identify those
elements of X which describe A at z = 0. This is the first row of the matrix A,
corresponding to the elements
{1, N + 1, 2N + 1, . . . , (M − 1)N + 1},
(E.39)
of X. The left hand side of (E.38) is then implemented in two steps. First, we set
the rows of L with row indices given by (E.39) to zero. Second, we set the diagonal
elements of L, with both indices given by (E.39), equal to one. The resulting modified
E.6 Spectral Methods for Two Dimensions
469
matrix has the property that it maps the signal field at the boundary z = 0 to itself.
The right hand side of (E.38) is realized by introducing a vector C into the right
hand side of (E.37), with all elements zero, except for those with indices given by
(E.39), which are set equal to the discretized boundary condition,
C1 = Ain (τ1 ),
(E.40)
CN +1 = Ain (τ2 ),
(E.41)
..
.
C(M −1)N +1 = Ain (τM ).
(E.42)
(E.43)
The same procedure is used to implement the boundary condition Bin for the atomic
excitations. This time the boundary is the first column of B, which is indices 1 to
N of B, and therefore indices N M + 1 to N M + N of X. Finally, we are left with
a modified equation
Lb X = C,
(E.44)
where the superscript b distinguishes the modified matrix, and where C contains all
the relevant boundary conditions. The formal solution is then
X = Lb−1 C.
(E.45)
As mentioned previously, this is calculated much more efficiently if Gaussian elimination, rather than explicit matrix inversion, is used. In Matlab, we invoke the
E.6 Spectral Methods for Two Dimensions
470
backslash operator. This completes the numerics. We are now in possession of
approximations to both A and B at all spatial and temporal collocation points.
Polynomial interpolation, or piecewise spline interpolation, provides us with an accurate representation of the solution for arbitrary z and τ . In Figure E.6 we plot
the solutions for A and B found using this method.
Figure E.6 Spectral methods in two dimensions. We use N = 5
spatial collocation points, and M = 15 temporal collocation points, to
solve the example system (E.1), with the same boundary conditions
as shown previously in Figure E.4. The code runs in 0.01 seconds on
a 3 GHz machine, and the interpolated solutions shown are identical
to those generated using spectral collocation and RK2 time stepping.
Appendix F
Atomic Vapours
In this final Appendix we cover some results pertaining to atomic vapours, and in
particular to our ongoing experiments with cesium vapour. We review the behaviour
of vapour pressure with temperature, the concept of oscillator strength and its relation to the dipole moment associated with a transition, and the various mechanisms
of line broadening. We finish with an analysis of the polarization of Stokes scattering
in cesium.
F.1
Vapour pressure
No solid is infinitely sticky. Although the constituents may be bound tightly, there
remains a finite probability that an atom will escape. Every solid is therefore accompanied by a diffuse cloud of free atoms, the pressure of which is known as its
vapour pressure. It is this diffuse vapour that is used for the storage medium in our
experiments (and many others). The density of the vapour determines its optical
F.1 Vapour pressure
472
depth, and with it the achievable storage efficiency (see Chapter 5). To estimate
this density, we treat the atoms as a classical statistical ensemble. The probability
that an atom occupies a state with energy E is given by the Boltzmann distribution,
p(E ) ∼ e−E /kB T .
(F.1)
The energy E required for an atom to escape depends on the potentials within the
solid, and is largely independent of temperature. The ratio between the probabilities
for escape at two temperatures T0 , T , is therefore
p
− E (1/T −1/T0 )
= e kB
.
p0
(F.2)
This is, in fact, a good approximation for the vapour pressure above a solid, if p is
now taken to be the pressure. The relation can be derived much more rigourously
from the Clausius-Clapeyron equation [204] . Generally the reference state temperature T0 is taken to be the boiling point of the material, since then the vapour
pressure p0 must be equal to the ambient pressure at this temperature. The energy
E is commonly given as the molar enthalpy of sublimation ∆H. Boltzmann’s constant kB is then switched for the molar gas constant R = 8.31 JK−1 mole−1 . Part
(a) of Figure F.1 shows the vapour pressure of thallium — the storage medium used
in our preliminary experiments — using the values ∆H = 182 kJmole−1 , T0 = 1746
K and p0 = 5.8 × 105 Pa. Application of the ideal gas law p = nkB T allows us to
convert the vapour pressure into the atomic number density, which is also shown in
F.1 Vapour pressure
473
the plot.
Calculation of the vapour pressure of cesium, which is shown in part (b), is
complicated by the fact that cesium melts at 28◦ C. The qualitative behaviour is
unchanged, but we use an empirical formula taken from the excellent document by
Daniel Steck [184] , which is available online.
(a)
(b)
0
Pressure (Pa)
15
10
2
10
Cesium
22
10
10
10
−10
10
20
0
10
10
5
10
−20
0
10
5
−30
10
300
18
−2
10
350
400
450
Temperature (K)
10
500
10
10
16
−4
10
300
350
400
450
Temperature (K)
10
500
Number density (m -3 )
10
Thallium
Figure F.1 The vapour pressure of (a) thallium, and (b) cesium.
The black curves show the vapour pressure in Pascals (left axis) as
a function of temperature, measured in Kelvin. The red curves show
the corresponding number density (right axis), calculated using the
ideal gas law.
As described in §10.9.1 of Chapter 10, a rough estimate of the optical depth
of an ensemble is given by d = nλ2 L ∼ n × 10−14 , assuming optical wavelengths
and a vapour cell a few centimeters long. From this it is clear that thallium has
insufficient density for efficient storage at reasonable temperatures. Raising the
temperature further starts to introduce a thermal background into the signal field
from blackbody radiation! On the other hand, cesium has an optical depth of order
100, even at room temperature.
F.2 Oscillator strengths
F.2
474
Oscillator strengths
The oscillator strength fjk is a dimensionless measure of the dominance of an atomic
transition |ji ↔ |ki, compared to all other possible transitions. The oscillator
strengths for all transitions from the state |ji sum to unity,
X
fjk = 1,
∀j.
(F.3)
k
The relation between the fjk and the dipole matrix elements djk = hj|er|ki may be
derived solely from this, and one other condition, which is that
fjk ∝ |djk |2 .
(F.4)
This simply relates the f ’s to the transition probabilities. The following derivation is
due to Charles Thiel at Montana State University. To determine the proportionality
constant, we insert (F.4) into (F.3).
X
fjk = 1 =
X
=
X
k
αjk |hj|er|ki|2
k
k
e2 αjk ×
1
[hj|r|kihk|r|ji + hj|r|kihk|r|ji] .
2
(F.5)
The development in the second line seems a little obtuse, but we now make use of
the following relation between the matrix elements of the position and momentum
F.2 Oscillator strengths
475
operators,
pjk = hj|m∂t r|ki
= imhj|[r, H]|ki
= −imωjk rjk ,
(F.6)
where m is the electron mass, and where ωjk is the frequency splitting between
the states |ji, |ki, which are taken to be energy eigenstates of the Hamiltonian H.
Substitution of this relation into (F.5) yields
1 =
X ie2 αjk
[hj|r|kihk|p|ji − hj|p|kihk|r|ji]
2mωjk
k
=
=
−ie2 α
hj| [r, p] |ji
2m
e2 ~α
,
2m
where in the penultimate line we used the decomposition of the identity I =
(F.7)
P
j
|jihj|,
and where we somewhat heuristically set αjk = αωjk . In the final line we made use
of the canonical commutation relation [r, p] = i~. The resulting expression for the
oscillator strength is
fjk =
2mωjk
|djk |2 .
~e2
(F.8)
The inverse relation is useful for calculating the dipole moment from the oscillator
strengths listed in data tables.
F.3 Line broadening
F.3
476
Line broadening
There are three processes that increase the absorption linewidth in our cesium
vapour: Doppler broadening, pressure broadening and power broadening. An excellent treatment of these effects can be found in The Quantum Theory of Light by
Loudon [107] . Here we briefly review the physics of line broadening.
F.3.1
Doppler broadening
Doppler broadening refers to the variation in the resonance frequencies of atoms
moving with different velocities. Consider an atom emitting light of wavelength λ0
while moving at a velocity v (see Figure F.2). Over the course of an optical period
T0 , the atom moves so that the wavelength appears compressed, λ = λ0 −vT0 . Using
the relations λ0 = 2πc/ω0 , λ = 2πc/ω and T0 = 2π/ω0 , we derive the Doppler shift
v
δω = −ω0 ,
c
(F.9)
where δω = ω − ω0 is the shift in angular frequency caused by the motion of the
atom. The spectral intensity is then given by the distribution of atomic velocities
in the vapour. An atom with velocity v and mass M has a kinetic energy
E = 12 M v 2 .
(F.10)
F.3 Line broadening
477
Using the Boltzmann distribution (F.1), and substituting for v using (F.9) gives the
spectrum
2
I(δω) ∝ e−(δω/γd ) ,
(F.11)
where the Doppler linewidth is given by
r
γd =
ω0
2kB T
× .
M
c
(F.12)
Figure F.2 The Doppler shift. An atom moving with velocity
v ‘catches up’ with the light it emits, so that the wave appears
squashed.
F.3.2
Pressure broadening
Collisions between atoms in a vapour cause disruptions to the wavetrain of light
emitted by them. In particular, the phase of the light is randomized by each collision.
To understand the effect on the spectrum of the light we should consider the statistics
of collisions.
An atom with cross sectional area σ travelling at velocity v sweeps out a volume
F.3 Line broadening
478
Figure F.3 Collisions in a vapour. An atom with cross section σ
travelling with velocity v sweeps out a volume vσdτ in a short time
dτ . Any other atom within this volume gets hit!
σvdτ in a short time dτ (see Figure F.3). The probability of a collision is just the
probability of finding another atom within this volume, which is given by n × σvdτ ,
where n is the number density of atoms in the vapour. Now consider ps (τ ), the
probability that an atom ‘survives’ without colliding for a time τ . We do not know
this probability yet, but we can say the following,
ps (τ + dτ ) = ps (τ ) × (1 − nσvdτ ) .
(F.13)
That is, the probability of surviving a further time dτ after τ is given by the probability of surviving for a time τ , and then not colliding during dτ . Taylor expanding
(F.13) gives
ps (τ ) + ∂τ ps (τ )dτ
= ps (τ ) − nσvps (τ )dτ,
⇒ ∂τ ps (τ ) = −nσvps (τ ).
(F.14)
F.3 Line broadening
479
The survival probability is therefore given by an exponential distribution,
ps (τ ) = γp e−γp τ ,
(F.15)
where γp = nσv, and where the preceeding factor of γp ensures that the distribution
is correctly normalized.
We are now in a position to derive the form of a collision-broadened spectrum.
The spectral intensity profile of the emitted light is given by
e 2
I(ω) = E(ω)
Z
1
∗
= Fτ √
E (t)E(t + τ ) dt (ω),
2π
(F.16)
e
where E(ω)
and E(t) = E0 eiω0 t+φ(t) are the spectral and temporal electric field
amplitudes of the light, and where in the second line we have used the convolution
theorem (see §D.3.5 in Appendix D). The convolution inside the curly braces is the
first order correlation function of E(t). In the absence of a collision, there is no
change to E, and this correlation function is constant. After a collision, the phase
φ is randomized, and when this is averaged over all atoms, the correlation function
vanishes. The average correlation function is thus proportional to the probability
that there is no collision in the interval [0, τ ]. Factorizing out the carrier frequency
F.3 Line broadening
480
ω0 using the shift theorem (see §D.3.4 in Appendix D), the spectrum is given by
I(δω) =
=
=
Z τ
|E0 |2
0
0
√
ps (τ )dτ (δω)
× Fτ 1 −
2π
0
n
o
|E0 |2
√
× Fτ e−γp |τ | (δω)
2π
γp
|E0 |2
,
× 2
2π
γp + δω 2
(F.17)
where in the second line the modulus sign indicates that the correlation function is
symmetric in time. (F.17) describes a Lorentzian lineshape with a width γp . Although this spectrum was derived by considering emission, the absorption spectrum
is identical, since absorption is simply the time reverse of emission. An estimate of
the collision-broadened linewidth is given by
r
γp ≈
n(πd2atom )
×
2kB T
.
M
(F.18)
The number density n can be obtained from the vapour pressure as described in the
start of this Appendix. The term in brackets approximates the collision cross-section
σ as the area of a circle with radius given by the atomic width datom . The last factor
is found by setting the kinetic energy (F.10) equal to the thermal energy kB T and
solving for the average atomic velocity.
F.3.3
Power broadening
In the presence of an intense laser field, an atomic absorption line broadens. One
way to understand this effect is to consider that the presence of light at the atomic
F.4 Raman polarization
481
resonance frequency stimulates emission from the excited state, and this reduces
the lifetime of the state, which introduces a concomitant broadening. Alternatively,
consider that the atomic dynamics are governed by Hamiltonian evolution. In the
loosest possible sense, the resulting behaviour can be characterized as a mixing of all
the frequencies in the Hamiltonian. Thus if there exists an optical frequency ω and
a Rabi frequency Ω, there will be driven oscillations at frequencies of ω ± Ω. Such
oscillations may be resonant with the atomic transition at ω0 , even when neither
frequency is separately. Therefore absorption is possible at large detunings, provided
the laser is sufficiently intense. The atomic transition is rendered broader by the laser
field. In fact, the power-broadened linewidth is indeed given by γpb = Ω, as a more
detailed derivation from the damped optical Bloch equations shows [107] . This effect
is essentially the same as the dynamic Stark effect that generates the instantaneous
frequency shift in (5.74) in Chapter 5, and the Autler-Townes splitting discussed in
§2.3.1 in Chapter 2.
F.4
Raman polarization
In Chapter 10 we describe the preliminary steps taken toward building a Raman
quantum memory in cesium vapour. Here we explain how we arrive at the conclusion
that the Stokes light scattered from a linearly polarized pump will be polarized
orthogonally to the pump, when far detuned.
We consider Raman scattering from the 6S1/2 ↔ 6P3/2 D2 line in cesium, at 852
nm. There are two hyperfine levels in the ground state, with F = 3 and F = 4.
F.4 Raman polarization
482
Atoms prepared in one of these levels are transferred by the Raman interaction to
the other level, with the transfer mediated by one of the excited states in the 6P3/2
manifold. In the process, a photon from the Raman pump is absorbed, and a Stokes
photon is emitted1 . Each photon carries only one quantum of angular momentum,
so the intermediate state cannot differ in its angular momentum quantum number
by more than 1 from the initial and final states. Of the four hyperfine levels —
F = 2, 3, 4, 5 — in the excited state manifold, only the central two with F = 3, 4 are
compatible with this requirement, so we consider only these two intermediate levels.
Within each hyperfine level, there are 2F + 1 Zeeman sublevels, with magnetic
quantum numbers m = −F, −F +1, . . . , F −1, F . We specify one of these states with
the notation |F, mi. The magnetic quantum numbers correspond to the component
of the atomic angular momentum along some axis, known as the quantization axis.
We must choose the direction for this axis before proceeding further. If a magnetic
field is applied to the atoms, the only sensible choice is to align the quantization
axis with the field. But in the absence of a magnetic field, the choice is arbitrary,
and we may choose the quantization direction in whichever way is convenient for
our calculations. As mentioned above, a single photon carries one unit of angular
momentum, and so it can change the magnetic quantum number by at most ±1, or
indeed not at all, if the spin of the photon has no component along the quantization
1
Strictly the term ‘Stokes’ refers to Raman scattered photons with less energy than the pump
photons. When they have more energy than the pump, they are commonly termed anti-Stokes
photons. In the present case, Stokes scattering only occurs if the atoms are prepared in the lower of
the two ground states, with F = 3. The distinction between Stokes and anti-Stokes is not important
for us, however. For simplicity, we always refer to the light emitted in Raman scattering as Stokes
light, regardless of whether or not its frequency is lower than that of the Raman pump.
F.4 Raman polarization
483
axis. Figure F.4 shows how the direction and polarization of a photon with respect to
the quantization axis is connected with the changes in m induced by its absorption.
quantization axis
Figure F.4 Polarization selection rules. Right and left circular
polarizations propagating along the quantization axis change m by
+1 and −1; they are known as σ+ and σ− polarizations, respectively.
It is intuitive that as the electric field follows a corkscrew trajectory, it
applies a torque to the atoms about the quantization axis, increasing
or decreasing m appropriately. π polarized light does not change m.
It is polarized linearly along the quantization axis and propagates
at right angles to it, so it cannot induce a turning moment. The
orthogonal linear polarization can, however, change m by either +1
or −1.
We choose the quantization axis so that the linearly polarized Raman pump
is π-polarized. With this choice, absorption of a pump photon cannot change the
magnetic quantum number m. Given the initial state |Fi , mi i of an atom, this
narrows down the number of possible intermediate states to two: either |Fint =
3, mint = mi i or |Fint = 4, mint = mi i (the only exception is if mi = 0 initially. Then
F must change, so there is only one possible intermediate state, with Fint 6= Fi ). The
atom now decays from the intermediate excited state into the ‘final’ ground hyperfine
level with Ff 6= Fi . There is no restriction on the polarization of the emitted Stokes
photon, so there are three possibilities for this latter part of the interaction. The
F.4 Raman polarization
484
possible final states are |Ff , mf = mi i and |Ff , mf = mi ±1i. As shown in Figure F.5,
each final state |Ff , mf i can be reached from the given initial state |Fi , mi i by two
alternative ‘paths’. The first via the excited state with Fm = 3, and the second with
Fm = 4. When two different paths connect the same starting and ending points,
there is the possibility of interference. The present case is archetypical: π-polarized
Stokes emission is forbidden because of destructive interference between the two
interaction pathways. To see this, we explicitly evaluate the quantum mechanical
amplitude Aππ for π-polarized Stokes emission given a π-polarized pump, given the
initial state |Fi , mi i,
Aππ (Fi , mi ) =
X
d(J = 1/2, Fi , mi → J = 3/2, Fm , mm = mi ) × (F.19)
Fm =3,4
d(J = 3/2, Fm , mm = mi → J = 1/2, Ff 6= Fi , mf = mi ).
Here d denotes a dipole matrix element connecting the states with angular momentum quantum numbers indicated by its arguments. Recall that J = 3/2 describes
the intermediate excited states. The selection rules set by fixing both photon polarizations to π keep the magnetic quantum number equal to its initial value throughout
the interaction. We have factorized out any dependence of the dipole matrix elements on the radial (or principle) quantum number. That this is always possible
is a consequence of the Wigner-Eckhart theorem. There exists a sophisticated and
confusing mathematical apparatus for dealing with the resulting ‘reduced’ matrix
F.4 Raman polarization
485
elements. In standard notation, we can write [184]
d(J, F, m → J 0 , F 0 , m0 ) = (−1)F +J+I+1
p
(2F + 1)(2J + 1)



 J
J0



1 


 F0
F


I 
0
1F F
C(m
0 −m)mm0 ,
(F.20)
j1 j2 J
where the curly braces denote a Wigner 6j symbol, and where Cm
is a Clebsch1 m2 M
Gordan coefficient. Here I = 7/2 is the cesium nuclear spin. The 1’s appearing
in various places represent the angular momentum of the photon involved in the
transition. Angular momentum conservation requires that the lower indices of the
Clebsch-Gordan coefficient obey the sum rule m1 + m2 = M . The symbols in (F.20)
can be evaluated by standard routines in Mathematica. Matlab routines have also
been written, and are available freely on the web. Performing the sum in (F.19)
using (F.20), we find that Aππ is identically zero for all choices of the initial state
(i.e. either Fi = 3 or 4, and any value of mi ). No doubt there is a deep grouptheoretical reason for this, but I do not know what it is! Nonetheless the calculation
shows that Stokes scattered photons emerge with the orthogonal linear polarization
to that of the pump. A very fortunate outcome for the experimenter who wishes to
filter the weak Raman signal out from the bright pump.
In evaluating the scattering amplitude, we have given equal weight to both pathways. This implicitly assumes that the optical fields are sufficiently far detuned from
the excited state manifold that the small energy splitting between the intermediate
hyperfine levels — around 200 MHz — makes a negligible difference to the Raman
coupling. If instead we tune into resonance with one of the levels, we can neglect
F.4 Raman polarization
486
-3 -2 -1 0 1 2 3
-4 -3 -2 -1 0 1 2 3 4
-3 -2 -1 0 1 2 3
-4 -3 -2 -1 0 1 2 3 4
Figure F.5 Alternative scattering pathways. Each initial state is
coupled by the Raman interaction to a final state via two possible
interaction pathways; the first involves an intermediate excited state
with Fm = 3 (black-blue), the second with Fm = 4 (black-red). The
figure illustrates this with the example of an atom initially prepared
in the state Fi = 4, mi = 1. Absorption of a π-polarized Raman pump
photon couples this state to the intermediate states |Fm = 3, mm = 1i
and |Fm = 4, mm = 1i. Finally emission of a Stokes photon reduces
the magnetic quantum number by one: both paths end with the final
state |Ff = 3, mf = 0i.
scattering involving the other level, and there is no longer any interference. Isolating
just a single term from the sum in (F.19), we calculate a Stokes π-polarization of
40% on resonance with the Fm = 3 state, averaged over all possible initial states.
This averaging assumes an unpolarized ensemble, with atoms populating all Zeeman
substates in the initial hyperfine level equally. If we tune into resonance with the
Fm = 4 state, on average only 26% of the Stokes light is π-polarized. These calculations show that even on resonance, there remains a significant proportion of the
Stokes scattering that is polarized orthogonally to the pump, so that polarization
filtering is always feasible.
Note that these conclusions hold equally well for the polarizations of the the
signal and control fields in a cesium quantum memory, which has at its heart the very
F.4 Raman polarization
487
same Raman interaction. Therefore a vertically polarized signal field can be stored
by a horizontally polarized control pulse, when both are tuned far from resonance.
If we choose a circularly polarized Raman pump — this time aligning the quantization axis with the pump propagation direction for convenience — we find that, far
from resonance, all of the Stokes light is emitted into the same circular polarization
as the pump. That is, right-circular scatters into right-circular, and left-circular into
left-circular. This is rather unfortunate, because it precludes the use of what might
otherwise have been a rather ingenious ‘trick’2 : Suppose that the atomic ensemble
is prepared in the the state |Fi = 4, mi = 4i. Then a σ+ -polarized photon can
only couple to a state with mm = 5, which lies within the Fm = 5 hyperfine level
in the excited state. This level cannot participate in the Raman interaction (see
the discussion at the start of this section), and so a σ+ -polarized photon cannot
act as a Raman pump. If the ensemble is used as a quantum memory, then a σ+ polarized control pulse cannot cause spontaneous Stokes scattering. This eliminates
the unwanted ‘noise’ process mentioned in §4.6 of Chapter 4 (see part (b) of Figure
4.3). Unfortunately, the signal photon must be σ− -polarized, and then the memory
coupling vanishes because quantum interference between the two possible scattering pathways destroys any coupling between orthogonal circular polarizations. This
explains why we make no attempt to spin-polarize our cesium ensemble: it is more
trouble than it is worth!
2
This was suggested by Jean Dalibard when he came to Oxford to give a colloquium.
Bibliography
[1] CE Shannon. A mathematical theory of communication. ACM SIGMOBILE
Mobile Computing and Communications Review, 5(1):3–55, 2001.
[2] G.E. Moore. Cramming more components on to integrated circuits. Electronics
Journal, 38(8):114–117, 1965.
[3] R.P. Feynman. Feynman Lectures on Computation. Da Capo Press, 1996.
[4] D. Deutsch. Quantum theory, the church-turing principle and the universal
quantum computer. Proceedings of the Royal Society of London. Series A,
Mathematical and Physical Sciences (1934-1990), 400(1818):97–117, 1985.
[5] Benjamin Schumacher. Quantum coding. Phys. Rev. A, 51(4):2738–2747, Apr
1995.
[6] A. M. Steane. Error correcting codes in quantum theory. Phys. Rev. Lett.,
77(5):793–797, Jul 1996.
[7] A. R. Calderbank and Peter W. Shor. Good quantum error-correcting codes
exist. Phys. Rev. A, 54(2):1098–1105, Aug 1996.
[8] W.K. Wootters and W.H. Zurek. A single quantum cannot be cloned. Nature,
299(5886):802–803, 1982.
[9] M.A. Nielsen and I.L. Chuang. Quantum Computation and Quantum Information. Cambridge University Press, 2000.
[10] A. Steane. The ion trap quantum information processor. Applied Physics B:
Lasers and Optics, 64(6):623–643, 1997.
[11] M. V. Gurudev Dutt, L. Childress, L. Jiang, E. Togan, J. Maze, F. Jelezko,
A. S. Zibrov, P. R. Hemmer, and M. D. Lukin. Quantum register based on individual electronic and nuclear spin qubits in diamond. Science, 316(5829):1312–
1316, 2007.
[12] L. Childress, MV Gurudev Dutt, JM Taylor, AS Zibrov, F. Jelezko,
J. Wrachtrup, PR Hemmer, and MD Lukin. Coherent Dynamics of Coupled
Electron and Nuclear Spin Qubits in Diamond, 2006.
BIBLIOGRAPHY
489
[13] E. Knill, R. Laflamme, and G. J. Milburn. A scheme for efficient quantum
computation with linear optics. Nature, 409(6816):46–52, 2001.
[14] N. Yoran and B. Reznik. Deterministic linear optics quantum computation
with single photon qubits. Phys. Rev. Lett., 91(3):037903, Jul 2003.
[15] Michael A. Nielsen. Optical quantum computation using cluster states. Physical Review Letters, 93(4):040503, 2004.
[16] J. M. Raimond, M. Brune, and S. Haroche. Manipulating quantum entanglement with atoms and photons in a cavity. Rev. Mod. Phys., 73(3):565–582,
Aug 2001.
[17] Karl Tordrup, Antonio Negretti, and Klaus Molmer. Holographic quantum
computing. Physical Review Letters, 101(4):040501, 2008.
[18] C.A. Muschik, I. de Vega, D. Porras, and J.I. Cirac. Quantum processing
photonic states in optical lattices. Arxiv preprint quant-ph/0611093, 2006.
[19] S. D. Barrett, P. P. Rohde, and T. M. Stace. Scalable quantum computing
with atomic ensembles, 2008.
[20] N. Gisin, G. Ribordy, W. Tittel, and H. Zbinden. Quantum cryptography.
Reviews of Modern Physics, 74(1):145–195, 2002.
[21] RL Rivest, A. Shamir, and L. Adleman. A method for obtaining digital signatures and public-key cryptosystems. Communications of the ACM, 21(2):120–
126, 1978.
[22] M.A. Nielsen and I.L. Chuang. Quantum Computation and Quantum Information. Cambridge University Press, 2000.
[23] P.W. Shor. Algorithms for quantum computation: Discrete logarithms and
factoring. Proceedings of the 35th Annual Symposium on Foundations of Computer Science, 124, 1994.
[24] C.Y. Lu, D.E. Browne, T. Yang, and J.W. Pan.
Demonstration of
Shor’s quantum factoring algorithm using photonic qubits. Arxiv preprint
arXiv:0705.1684, 2007.
[25] BP Lanyon, TJ Weinhold, NK Langford, M. Barbieri, DFV James,
A. Gilchrist, and AG White. Experimental Demonstration of a Compiled
Version of Shor’s Algorithm with Quantum Entanglement. Physical Review
Letters, 99(25):250505, 2007.
[26] L.M.K. Vandersypen, M. Steffen, G. Breyta, C.S. Yannoni, R. Cleve, and I.L.
Chuang. Experimental Realization of an Order-Finding Algorithm with an
NMR Quantum Computer. Physical Review Letters, 85(25):5452–5455, 2000.
BIBLIOGRAPHY
490
[27] C.H. Bennett, G. Brassard, et al. Quantum cryptography: Public key distribution and coin tossing. Proceedings of IEEE International Conference on
Computers, Systems, and Signal Processing, 175, 1984.
[28] W. Heisenberg. The Physical Principles of the Quantum Theory. Courier
Dover Publications, 1949.
[29] JS Bell. On the Einstein-Podolsky-Rosen paradox. Physics (US) Discontinued
with Vol. 4, no. 1, 1968, 1, 1964.
[30] Artur K. Ekert. Quantum cryptography based on bell’s theorem. Phys. Rev.
Lett., 67(6):661–663, Aug 1991.
[31] Jian-Wei Pan, Dik Bouwmeester, Harald Weinfurter, and Anton Zeilinger. Experimental entanglement swapping: Entangling photons that never interacted.
Phys. Rev. Lett., 80(18):3891–3894, May 1998.
[32] D. Deutsch, A. Ekert, R. Jozsa, C. Macchiavello, S. Popescu, and A. Sanpera.
Quantum Privacy Amplification and the Security of Quantum Cryptography
over Noisy Channels. Physical Review Letters, 77(13):2818–2821, 1996.
[33] W. Dür, H.-J. Briegel, J. I. Cirac, and P. Zoller. Quantum repeaters based on
entanglement purification. Phys. Rev. A, 59(1):169–181, Jan 1999.
[34] Charles H. Bennett, Herbert J. Bernstein, Sandu Popescu, and Benjamin Schumacher. Concentrating partial entanglement by local operations. Phys. Rev.
A, 53(4):2046–2052, Apr 1996.
[35] Jian-Wei Pan, Christoph Simon, Caslav Brukner, and Anton Zeilinger. Entanglement purification for quantum communication. Nature, 410(6832):1067–
1070, 2001.
[36] Christoph Simon and Jian-Wei Pan. Polarization entanglement purification
using spatial entanglement. Phys. Rev. Lett., 89(25):257901, Dec 2002.
[37] L.M. Duan, MD Lukin, JI Cirac, and P. Zoller. Long-distance quantum communication with atomic ensembles and linear optics. Nature, 414:413–418,
2001.
[38] Nicolas Sangouard, Christoph Simon, Bo Zhao, Yu-Ao Chen, Hugues de Riedmatten, Jian-Wei Pan, and Nicolas Gisin. Robust and efficient quantum repeaters with atomic ensembles and linear optics. Physical Review A (Atomic,
Molecular, and Optical Physics), 77(6):062301, 2008.
[39] L. Jiang, J. M. Taylor, and M. D. Lukin. Fast and robust approach to longdistance quantum communication with atomic ensembles. Physical Review A
(Atomic, Molecular, and Optical Physics), 76(1):012301, 2007.
BIBLIOGRAPHY
491
[40] Zeng-Bing Chen, Bo Zhao, Yu-Ao Chen, Jorg Schmiedmayer, and Jian-Wei
Pan. Fault-tolerant quantum repeater with atomic ensembles and linear optics.
Physical Review A (Atomic, Molecular, and Optical Physics), 76(2):022329,
2007.
[41] Nicolas Sangouard, Christoph Simon, Jiri Minar, Hugo Zbinden, Hugues
de Riedmatten, and Nicolas Gisin. Long-distance entanglement distribution
with single-photon sources. Physical Review A (Atomic, Molecular, and Optical Physics), 76(5):050301, 2007.
[42] Bo Zhao, Zeng-Bing Chen, Yu-Ao Chen, Jorg Schmiedmayer, and Jian-Wei
Pan. Robust creation of entanglement between remote memory qubits. Physical Review Letters, 98(24):240502, 2007.
[43] L.M. Duan, JI Cirac, and P. Zoller. Three-dimensional theory for interaction between atomic ensembles and free-space light. Physical Review A,
66(2):23818, 2002.
[44] Peter J. Mosley, Jeff S. Lundeen, Brian J. Smith, Piotr Wasylczyk, Alfred B.
U’Ren, Christine Silberhorn, and Ian A. Walmsley. Heralded generation of
ultrafast single photons in pure quantum states. Physical Review Letters,
100(13):133601, 2008.
[45] Axel Kuhn, Markus Hennrich, and Gerhard Rempe. Deterministic singlephoton source for distributed quantum networking.
Phys. Rev. Lett.,
89(6):067901, Jul 2002.
[46] Christian Kurtsiefer, Sonja Mayer, Patrick Zarda, and Harald Weinfurter. Stable solid-state source of single photons. Phys. Rev. Lett., 85(2):290–293, Jul
2000.
[47] A. Beveratos, S. Kühn, R. Brouri, T. Gacoin, J. P. Poizat, and P. Grangier. Room temperature stable single-photon source. The European Physical
Journal D - Atomic, Molecular, Optical and Plasma Physics, 18(2):191–196,
2002.
[48] I. I. Rabi, J. R. Zacharias, S. Millman, and P. Kusch. A new method of
measuring nuclear magnetic moment. Phys. Rev., 53(4):318, Feb 1938.
[49] H. Mabuchi and A. C. Doherty. Cavity quantum electrodynamics: Coherence
in context. Science, 298(5597):1372–1377, 2002.
[50] P. Rabl, D. DeMille, J. M. Doyle, M. D. Lukin, R. J. Schoelkopf, and P. Zoller.
Hybrid quantum processors: Molecular ensembles as quantum memory for
solid state circuits. Physical Review Letters, 97(3):033003, 2006.
[51] J. I. Cirac, P. Zoller, H. J. Kimble, and H. Mabuchi. Quantum state transfer
and entanglement distribution among distant nodes in a quantum network.
Phys. Rev. Lett., 78(16):3221–3224, Apr 1997.
BIBLIOGRAPHY
492
[52] M. Hennrich, T. Legero, A. Kuhn, and G. Rempe. Vacuum-stimulated raman
scattering based on adiabatic passage in a high-finesse optical cavity. Phys.
Rev. Lett., 85(23):4872–4875, Dec 2000.
[53] EM Purcell. Spontaneous emission probabilities at radio frequencies. Phys.
Rev, 69(681):166, 1946.
[54] AD Boozer, A. Boca, R. Miller, TE Northup, and HJ Kimble. Reversible State
Transfer between Light and a Single Trapped Atom. Physical Review Letters,
98(19):193601, 2007.
[55] J. Bochmann, M. Muecke, G. Langfahl-Klabes, C. Erbel, B. Weber, H. P.
Specht, D. L. Moehring, and G. Rempe. Fast Excitation and Photon Emission
of a Single-Atom-Cavity System. Phys. Rev. Lett., 67(13):1727–1730, Sep 2008.
[56] J. P. Reithmaier, G. Sek, A. Loffler, C. Hofmann, S. Kuhn, S. Reitzenstein,
L. V. Keldysh, V. D. Kulakovskii, T. L. Reinecke, and A. Forchel. Strong
coupling in a single quantum dot-semiconductor microcavity system. Nature,
432(7014):197–200, 2004.
[57] A. Kiraz, P. Michler, C. Becher, B. Gayral, A. Imamoğlu, L. Zhang, E. Hu,
WV Schoenfeld, and PM Petroff. Cavity-quantum electrodynamics using a
single InAs quantum dot in a microdisk structure. Applied Physics Letters,
78:3932, 2001.
[58] A. ImamogŻlu, DD Awschalom, G. Burkard, DP DiVincenzo, D. Loss,
M. Sherwin, and A. Small. Quantum Information Processing Using Quantum Dot Spins and Cavity QED. Physical Review Letters, 83(20):4204–4207,
1999.
[59] R.J. Young, S.J. Dewhurst, R.M. Stevenson, P. Atkinson, A.J. Bennett, M.B.
Ward, K. Cooper, D.A. Ritchie, and A.J. Shields. Single electron-spin memory
with a semiconductor quantum dot. Arxiv preprint arXiv:0706.2143, 2007.
[60] A. André, D. Demille, JM Doyle, MD Lukin, SE Maxwell, P. Rabl,
RJ Schoelkopf, and P. Zoller. A coherent all-electrical interface between polar molecules and mesoscopic superconducting resonators. Nature Physics,
2(9):636–642, 2006.
[61] M.K. Tey, S.A. Aljunid, F. Huber, B. Chng, Z. Chen, G. Maslennikov, and
C. Kurtsiefer. Interfacing light and single atoms with a lens. Arxiv preprint
arXiv:0804.4861, 2008.
[62] M.K. Tey, Z. Chen, S.A. Aljunid, B. Chng, F. Huber, G. Maslennikov, and
C. Kurtsiefer. Strong interaction between light and a single trapped atom
without a cavity. Arxiv preprint arXiv:0802.3005, 2008.
BIBLIOGRAPHY
493
[63] M. Sondermann, R. Maiwald, H. Konermann, N. Lindlein, U. Peschel, and
G. Leuchs. Design of a mode converter for efficient light-atom coupling in free
space. Applied Physics B: Lasers and Optics, 89(4):489–492, 2007.
[64] K.-J. Boller, A. Imamolu, and S. E. Harris. Observation of electromagnetically
induced transparency. Phys. Rev. Lett., 66(20):2593–2596, May 1991.
[65] Michael Fleischhauer, Atac Imamoglu, and Jonathan P. Marangos. Electromagnetically induced transparency: Optics in coherent media. Reviews of
Modern Physics, 77(2):633, 2005.
[66] J. H. Eberly, M. L. Pons, and H. R. Haq. Dressed-field pulses in an absorbing
medium. Phys. Rev. Lett., 72(1):56–59, Jan 1994.
[67] S. H. Autler and C. H. Townes. Stark effect in rapidly varying fields. Phys.
Rev., 100(2):703–722, Oct 1955.
[68] Lene Vestergaard Hau, S. E. Harris, Zachary Dutton, and Cyrus H. Behroozi.
Light speed reduction to 17 metres per second in an ultracold atomic gas.
Nature, 397(6720):594–598, 1999.
[69] M. Fleischhauer and M. D. Lukin. Dark-state polaritons in electromagnetically
induced transparency. Phys. Rev. Lett., 84(22):5094–5097, May 2000.
[70] Chien Liu, Zachary Dutton, Cyrus H. Behroozi, and Lene Vestergaard Hau.
Observation of coherent optical information storage in an atomic medium using
halted light pulses. Nature, 409(6819):490–493, 2001.
[71] K. S. Choi, H. Deng, J. Laurat, and H. J. Kimble. Mapping photonic entanglement into and out of a quantum memory. Nature, 452:67–71, March
2008.
[72] Julien Laurat, Chin wen Chou, Hui Deng, Kyung Soo Choi, Daniel Felinto,
Hugues de Riedmatten, and H J Kimble. Towards experimental entanglement
connection with atomic ensembles in the single excitation regime. New Journal
of Physics, 9(6):207, 2007.
[73] T. Chaneliere, D. N. Matsukevich, S. D. Jenkins, S.-Y. Lan, R. Zhao, T. A. B.
Kennedy, and A. Kuzmich. Quantum interference of electromagnetic fields
from remote quantum memories. Physical Review Letters, 98(11):113602, 2007.
[74] Irina Novikova, Nathaniel B. Phillips, and Alexey V. Gorshkov. Optimal light
storage with full pulse shape control, 2008.
[75] CV Raman and KS Krishnan. A new type of secondary radiation. Nature,
121(3048):501, 1928.
[76] A. E. Kozhekin, K. Mølmer, and E. Polzik. Quantum memory for light. Phys.
Rev. A, 62:033809, 2000.
BIBLIOGRAPHY
494
[77] J. Nunn, I. A. Walmsley, M. G. Raymer, K. Surmacz, F. C. Waldermann,
Z. Wang, and D. Jaksch. Mapping broadband single-photon wave packets
into an atomic memory. Physical Review A (Atomic, Molecular, and Optical
Physics), 75(1):011401, 2007.
[78] Mattias Nilsson and Stefan Krll. Solid state quantum memory using complete
absorption and re-emission of photons by tailored and externally controlled
inhomogeneous absorption profiles. Optics Communications, 247(4-6):393–
403, 2005.
[79] S. A. Moiseev and S. Kröll. Complete reconstruction of the quantum state of a
single-photon wave packet absorbed by a doppler-broadened transition. Phys.
Rev. Lett., 87(17):173601, Oct 2001.
[80] SA Moiseev, VF Tarasov, and BS Ham. Quantum memory photon echo-like
techniques in solids. Journal of Optics B: Quantum and Semiclassical Optics,
5(4):S497–S502, 2003.
[81] S. A. Moiseev and B. S. Ham. Photon-echo quantum memory with efficient
multipulse readings. Phys. Rev. A, 70(6):063809, Dec 2004.
[82] B. Kraus, W. Tittel, N. Gisin, M. Nilsson, S. Kroll, and J. I. Cirac. Quantum memory for nonstationary light fields based on controlled reversible inhomogeneous broadening. Physical Review A (Atomic, Molecular, and Optical
Physics), 73(2):020302, 2006.
[83] AL Alexander, JJ Longdell, MJ Sellars, and NB Manson. Photon Echoes
Produced by Switching Electric Fields. Physical Review Letters, 96(4):43602,
2006.
[84] G. Hétet, JJ Longdell, AL Alexander, PK Lam, and MJ Sellars. Electro-Optic
Quantum Memory for Light Using Two-Level Atoms. Physical Review Letters,
100(2):23601, 2008.
[85] M. U. Staudt, S. R. Hastings-Simon, M. Nilsson, M. Afzelius, V. Scarani,
R. Ricken, H. Suche, W. Sohler, W. Tittel, and N. Gisin. Fidelity of an
optical memory based on stimulated photon echoes. Physical Review Letters,
98(11):113601, 2007.
[86] E. Desurvire, C.R. Giles, and J.L. Zyskind. Erbium-doped fiber amplifier,
June 25 1991. US Patent 5,027,079.
[87] Nicklas Ohlsson, R. Krishna Mohan, and Stefan Krll. Quantum computer
hardware based on rare-earth-ion-doped inorganic crystals. Optics Communications, 201(1-3):71–77, 2002.
[88] E. Fraval, M. J. Sellars, and J. J. Longdell. Dynamic decoherence control of
a solid-state nuclear-quadrupole qubit. Physical Review Letters, 95(3):030506,
2005.
BIBLIOGRAPHY
495
[89] B. R. Judd. Optical absorption intensities of rare-earth ions. Phys. Rev.,
127(3):750–761, Aug 1962.
[90] G. Hétet, JJ Longdell, MJ Sellars, PK Lam, and BC Buchler. Bandwidth and
Dynamics of the Gradient Echo Memory. Arxiv preprint arXiv:0801.3860,
2008.
[91] Mikael Afzelius, Christoph Simon, Hugues de Riedmatten, and Nicolas Gisin.
Multi-Mode Quantum Memory based on Atomic Frequency Combs, 2008.
[92] Luı́s E. E. de Araujo, Ian A. Walmsley, and C. R. Stroud. Analytic solution
for strong-field quantum control of atomic wave packets. Phys. Rev. Lett.,
81(5):955–958, Aug 1998.
[93] L. E. E. de Araujo and I. A. Walmsley. Quantum control of molecular
wavepackets: An approximate analytic solution for the strong-response regime.
Journal of Physical Chemistry A, 103(49):10409–10416, 1999.
[94] Luı́s E. E. de Araujo and Ian A. Walmsley. Quantum control of rydberg wave
packets in the strong-response regime. Phys. Rev. A, 63(2):023401, Jan 2001.
[95] L.E.E. de Araujo and I.A. Walmsley. Analytic solution for quantum control of
atomic and molecular wavepackets. JOURNAL OF OPTICS B QUANTUM
AND SEMICLASSICAL OPTICS, 5(1):27–42, 2003.
[96] J. Oreg, G. Hazak, and JH Eberly. Multilevel inversion schemes in and beyond
the adiabatic limit. Physical Review A, 32(5):2776–2783, 1985.
[97] S. Guérin, S. Thomas, and H. R. Jauslin. Optimization of population transfer
by adiabatic passage. Phys. Rev. A, 65(2):023409, Jan 2002.
[98] B. W. Shore, K. Bergmann, A. Kuhn, S. Schiemann, J. Oreg, and J. H. Eberly.
Laser-induced population transfer in multistate systems: A comparative study.
Phys. Rev. A, 45(7):5297–5300, Apr 1992.
[99] B. Lauritzen, S. R. Hastings-Simon, H. de Riedmatten, M. Afzelius, and
N. Gisin. State Preparation by Optical Pumping in Erbium Doped Solids using Stimulated Emission and Spin Mixing. Messenger of Mathematics, 19:1–5,
2008.
[100] Hugues de Riedmatten, Mikael Afzelius, Matthias U. Staudt, Christoph Simon,
and Nicolas Gisin. A solid-state light-matter interface at the single-photon
level. Nature, 456(7223):773–777, 12 2008.
[101] S.L. Braunstein and P. van Loock. Quantum information with continuous
variables. Reviews of Modern Physics, 77(2):513–577, 2005.
[102] Samuel L. Braunstein and Peter van Loock. Quantum information with continuous variables. Reviews of Modern Physics, 77(2):513, 2005.
BIBLIOGRAPHY
496
[103] T. C. Ralph. Continuous variable quantum cryptography. Phys. Rev. A,
61(1):010303, Dec 1999.
[104] C. Silberhorn, TC Ralph, N. Lütkenhaus, and G. Leuchs. Continuous Variable Quantum Cryptography: Beating the 3 dB Loss Limit. Physical Review
Letters, 89(16):167901, 2002.
[105] Stefano Pirandola, Stefano Mancini, Seth Lloyd, and Samuel L. Braunstein.
Continuous-variable quantum cryptography using two-way quantum communication. Nat Phys, advanced online publication:–, 2008.
[106] Lijian Zhang, Christine Silberhorn, and Ian A. Walmsley. Secure quantum
key distribution using continuous variables of single photons. Physical Review
Letters, 100(11):110504, 2008.
[107] Rodney Loudon. The Quantum Theory of Light. Oxford University Press,
2004.
[108] E. Wigner. On the quantum correction for thermodynamic equilibrium. Phys.
Rev., 40(5):749–759, Jun 1932.
[109] T. Hirano, K. Kotani, T. Ishibashi, S. Okude, and T. Kuwamoto. 3 dB squeezing by single-pass parametric amplification in a periodically poled KTiOPO4
crystal. Optics Letters, 30(13):1722–1724, 2005.
[110] RJ Senior, GN Milford, J. Janousek, AE Dunlop, K. Wagner, H.A. Bachor,
TC Ralph, EH Huntington, and CC Harb. Observation of a comb of optical
squeezing over many gigahertz of bandwidth. Optics Express, 15(9):5310–5317,
2007.
[111] Henning Vahlbruch, Moritz Mehmet, Simon Chelkowski, Boris Hage, Alexander Franzen, Nico Lastzka, Stefan Gossler, Karsten Danzmann, and Roman
Schnabel. Observation of squeezed light with 10-db quantum-noise reduction.
Physical Review Letters, 100(3):033602, 2008.
[112] Y. Eto, T. Tajima, Y. Zhang, and T. Hirano. Observation of quadrature
squeezing in a?ˆ(2) nonlinear waveguide using a temporally shaped local oscillator pulse. Optics Express, 16(14):10650–16657, 2008.
[113] Alex Abramovici, William E. Althouse, Ronald W. P. Drever, Yekta Gursel,
Seiji Kawamura, Frederick J. Raab, David Shoemaker, Lisa Sievers, Robert E.
Spero, Kip S. Thorne, Rochus E. Vogt, Rainer Weiss, Stanley E. Whitcomb,
and Michael E. Zucker. Ligo: The laser interferometer gravitational-wave
observatory. Science, 256(5055):325–333, 1992.
[114] Carlton M. Caves. Quantum-mechanical noise in an interferometer. Phys.
Rev. D, 23(8):1693–1708, Apr 1981.
BIBLIOGRAPHY
497
[115] H. Yuen and J. Shapiro. Optical communication with two-photon coherent
states–part i: Quantum-state propagation and quantum-noise. Information
Theory, IEEE Transactions on, 24(6):657–668, Nov 1978.
[116] T. Fernholz, H. Krauter, K. Jensen, J. F. Sherson, A. S. Sorensen, and E. S.
Polzik. Spin squeezing of atomic ensembles via nuclear-electronic spin entanglement. Physical Review Letters, 101(7):073601, 2008.
[117] Vivi Petersen, Lars Bojer Madsen, and Klaus Molmer. Magnetometry with
entangled atomic samples. Physical Review A (Atomic, Molecular, and Optical
Physics), 71(1):012312, 2005.
[118] D. Leibfried, M. D. Barrett, T. Schaetz, J. Britton, J. Chiaverini, W. M.
Itano, J. D. Jost, C. Langer, and D. J. Wineland. Toward heisenberg-limited
spectroscopy with multiparticle entangled states. Science, 304(5676):1476–
1478, 2004.
[119] C. F. Roos, M. Chwalla, K. Kim, M. Riebe, and R. Blatt. /‘designer atoms/’
for quantum metrology. Nature, 443(7109):316–319, 2006.
[120] P. J. Windpassinger, D. Oblak, P. G. Petrov, M. Kubasik, M. Saffman,
C. L. Garrido Alzar, J. Appel, J. H. Muller, N. Kjaergaard, and E. S. Polzik.
Nondestructive probing of rabi oscillations on the cesium clock transition near
the standard quantum limit. Physical Review Letters, 100(10):103601, 2008.
[121] Masahiro Kitagawa and Masahito Ueda. Nonlinear-interferometric generation
of number-phase-correlated fermion states. Phys. Rev. Lett., 67(14):1852–1854,
Sep 1991.
[122] Brian Julsgaard, Jacob Sherson, J. Ignacio Cirac, Jaromir Fiurasek, and Eugene S. Polzik. Experimental demonstration of quantum memory for light.
Nature, 432(7016):482–486, 2004.
[123] K. Hammerer, A.S. Sorensen, and E.S. Polzik. Quantum interface between
light and atomic ensembles. Not yet published, 2008.
[124] Jun Liu, Songcan Chen, and Xiaoyang Tan. Fractional order singular value
decomposition representation for face recognition. Pattern Recogn., 41(1):378–
395, 2008.
[125] T. N. Palmer, R. Buizza, F. Molteni, Y.-Q. Chen, and S. Corti. Singular vectors and the predictability of weather and climate. Philosophical Transactions:
Physical Sciences and Engineering, 348(1688):459–475, 1994.
[126] Guifré Vidal. Entanglement of pure states for a single copy. Phys. Rev. Lett.,
83(5):1046–1049, Aug 1999.
[127] E. Beltrami. Sulle funzioni bilineari. Giornale di Matematiche ad Uso degli
Studenti Delle Universita, 11:98–106, 1873.
BIBLIOGRAPHY
[128] C. Jordan. Sur la réduction des formes bilinéaires.
l’Academie des Sciences, Paris, 78:614–617, 1874.
498
Comptes Rendus de
[129] JJ Sylvester. A new proof that a general quadric may be reduced to its canonical form (that is, a linear function of squares) by means of a real orthogonal
substitution. Messenger of Mathematics, 19:1–5, 1889.
[130] E. Schmidt. Zur Theorie Der Linearen Und Nichtlinearen Integralgleichungen.
I Teil: Entwicklung WillkuLicher Funktion Nach Systemen Vorgeschriebener.
Mathenatische Annalen, 63:433–476, 1907.
[131] H. Bateman. A formula for the solving function of a certain integral equation of
the second kind. Transactions of the Cambridge Philosophical Society, 20:179–
187, 1908.
[132] C. Eckart and G. Young. The approximation of one matrix by another one of
lower rank. Psychometrika, 1:211–218, 1936.
[133] Alexey V. Gorshkov, Axel Andre, Mikhail D. Lukin, and Anders S. Sorensen.
Photon storage in lambda-type optically dense atomic media. ii. free-space
model. Physical Review A, 76:033805, 2007.
[134] J.M. Supplee. Optical Bloch equations: Informal motivation without the
Schrödinger equation. American Journal of Physics, 68:180, 2000.
[135] T.H. Boyer. Electrostatic potential energy leading to a gravitational mass
change for a system of two point charges. American Journal of Physics,
47(2):129–31, 1979.
[136] B.R. Holstein and A.R. Swift. Elementary derivation of the radiation field
from an accelerated charge. American Journal of Physics, 49:346, 1981.
[137] R.P. Feynman, R.B. Leighton, M. Sands, and EM Hafner. The Feynman
Lectures on Physics; Vol. I. American Journal of Physics, 33:750, 1965.
[138] A. Kalachev. Quantum storage on subradiant states in an extended atomic
ensemble. Physical Review A, 76(4):43812, 2007.
[139] M. G. Raymer and J. Mostowski. Stimulated Raman scattering: Unified
treatment of spontaneous initiation and spatial propagation. Phys. Rev. A,
24(4):1980–1993, 1981.
[140] Alexey V. Gorshkov, Axel Andre, Mikhail D. Lukin, and Anders S. Sorensen.
Photon storage in lambda-type optically dense atomic media. ii. free-space
model.
Physical Review A (Atomic, Molecular, and Optical Physics),
76(3):033805, 2007.
[141] R.W. Boyd. Nonlinear Optics. Academic Press, 2003.
BIBLIOGRAPHY
499
[142] N. Rosen and C. Zener. Double Stern-Gerlach Experiment and Related Collision Phenomena. Physical Review, 40(4):502–507, 1932.
[143] P. Pechukas and J.C. Light. On the Exponential Form of Time-Displacement
Operators in Quantum Mechanics. The Journal of Chemical Physics, 44:3897,
2004.
[144] RT Robiscoe. Extension of the Rosen-Zener solution to the two-level problem.
Physical Review A, 17(1):247–260, 1978.
[145] S. Klarsfeld and JA Oteo. Magnus approximation in the adiabatic picture.
Physical Review A, 45(5):3329–3332, 1992.
[146] WR Salzman. An alternative to the magnus expansion in time-dependent
perturbation theory. The Journal of Chemical Physics, 82:822, 1985.
[147] Samuel L. Braunstein. Squeezing as an irreducible resource. Physical Review
A (Atomic, Molecular, and Optical Physics), 71(5):055801, 2005.
[148] W. Wasilewski and MG Raymer. Pairwise entanglement and readout of
atomic-ensemble and optical wave-packet modes in traveling-wave Raman interactions. Physical Review A, 73(6):63816, 2006.
[149] W. Wasilewski, AI Lvovsky, K. Banaszek, and C. Radzewicz. Pulsed
squeezed light: Simultaneous squeezing of multiple modes. Physical Review
A, 73(6):63819, 2006.
[150] Irina Novikova, Alexey V. Gorshkov, David F. Phillips, Anders S. Sorensen,
Mikhail D. Lukin, and Ronald L. Walsworth. Optimal control of light pulse
storage and retrieval. Physical Review Letters, 98(24):243602, 2007.
[151] Nathaniel B. Phillips, Alexey V. Gorshkov, and Irina Novikova. Optimal light
storage in atomic vapor, 2008.
[152] K. Surmacz, J. Nunn, K. Reim, K. C. Lee, V. O. Lorenz, B. Sussman,
I. A. Walmsley, and D. Jaksch. Efficient spatially resolved multimode quantum memory. Physical Review A (Atomic, Molecular, and Optical Physics),
78(3):033806, 2008.
[153] C. Simon, H. de Riedmatten, M. Afzelius, N. Sangouard, H. Zbinden, and
N. Gisin. Quantum repeaters with photon pair sources and multimode memories. Phys. Rev. Lett., 98:190503, 2007.
[154] O. A. Collins, S. D. Jenkins, A. Kuzmich, and T. A. B. Kennedy. Multiplexed
memory-insensitive quantum repeaters. Physical Review Letters, 98(6):060502,
2007.
BIBLIOGRAPHY
500
[155] J. Nunn, K. Reim, K. C. Lee, V. O. Lorenz, B. J. Sussman, I. A. Walmsley,
and D. Jaksch. Multimode memories in atomic ensembles. Physical Review
Letters, 101(26):260502, 2008.
[156] J. H. Eberly. Schmidt Analysis of Pure-State Entanglement. ArXiv Quantum
Physics e-prints, August 2005.
[157] R. Grobe, K. Rzazewski, and JH Eberly. Measure of electron-electron correlation in atomic physics. Journal of Physics B: Atomic, Molecular, and Optical
Physics, 27(16):L503–L508, 1994.
[158] M. A. Nielsen and I. L. Chuang. Quantum Computation and Quantum Information. Cambridge University Press, 2004.
[159] G.N. Watson. A Treatise on the Theory of Bessel Functions. Cambridge
University Press, 1995.
[160] Nicolas Sangouard, Christoph Simon, Mikael Afzelius, and Nicolas Gisin.
Analysis of a quantum memory for photons based on controlled reversible
inhomogeneous broadening. Physical Review A, 75:032327, 2007.
[161] JJ Longdell, G. Hétet, MJ Sellars, and PK Lam. Analytic treatment of
CRIB Quantum Memories for Light using Two-level Atoms. Arxiv preprint
arXiv:0807.0067, 2008.
[162] G. Hetet, J. J. Longdell, M. J. Sellars, P. K. Lam, and B. C. Buchler. Multimodal properties and dynamics of gradient echo quantum memory. Physical
Review Letters, 101(20):203601, 2008.
[163] Alexey V. Gorshkov, Axel Andre, Mikhail D. Lukin, and Anders S. Sorensen.
Photon storage in lambda-type optically dense atomic media. iii. effects of
inhomogeneous broadening. Physical Review A, 76:033806, 2007.
[164] G. Hétet, M. Hosseini, B. M. Sparkes, D. Oblak, P. K. Lam, and B. C. Buchler.
Photon echoes generated by reversing magnetic field gradients in a rubidium
vapor. Opt. Lett., 33(20):2323–2325, 2008.
[165] S.A. Moiseev and W. Tittel. Raman-Echo Quantum Memory. 2008.
[166] Alexey V. Gorshkov, Tommaso Calarco, Mikhail D. Lukin, and Anders S.
Sorensen. Photon storage in lambda-type optically dense atomic media. iv.
optimal control using gradient ascent. Physical Review A (Atomic, Molecular,
and Optical Physics), 77(4):043806, 2008.
[167] C. Kittel and P. McEuen. Introduction to solid state physics. Wiley New York,
1986.
[168] PG Klemens. Anharmonic Decay of Optical Phonons.
148(2):845–848, 1966.
Physical Review,
BIBLIOGRAPHY
501
[169] A. Debernardi, S. Baroni, and E. Molinari. Anharmonic Phonon Lifetimes
in Semiconductors from Density-Functional Perturbation Theory. Physical
Review Letters, 75(9):1819–1822, 1995.
[170] M.S. Liu, L.A. Bursill, S. Prawer, and R. Beserman. Temperature dependence of the first-order Raman phonon line of diamond. Physical Review B,
61(5):3391–3395, 2000.
[171] FC Waldermann, B.J. Sussman, J. Nunn, VO Lorenz, KC Lee, K. Surmacz,
KH Lee, D. Jaksch, IA Walmsley, P. Spizziri, et al. Measuring phonon dephasing with ultrafast pulses using Raman spectral interference. Physical Review
B, 78(15), 2008.
[172] A. Cantarero, C. Trallero-Giner, and M. Cardona. Excitons in one-phonon
resonant raman scattering: Deformation-potential interaction. Phys. Rev. B,
39(12):8388–8397, Apr 1989.
[173] W. Pötz and P. Vogl. Theory of optical-phonon deformation potentials in
tetrahedral semiconductors. Physical Review B, 24(4):2025–2037, 1981.
[174] GD Whitfield. Theory of Electron-Phonon Interactions. Physical Review Letters, 2(5):204–205, 1959.
[175] H. Frohlich. Interaction of electrons with lattice vibrations. Proceedings of
the Royal Society of London. Series A, Mathematical and Physical Sciences
(1934-1990), 215(1122):291–298, 12 1952.
[176] Kun Huang. On the interaction between the radiation field and ionic crystals. Proceedings of the Royal Society of London. Series A, Mathematical and
Physical Sciences (1934-1990), 208(1094):352–365, 09 1951.
[177] K. Rzazewski and RW Boyd. Equivalence of interaction Hamiltonians in the
electric dipole approximation. Journal of Modern Optics, 51(8):1137–1147,
2004.
[178] M. O. Scully and M. S. Zubairy. Quantum Optics. Cambridge University
Press, 1997.
[179] R. P. Feynman. Forces in molecules. Phys. Rev., 56(4):340–343, Aug 1939.
[180] W. Hayes, R. Loudon, and J.F. Scott. Scattering of Light by Crystals. American Journal of Physics, 47:571, 1979.
[181] R. Loudon. Theory of the First-Order Raman Effect in Crystals. Proceedings
of the Royal Society of London. Series A, Mathematical and Physical Sciences
(1934-1990), 275(1361):218–232, 1963.
[182] R. Loudon. Theory of Stimulated Raman Scattering from Lattice Vibrations.
Proceedings of the Physical Society, 82:393–400, 1963.
BIBLIOGRAPHY
502
[183] A.D. Greentree, P. Olivero, M. Draganski, E. Trajkov, J.R. Rabeau, P. Reichart, B.C. Gibson, S. Rubanov, S.T. Huntington, D.N. Jamieson, et al. Critical components for diamond-based quantum coherent devices. JOURNAL OF
PHYSICS CONDENSED MATTER, 18(21):825, 2006.
[184] Daniel A. Steck. Cesium D line data. 2008.
[185] J.L. Picque. Hyperfine optical pumping of cesium with a CW GaAs laser.
Quantum Electronics, IEEE Journal of, 10(12):892–897, 1974.
[186] Zhaoming Zhu, Daniel J. Gauthier, and Robert W. Boyd. Stored Light in an
Optical Fiber via Stimulated Brillouin Scattering. Science, 318(5857):1748–
1750, 2007.
[187] Ashok K. Mohapatra, Mark G. Bason, Bjorn Butscher, Kevin J. Weatherill,
and Charles S. Adams. A giant electro-optic effect using polarizable dark
states. Nat Phys, 4(11):890–894, 11 2008.
[188] D. E. Chang, V. Gritsev, G. Morigi, V. Vuletic, M. D. Lukin, and E. A.
Demler. Crystallization of strongly interacting photons in a nonlinear optical
fibre. Nat Phys, 4(11):884–889, 11 2008.
[189] G. Golub and W. Kahan. Calculating the singular values and pseudo-inverse
of a matrix. Milestones in Matrix Computation: The Selected Works of Gene
H. Golub with Commentaries, 2007.
[190] GW Stewart. On Scaled Projections and Pseudo-inverses. University of Maryland, 1988.
[191] A.E. Albert. Regression and the Moore-Penrose pseudoinverse. Academic
Press, New York, 1972.
[192] TNE Greville. Some Applications of the Pseudoinverse of a Matrix. SIAM
Review, 2:15, 1960.
[193] M. Born and P. Jordan. On quantum mechanics. Z. Phys, 34(11-12):858–888,
1925.
[194] P.A.M. Dirac. The Principles of Quantum Mechanics. Oxford University
Press, 1958.
[195] M. Born. Quantenmechanik der Stoßvorgänge.
Hadrons and Nuclei, 38(11):803–827, 1926.
Zeitschrift für Physik A
[196] D.J. Amit and V. Martin-Mayor. Field theory, the renormalization group, and
critical phenomena. World Scientific Singapore, 1984.
[197] RG Woolley. Gauge invariant wave mechanics and the Power-Zienau-Woolley
transformation. Journ. Phys. A, 13:2795, 1980.
BIBLIOGRAPHY
503
[198] M. Babiker and R. Loudon. Derivation of the power-zienau-woolley hamiltonian in quantum electrodynamics by gauge transformation. Proceedings of
the Royal Society of London. Series A, Mathematical and Physical Sciences,
385(1789):439–460, 1983.
[199] S.C. Reddy and L.N. Trefethen. Stability of the method of lines. Numerische
Mathematik, 62(1):235–267, 1992.
[200] L.N. Trefethen. Spectral Methods in MATLAB. Society for Industrial Mathematics, Philadelphia, 2000.
[201] D. Gottlieb and S.A. Orszag. Numerical Analysis of Spectral Methods: Theory
and Applications. Society for Industrial Mathematics, 1977.
[202] J.P. Boyd. Chebyshev and Fourier Spectral Methods. Courier Dover Publications, 2001.
[203] J.W. Cooley and J.W. Tukey. An algorithm for the machine computation of
complex Fourier series. Mathematics of Computation, 19(90):297–301, 1965.
[204] RG Gareev, FA Arslanov, and GG Telyashev. THEORETICAL PRINCIPLES. Chemistry and Technology of Fuels and Oils, 37(4), 2001.