* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
Download Quantum Memory in Atomic Ensembles - Oxford Physics
Probability amplitude wikipedia , lookup
Renormalization wikipedia , lookup
Quantum field theory wikipedia , lookup
Wave function wikipedia , lookup
Matter wave wikipedia , lookup
Quantum machine learning wikipedia , lookup
Double-slit experiment wikipedia , lookup
Quantum entanglement wikipedia , lookup
Interpretations of quantum mechanics wikipedia , lookup
X-ray fluorescence wikipedia , lookup
Quantum teleportation wikipedia , lookup
EPR paradox wikipedia , lookup
Coherent states wikipedia , lookup
Path integral formulation wikipedia , lookup
Quantum group wikipedia , lookup
Atomic theory wikipedia , lookup
Renormalization group wikipedia , lookup
Bohr–Einstein debates wikipedia , lookup
Hidden variable theory wikipedia , lookup
Hydrogen atom wikipedia , lookup
Density matrix wikipedia , lookup
Quantum electrodynamics wikipedia , lookup
Relativistic quantum mechanics wikipedia , lookup
Tight binding wikipedia , lookup
Scalar field theory wikipedia , lookup
Quantum state wikipedia , lookup
Wave–particle duality wikipedia , lookup
History of quantum field theory wikipedia , lookup
Delayed choice quantum eraser wikipedia , lookup
Canonical quantization wikipedia , lookup
Quantum key distribution wikipedia , lookup
Symmetry in quantum mechanics wikipedia , lookup
Ultrafast laser spectroscopy wikipedia , lookup
Theoretical and experimental justification for the Schrödinger equation wikipedia , lookup
Quantum Memory in Atomic Ensembles Joshua Nunn St. John’s College, Oxford Submitted for the degree of Doctor of Philosophy Hilary Term 2008 Supervised by Prof. Ian A. Walmsley Clarendon Laboratory University of Oxford United Kingdom For glory. I mean Gloria. Abstract This thesis is a predominantly theoretical study of light storage in atomic ensembles. The efficiency of ensemble quantum memories is analyzed and optimized using the techniques of linear algebra. Analytic expressions describing the memory interaction in both EIT and Raman regimes are derived, and numerical methods provide solutions where the analytic expressions break down. A three dimensional numerical model of off-axis retrieval is presented. Multimode storage is considered, and the EIT, Raman, CRIB and AFC protocols are analyzed. It is shown that inhomogeneous broadening improves the multimode capacity of a memory. Raman storage in a diamond crystal is shown to be feasible. Finally, experimental progress toward implementing a Raman quantum memory in cesium vapour is described. Acknowledgements I have been very lucky to work with a fantastic set of people. I owe a debt of gratitude to my supervisor Ian Walmsley, whose implacable good humour always transmutes frustration into comedy, and who has overseen my experimental failures with only mild panic. Much of the theory was conceived in the course of meetings with Karl Surmacz, who has been Dr. Surmacz for a year already. I fried my first laser in the lab with Felix Waldermann, and his patience and subtle sense of humour are missed — he is also qualified and long-gone! My current postdocs Virginia Lorenz and Ben Sussman have been a continual source of exciting discussion, and North American optimism. We’re gonna make it work guys! And no misunderstanding can survive a keen frisking at the hands of Klaus Reim, who is currently completing his D.Phil — and mine — within our group. KC Lee has made the transition from theorist to experimentalist, without a blip in his coffee intake, and he remains an inspiration. The help and encouragement of Dieter Jaksch, and more recently Christoph Simon, are greatly appreciated. Thanks must also be due to my office neighbours: Pete Mosley, who introduced me to the concept of a progress chart, and with it a reification of inadequacy, and Dave Crosby, who go-karts better than he sings. The rest of the ultrafast group divides cleanly into those who drink tea and those who do not. A great big thank you to the tea drinkers: you understand that a tea break is more than a bathroom break with a drink. It lies at the heart of what it means to prevaricate. Adam Wyatt knows this. He is a tea soldier. As for the tea-less philistines (you know who you are), I have nothing to say to you. (dramatic pause). Nothing. Thanks to my Oxford massif, Andy Scott and Tom Rowlands-Rees, who know a good lunch when they see one. And in that vein, thanks to Matthijs Branderhorst, who along with Ben, introduced me to the burrito. There is a growing ultrafast diaspora — good people in far-off places — and of these I should like to big-up Daryl Achilles and Jeff Lundeen, who are awethome even when in errr-rr. I could not have made it this far without the support of Sonia: there can be few less attractive prospects than the unshaven maniac that is a D.Phil student writing up. Thank you for keeping me sane! And lastly my parents. Thanks mum and thanks dad! Next up: driving license. Contents 1 Introduction 1 1.1 Classical Computation . . . . . . . . . . . . . . . . . . . . . . . . . . 3 1.2 Quantum Computation . . . . . . . . . . . . . . . . . . . . . . . . . 4 1.2.1 Qubits . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4 1.2.2 Noise . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6 1.2.3 No cloning . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7 1.2.4 Universality . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9 1.3 Quantum Memory . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10 1.4 Linear Optics Quantum Computing . . . . . . . . . . . . . . . . . . 11 1.5 Quantum Communication . . . . . . . . . . . . . . . . . . . . . . . . 13 1.6 Quantum Repeaters . . . . . . . . . . . . . . . . . . . . . . . . . . . 19 1.6.1 The Ekert protocol . . . . . . . . . . . . . . . . . . . . . . . . 21 1.6.2 Entanglement Swapping . . . . . . . . . . . . . . . . . . . . . 22 1.6.3 Entanglement Purification . . . . . . . . . . . . . . . . . . . . 24 1.6.4 The DLCZ protocol and number state entanglement . . . . . 24 CONTENTS 1.7 v Modified DLCZ with Quantum Memories . . . . . . . . . . . . . . . 2 Quantum Memory: Approaches 31 35 2.1 Cavity QED . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38 2.2 Free space coupling . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39 2.3 Ensembles . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40 2.3.1 EIT . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41 2.3.2 Raman . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46 2.3.3 CRIB . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 50 2.3.4 AFC . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 55 Continuous Variables . . . . . . . . . . . . . . . . . . . . . . . . . . . 58 2.4 3 Optimization 3.1 67 The Singular Value Decomposition . . . . . . . . . . . . . . . . . . . 70 3.1.1 Unitary invariance . . . . . . . . . . . . . . . . . . . . . . . . 75 3.1.2 Connection with Eigenvalues . . . . . . . . . . . . . . . . . . 75 3.1.3 Hermitian SVD . . . . . . . . . . . . . . . . . . . . . . . . . . 76 3.1.4 Persymmetry . . . . . . . . . . . . . . . . . . . . . . . . . . . 76 3.2 Norm maximization . . . . . . . . . . . . . . . . . . . . . . . . . . . 78 3.3 Continuous maps . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 79 3.3.1 Normally and Anti-normally ordered kernels. . . . . . . . . . 81 3.3.2 Memory Optimization. . . . . . . . . . . . . . . . . . . . . . . 81 3.3.3 Unitary invariance . . . . . . . . . . . . . . . . . . . . . . . . 82 CONTENTS vi 3.4 Optimizing storage followed by retrieval . . . . . . . . . . . . . . . . 85 3.5 A Simple Example . . . . . . . . . . . . . . . . . . . . . . . . . . . . 86 4 Equations of motion 92 4.1 Interaction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 93 4.2 Electric Field . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 94 4.3 Dipole Operator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 97 4.3.1 98 Parity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.4 Hamiltonian . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 100 4.5 Linear approximation (1) . . . . . . . . . . . . . . . . . . . . . . . . 102 4.6 Rotating Wave Approximation . . . . . . . . . . . . . . . . . . . . . 103 4.7 Unwanted Coupling 4.8 Linear Approximation (2) . . . . . . . . . . . . . . . . . . . . . . . . 106 4.9 Propagation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 108 . . . . . . . . . . . . . . . . . . . . . . . . . . . 105 4.10 Paraxial and SVE approximations . . . . . . . . . . . . . . . . . . . 110 4.11 Continuum Approximation . . . . . . . . . . . . . . . . . . . . . . . 112 4.12 Spontaneous Emission and Decoherence . . . . . . . . . . . . . . . . 116 5 Raman & EIT Storage 120 5.1 One Dimensional Approximation . . . . . . . . . . . . . . . . . . . . 120 5.2 Solution in k-space . . . . . . . . . . . . . . . . . . . . . . . . . . . . 123 5.2.1 Boundary Conditions . . . . . . . . . . . . . . . . . . . . . . 123 5.2.2 Transformed Equations . . . . . . . . . . . . . . . . . . . . . 124 CONTENTS 5.3 5.4 5.2.3 Optimal efficiency . . . . . . . . . . . . . . . . . . . . . . . . 125 5.2.4 Solution in Wavelength Space . . . . . . . . . . . . . . . . . . 129 5.2.5 Including the Control . . . . . . . . . . . . . . . . . . . . . . 134 5.2.6 An Exact Solution: The Rosen-Zener case . . . . . . . . . . . 137 5.2.7 Adiabatic Limit . . . . . . . . . . . . . . . . . . . . . . . . . . 145 5.2.8 Reaching the optimal efficiency . . . . . . . . . . . . . . . . . 152 5.2.9 Adiabatic Approximation . . . . . . . . . . . . . . . . . . . . 155 Raman Storage . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 159 5.3.1 Validity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 163 5.3.2 Matter Biased Limit . . . . . . . . . . . . . . . . . . . . . . . 167 5.3.3 Transmitted Modes. . . . . . . . . . . . . . . . . . . . . . . . 168 Numerical Solution . . . . . . . . . . . . . . . . . . . . . . . . . . . . 182 5.4.1 5.5 Dispersion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 187 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 188 6 Retrieval 6.1 vii 190 Collinear Retrieval . . . . . . . . . . . . . . . . . . . . . . . . . . . . 190 6.1.1 Forward Retrieval . . . . . . . . . . . . . . . . . . . . . . . . 191 6.2 Backward Retrieval . . . . . . . . . . . . . . . . . . . . . . . . . . . . 197 6.3 Phasematched Retrieval . . . . . . . . . . . . . . . . . . . . . . . . . 207 6.4 6.3.1 Dispersion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 210 6.3.2 Scheme . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 211 Full Propagation Model . . . . . . . . . . . . . . . . . . . . . . . . . 213 CONTENTS viii 6.4.1 Diffraction . . . . . . . . . . . . . . . . . . . . . . . . . . . . 215 6.4.2 Control Field . . . . . . . . . . . . . . . . . . . . . . . . . . . 216 6.4.3 Boundary Conditions 6.4.4 Read out . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 220 6.4.5 Efficiency . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 222 . . . . . . . . . . . . . . . . . . . . . . 218 6.5 Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 223 6.6 Angular Multiplexing . . . . . . . . . . . . . . . . . . . . . . . . . . . 228 6.6.1 Optimizing the carrier frequencies . . . . . . . . . . . . . . . 229 6.6.2 Capacity 7 Multimode Storage 7.1 7.2 234 Multimode Capacity from the SVD . . . . . . . . . . . . . . . . . . . 235 7.1.1 Schmidt Number . . . . . . . . . . . . . . . . . . . . . . . . . 237 7.1.2 Threshold multimode capacity . . . . . . . . . . . . . . . . . 239 Multimode scaling for EIT and Raman memories . . . . . . . . . . . 241 7.2.1 7.3 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 231 A spectral perspective . . . . . . . . . . . . . . . . . . . . . . 242 CRIB . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 245 7.3.1 lCRIB . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 246 7.3.2 Simplified Kernel . . . . . . . . . . . . . . . . . . . . . . . . . 249 7.3.3 tCRIB . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 254 7.4 Broadened Raman . . . . . . . . . . . . . . . . . . . . . . . . . . . . 262 7.5 AFC . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 268 CONTENTS ix 8 Optimizing the Control 276 8.1 Adiabatic shaping . . . . . . . . . . . . . . . . . . . . . . . . . . . . 277 8.2 Non-adiabatic shaping . . . . . . . . . . . . . . . . . . . . . . . . . . 279 9 Diamond 286 9.1 Diamond Scheme . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 286 9.2 Quantization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 288 9.3 Acoustic and Optical Phonons 9.4 9.5 9.6 . . . . . . . . . . . . . . . . . . . . . 290 9.3.1 Decay . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 291 9.3.2 Energy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 293 Raman interaction . . . . . . . . . . . . . . . . . . . . . . . . . . . . 294 9.4.1 Excitons . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 294 9.4.2 Deformation Potential . . . . . . . . . . . . . . . . . . . . . . 296 Propagation in Diamond . . . . . . . . . . . . . . . . . . . . . . . . . 299 9.5.1 Hamiltonian . . . . . . . . . . . . . . . . . . . . . . . . . . . . 300 9.5.2 Electron-radiation interaction . . . . . . . . . . . . . . . . . . 301 9.5.3 Electron-lattice interaction . . . . . . . . . . . . . . . . . . . 307 9.5.4 Crystal energy . . . . . . . . . . . . . . . . . . . . . . . . . . 309 Heisenberg equations . . . . . . . . . . . . . . . . . . . . . . . . . . . 310 9.6.1 Adiabatic perturbative solution . . . . . . . . . . . . . . . . . 311 9.7 Signal propagation . . . . . . . . . . . . . . . . . . . . . . . . . . . . 314 9.8 Coupling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 318 9.9 Selection Rules . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 321 CONTENTS x 9.10 Noise . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 322 10 Experiments 323 10.1 Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 323 10.2 Thallium 10.3 Cesium . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 325 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 326 10.4 Cell . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 328 10.4.1 Temperature control . . . . . . . . . . . . . . . . . . . . . . . 329 10.4.2 Magnetic shielding . . . . . . . . . . . . . . . . . . . . . . . . 329 10.5 Buffer gas . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 331 10.6 Control pulse . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 333 10.6.1 Pulse duration . . . . . . . . . . . . . . . . . . . . . . . . . . 333 10.6.2 Tuning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 336 10.6.3 Shaping . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 337 10.7 Pulse picker . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 337 10.8 Stokes scattering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 338 10.9 Coupling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 342 10.9.1 Optical depth . . . . . . . . . . . . . . . . . . . . . . . . . . . 342 10.9.2 Rabi frequency . . . . . . . . . . . . . . . . . . . . . . . . . . 344 10.9.3 Raman memory coupling . . . . . . . . . . . . . . . . . . . . 346 10.9.4 Focussing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 347 10.10Line shape . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 349 10.11Effective depth . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 351 CONTENTS xi 10.12Optical pumping . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 353 10.12.1 Pumping efficiency . . . . . . . . . . . . . . . . . . . . . . . . 355 10.13Filtering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 356 10.13.1 Polarization filtering . . . . . . . . . . . . . . . . . . . . . . . 358 10.13.2 Lyot filter . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 358 10.13.3 Etalons . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 359 10.13.4 Spectrometer . . . . . . . . . . . . . . . . . . . . . . . . . . . 362 10.13.5 Spatial filtering . . . . . . . . . . . . . . . . . . . . . . . . . . 362 10.14Signal pulse . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 364 10.15Planned experiment . . . . . . . . . . . . . . . . . . . . . . . . . . . 366 11 Summary 369 11.1 Future work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 372 A Linear algebra 374 A.1 Vectors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 375 A.1.1 Adjoint vectors . . . . . . . . . . . . . . . . . . . . . . . . . . 377 A.1.2 Inner product . . . . . . . . . . . . . . . . . . . . . . . . . . . 378 A.1.3 Norm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 379 A.1.4 Bases . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 380 A.2 Matrices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 381 A.2.1 Outer product . . . . . . . . . . . . . . . . . . . . . . . . . . 384 A.2.2 Tensor product . . . . . . . . . . . . . . . . . . . . . . . . . . 385 CONTENTS xii A.3 Eigenvalues . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 388 A.3.1 Commutators . . . . . . . . . . . . . . . . . . . . . . . . . . . 389 A.4 Types of matrices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 391 A.4.1 The identity matrix . . . . . . . . . . . . . . . . . . . . . . . 391 A.4.2 Inverse matrix . . . . . . . . . . . . . . . . . . . . . . . . . . 392 A.4.3 Hermitian matrices . . . . . . . . . . . . . . . . . . . . . . . . 393 A.4.4 Diagonal matrices . . . . . . . . . . . . . . . . . . . . . . . . 394 A.4.5 Unitary matrices . . . . . . . . . . . . . . . . . . . . . . . . . 396 B Quantum mechanics 399 B.1 Postulates . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 400 B.1.1 State vector . . . . . . . . . . . . . . . . . . . . . . . . . . . . 400 B.1.2 Observables . . . . . . . . . . . . . . . . . . . . . . . . . . . . 400 B.1.3 Measurements . . . . . . . . . . . . . . . . . . . . . . . . . . 400 B.1.4 Dynamics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 401 B.2 The Heisenberg Picture . . . . . . . . . . . . . . . . . . . . . . . . . 403 B.2.1 The Heisenberg interaction picture . . . . . . . . . . . . . . . 405 C Quantum optics 407 C.1 Modes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 407 C.2 Quantum states of light . . . . . . . . . . . . . . . . . . . . . . . . . 410 C.2.1 Fock states . . . . . . . . . . . . . . . . . . . . . . . . . . . . 410 C.2.2 Creation and Annihilation operators . . . . . . . . . . . . . . 411 CONTENTS xiii C.3 The electric field . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 414 C.4 Matter-Light Interaction . . . . . . . . . . . . . . . . . . . . . . . . . 415 C.4.1 The A.p Interaction . . . . . . . . . . . . . . . . . . . . . . . 415 C.4.2 The E.d Interaction . . . . . . . . . . . . . . . . . . . . . . . 418 C.5 Dissipation and Fluctuation . . . . . . . . . . . . . . . . . . . . . . . 422 D Sundry Analytical Techniques 427 D.1 Contour Integration . . . . . . . . . . . . . . . . . . . . . . . . . . . 427 D.1.1 Cauchy’s Integral Formula . . . . . . . . . . . . . . . . . . . . 429 D.1.2 Typical example . . . . . . . . . . . . . . . . . . . . . . . . . 430 D.2 The Dirac Delta Function . . . . . . . . . . . . . . . . . . . . . . . . 432 D.3 Fourier Transforms . . . . . . . . . . . . . . . . . . . . . . . . . . . . 434 D.3.1 Bilateral Transform . . . . . . . . . . . . . . . . . . . . . . . 434 D.3.2 Unitarity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 434 D.3.3 Inverse . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 435 D.3.4 Shift . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 435 D.3.5 Convolution . . . . . . . . . . . . . . . . . . . . . . . . . . . . 436 D.3.6 Transform of a Derivative . . . . . . . . . . . . . . . . . . . . 436 D.4 Unilateral Transform . . . . . . . . . . . . . . . . . . . . . . . . . . . 437 D.4.1 Shift . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 438 D.4.2 Convolution . . . . . . . . . . . . . . . . . . . . . . . . . . . . 438 D.4.3 Transform of a Derivative . . . . . . . . . . . . . . . . . . . . 439 D.4.4 Laplace Transform . . . . . . . . . . . . . . . . . . . . . . . . 440 CONTENTS xiv D.5 Bessel Functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 440 D.5.1 Orthogonality . . . . . . . . . . . . . . . . . . . . . . . . . . . 441 D.5.2 Memory Propagator . . . . . . . . . . . . . . . . . . . . . . . 442 D.5.3 Optimal Eigenvalue Kernel . . . . . . . . . . . . . . . . . . . 445 E Numerics 447 E.1 Spectral Collocation . . . . . . . . . . . . . . . . . . . . . . . . . . . 449 E.1.1 Polynomial Differentiation Matrices . . . . . . . . . . . . . . 452 E.1.2 Chebyshev points . . . . . . . . . . . . . . . . . . . . . . . . . 453 E.2 Time-stepping . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 455 E.3 Boundary Conditions . . . . . . . . . . . . . . . . . . . . . . . . . . . 457 E.4 Constructing the Solutions . . . . . . . . . . . . . . . . . . . . . . . . 459 E.5 Numerical Construction of a Green’s Function . . . . . . . . . . . . . 462 E.6 Spectral Methods for Two Dimensions . . . . . . . . . . . . . . . . . 465 F Atomic Vapours F.1 Vapour pressure 471 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 471 F.2 Oscillator strengths . . . . . . . . . . . . . . . . . . . . . . . . . . . . 474 F.3 Line broadening . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 476 F.3.1 Doppler broadening . . . . . . . . . . . . . . . . . . . . . . . 476 F.3.2 Pressure broadening . . . . . . . . . . . . . . . . . . . . . . . 477 F.3.3 Power broadening . . . . . . . . . . . . . . . . . . . . . . . . 480 F.4 Raman polarization . . . . . . . . . . . . . . . . . . . . . . . . . . . 481 List of Figures 1.1 The state space of a qubit . . . . . . . . . . . . . . . . . . . . . . . . 5 1.2 The BB84 protocol . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17 1.3 Entanglement swapping . . . . . . . . . . . . . . . . . . . . . . . . . 23 1.4 Single-rail entanglement swapping . . . . . . . . . . . . . . . . . . . 26 1.5 QKD with single-rail entanglement . . . . . . . . . . . . . . . . . . . 27 1.6 Λ-level structure of atoms for DLCZ . . . . . . . . . . . . . . . . . . 29 1.7 Generation of number state entanglement in DLCZ . . . . . . . . . . 31 1.8 Modification to DLCZ with photon sources and quantum memories . 33 2.1 The simplest quantum memory . . . . . . . . . . . . . . . . . . . . . 36 2.2 Adding a dark state . . . . . . . . . . . . . . . . . . . . . . . . . . . 37 2.3 Cavity QED . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39 2.4 Confocal coupling in free space . . . . . . . . . . . . . . . . . . . . . 40 2.5 Atomic ensemble memory . . . . . . . . . . . . . . . . . . . . . . . . 41 2.6 EIT. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 42 2.7 Stopping light with EIT. . . . . . . . . . . . . . . . . . . . . . . . . . 45 LIST OF FIGURES xvi 2.8 Raman storage. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48 2.9 CRIB storage. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52 2.10 tCRIB vs. lCRIB. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54 2.11 AFC storage. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 56 2.12 Wigner distributions. . . . . . . . . . . . . . . . . . . . . . . . . . . . 61 2.13 Atomic quadratures. . . . . . . . . . . . . . . . . . . . . . . . . . . . 64 2.14 QND memory. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65 2.15 Level scheme for a QND memory. . . . . . . . . . . . . . . . . . . . . 66 3.1 Storage map. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 68 3.2 Linear transformation . . . . . . . . . . . . . . . . . . . . . . . . . . 72 3.3 Persymmetry. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 77 4.1 The Λ-system again. . . . . . . . . . . . . . . . . . . . . . . . . . . . 93 4.2 Time-ordering. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 103 4.3 Useful and nuisance couplings. . . . . . . . . . . . . . . . . . . . . . 105 5.1 Quantum memory boundary conditions. . . . . . . . . . . . . . . . . 124 5.2 Bessel zeros. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 132 5.3 Optimal storage efficiency. . . . . . . . . . . . . . . . . . . . . . . . . 133 5.4 The Rosen-Zener model. . . . . . . . . . . . . . . . . . . . . . . . . . 140 5.5 Raman efficiency. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 162 5.6 Raman storage as a beamsplitter. . . . . . . . . . . . . . . . . . . . . 181 5.7 Modified DLCZ protocol with partial storage. . . . . . . . . . . . . . 182 LIST OF FIGURES 5.8 xvii Comparison of predictions for the optimal input modes in the adiabatic limit. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 184 5.9 Comparison of predictions for the optimal input modes outside the adiabatic limit. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 185 5.10 Broadband Raman storage. . . . . . . . . . . . . . . . . . . . . . . . 187 5.11 Broadband EIT storage. . . . . . . . . . . . . . . . . . . . . . . . . . 188 6.1 Forward retrieval. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 197 6.2 Phasematching considerations for backward retrieval. . . . . . . . . . 200 6.3 Backward Retrieval. . . . . . . . . . . . . . . . . . . . . . . . . . . . 206 6.4 Non-collinear phasematching. . . . . . . . . . . . . . . . . . . . . . . 209 6.5 Efficient, phasematched memory for positive and negative phase mismatches. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 209 6.6 Focussed beams. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 216 6.7 Effectiveness of our phasematching scheme. . . . . . . . . . . . . . . 224 6.8 Comparing phasematched and collinear efficiencies. . . . . . . . . . . 225 6.9 Angular multiplexing. . . . . . . . . . . . . . . . . . . . . . . . . . . 230 6.10 Minimum momentum mismatch. . . . . . . . . . . . . . . . . . . . . 232 7.1 Bright overlapping modes are distinct. . . . . . . . . . . . . . . . . . 236 7.2 Visualizing the multimode capacity. 7.3 The appearance of a multimode Green’s function. . . . . . . . . . . . 239 7.4 Multimode scaling for Raman and EIT memories. . . . . . . . . . . . 244 . . . . . . . . . . . . . . . . . . 237 LIST OF FIGURES xviii 7.5 Scaling of Schmidt number with broadening. . . . . . . . . . . . . . 252 7.6 Comparison of the predictions of the kernels (7.23) and (7.18). . . . 253 7.7 Understanding the linear multimode scaling of lCRIB. . . . . . . . . 254 7.8 Multimode scaling for CRIB memories. . . . . . . . . . . . . . . . . . 261 7.9 The multimode scaling of a broadened Raman protocol. . . . . . . . 268 7.10 The multimode scaling of the AFC memory protocol. . . . . . . . . . 275 8.1 Adiabatic control shaping. . . . . . . . . . . . . . . . . . . . . . . . . 284 8.2 Non-adiabatic control shaping. . . . . . . . . . . . . . . . . . . . . . 285 9.1 The crystal structure of diamond. . . . . . . . . . . . . . . . . . . . . 287 9.2 Phonon aliasing. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 290 9.3 Phonon dispersion. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 292 9.4 Band structure. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 295 9.5 An exciton. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 296 9.6 The Raman interaction in diamond. . . . . . . . . . . . . . . . . . . 298 10.1 Observing Stokes scattering as a first step. . . . . . . . . . . . . . . . 325 10.2 Thallium atomic structure. . . . . . . . . . . . . . . . . . . . . . . . 326 10.3 Cesium atomic structure. . . . . . . . . . . . . . . . . . . . . . . . . 327 10.4 First order autocorrelation. . . . . . . . . . . . . . . . . . . . . . . . 334 10.5 Second order interferometric autocorrelation. . . . . . . . . . . . . . 336 10.6 Stokes scattering efficiency. . . . . . . . . . . . . . . . . . . . . . . . 342 10.7 Cesium optical depth. . . . . . . . . . . . . . . . . . . . . . . . . . . 345 LIST OF FIGURES xix 10.8 Cesium D2 absorption spectrum. . . . . . . . . . . . . . . . . . . . . 350 10.9 Absorption linewidth. . . . . . . . . . . . . . . . . . . . . . . . . . . 352 10.10Equal populations destroy quantum memory. . . . . . . . . . . . . . 354 10.11Optical pumping. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 355 10.12Verifying efficient optical pumping. . . . . . . . . . . . . . . . . . . . 357 10.13Lyot filter. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 360 10.14Stokes filtering. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 361 10.15Backward Stokes scattering. . . . . . . . . . . . . . . . . . . . . . . . 364 10.16A possible design for demonstration of a cesium quantum memory. . 368 A.1 A vector. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 376 A.2 The inner product of two vectors. . . . . . . . . . . . . . . . . . . . . 380 A.3 A matrix acting on a vector. . . . . . . . . . . . . . . . . . . . . . . . 383 A.4 Eigenvectors and eigenvalues. . . . . . . . . . . . . . . . . . . . . . . 389 A.5 Non-commuting operations. . . . . . . . . . . . . . . . . . . . . . . . 390 A.6 A unitary transformation. . . . . . . . . . . . . . . . . . . . . . . . . 396 C.1 Symmetrized photons. . . . . . . . . . . . . . . . . . . . . . . . . . . 413 D.1 Contour integrals. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 429 D.2 Upper closure. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 432 D.3 Integration limits. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 439 D.4 Lower closure. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 444 E.1 The method of lines. . . . . . . . . . . . . . . . . . . . . . . . . . . . 449 LIST OF FIGURES xx E.2 Periodic extension. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 451 E.3 Chebyshev Points. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 455 E.4 Example solutions. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 462 E.5 A numerically constructed Green’s function. . . . . . . . . . . . . . . 464 E.6 Spectral methods in two dimensions. . . . . . . . . . . . . . . . . . . 470 F.1 Vapour pressure. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 473 F.2 The Doppler shift. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 477 F.3 Collisions in a vapour. . . . . . . . . . . . . . . . . . . . . . . . . . . 478 F.4 Polarization selection rules. . . . . . . . . . . . . . . . . . . . . . . . 483 F.5 Alternative scattering pathways. . . . . . . . . . . . . . . . . . . . . 486 Chapter 1 Introduction The prospect of building a quantum computer, with speed and power far outstripping the best possible classical computers, has motivated an enormous and sustained research effort over the last two decades. In this thesis we explore a number of candidates for the ‘memory’ that would be required by such a device. As we will see, building a quantum memory is considerably harder than fabricating the RAM chips used by modern computers. For instance, it is not possible to copy quantum information, nor can quantum information be digitized. These facts make quantum storage particularly vulnerable to noise, and loss — problems for which solutions must be found before quantum computation can mature into a viable technology. The bulk of this thesis is concerned with optimization of the efficiency and storage capacity of a quantum memory. We focus on optical memories, in which a pulse of light is ‘stopped’ for a controllable period, before being re-released. The structure of the thesis is as follows. In this chapter we introduce the concepts 2 of quantum computing and quantum communication, and we discuss the context and motivation for the present work on quantum memories. In Chapter 2 we survey the various approaches to quantum memory, and we describe the principles behind the memory protocols analyzed later. Chapter 3 introduces the mathematical basis for our approach to analyzing and optimizing ensemble memories — the Green’s function and its singular value decomposition. In Chapter 4 we derive the equations of motion describing the quantum memory interaction in an ensemble of Λ-type atoms. In Chapter 5 we apply the techniques of Chapter 3 to this interaction. Several new results are derived, and connections are made with previous work. Chapter 6 is concerned with retrieval of the stored excitations from a Λ-ensemble. It is shown that both forward and backward retrieval are problematic. A numerical model is presented that confirms the efficacy of an off-axis geometry, which solves these problems. In Chapter 7 we move on to consider multimode storage. Our formalism provides a natural way to calculate the multimode capacity of a memory, and we study the multimode scaling of all the memory protocols introduced in Chapter 2. Chapter 8 describes how to optimize a Λ-type memory by shaping the ‘control pulse’. In Chapter 9 we study the Raman interaction in a diamond crystal: we show that a diamond Raman quantum memory is feasible. Finally in Chapter 10 we review our attempts to implement a Raman quantum memory in the laboratory, using cesium vapour. But let us begin at the beginning. 1.1 Classical Computation 1.1 3 Classical Computation Classical computers are conventional computers, like the one I am using to typeset this document. Their importance as enablers of technological progress, as well as their utility as a technology in their own right, attest to the fantastic potential of classical computation. They are typified by the use of bit1 strings — sequences of 1’s and 0’s — to encode information. Information is processed by application of binary logic to the bits. That is, Boolean operations such as OR, AND or not-AND (NAND). This last operation is a universal gate, because any logic operation can be constructed using only NAND gates. Such a gate can be implemented electronically using a pair of transistors, millions of which can be combined on a single silicon chip. The rest is history. Computers have progressed in leaps and bounds over the last fifty years. In 1965 computers were developing so fast that Gordon Moore, a founder of the industrial giant Intel, proposed a ‘law’, stipulating that the number of transistors comprising a processor would double every year [2] . Incredibly, this exponential improvement in computing power has persisted for over 40 years. But improvement by miniaturization cannot continue indefinitely. The reason for this is that the physics of electronic components undergoes a qualitative change at small scales: classical physics becomes quantum physics. In his 1983 lectures on computation [3] , Richard Feynman consid1 ‘Bit’ first appeared in Claude Shannon’s 1948 paper on the theory of communication as a contraction of ‘binary digit’ [1] ; the name is apposite, since one bit is the smallest ‘piece’ or ‘chunk’ of information there can be: one bit of information is one bit of information. Shannon attributes the term to John Tukey, a creator of the digital Fourier transform, who is also credited with coining the word ‘software’. 1.2 Quantum Computation 4 ers the fate of classical computation as shrinking dimensions bring quantum effects into play. Two years later David Deutsch published the first explicitly quantum algorithm [4] , demonstrating how quantum physics actually permits more powerful computation than classical physics allows. A quantum computer, capable of harnessing this greater power, must process quantum information, encoded not with ordinary bits, but with quantum bits. 1.2 1.2.1 Quantum Computation Qubits A quantum bit — a qubit 2 — is an object with two mutually exclusive states, 0 and 1, say. The only difference with a classical bit is that the object is described by quantum mechanics. Accordingly, we label the two states by the kets |0i and |1i (see Appendix B). These kets are to be thought of as vectors in a two dimensional space: the state space of the qubit (see Figure 1.1). The classical property of mutual exclusivity is manifested in the quantum formalism by requiring that |0i and |1i are perpendicular to one another in the state space. In general, the state of the qubit can be any vector, of length 1, in the state space. Since both the kets |0i and |1i have length 1, and since they point in perpendicular directions, an arbitrary qubit state |ψi can always be written as a linear combination of them, |ψi = α|0i + β|1i, 2 (1.1) The term ‘qubit’ first appears in a paper by Benjamin Schumacher in 1995 [5] ; he credits its invention to a conversation with William Wootters. 1.2 Quantum Computation 5 where α and β are two numbers which must satisfy the normalization condition |α|2 + |β|2 = 1. States like (1.1), which are a combination of the two mutually exclusive states |0i and |1i, are called superposition states, or just superpositions. What does it mean to say that a qubit is in a superposition between its two mutually exclusive states? Somehow it is both 0 and 1 at the same time. Physically, this is like saying that a switch is both ‘open’ and ‘closed’, or that a lamp is both ‘on’ and also ‘off’. Already, for the simplest possible system, without any real dynamics — no interactions, nothing happening — we see that the basic structure of quantum mechanics does not sit well with our intuition. Despite these interpretational diffi- stat e sp ace Figure 1.1 A visual representation of the state space of a qubit culties, superposition is central to the success of quantum mechanics. Atomic and molecular physics, nuclear and particle physics, optics and electronics all make use of superpositions to successfully explain processes and interactions. From the point of view of computation, the existence of states like (1.1) provides a clue to the greater capabilities of a quantum computer. Each qubit has two ‘parts’, the |0i part and the 1.2 Quantum Computation 6 |1i part; logical operations on qubits act on both parts together, and the output of a calculation also has these two parts. So there’s some sense in which a qubit plays the role of two classical bits, stuck ‘on top of eachother’. David Deutsch coined the term quantum parallelism for this property — he considers it to be the strongest evidence for the existence of parallel universes. The B-movie-esque connotations of this ‘many-worlds’ view make it generally unpopular among physicists, but the appeal of quantum computing remains, independently of how it is understood. 1.2.2 Noise A potential difficulty associated with quantum computing is also apparent from (1.1): the numbers α and β can be varied continuously (subject to the normalization constraint). So the number of possible states |ψi of a qubit is infinite! This also hints at their greater information carrying capacity, but it means that they must be carefully protected from the influence of noise. Classical bits have exactly two states; if noise introduces some distortions, it is usually possible to correct these simply by comparing the distorted bit to an ideal one. Only very large fluctuations can make a 0 look like a 1, so the discrete structure of classical bits makes them very robust. By contrast, a perturbed qubit state is also a valid qubit state. In this respect, the difference between bits and qubits can be likened to the difference between digital and analogue musical recordings: The quality of music reproduced by a CD does not degrade gradually with time, whereas old cassettes sound progressively worse as distortions creep into the waveform imprinted on the tape. In fact, 1.2 Quantum Computation 7 it is possible to correct errors by constructing codes involving bunches of qubits. The invention of these codes in 1996 by Calderbank, Shor and Steane [6,7] was a major milestone in demonstrating the practical viability of quantum computation. Nonetheless these error correcting schemes currently require that noise is suppressed below thresholds of a few percent, which makes techniques for isolating qubits from noise a technological sine qua non. 1.2.3 No cloning Another problematic aspect of quantum information is that it cannot be copied. The proof of this fact is known as the no-cloning theorem [8] . Suppose that we have a device which can copy a qubit. If we give it a qubit in state |ψi, and also a ‘blank’ qubit in some standard initial state |blanki, this machine spits out our original qubit, plus a clone, both in the state |ψi. In Dirac notation, using kets, the action of our qubit photocopier is written as U |blanki|ψi = |ψi|ψi. (1.2) Here U is the unitary transformation implemented by our machine. Unitary transformations are those which preserve the lengths of the kets upon which they act. Since all physical states have length 1, and any process must produce physical states from physical states, it follows that all processes are described by length-preserving — unitary — transformations (see §B.1.4 in Appendix B). Had we fed our machine 1.2 Quantum Computation 8 a different state, for example |φi, we would have U |blanki|φi = |φi|φi. (1.3) The length of a ket |ϕi is defined by taking the scalar product hϕ|ϕi of |ϕi with itself (see §A.1.2 in Appendix A). To prove the impossibility of cloning, we take the scalar product of the first relation (1.2) with the second, (1.3). hψ|hblank|U † U |blanki|φi = hψ|hψ||φi|φi. (1.4) The U acting on |blanki on the left hand side does not change its length, which is just equal to 1, so the result simplifies to hψ|φi = hψ|φi2 . (1.5) Clearly, this expression does not hold for arbitrary choices of |ψi and |φi, and therefore cloning an arbitrary qubit is impossible. In fact, (1.5) is only true when hψ|φi is either 1 or 0. The first case corresponds to |ψi = |φi, which says that it is possible to build a machine that can make copies of one particular, pre-determined state. The second case occurs only when |ψi and |φi are perpendicular, as is the case for |0i and |1i. This says that it is possible to clone mutually exclusive states. Indeed, this is precisely what classical computers are doing when they copy digitized information. An immediate consequence of the no-cloning theorem is that a quantum memory 1.2 Quantum Computation 9 must work in a qualitatively different way to a classical computer memory. To store quantum information, that information must be transferred to the memory, rather than simply copied to it. It is never possible to ‘save a back-up’, as we routinely do with classical computers. To build a quantum memory, we must find an interaction between information carrier and storage medium which ‘swaps’ their quantum states, so that the storage medium ends up with all the information, with nothing left behind. In this thesis we will examine various ways of accomplishing this optically, by considering collections of atoms which essentially ‘swallow’ a photon in a controlled way, completely transferring the quantum state of an optical field to that of the atoms. 1.2.4 Universality Any classical computation is possible provided that one is able to apply NAND gates to pairs of bits. What is required to perform arbitrary quantum computations? This question is not trivial, but the answer is fortuitously simple [9] . Any quantum computation can be performed, provided that one is able to arbitrarily control the state of any qubit (single-qubit rotations), and provided that one can make pairs of qubits interact with one another (two-qubit gates). It is generally sufficient to have only a single type of interaction, so long as the final state of both interacting qubits depends in some way on the initial state of both qubits. Such a gate is known as an entangling gate, and they are notoriously difficult to implement. 1.3 Quantum Memory 1.3 10 Quantum Memory In the light of the preceding discussion, a quantum memory can be understood as a physical system that is well protected from noise, and that can be made to interact with information carriers so that their quantum state is transferred into, or out of, the memory. Note how we have distinguished the system comprising the memory from the information carriers. In many cases, this distinction is artificial. For instance, in ion trap quantum computing [10] , the hyperfine states of calcium ions are used as qubits. These ions are isolated from their noisy environment by trapping them with oscillating electric fields; the quantum states of the qubits therefore remain un-distorted for long periods (on the order of seconds), and so there is no need to transfer these states into a separate memory. But there are other proposals for quantum computing that make explicit use of quantum memories. An example is the use of nitrogen-vacancy centres in diamond for quantum computing [11,12] . Here a single electron from a nitrogen atom, lodged in a diamond crystal and surrounded by carbon atoms, is used as a qubit. The electron qubit can be controlled with laser pulses to perform computations, but this very sensitivity to light makes it susceptible to damage from noise. Therefore a scheme was devised to transfer the quantum state of the electron to that of a nearby carbon nucleus. The carbon nucleus is de-coupled from the optical field, and it can be used to store quantum information for many minutes. A common theme among such computation schemes is an antagonism between controllability and noise-resilience. That is, systems which are easily manipulated 1.4 Linear Optics Quantum Computing 11 and controlled with external fields are susceptible to noise from those same fields, while well-isolated systems that are not badly affected by noise are generally hard to access and control in order to perform computations. This trade-off leads to a natural division of labour between systems that are easily manipulated, but shortlived, and systems that are not easily controlled, but long-lived. Many quantum computing architectures put both types of system to use, the former as quantum processor, the latter as quantum memory. 1.4 Linear Optics Quantum Computing Since James Clerk Maxwell wrote down the equations of electromagnetism in 1873, physics has undergone profound upheavals at least twice, with the development of both Relativity and Quantum Mechanics in the early twentieth century. Maxwell’s equations have weathered these storms with astonishing fortitude, being both relativistically covariant and directly applicable in quantum field theory. They are probably the oldest correct equations in physics. Implicit within them is a description of the photon, the quantum of the electromagnetic field. Photons come with one of two polarizations, and superposition states of these polarizations are readily prepared in the lab. In addition, they are themselves discrete entities, and it is possible to generate superpositions of different numbers of photons. Photons therefore embody the archetypal qubit, and for this reason Maxwell’s equations remain as central to the emerging discipline of quantum information processing as they were to the pioneers of telegraphy and radio. 1.4 Linear Optics Quantum Computing 12 Photons occupy a frustrating territory on the balance sheet of usefulness for quantum computation. They are ideal qubits, and arbitrary manipulation of their polarization and number states can be accomplished with simple waveplates, beamsplitters and phase-shifters. That is, single-qubit rotations are ‘cheap’. Unfortunately, entangling gates between photons are much more difficult to realise. This is unsurprising, since such a gate requires that two photons be made to interact with one another, and it is well known that light does not generally interact with light: torch beams do not ‘bounce off’ each other; rather they pass through each other unaffected. In 2001 Emanuel Knill, Raymond Laflamme and Gerard Milburn showed how to overcome these difficulties by careful use of measurements [13] , making universal quantum computation possible with only ‘linear optics’. Further developments [14,15] have cemented linear optics quantum computing (LOQC) as an important paradigm for the future of quantum computation. However the two-qubit gates proposed are generally non-deterministic. As the number of gates required in a computational step increases, the probability that all gates are implemented successfully decreases exponentially, so that large computations must be repeated many times for yielding reliable answers. This problem of scalability can be mitigated if the photons output from successful gates can be stored until all the required gates succeed. But photons generally have a short lifetime because they travel at the speed of light: if they are confined in a laboratory they must be trapped by mirrors (in a cavity) or by a waveguide (optical fibre), and absorption or scattering losses are inevitable on time scales of milliseconds or greater [16] . Therefore the ability to 1.5 Quantum Communication 13 transfer the quantum state of a photon into a quantum memory would be a boon to LOQC. Another possibility for quantum computing with photons is to implement twoqubit gates inside a quantum memory. Single-qubit operations are easily performed on the photons directly; when interactions are needed, the photons are transferred to atomic excitations which can be manipulated with external fields to accomplish the entangling gates [17–19] . Applications such as these constitute the most ambitious motivation for the study of optical quantum memories. In the next section we will see that quantum memories are also required in extending the range of so-called quantum communication protocols, which provide guaranteed security from eavesdroppers. 1.5 Quantum Communication Although practical quantum computing remains beyond the reach of current technology, another application of quantum mechanics has already made the leap into the commercial sector. Quantum Key Distribution (QKD) is a technique which allows two communicating parties to be absolutely certain, in so far as the laws of physics are known to be correct, that their messages have not been intercepted [20] . It is possible to purchase QKD systems from two companies: MagiQ based in New York, and ID Quantique in Geneva; many other businesses are incumbent, and the market for such guaranteed-secure communication is estimated at around a billion dollars annually. The idea behind QKD is simple: if Alice sends a message to Bob 1.5 Quantum Communication 14 in which she has substituted each letter for a different one, in a completely random way, neither Bob, nor anyone else, can decode the message, unless Alice tells them how she did the substitution. This information is known as the key, and only someone in possession of the full key has access to the contents of Alice’s message. Encrypting messages in this way is the oldest and simplest method of encryption. It is absolutely and completely secure, provided that only the intended recipient has access to the key. Once the key has been used, it should not be used again, since with repeated use an eavesdropper, conventionally called Eve, might start to see patterns in the encrypted messages and begin to guess the substitution rule. For this reason this encryption protocol is known as the one time pad. For each message sent, a new, completely random key must be used by both Alice and Bob. How does Alice send the keys to Bob? If these are encrypted, she will need to send another key beforehand, and our perfect security is swallowed by an infinite regression. If she sends the keys unencrypted, can she be sure that Eve has not intercepted them? If she has, then Eve has access to all of Alice’s subsequent messages, and there’s no way for Alice or Bob to know their code has been cracked until the paparazzi arrive. These issues are eliminated by public key cryptography. Here, Bob tells Alice the substitution rule she should use for her message. The encryption is done in such a way that Alice’s message cannot be decoded using this rule, so it doesn’t matter if Eve discovers it. Alice then sends her coded message to Bob, who knows how to decrypt the message. An implementation of this idea using the mathematics of large prime numbers was developed in 1978 by Ron Rivest, Adi Shamir and 1.5 Quantum Communication 15 Leonard Adleman [21] . The RSA cryptosystem is the industry standard for secure communication over the internet; much of modern finance relies on its security. But unlike the one time pad, no-one has proved that it is secure. The RSA algorithm relies on the empirical fact that it is computationally very demanding to find the prime factors of a large number. That is, if two large prime numbers p and q are multiplied together to give their product n, it is not practically possible to find p and q, given knowledge of n alone. The best algorithm, the number field sieve, can find the factors of n in roughly eN 1/3 2/3 log2 N computational steps [22] , where N is the number of bits needed to represent n. This exponential scaling means that it is easy to make the calculation intractably long by making n just a little larger. But this is not the whole story. In 1994 Peter Shor showed how a quantum computer could be used to perform this factorization much faster [23] . Shor’s algorithm requires just N 2 (log2 N ) log2 (log2 N ) steps, an exponential improvement over the best conventional methods. A quantum computer that can implement this algorithm efficiently does not yet exist, although proof-of-principle experiments using LOQC have been performed [24–26] . But it is now known that the RSA cryptosystem is living on borrowed time: if a practical quantum computer is ever made, modern secure communications will be spectacularly compromised. Enter QKD. QKD makes use of quantum mechanics to distribute the keys required for a one time pad protocol in a secure way, avoiding the infinite regress arising from a classical protocol, and obviating the need to rely on the fatally flawed RSA system. The goal is to provide both Alice and Bob with an identical string 1.5 Quantum Communication 16 of completely random bits, which they can use as keys to encrypt and decrypt a message. The most widely known protocol used to do this is known as BB84, after the 1984 paper by Bennett and Brassard [27] . Alice sends photons, one at a time, to Bob. Alice can choose the polarization of each photon to point in one of four possible directions: horizontal, vertical, anti-diagonal or diagonal (|Hi, |V i, |Ai or |Di; see Figure 1.2). These directions form a pair of perpendicular polarizations, with a quarter-turn between them. Each of these pairs is known as a basis. The unrotated basis contains the |Hi and |V i polarizations, and is known as the rectilinear basis. The rotated basis contains the |Ai and |Di polarizations, and is referred to as the 45-degree basis. Bob can measure the polarization of the photons he receives from Alice, but to do so he has to line up his detector with either the rectilinear or the 45-degree basis — he has to choose. The measurement gives one of two possible results, either 0 or 1, but these results mean nothing unless the photon polarization belonged to the basis that Bob chose to measure. For example, a |Di photon polarized in the 45-degree basis will give a completely random result, either 0 or 1 with equal probability, if Bob aligns his detector with the rectilinear basis. So if he gets the basis wrong, his measurement results are useless. If he chooses correctly however, and the photon is polarized in the same basis that he measures in, then the 0 or 1 results tell him to which of the two possible directions in that basis the photon polarization belonged. So for instance if Bob aligns his detector with the 45-degree basis, the |Di photon will always give a 1 result. An |Ai photon would give a 0 result for this measurement, while either |Hi or |V i photons would give random results. 1.5 Quantum Communication 17 This strange property of photons, that measurements give useful or uncertain results depending on the measurement basis, is a manifestation of Heisenberg’s uncertainty principle [28] . It is uniquely quantum mechanical. 45 -d eg re e rectilinear Figure 1.2 BB84 protocol. Alice sends photons to Bob with polarizations chosen randomly from the four possible directions |Hi/|V i and |Ai/|Di, represented here as qubit states. To measure the polarization, Bob (or Eve) must choose a basis, rectilinear or 45-degree, for their measurement. Only photons polarized in this basis will yield useful information. To proceed with QKD, Alice generates two completely random, unrelated, bit strings. For each photon, she uses the first bit string to decide in which basis to polarize her photon. For instance, 0 could signify rectilinear and a 1 would mean 45degree. Then she uses the second bit string to decide which of the two perpendicular polarizations in that basis to use. When Bob receives Alice’s photons, he records the results of his measurements, and the basis he used for each measurement. Alice then sends Bob her first bit string. This tells him the bases each photon belonged to. He knows that his results are useless every time he chose the wrong basis for his measurement. So he discards these results. The remaining results tell him the 1.5 Quantum Communication 18 correct polarization of each photon. That is, Bob’s remaining results now tally exactly with Alice’s second bit string. Alice and Bob now share a cryptographic key they can use for a one-time pad. But what about Eve? Well, Eve may have intercepted Alice’s first bit string, but this only contains information about which results to discard, it tells Eve nothing about what those measurement results were, so this does not help her in cracking Alice and Bobs’ code. She could also have intercepted the photons that Alice sent, and tried to measure their polarizations. But, just like Bob, she has to guess at which basis to measure in. She gets a result, but she has no idea whether the basis she chose is correct. She has to send photons on to Bob, otherwise he will receive nothing and get suspicious. But Eve does not know what polarization to give her photons, because she doesn’t know whether her measurements are reliable or not. Suppose she just decides to send photons polarized according to her measurement results in the basis she chose to measure in. Bob receives these photo ns and measures them, none the wiser. But after Alice sends her first bit string, and Bob discards his unreliable measurements, Bob’s remaining results may not tally perfectly with Alice’s anymore. This is because sometimes Eve will have chosen a different basis to Alice, obtained useless measurement results, and sent photons to Bob with the wrong polarization. So Bob and Alice can compare their keys, or a small part of them, and if they do not match up, they know that Eve has been tampering with their photons. There is no way for Eve to listen in without Alice and Bob discovering her presence. In quantum mechanics, measurements affect the system being measured, and the BB84 protocol exploits this fact to guarantee 1.6 Quantum Repeaters 19 secure communication. Quantum memories are not needed in the above protocol, provided Alice’s photons survive to reach Bob. As mentioned in Section 1.4 in the context of LOQC, photons generally do not survive for longer than around 1 ms in an optical fibre, so Bob should not be further than around 200 miles from Alice, otherwise her photons will be scattered or absorbed before they reach him. In order to extend the distance over which QKD is possible, some kind of amplifier is needed, which can give the photons a ‘boost’, while maintaining their quantum state — that is, their polarization. But the photons are qubits. They cannot simply be copied; we know this from the no-cloning theorem (see Section 1.2.3). What is required is a modification to the protocol described above, and a device known as a quantum repeater. Such a device requires a quantum memory. If quantum memories can be made efficient, with storage times on the order of 1 s, intercontinental quantum communication becomes possible. In the next section we introduce the quantum repeater, and discuss the usefulness of quantum memories in this context. 1.6 Quantum Repeaters A quantum repeater is a device designed to extend entanglement. Entanglement is a purely quantum mechanical property that can be used as a resource to perform QKD. In this section we will introduce entanglement, examine how it degrades over long distances, and how quantum repeaters ameliorate this degradation. Entanglement is a property of composite quantum systems. As an example, 1.6 Quantum Repeaters 20 consider two qubits. Classically, a system composed of two parts could be described by the states of each part. In quantum mechanics this is not always true. Just as a qubit can exist in a superposition of different states, so a system comprising two qubits can exist in a superposition of different combined states. Such states arise from a blend of correlation and indeterminism. To see this, suppose that we have a machine that produces two photons, each with the same polarization. Now suppose that the direction of this polarization is not fixed. It might polarize both photons horizontally, we’ll label this polarization |0i, or vertically, |1i. We have no way of knowing which of these two polarizations the machine uses, we only know that both photons will have the same polarization. Such a state cannot be described by talking about each photon in turn, as is clear from the language we used to describe the set up. Using subscripts to denote the two photons, the state is written as 1 |ψi = √ (|0i1 |0i2 + |1i1 |1i2 ) . 2 (1.6) √ The factor of 1/ 2 appears simply to fix the length of the state vector |ψi to 1. This state is an entangled state, because there is no way to write it as a product of states of the first photon with states of the second. It expresses the two properties of our machine: first, that the two photons always have the same polarization, and second, that it is not certain which of the two polarizations will be produced. In fact, because the two possible states |0i and |1i are mutually exclusive, (1.6) is a maximally entangled state, sometimes known as a Bell state. Bell states represent much of 1.6 Quantum Repeaters 21 what is counter-intuitive about quantum mechanics. Their name derives from John Bell’s famous 1964 paper [29] in which he proves that these states are incompatible with local realism. A ‘local’ world is one in which no effect can propagate faster than the speed of light; a ‘real’ world is one in which all properties can be assigned definite values at all times. That modern physics describes states which do not admit a local realistic interpretation is intriguing and controversial. Below we will see that these states are also a resource for quantum communication. If their use becomes widespread, we will be in the awkward position of deriving practical benefits from a technology based on a philosophical conundrum! 1.6.1 The Ekert protocol In 1991 Artur Ekert proposed a modification of the BB84 QKD protocol based on the use of Bell states like (1.6) [30] . In this protocol, our machine for generating entangled photon pairs is used. One photon from each pair is sent to Alice, the other to Bob. Now Alice and Bob both have polarization detectors; they each have to choose a basis to measure their photons in. Sometimes they will choose the same basis as eachother, sometimes they will choose different bases. When they choose differently, their results are meaningless, but when they choose the same basis, their results are perfectly correlated. This is obvious for the rectilinear basis by inspection of the form of (1.6). A little algebra shows that the same perfect correlations also hold if both Alice and Bob measure in the 45-degree basis. The QKD is accomplished in the same way as for the BB84 protocol: Alice tells Bob the measurement bases 1.6 Quantum Repeaters 22 she used; Bob discards the results of all measurements where his basis differed from Alice’s. Alice and Bob then compare part of the remaining results to check that they are correlated, as they should be. Poor correlations signify the presence of an eavesdropper. The importance of this modified protocol is that entanglement is a transferrable resource. Below we will see how entanglement can be swapped between photons to extend the range of quantum communication. 1.6.2 Entanglement Swapping Entanglement swapping allows one to entangle two photons that have never encountered eachother. The situation is sketched in Figure 1.3. Two sources each emit a pair of entangled photons in the state (1.6). One photon from each pair is sent into a polarizing beam splitter, which transmits horizontally polarized photons, and reflects vertically polarized photons. Behind the beamsplitter are a pair of photon detectors. The beamsplitter has the effect of limiting the information we can learn about the photons from the detectors. For instance, if both photon detectors D1 and D2 fire together, it could be that photons (2) and (3) were both vertically polarized, or that they were both horizontally polarized. That is, a ‘coincidence count’ from D1 and D2 only tells us that photons (2) and (3) had the same polarization; it reveals nothing about what that polarization was. But we know from the state (1.6) that photon (1) has the same polarization as photon (2), and similarly that photon (4) has the same polarization as photon (3). So if photons (2) and (3) have 1.6 Quantum Repeaters 23 the same polarization, so do photons (1) and (4). Their polarization is unknown, but correlated. Therefore, after a coincidence count, the two remaining photons, (1) and (4), are in a Bell state. The entanglement between photons (1)-(2) and (3)-(4) has been swapped to photons (1)-(4). This procedure was first demonstrated experimentally by Jian Wei-Pan et al. in 1998 [31] , and is now an essential tool for LOQC. & D1 D2 PBS 1 2 S1 3 4 S2 Figure 1.3 Entanglement swapping. Two independent sources, S1 and S2, emit pairs (1)-(2) and (3)-(4) of polarization entangled photons. Photons (2) and (3) are directed into a polarizing beam splitter (PBS). When both detectors D1 and D2 fire behind the PBS, photons (1) and (4), which have never met, become entangled. It’s clear from the above arguments that entanglement swapping is not much more than a re-assignment of our knowledge regarding correlations, in the light of a measurement carefully designed to reveal only partial information. Nonetheless, a real resource — entanglement — has been extended over a larger distance by this procedure. And QKD can now be performed using photons (1) and (4). 1.6 Quantum Repeaters 1.6.3 24 Entanglement Purification So far we have shown how entanglement can be extended over large distances by swapping perfect Bell states, each distributed over shorter distances. In practice, however, propagating even over short distances can distort the polarizations of the photons. Small distortions do not completely destroy the entanglement; rather there is a smooth degradation in the usefulness of the photons for QKD as the distortions become worse. Nevertheless, with each entanglement swap, these deleterious effects are compounded, so that the entanglement vanishes after only a few swapping operations. However, it is possible to transform several poorly entangled photon pairs into one photon pair with near-perfect entanglement using an ‘entanglement purification protocol’. Several of these exist [32–36] , generally they involve mixing and measuring photons in a way that strengthens the correlations between the remaining photons. Such procedures allow entanglement to be ‘topped up’, at the expense of having to use more photons. QKD across distances much larger than those over which distortions affect photons can then be implemented. Below we will introduce a paradigm for constructing a quantum repeater which does not explicitly make use of polarization entanglement, but which combines entanglement swapping and entanglement purification into a single step. 1.6.4 The DLCZ protocol and number state entanglement As mentioned previously, photons have several different degrees of freedom that can be used to encode qubits. We have mostly focussed on polarization qubits for QKD, 1.6 Quantum Repeaters 25 but in the following protocol qubits are encoded in the number of photons occupying a given optical mode, so-called single-rail encoding. Consider a machine similar to the Bell state sources discussed above, which emits one photon either to the left, or to the right. Using subscripts L and R for these directions, the state produced by this machine is 1 |ψi = √ (|0iL |1iR + |1iL |0iR ) . 2 (1.7) Note that the states |0i and |1i now refer to the number of photons, rather than the photon polarization as before. This state is also a maximally entangled state. Like (1.6) it is also a Bell state. It expresses the correlation that a photon in one mode always signifies the absence of a photon in the other mode. And it expresses the indeterminacy, built into our device, that there is no way of knowing whether the emitted photon will be found in the left or the right mode. How does entanglement swapping work on such a state? A slightly different set-up is used (see Figure 1.4). The action of the measurement is particularly clear in this example: the beamsplitter (BS) mixes the two modes (2) and (3), so that a detection at D1 or D2 tells us only that one of those modes contained a photon, but not which one. A bit of epistemic book-keeping reveals the entanglement swap: if (2) contained a photon but (3) did not, that means (1) had no photon while (4) carries a photon. Similarly if (2) was empty but (3) contained a photon, (1) must carry a photon while (4) does not. There is no way to distinguish these possibilities, and therefore a single detection behind the BS puts the modes (1) and (4) into the entangled state (1.7). For single-rail encoding, the most damaging effect of propagation is the pos- 1.6 Quantum Repeaters 26 | D1 D2 BS 1 2 S1 3 4 S2 Figure 1.4 Single-rail entanglement swapping. Two independent sources, S1 and S2, emit single photons into modes (1)-(2) and (3)-(4) in the state (1.7). Modes (2) and (3) are mixed on a beam splitter (BS). When one of the detectors D1 or D2 (but not both) fires behind the BS, modes (1) and (4) become entangled. sibility of photon loss, through absorption or scattering. This has the effect of introducing a third term into the state (1.7) of the form |0iL |0iR , corresponding to no photons in either mode (they’ve all been lost!). With the addition of this term, the quality of the entanglement is reduced. But this ‘vacuum’ component can never cause any detection events. Therefore if either of the detectors D1 or D2 fire, signaling a successful entanglement swap, the vacuum component is removed, since the detection of a photon renders it counterfactual. For this type of state, entanglement swapping also accomplishes a degree of entanglement purification. The ability to perform both swapping and purification using such a simple measurement makes this type of encoding attractive as a means of distributing entanglement over large distances for quantum communication. However, it is not immediately obvious how to generalize the Ekert QKD protocol to states encoded in this way. For example, to perform a measurement in the analogue of the 45-degree basis, one would require detectors that are sensitive to superpositions of photon 1.6 Quantum Repeaters 27 number states. These issues are avoided by combining two states of the form (1.7); the measurement scheme is shown below in Figure 1.5. Two entangled states of the & 1 2 D2 D1 BS BS D3 D4 4 3 & Figure 1.5 QKD with single-rail entanglement. Two entangled states are distributed between Alice — detectors D1 and D3 on the left — and Bob — detectors D2 and D4 on the right. The photons on each side are mixed on a beamsplitter (BS). A ‘polarization measurement’ is made when a single detector fires on each side. Both D1 and D2 firing is a ‘0’. D3 and D4 firing is a ‘1’. Adjusting the phases φA,B allows to select the measurement bases. form (1.7), are distributed between Alice and Bob. Alice, on the left hand side, has two detectors, D1 and D3. Bob, on the right, has detectors D2 and D4. If Alice only records measurements when one of her detectors fires, she knows that one photon came to her, while the other went to Bob. Similarly a single detection at Bob’s side tells him that Alice received the other photon. The phases φA and φB are independently chosen by Alice and Bob from the set {0, π/2}. These phases, in combination with the beamsplitters, allow Alice and Bob to control the basis in which their detectors measure, in direct analogy with the rectilinear and 45-degree bases of the Ekert protocol. When they choose the same phases, their measurements 1.6 Quantum Repeaters 28 should be correlated, with D3 and D4 firing together for a ‘1’ result, and D1 and D2 firing together for a ‘0’ result. Their is no correlation if they choose different phases. Alice and Bob publicly announce their basis choices, and then compare some of their results to check for the presence of Eve. The above discussion shows how number state entanglement can be used for quantum communication over long distances. A specific proposal for implementing this protocol using atomic ensembles to generate the entangled states was first made in 2001 by Lu-Ming Duan, Michael Lukin, Ignacio Cirac and Peter Zoller [37] . The DLCZ protocol is exciting because it is firmly grounded in feasible technology. Several improvements have since been suggested [38–42] , which make the scheme more robust to phase instability, photon loss, detector noise and inefficiency. Below we briefly introduce the principle behind the original protocol, since this will serve as a useful introduction to the uses of atomic ensembles in quantum information processing technologies. We consider two clouds of atoms L and R, each with internal energy levels arranged as depicted in Figure 1.6. This type of Λ-structure is ubiquitous in quantum information protocols. We will encounter it many times in our survey of quantum memory protocols. It also arises in all but the simplest quantum systems: not just atoms, but crystals, quantum dots and molecules, as we will see. The atomic ensembles L and R will together play the role of the entangled photon source introduced earlier. Here’s how it works. Both ensembles are pumped simultaneously by a laser pulse. The pump pulse is tuned out of resonance with the atomic state |2i, so that 1.6 Quantum Repeaters 29 Readout Generation (a) (b) Figure 1.6 Λ-level structure of atoms used in the DLCZ protocol. (a): A pump pulse excites an atom out of the ground state |1i, which decays down to a long-lived metastable state |3i, emitting a ‘Stokes photon’. (b): A readout pulse brings the excited atom back to the ground state, which emits an ‘anti-Stokes photon’ in the process. most of the time nothing happens. But provided that the number of atoms in each cloud is large enough, there is a small probability that the pump pulse will cause a two-photon Raman transition in one of the atoms (see §2.3.2 in Chapter 2), exciting it to a long-lived metastable state |3i, as shown in Figure 1.6 (a). Such an excitation is always accompanied by the emission of a Stokes photon. The probability of two atoms being excited this way is negligibly small. Therefore at most one Stokes photon is emitted from the ensembles. Due to the extended, pencil-like geometry of the ensembles, any Stokes photons tend to be emitted in the forward direction, so that they can be captured and directed as required [43] . Detectors D1 and D2 are placed behind a beamsplitter in front of the ensembles, as shown in Figure 1.7. The beamsplitter mixes the optical modes from the two ensembles, so that if one of the detectors fires, we know that a Stokes photon was emitted from one of the ensembles, but we don’t know which. After a detection then, the state of the atomic 1.6 Quantum Repeaters 30 ensembles is of precisely the form (1.7), where |1i and |0i now refers to the presence or absence of an excited atom in an ensemble. We have now generated the number state entanglement required for the repeater protocol. A difference with the preceding discussion is that the entanglement is between the atomic ensembles, rather than optical modes. But this is an advantage, because the atomic excitations are stationary, and can last for a long time, while photons are generally short lived, as mentioned earlier. Since all the steps of the protocol involve waiting for particular detectors to fire, it is essential that the entanglement can be preserved until all the required steps are completed. When the entanglement is ‘needed’, for instance to perform an entanglement swapping operation, the atomic excitations can be converted back into photons. This is done by applying a ‘readout’ pulse to the ensembles, which returns any excited atoms to their ground states via an anti-Stokes transition, as shown in Figure 1.6 (b). The anti-Stokes modes inherit the entanglement from the ensembles. If the anti-Stokes mode from ensemble L is sent off in one direction (left), and the anti-Stokes mode from the R ensemble is sent in the other direction (right), we now have a device that can emit entangled states on-demand. The sources S1 and S2 appearing in Figure 1.4 should now be thought of as each comprised of a pair of entangled atomic ensembles, waiting to be ‘read-out’ when desired. So far we have seen how atomic ensembles show promise for quantum communication protocols, but we have not encountered a specific need for quantum optical memories. To conclude this section, we describe how the DLCZ protocol can be 1.7 Modified DLCZ with Quantum Memories 31 D1 L BS R D2 Figure 1.7 Generation of number state entanglement using atomic ensembles in the DLCZ quantum repeater protocol. A click in one of the detectors D1 or D2 tells us that one atom, in either of the ensembles L or R, is excited, but because of the beamsplitter (BS), we don’t know to which ensemble this atom belongs. Therefore the ensembles are left in the entangled state (1.7). improved by using memories in combination with single photon sources [41] . 1.7 Modified DLCZ with Quantum Memories Recall that we claimed the probability for two atoms in the ensembles L and R to be excited by the pump pulses was so small as to be insignificant. If two atoms could be excited, it would mean that sometimes one Stokes photon from L and one from R is emitted simultaneously. Suppose that one of these photons is lost somehow. Perhaps it is absorbed in an optical fibre. Then only one of the detectors D1 or D2 behind the beamsplitter will fire, but both ensembles are in fact excited. To account for this possibility, we should add a term to the state (1.7) of the form |1iL |1iR — that is, the ensembles are no longer in a maximally entangled state. Of course the probability of a double excitation can always be made arbitrarily small by making the pump pulses weaker. But this also reduces the probability of even 1.7 Modified DLCZ with Quantum Memories 32 a single excitation, so that it is necessary to wait a long time before one of the detectors fires. If the waiting time becomes too long, the entanglement stored in the ensembles starts to degrade as the atoms drift and collide, so this limits the distance over which entanglement can be distributed. This issue can be resolved if the entanglement is generated using a single-photon source and a beamsplitter. The modified set up is shown in Figure 1.8. The atomic ensembles L and R are now used as quantum memories, which we label QML and QMR. We will explore the details of atomic ensemble quantum memories in Chapters 4–8. For now we only need to know that a photon entering a quantum memory is trapped until it is needed. We have assumed that we have access to single photon sources (SL and SR) that each emit one and only one photon, on-demand. Sources like this are actually rather difficult to make, but research into this technology has advanced greatly over the last few years [44–47] , and including such sources into the design is no less realistic than the inclusion of quantum memories. To generate number-state entanglement, we start by triggering both sources SL and SR. Each emits a photon, which then encounters a beamsplitter (BS). At the beamsplitters the photons can either be reflected into a quantum memory, or transmitted, in which case they are brought together on a final BS placed in front of detectors D1 and D2. In the case that just one of the detectors fires, we know that only one photon was transmitted at the first BS; the other photon must have been reflected, in which case it is now stored in a quantum memory. But the final BS prevents us knowing which quantum memory contains the photon. We therefore have a superposition 1.7 Modified DLCZ with Quantum Memories 33 state, with either QML excited and QMR empty, or vice versa. This is the desired entangled state (1.7). What if both photons are transmitted, but one of them is D1 BS BS QML SL D2 BS QMR SR Figure 1.8 Modification to the DLCZ protocol using single photon sources (SL and SR) and quantum memories (QML and QMR). The number state entanglement is generated by beamsplitters. somehow lost? In that case only one detector fires, but neither quantum memory is excited. We therefore have to add a vacuum component of the form |0iL |0iR to the state (1.7). This certainly degrades the entanglement in the same way as the |1iL |1iR component did for the corresponding error in the DLCZ protocol. But when we perform entanglement swapping to extend the entanglement, we always wait until the appropriate detector fires (see Section 1.6.2), and this purifies the entanglement, removing the vacuum term (see Section 1.6.3). Therefore vacuum errors do not damage the protocol, and we are able to run it faster and further than the original DLCZ scheme permits. Hopefully the possible uses for quantum optical memories is now clear. Presum- 1.7 Modified DLCZ with Quantum Memories 34 ably future proposals will develop further applications for them. In the next chapter, we will review the various techniques used for quantum storage. Chapter 2 Quantum Memory: Approaches The aim of a quantum optical memory is to convert a flying qubit — an incident photon — into a stationary qubit — an atomic excitation. This conversion should be reversible, so that the photon may be re-emitted some time later, at the behest of the user. In this chapter we review a number of approaches for achieving this kind of quantum storage. More detailed calculations are deferred until the next chapter. The simplest system one could imagine for storing a photon would consist of a single atom, coupled to a single optical mode via an electric dipole transition (see Figure 2.1). An incident photon, resonant with the transition, is absorbed, promoting the atom from its ground state |1i to its excited state |2i. The photon is now ‘trapped’ as a stationary excitation of the atom, and the quantum storage is complete. However, quantum mechanics is always invariant under time-reversal. That is to say, the electric dipole interaction (see §C.4 in Appendix C), which allows the photon to be absorbed by the atom, also causes the atom to re-emit the photon, 36 (b) (c) 1 Excitation probability (a) 0.5 0 0 time/ Figure 2.1 The simplest quantum memory. (a): A single atom is coupled to a single optical mode. (b): An incident photon is absorbed, exciting the atom. (c): Unfortunately, time reversal symmetry requires that the photon is immediately re-emitted. almost immediately afterwards. In fact, these two processes compete continuously, so that over time the population of the excited state oscillates back and forth as the atom absorbs and re-emits the photon; see Figure 2.1 (c). This behaviour is known as Rabi flopping, after Isidor Isaac Rabi, who first used the phenomenon in the context of nuclear magnetic resonance [48] . The Rabi frequency Ω of these oscillations is proportional to the strength of the coupling between the atom and the electromagnetic field. The above scheme needs some modification if it is to store a photon in a controlled way. One solution is to introduce a third state |3i — a dark state — that is not coupled to the photon mode we want to store. We should have some control field we can apply that transfers the atomic state from |2i to |3i once the storage is complete. In this way the Rabi oscillations are ‘frozen’, and the photon remains trapped as an excitation of the dark state for as long as is desired. Provided the atom is well isolated, the dark state will persist for as long as is needed. To retrieve 37 the photon, the control field is applied again, transferring the atomic state from |3i back to |2i. The Rabi flopping of Figure 2.1 (c) continues and the photon is re-emitted. The scheme is shown in Figure 2.2. This typifies the approach taken in (a) (b) storage (c) retrieval input control Figure 2.2 Adding a dark state. (a): We address the atom with an auxiliary control. (b): After the input photon is absorbed, the control transfers the atom to |3i. (c): When the photon is needed, the control is re-applied. many quantum optical memories, although the details of the protocols differ widely. While single atoms have been used, ensembles of many atoms, each with the Λ-type structure of Figure 2.2, are also commonly employed. The atom(s) can be enclosed in an optical cavity, trapped at the centre of a confocal microscope, or addressed by collimated beams. The optical fields may be resonant, or off-resonant with the atomic transitions; the shape and timing of the control fields can vary, and indeed non-optical controls, such as magnetic or electric fields, may be used. We will briefly review this menagerie of memory protocols in the following sections. 2.1 Cavity QED 2.1 38 Cavity QED In the forerunning discussion we presumed that a single atom could be coupled to a single optical mode. This is a highly unnatural state of affairs, since the electromagnetic field pervades all of space, so that atoms are generally surrounded by a bath of electromagnetic field modes. The most dramatic effect of this ‘reservoir’ of modes is that any atomic excitations coupled to the field tend to leak away rather quickly: an emitted photon is very unlikely to couple back to the atom, so the atomic population does not exhibit Rabi flopping; rather it decays exponentially. This is known as spontaneous emission — the stronger the coupling between an atomic state and the field, the shorter the lifetime of that state. An associated consequence is that an incident photon is very unlikely to couple to the atom. Therefore a single atom in free space cannot be used for quantum storage. In order to recover strong coupling between an atom and an optical mode, it is necessary to suppress the interaction with all unwanted field modes. This can be done interferometrically, but introducing a highly reflective cavity around the atom (see Figure 2.3). Any fields inside the cavity are reflected back on themselves by the cavity mirrors. Only those fields with a wavelength equal to a half integer multiple of the cavity length add constructively when folded back on themselves; all other wavelengths interfere destructively. The optical modes supported by the cavity are therefore spaced regularly in frequency. If the resonant frequency of the atomic transition |1i ↔ |2i is close to one of these supported modes, the atomic coupling will be confined to this single mode. All other cavity modes being too far from resonance 2.2 Free space coupling 39 to contribute significantly. If the volume of the cavity is sufficiently small, and the Figure 2.3 Cavity QED. An atom is confined in a high-finesse optical cavity, which supports only a discrete set of optical frequencies. Only one optical mode, resonant with the atomic transition, couples to the atom. cavity mirrors sufficiently reflective, it is possible to bring an atom into the so-called strong-coupling regime [49–53] , where the light matter interaction behaves broadly as described in Figures 2.1 and 2.2. This so-called Cavity QED approach to light-matter interfaces has been widely applied. Both trapped and moving beams of atoms are used [16,45,54,55] , as well as quantum dots [56–59] and molecules [60] . However, cavities are rather difficult to fabricate; approaches which do away with this requirement would be easier to scale up. 2.2 Free space coupling Another possibility is to focus an incident photon onto an atom in a way that maximizes the probability of it being absorbed. The cavity is replaced with a pair of microscope objectives, as shown in Figure 2.4. The rationale behind this approach is that good coupling should be possible if the spatial shape of an incoming photon matches the spatial pattern of an outgoing spontaneously emitted photon. This time-reversal argument suggests that an incident photon should be converted into a 2.3 Ensembles 40 collapsing dipole pattern. The dipole pattern of an atom is very far from a narrow collimated beam — it is nearly isotropic, covering almost 4π steradians. Therefore the lenses should be as wide as possible: this in itself represents a technical challenge. Only preliminary experiments have been done [61] , but theoretical work [62,63] suggests that efficient coupling can be achieved if the numerical aperture of the objective lenses is made large enough. Figure 2.4 Confocal coupling in free space. An atom is trapped at the focus of a pair of microscope objectives. If the solid angle subtended by an incident photon is sufficiently large, the coupling efficiency is predicted to approach unity. 2.3 Ensembles In addition to the practical difficulties of building a high-finesse cavity, or indeed a high numerical aperture confocal microscope, there are technological hurdles associated with trapping and placing single atoms. Another possibility for quantum storage is to use ensembles of atoms, as shown in Figure 2.5. An incident photon may have a small probability with interacting with any given atom, but as the number of atoms in the ensemble increases, the probability that the photon fails to interact with all the atoms decreases exponentially. Therefore it is possible to asymptotically approach unit interaction probability simply by adding more atoms. 2.3 Ensembles 41 This approach is the main focus of this thesis. As we will see, introducing many atoms makes it possible to store more than one photon, or more than one optical mode. Complicated optical arrangements and cavities are not required, and a wide range of possible storage media — from atomic vapours to Bose-Einstein condensates, quantum dots, crystals and fibres — are well-suited for protocols of this kind. We are primarily concerned with schemes based on the absorption, and subsequent Figure 2.5 Atomic ensemble memory. If an incident photon encounters enough atoms, it is almost certain to be absorbed. re-emission, of a freely propagating photon by an atomic ensemble. These schemes divide broadly into four protocols, all closely related, which we will now introduce. 2.3.1 EIT EIT stands for Electromagnetically Induced Transparency. It was first observed by Boller et al. in 1991 [64] . In EIT, a strong laser — the control — is shone into an atomic ensemble with a Λ-structure, as shown below in Figure 2.6 (a). Ordinarily, a weak probe beam would be absorbed by the atoms, but the interaction with the control laser causes the ensemble to become transparent to the probe. This can be understood by considering the dressed states of the atom, under the influence of the control [65,66] . The control field couples the states |2i and |3i, so the Hamiltonian for the electric dipole interaction of an atom with the control is of the (a) 42 (b) (c) Susceptibility (arb. units) 2.3 Ensembles 0 Detuning (arb. units) Figure 2.6 EIT. (a): a weak probe beam propagates through an ensemble of Λ-type atoms, while a strong control field couples the excited and metastable states. (b): the control mixes states |2i and |3i to produce an Autler-Townes doublet. (c): The imaginary part of the probe susceptibility (solid line) exhibits a double resonance, with a transparency window at the atomic |1i ↔ |2i transition frequency. The real part (dotted line) changes quickly within this window, causing marked dispersion. form H = Ω|2ih3| + h.c., where Ω is the Rabi frequency of the control laser on the |2i ↔ |3i transition. This can be simply re-written as H = Ω (|+ih+| − |−ih−|), √ where the dressed states are defined by |±i = (|2i ± |3i) / 2. These dressed states are simply rotated versions of the natural atomic states |2i and |3i. In fact, although we have changed our notation, the relationship between these states is the same as between the 45-degree and rectilinear bases discussed in the context of QKD in chapter 1. The Hamiltonian is diagonal in the dressed basis, and we can read off the energies of the dressed states as ±Ω. That is, the combination of the control with the atom produces a system with a double-resonance — known as an AutlerTownes doublet [67] — with a splitting between the dressed states set by Ω. This is a manifestation of the dynamic Stark effect: the control brings the energy of state |3i up to that of |2i, and the dipole interaction induces an anti-crossing. From the above arguments, we might already expect the probe absorption spec- 2.3 Ensembles 43 trum to divide into two peaks, with a transparency window in between. Figure 2.6 (c) shows the results of a steady-state calculation of the linear susceptibility for the probe field as a function of its detuning from the |2i ↔ |3i resonance. The probe absorption is proportional to the imaginary part, which shows the expected doublet structure. But the depth of the transparency window cannot be explained simply by superposing two identical resonances. As is clear from the plot, the absorption actually vanishes completely in the centre of the transparency window. This total transparency is due to quantum interference: the contributions to the susceptibility from the two dressed states are of opposite sign, because one is shifted above, and the other below the original resonance. There is therefore an exact cancellation at this resonance, and the susceptibility is identically zero at this point. In addition to propagating without absorption, the probe beam also propagates extremely slowly. More precisely, the group velocity of the probe is reduced by the strong dispersion within the transparency window. The refractive index n of the atomic ensemble is given by n = p 1 + Re(χ), where χ is the probe suscep- tibility. Inspection of part (c) of Figure 2.6 shows that the real part of χ varies extremely rapidly across the transparency window; therefore the refractive index is also changing quickly in this spectral region. To see why this slows the group velocity, consider a pulsed probe. A pulse is composed of a range of frequencies, covering a spectral bandwidth δω, all interfering constructively, so that the phase variation over the pulse bandwidth is roughly zero. Each frequency component of the pulse accumulates an optical phase kδz − ωδt over spatial and temporal increments δz, 2.3 Ensembles 44 δt, where k = nω/c is the wavevector of each component. The trajectory of the pulse is the locus of those points at which all the components of the pulse remain in phase, so that δkδz = δωδt, with δk the range of wavevectors spanned by the pulse. The group velocity — the velocity of the pulse — is therefore defined by vg = dz/dt = dω/dk = c/(n + ωn0 ), where n0 = dn/dω. The steep increase in Re(χ) across the transparency window makes n0 large, and therefore vg is small. Group velocities as low as 17 ms−1 have been demonstrated in the laboratory [68] . The use of EIT as a method for storing light is an elegant application of these effects. It was first described by Misha Lukin and Michael Fleischhauer in 2000 [69] , and has since been demonstrated experimentally many times [70–74] . The protocol works as follows. An atomic ensemble of Λ-type atoms is illuminated by a control field, preparing a transparency window. A probe pulse — to be stored — is now directed into the ensemble, tuned to the centre of the transparency window. As described above, it propagates slowly, but without loss. Even if the spatial extent of the pulse is much longer than the ensemble initially, the sudden slowing of the pulse as it enters the ensemble causes it to bunch up, so that it fits within the ensemble as it propagates (see Figure 2.7). As soon as the entire pulse is inside the ensemble, the control beam is attenuated. That is, the Rabi frequency Ω is reduced, so that the splitting of the Autler-Townes doublet decreases. The transparency window gets smaller, and the variation in Re(χ) becomes steeper, so the group velocity of the probe falls. This process continues until the control is switched off entirely, at which point the transparency window completely collapses; the dispersion diverges, and 2.3 Ensembles 45 the group velocity vanishes: the probe has been brought to a complete stop! In fact, the quantum state of the optical field has been transferred to the atoms, and the state can be stored for as long as the coherence of the atoms survives without distortions. If the control field is switched back on, the probe field is re-accelerated, and emerges — hopefully — unchanged, as if the ensemble had not been present. Figure 2.7 Stopping light with EIT. A long probe pulse bunches up as it enters the EIT medium, due to the slow group velocity at the centre of the transparency window. The arguments just given provide a useful physical picture for the mechanism behind light-stopping by EIT. In isolation they don’t provide a satisfactory explanation for why the probe is not simply absorbed as the transparency window is made narrower, nor for precisely how the probe light is transferred to an atomic excitation. A more complete discussion can be found in the paper by Lukin and Fleischhauer [69] . Nonetheless we can still draw some conclusions about the circumstances under which this procedure will work. The probe spectrum should not be wider than the transparency window, otherwise it will be partially absorbed, so the control must be sufficiently intense when the probe enters the ensemble. The control cannot be turned off too quickly, since the dressed states producing the transparency 2.3 Ensembles 46 window are only a good approximation in the steady, or nearly steady state. The control intensity must be reduced adiabatically; we’ll examine this condition more closely in Chapter 5 (see §5.2.9). Finally, the control should be turned off before the probe escapes from the ensemble, so the initial group velocity should be sufficiently slow — this means the ensemble density should not be too low. These considerations tell us that the bandwidth of probe pulses that can be stored via EIT, and the efficiency of the storage, is limited by the density of the ensemble, and the available control intensity. We now introduce a related protocol, with the aim of circumventing some of these limitations. As more detailed calculations show, this attempt is only partially successful, but the flexibility of the new protocol makes it an attractive alternative. 2.3.2 Raman In 1928 Chandrasekhara Venkata Raman was the first to observe the weak inelastic scattering of light from the internal excitations of vapours and liquids. No lasers were available, and he used a focussed beam of sunlight to generate the required intensity [75] . His work earned him the Nobel Prize in 1930; the eponymous Raman scattering is now used routinely in spectroscopy, industrial sensing and indeed quantum optics. Raman scattering can be understood rather simply in the context of an atomic Λ-system. It is represented schematically in Figure 1.6 (a) from Chapter 1 — the interaction used in the DLCZ repeater protocol is precisely Raman scattering. It 2.3 Ensembles 47 is a two-photon process — that is, second-order in the electric dipole interaction — in which a pump photon (green arrow in the diagram) is absorbed, and at the same time a Stokes photon (blue wavy arrow) is emitted. Energy conservation is satisfied if the frequency difference between the pump and Stokes photons is equal to the frequency splitting between the states |1i and |3i in the atoms. The optical fields need not be tuned into resonance with the |1i ↔ |3i transition: the likelihood of Raman scattering decreases as the fields are tuned further away from resonance, but given a sufficiently intense pump field, and sufficiently many atoms, the Raman interaction can be made rather strong (see §§10.8, 10.9 in Chapter 10). In a Raman quantum memory, the Raman interaction is turned on its head: a signal photon is absorbed, and a strong control field stimulates the emission of a photon into the control beam (see Figure 2.8). The signal and control fields are tuned into two-photon resonance, that is, the difference in their frequencies is equal to the frequency splitting between |1i and |3i. The process is conceptually very similar to that outlined at the beginning of this chapter (c.f. Figure 2.2). But in a Raman memory, no atoms are ever excited into the state |2i. This is because the fields are tuned out of resonance with this state, with a common detuning ∆. Instead, the control field creates a virtual state — represented by the dotted lines in Figure 2.8 — and an atom is excited into this virtual state by the signal photon. The control field then transfers the atom into the state |3i for storage. Of course, all possible time-orderings for this process contribute to the interaction, but from this perspective it is clear that no storage is 2.3 Ensembles 48 Storage (a) Retrieval (b) Figure 2.8 Raman storage. Both fields are off-resonant, with a detuning ∆ from the excited state. (a): A signal photon is absorbed, and at the same time a photon is emitted into the control beam. This two-photon Raman transition promotes one atom in the ensemble from |1i to |3i. (b): The strong control is applied again, and the inverse Raman process returns the excited atom to its ground state |1i, re-emitting the signal photon. possible without the presence of the control field. Hence the name. We might expect the physics of Raman storage to relate very closely to that of EIT storage, and indeed they are confusingly similar. To connect Raman storage with our previous discussion of EIT, we can consider the dressed states of the atoms in a Raman memory, under the influence of the control. In fact, we have already done so. The virtual state into which the signal photon is absorbed is precisely the dressed state produced from the atomic state |3i when the control is present. Just as in the case of the |−i state of EIT, the virtual state is really a combination of the states |2i and |3i. The difference is that it contains a very small amount of the excited state |2i, because of the large detuning ∆. The other dressed state, the equivalent of |+i for the Raman memory, is not shown in Figure 2.8, since it is so close to the bare atomic state |2i as to be indiscernible. Again, the large detuning means that this state is made up almost entirely of |2i, with almost no contribution 2.3 Ensembles 49 from |3i. We can therefore frame the difference between EIT and Raman storage in the following way. In an EIT memory, the photon to be stored is tuned between the dressed states; it is brought to a halt by turning the control field off. In a Raman memory, the signal photon is tuned into resonance with one of the dressed states, which is made up almost entirely of the storage state |3i. Once it has been absorbed, the control field is turned off, and the state becomes ‘dark’ — decoupled from the electromagnetic field. One motivation for studying Raman storage is the possibility of broadband storage. That is, the storage of temporally short photons, which necessarily comprise a large range of frequencies. EIT storage requires that the spectral bandwidth of the input photon should fit within the transparency window. Raman storage does not rely on a transparency effect, so this limitation does not pertain. In addition, the spectral width of the virtual state into which the signal photon is absorbed is set by the spectral width of the control field. We might therefore expect that broadband photons can be stored with a broadband control. We should ensure that the excited state |2i is never populated, since this state will eventually decay and our photon will be lost. This means that the detuning ∆ must greatly exceed the bandwidth of the signal photon. Nonetheless, given sufficiently many atoms, the large detuning need not weaken the interaction, and efficient broadband storage should be possible. In fact, as discussed in §5.2.9 of Chapter 5, the Raman memory protocol is ideally suited to broadband storage, although its advantages over the EIT protocol in this respect are not entirely clear-cut. 2.3 Ensembles 50 A Raman memory was first proposed by Kozhekin et al. in 2000 [76] . They did not explicitly address the problem of storing a single photon; rather they considered the transfer of quadrature squeezing — a phenomenon we’ll describe in the next section on continuous variables memories — from light to atoms. The theoretical treatment of a Raman memory for photon storage was published in 2007 [77] , and forms part of this thesis. Raman storage is yet to be implemented experimentally, but we are currently attempting to demonstrate the protocol in cesium vapour; details are given in Chapter 10. 2.3.3 CRIB CRIB stands for Controlled Reversible Inhomogeneous Broadening. In this protocol, a spatially varying electric or magnetic field is applied to the ensemble. This shifts the resonant frequency of the |1i ↔ |2i transition in the atoms, with a frequency shift proportional to the strength of the field. Therefore, depending on their position, some atoms experience a large shift; others a smaller shift, or a negative shift. The net effect is to produce an inhomogeneous broadening of the atomic resonance. That is, each atom has a narrow resonance, but the ensemble as a whole absorbs light over a broad range of frequencies, as determined by the applied field (see Figure 2.9). Storage is accomplished via the general procedure sketched in Figure 2.2 (b). A signal photon is tuned into resonance with the broadened |1i ↔ |2i transition, and is completely absorbed. An important feature of the scheme is that broadband photons can be stored, because the ensemble resonance covers a wider spectral range 2.3 Ensembles 51 than the unbroadened transition. In particular, photons with a temporal duration much shorter than the spontaneous emission lifetime of the state |2i can be stored. Therefore the absorption process is finished long before the atoms have had time to decay, and losses due to spontaneous emission are minimal. Note that so far in the protocol we have not applied any optical control fields, so there are no dressed states — the signal photon has simply been resonantly absorbed. When the absorption is complete, a single atom in the ensemble has been excited into state |2i. The broadening field is now switched off, so that the atoms recover their natural resonance frequencies. Then a control field — a laser pulse tuned to the |2i ↔ |3i transition — transfers the excited atom to the storage state |3i. Storage has now been completed, but the stored excitation is rather mixed up. To understand why, it is easiest to drop our consideration of photons for a moment, and consider the memory as it would be described classically. The atoms can be thought of as a collection of dipoles (separated charges) that are set in motion by an impinging field — the signal field. Due to the applied inhomogeneous broadening, each dipole is oscillating at a slightly different frequency. Therefore, although they all begin oscillating together, driven by the signal pulse, they soon drift out of phase with one another. When the control pulse is applied, the dipoles are essentially frozen, since their motion is transferred to the dark state |3i. In being stored, the signal pulse has had its phase scrambled. That is, information about the time-ofarrival of the signal pulse has been lost. The crucial step in retrieving the signal is to recover this information, by reversing the dephasing process. This is the vital ‘R’ in 2.3 Ensembles 52 CRIB. Because the inhomogeneous broadening is man-made, we are able to switch it around, by changing the polarity of the external field. When the broadening is flipped in this way, every atom that was shifted to a higher frequency is now shifted to a correspondingly lower frequency, and vice versa. This provides us with a way to completely reverse the dynamics of the memory. Retrieval works like this: the control pulse is sent into the ensemble. This unfreezes the stored excitation, transferring the atoms from |3i to |2i. The atomic dipoles are still out of phase, but then the inhomogeneous broadening is re-applied, this time with the reverse polarity; see Figure 2.9 (b). The atoms that were red-shifted at storage are now blue-shifted, and those previously blue-shifted are now red-shifted. The atomic dipoles therefore eventually re-phase. The collective oscillation of the entire ensemble, in phase, then acts as a source for the electric field, and the signal pulse is re-emitted. Retrieval Storage (a) (b) Figure 2.9 CRIB storage. (a): A signal photon is absorbed resonantly by the broadened ensemble, exciting an atom to state |2i. Then a strong control pulse transfers the excited atom to the storage state. (b): To retrieve the stored photon, the control is applied again, and the inhomogeneous broadening is switched on, this time with reversed polarity. The optical dipoles eventually re-phase, and the signal photon is re-emitted. CRIB storage was first proposed by Nilsson and Kröll [78] , following analyses of 2.3 Ensembles 53 a generalized photon echo by Moiseev [79–81] ; the acronym appeared in a proposal by Kraus et al. [82] . The protocol has been implemented experimentally by several groups, all using rare-earth ions doped into solids [83–85] . The reasons for choosing these materials as storage media are manifold. First, use of atoms in the solid state eliminates limitations to the storage time arising from atomic collisions, or from atoms drifting out of the interaction region (see §10.5 in Chapter 10). Second, the rare-earth elements all share a rather peculiar feature in their electronic structure, namely that the radius of the 4f shell is smaller than the radii of the (filled) 5s and 5p shells [86] . The optically active electrons in the 4f shell are therefore shielded by the 5s and 5p shells. These filled shells are spherically symmetric, and can be thought of as metallic spheres that isolate the f electrons from external fields, much as would a Faraday cage. Therefore, even when doped into a solid, rare-earths maintain much of their electronic structure. Perturbations due to the surrounding ‘host’ material may induce frequency shifts, but noise and fluctuations, which might reduce the possible storage time, are effectively eliminated. Optical transitions between different states within the f shell are generally forbidden, since all these states have the same parity (see §4.3.1 in Chapter 4). This means that spontaneous emission from these states is greatly suppressed, making them ideal for use in a quantum memory. Spontaneous lifetimes of several hours have been measured [87] , and photon storage times of up to 30 s are realistic [88] . Of course, if no transitions are allowed at all, no incident photons can ever be absorbed. But one effect of the host is to alter the atomic potential so that the f shell acquires an 2.3 Ensembles 54 admixture of d orbital states, making electric dipole transitions possible [89] . Finally, the 4f 4 I15/2 ↔ 4 I13/2 transition in Erbium has a wavelength of 1.5 µm, which matches the wavelength at which optical fibres used for telecommunications have minimal absorption. Particular attention has therefore been paid to the possibilities for building a solid state quantum memory based on Erbium ions, since it would integrate extremely well with existing telecoms systems. We can distinguish two categories of CRIB, based on the direction of variation of the external field with respect to the propagation direction of the signal (see Figure 2.10). If these directions are perpendicular, so that the atomic frequencies are broadened across the ensemble, we call this tCRIB (transverse CRIB). If they are parallel, with the atoms broadened along the ensemble, we call this lCRIB (longitudinal CRIB). The latter of these is sometimes referred to as GEM (gradient echo memory [84,90] ). As we will see in Chapter 7, the performance of these two schemes for photon storage is essentially the same. (a) (b) Figure 2.10 tCRIB vs. lCRIB. (a): In tCRIB, an external field broadens the ensemble resonance in a direction transverse to the propagation direction of the light to be stored. (b) In lCRIB, the broadening field is applied parallel to the propagation direction. 2.3 Ensembles 2.3.4 55 AFC The atomic frequency comb memory protocol was proposed recently [91] by Afzelius et al.. Their research group in Geneva have focussed on the implementation of CRIB, and AFC takes a number of cues from their experience in this connection. In the AFC protocol, we suppose that it is possible to prepare an atomic ensemble with a large number of absorption lines equally spaced in frequency (see Figure 2.11). This atomic frequency comb plays a similar role to the inhomogeneous broadening in CRIB. It increases the bandwidth of the absorptive resonance, so that a temporally short signal is efficiently absorbed. As with CRIB, this means that the entire absorption process can be completed long before the excited state |2i has any time to decay. Again, for long-term storage, the excitation is mapped to the dark state |3i. The ‘trick’ in the design of the AFC protocol becomes clear at retrieval. Recall that, after transferring the stored excitation back to |2i, the atomic dipoles must be brought back into phase with one another before the signal can be re-emitted. In CRIB this is done by reversing the atomic detunings, and this would certainly work for AFC. But even if the atomic resonances are not altered in any way, the AFC memory still re-emits the signal! The reason is that the atomic frequency comb has a discrete spectral structure. Therefore, the optical polarization undergoes periodic revivals — re-phasings — at a rate given by the frequency of the beat note associated with the comb frequencies. If the frequency separation between adjacent comb teeth is ∆, the time between re-phasings is roughly 1/∆. Having introduced the principle behind AFC storage, a number of comments 2.3 Ensembles 56 (a) Storage (b) Retrieval Figure 2.11 AFC storage. (a): A broadband signal photon, with a bandwidth covering many comb teeth, is absorbed by the atomic frequency comb. The excitation is then transferred to state |3i by a control pulse. (b) To retrieve the signal, the control is re-applied to return the excitation to the frequency comb. The atomic dipoles re-phase, due to the discrete structure of the comb, and the signal is re-emitted. are in order. First, it is not obvious that an ensemble with a comb-like resonance will smoothly absorb the signal photon. One might expect that only those parts of the signal spectrum overlapping with the comb teeth would be absorbed, with the intervening frequencies simply transmitted and lost. In fact, the absorption between the comb teeth never vanishes completely. Provided the ensemble contains enough atoms, the combined absorption over the whole ensemble is enough to store all the frequencies in the signal. Second, it is not obvious how the control field can transfer all the atoms in the frequency comb from state |2i to state |3i and back again. This must be done in AFC so as to prevent the periodic revivals in polarization from re-emitting the signal too early. A sufficiently bright and short control pulse will accomplish the state transfer — the control pulse bandwidth should span the full spectral width of the comb. So-called coherent control can be used to shape the control pulse so as to 2.3 Ensembles 57 maximize the transfer efficiency [92–98] . Third, it is not obvious how to prepare an atomic frequency comb. The procedure suggested by Afzelius et al. is based around an implementation with rare-earths doped into solids. In these materials, natural variations in the position within the host occupied by the rare-earths causes the resonant frequencies to be shifted randomly over a broad spectral range. This natural inhomogeneous broadening is not useful for CRIB, since it cannot be ‘reversed’. Therefore in the CRIB protocol, it is necessary to remove all the atoms except those with the desired resonant frequency, before applying the external field to artificially broaden the resonances of only these atoms. To ‘remove’ undesired atoms, an optical pump is employed [99] (see §10.12 in Chapter 10). This is a series of laser pulses with frequencies tuned so as to transfer unwanted atoms from the ground state |1i into a new state |4i (not shown in any diagrams so far), where they are ‘shelved’ for the duration of the memory protocol. The shelf state can be chosen to have an extremely long lifetime (several hours, as mentioned above in Section 2.3.3); this is why we can consider the atoms as having been simply removed from the ensemble. To prepare the frequency comb for the AFC protocol, a similar optical pumping procedure is used: all atoms with resonant frequencies in between the comb teeth are shelved. A significant practical advantage of AFC over CRIB is now clear. CRIB requires that we pump out all but a single narrow resonance within the ensemble: we ‘throw away’ a lot of atoms. In AFC, we pump out all but N narrow resonances, where N is the number of comb teeth, which may be quite large. Therefore we throw away fewer atoms, and this allows for 2.4 Continuous Variables 58 a much stronger absorption — a much more efficient memory — using an ensemble with the same doping concentration as a less efficient CRIB protocol. In Chapter 7 we will show that AFC is also well-suited to the parallel storage of multiple signal fields, making it attractive for use in quantum repeaters. AFC storage has not yet been demonstrated in its entirety, but proof-of-principle experiments have been performed by de Riedmatten et al. [100] . 2.4 Continuous Variables So far, our discussion of quantum memories has focussed on the storage of single photons. For the purposes of QKD and computation discussed in Chapter 1, the quantum information encoded into these photons is of a discrete nature. Either the number of photons, or their polarization, can be used to represent the two basis states of a qubit. This is not the only paradigm for quantum information processing, however. Photons possess other degrees of freedom that are not discrete. The position of a photon, or its momentum, for example. These quantities don’t lend themselves to representation in terms of qubit states. But it is still possible to use such continous variables to encode information (as is done in classical analogue computing). And the specifically quantum features of qubits that confer their increased computational power — superposition; entanglement — all carry over to continuous variables. Therefore quantum computing protocols [101,102] , and indeed QKD protocols [103–106] , that take advantage of the continuous degrees of freedom possessed by photons, have all been developed. In general their relationship to the equivalent 2.4 Continuous Variables 59 qubit-based protocols is similar to that between analogue and digital classical computation. On the one hand, the continuous versions are robust to noise, in the sense that information is not completely erased in the presence of distortions. On the other hand, below a certain threshold, digital algorithms can be made essentially impervious to noise, producing ‘perfect’ outputs, whereas analogue computations are always subject to fluctuations — they never work perfectly. Anyone who has ever witnessed the catastrophic failure of ‘crisp and clear’ digital television with a poor signal will appreciate the fuzzy watchability of analogue television in the same conditions. In this section we briefly introduce the concepts of continuous variables, so as to understand a class of memory based on continuous variables storage. The most commonly used variables are the field quadratures. These are defined in terms of the electric field E associated with an optical mode, E = x cos(ωt) + p sin(ωt), (2.1) where ω is the optical carrier frequency, and t is the time. The coefficients x and p for the amplitude of the field are known as the in-phase and out-of-phase field quadratures, respectively [107] . By convention, the same symbols as would normally be associated with position and momentum are used, and this is motivated by their formal similarity. Like the position and momentum of a classical pendulum, the quadratures are continuous variables that describe the oscillation of the field. Even 2.4 Continuous Variables 60 more strikingly, in quantum mechanics, x and p are complementary in the same way as are the position and momentum of a harmonic oscillator. That is to say, a measurement of one quadrature ‘disturbs’ the other, so that it is impossible to precisely measure both simultaneously. In appropriately scaled units, the Heisenberg uncertainty principle applies [103] : ∆x∆p ≥ 1, (2.2) where ∆x, ∆p are the precisions with which the quadratures are simultaneously known. Just as in classical physics, the quantum state of an optical field can be completely described in terms of x and p. Due to the uncertainty principle (2.2), the state is not a single point in (x,p)-space, or phase space as it is more often called, but rather a kind of ‘blob’, whose spread represents the uncertainties ∆x and ∆p. This blob is known as the Wigner distribution [108] , and it is the representation of choice for states parameterized by continuous variables. The most ‘normal’ state of an optical field — the coherent state — is a monochromatic beam, like that produced by a laser. It is completely classical, in the sense that the formalism of quantum mechanics is not required to describe it. Classical electromagnetism is sufficient. Its Wigner distribution is a Gaussian ‘hill’, with equal uncertainties in x and p (see Figure 2.12 (a)). Clearly this is not the only type of distribution compatible with 2.2. Figure 2.12 (b) shows an example of a squeezed state, with a small uncertainty in x. The price for this increased precision in x is that the 2.4 Continuous Variables 61 uncertainty in p grows, but states like this can be extremely useful for reducing the noise on measurements associated with just one of the quadratures. Squeezed states of light would appear to be rather exotic, but they arise naturally in any process that converts one frequency into another. Such non-linear processes only require that the potential in which optically active electrons move is not exactly quadratic1 in the electrons’ displacement. This happens to some extent in nearly all materials, and the technology for generating efficient squeezing is now rather well developed [109–112] . (a) (b) Figure 2.12 Wigner distributions in quadrature phase space for two different states of an optical mode. (a): A classical coherent state, with equal uncertainties in x and p. (b) A squeezed state, with ∆x halved, but ∆p doubled. The lengths of the dotted lines give the brightness of the states; their angles give their relative phases. Research into squeezed light is being actively pursued as a means to improve the sensitivity of gravitational wave detectors [113,114] , and to toughen the noise tolerance of communication systems [115] . As might be expected from the ubiquity of uncer1 If the potential is not quadratic (i.e. anharmonic), the ‘restoring force’ on the electrons is non-linear. 2.4 Continuous Variables 62 tainty relations like (2.2) in quantum mechanics, the phenomenon of squeezing is not confined to light, and their are metrological benefits that accrue if atoms can be squeezed [116] . Increased sensitivity to magnetic fields [117] , enhanced spectroscopic precision [118] , and better atomic clocks [119–121] , as well as quantum computation and communication, are all enabled by the ability to manipulate non-classical states of light and matter. One of the most successful demonstrations of quantum memory has grown out of research in this area. The mechanism by which storage is accomplished is most transparent when couched in the language of continuous variables. The memory works by transferring the quadratures of a light beam into an ensemble of atoms, where the quadratures X, P associated with the atoms are their ‘coronal’ angular momenta Jx = X, Jy = P (see Figure 2.13). The storage interaction involves two steps. In the first, a control and signal beam — both tuned far off-resonance — are directed through an atomic ensemble. A strong magnetic field is applied to the ensemble, which aligns the atomic spins, so that all the atoms are initially in state |1i (see Figure 2.15). The control is polarized parallel to the field, so that it does not induce a turning moment. On the other hand, the signal is polarized perpendicular to the field. Two-photon Raman transitions, involving both the control and the signal, can therefore change the z-component of the collective atomic angular momentum, transferring atoms from |1i to |3i or vice versa. In terms of the atomic quadratures, 2.4 Continuous Variables 63 the passage of the optical fields through the ensemble induces the transformation e = X + p, X→X (2.3) where the tilde denotes the quadrature value at the end of the interaction. At the same time, the x quadrature of the light is also shunted — a manifestation of the Faraday effect, x→x e = x + P. (2.4) The other quadratures p, P of both the signal and the atoms are unchanged. For this reason the interaction is known as a quantum non-demolition (QND) measurement, since information about the p quadrature of the signal is transferred to the atoms, without altering it. In the second step of the storage protocol, the x quadrature of the signal is measured using balanced homodyne detection [107] . As shown in Figure 2.14, a polarizing beamsplitter, aligned in the 45-degree basis, mixes the control and signal beams. The difference in the intensities detected at the two output ports of the beamsplitter is directly proportional to x. This measurement result is then used to determine the strength of a radio frequency pulse that applies a controlled torque to the atoms, producing a shift P → Pe = P − x e = P − (x + P ) = −x. (2.5) The two maps (2.3) and (2.5) taken together almost constitute an ideal memory. 2.4 Continuous Variables 64 Figure 2.13 Atomic quadratures. In the presence of a strong magnetic field, the atomic spins in an ensemble align with the z-axis. The quadratures X and P for the atoms are then given by small displacements of their angular momenta in the coronal plane, normal to the z-axis. Apart from an unimportant minus sign in (2.5), the only difficulty is the presence of the initial atomic quadrature X in (2.3). The average value of X can be made to vanish, but the finite spread of the initial atomic Wigner distribution still introduces unwanted fluctuations. This spread can be reduced by squeezing the Wigner distribution of the atoms, so that ∆X → 0, before attempting the memory, and schemes for doing this are in development [116] . The above type of continuous variables memory, based on a QND interaction followed by measurement and feedback, was first implemented by Julsgaard et al. at the Niels Bohr intitute in Copenhagen [122] using an ensemble of cesium atoms. The same research group, led by Eugene Polzik, has since refined and extended the technology, and have demonstrated quantum teleportation of light onto atoms, deterministic entanglement generation, efficient spin squeezing of atomic ensembles 2.4 Continuous Variables 65 Figure 2.14 QND memory. A magnetic field aligns the atomic spins in an ensemble. A strong control, polarized parallel, and a weak signal, polarized perpendicular to the spins, are sent through the atoms. The control cannot rotate the atomic spins, but the signal does. A homodyne measurement extracts the resulting x quadrature of the signal, and a radio frequency pulse, generated by the coils shown, applies a final rotation. and many other continuous variables protocols. An excellent review of their progress in this area can be found in the recent review article by Hammerer et al [123] . The conceptual shift between the above description in terms of continuous variables, and the rather intuitive picture of ‘reversible absorption’ we have used for all the other memory protocols, makes comparisons difficult. Certainly the optimization of this protocol does not fit into the general scheme we apply to the optimization of the other memory protocols in this thesis. The scheme has enjoyed considerable success, despite its technical complexity, probably due in large part to the experience and prowess of those working at the Niels Bohr institute. Nonetheless, it does not perform as well as the other protocols as a component of a DLCZ-type quantum repeater. This is because, as described in Chapter 1, the entanglement purification 2.4 Continuous Variables 66 Figure 2.15 Level scheme for a QND memory. The atoms begin with their spins aligned with the external magnetic field, in state |1i. State |3i cannot be reached by interaction with the control alone (green arrows), since it is π-polarized — parallel to the B-field — and it cannot induce a spin flip (see Figure F.4 in Appendix F). Raman transitions involving both the control and signal (blue wavy arrows) can transfer atoms to state |3i. The intermediate excited states are collectively labelled |2i. Both types of Raman interaction are involved: Stokes scattering, as in the DLCZ protocol (see Figure 1.6 in Chapter 1), and the anti-Stokes scattering used in the Raman memory protocol (see Figure 2.8). in these repeaters is effective against photon loss, but not against photon addition, and the Raman transitions shown in Figure 2.15 can produce extra photons that contaminate the number state entanglement. Other applications for which photon loss is particularly damaging would benefit from a continuous variables memory, since loss can be essentially eliminated by increasing the power of the RF pulses applied in the feedback step, at the expense of distorting the stored state. In any case, we will not discuss continuous variables memory further. In the next chapter we introduce the optimization scheme relevant for the absorptionbased memories we have discussed. Chapter 3 Optimization Here we introduce the optimization scheme that we will apply to the Raman, EIT, CRIB and AFC ensemble memory protocols. In all these protocols, an input field, the signal field, is transferred to a stationary excitation inside an atomic ensemble. The aim of the optimization is to maximize the efficiency of this transfer, which amounts to maximizing the amount of stored excitation, given a fixed input. Fortunately, a technique borrowed from the mathematical toolbox of linear algebra makes this optimization extremely easy to perform. Suppose that the signal field is a pulse, with a time-dependent amplitude A(τ ), where τ is the time. The action of the memory is to absorb this input pulse, and convert the incident energy into some kind of long-lasting excitation within the ensemble (see Figure 3.1). We’ll denote the amplitude of this excitation by B(z), where z is the position along the ensemble. Here the z dependence allows for the possibility that the excitation may be distributed over the length of the ensemble 68 in a non-uniform way. This kind of collective, de-localized excitation is typically referred to as a spin wave, since in many cases the atomic states involved are states of different spin angular momentum (as in the QND memory discussed at the end of Chapter 2). We’ll use this term indiscriminately, to describe any distributed excitation relevant to quantum storage, regardless of the nature of the quantum numbers associated with the atomic states. In the next chapter, we’ll define the spin wave more precisely. For the purposes of optimization, it is sufficient to understand the spin wave as the stationary counterpart of the propagating signal field. That is, the storage process is a mapping A → B, and the retrieval process is the inverse map B → A. Figure 3.1 Storage map. An incident signal field A(τ ) is mapped to a stationary spin wave B(z) within an atomic ensemble. If the signal field initially contains NA = N photons, we would ideally like the spin wave at the end of the storage process to contain the same number of excited atoms NB = N , so that all the input light has been stored. In practice, some of the input light will pass through the ensemble and be lost, and some of the excited atoms will decay back down to their ground state, or drift out of the interaction region. The efficiency η of the storage interaction is simply the ratio of the number 69 of excited atoms to the number of input photons, η = NB /NA , where the number of quanta, either material of optical, is found by integrating the squared norm of the relevant amplitude, Z ∞ NA = |A(τ )|2 dτ, −∞ Z NB = L |B(z)|2 dz. (3.1) 0 Here L denotes the length of the ensemble. We note that in this type of memory there is no process that can produce extra excitations. At least in principle, there is no source of background noise. The failure mode of the memory is photon loss: photons directed into the memory are not stored, and so are not recovered in the retrieval process. This differs from the QND memory mentioned in the previous chapter, which need not suffer from photon loss, but which may introduce noise at retrieval. Therefore the performance of this type of memory is optimal when the efficiency is maximized. No other figure of merit, involving the suppression of atomic fluctuations for instance, is relevant. A key requirement of a quantum memory is linearity. That is, suppose that we store a signal field A = αA1 + βA2 built from two contributions A1 and A2 . This is a superposition state, like (1.1) in Chapter 1, and to preserve its encoded quantum information, the coefficients α, β should not be altered by the storage process. The resulting spin wave should be of the form B = αB1 +βB2 , where B1 , B2 are the spin waves generated by storage of A1 , A2 only. This property, required to faithfully store superpositions, restricts the storage map A → B to be a linear map (and similarly 3.1 The Singular Value Decomposition 70 for the retrieval process). Fortunately, as we will see, all the memory protocols we will analyze are indeed linear. In general, then, we can always write the storage map in the following way Z ∞ B(z) = K(z, τ )A(τ ) dτ. (3.2) −∞ The integral kernel K is known as the Green’s function, or the propagator for the storage interaction. It contains all the information about how the memory behaves. In the next chapter we will show how to derive expressions for K in some cases. Generally, it is possible to construct the form of K numerically, as we will show in Chapter 5 (see §5.4). For the moment, we suppose that we are able to write the memory interaction in the form (3.2). It is clear that the efficiency of the memory depends on achieving a good ‘match’ between K and the shape, in time, of A. For instance, no storage is possible if K ∼ 0 during the arrival time of the signal. K and A should ‘overlap’. Optimizing a quantum memory involves finding the shape of A that maximizes this overlap, so that all of A ends up in B. In the next section we introduce the singular value decomposition, a valuable analytical tool that can be used to find this optimal shape, and more besides. 3.1 The Singular Value Decomposition The SVD is most commonly encountered in the context of matrices. It is one of the most useful results in linear algebra, and it finds applications from face recognition [124] to weather prediction [125] . Under the name Schmidt decomposition it is 3.1 The Singular Value Decomposition 71 of critical importance in quantum information theory as a tool for the analysis of bi-partite entanglement [126] . It seems to have been discovered independently several times in the 19th century [127–129] . Erhard Schmidt applied the decomposition to integral operators (an example of which is (3.2)) in 1907 [130] , and Bateman coined the term ‘singular values’ in 1908 [131] . The proof of the decomposition for arbitrary matrices was given in 1936 by Eckart and Young [132] . At the heart of the SVD is a geometric interpretation for the action of a linear operator. A linear operator, or linear map, takes some vector a as input, and spits out another vector b as output. Representing the linear operator as a matrix M , we have b = M a. (3.3) Looking at Figure 3.2 (a), it is clear that such a transformation is equivalent to the following procedure. (i) rotate a until it lies along one of the coordinate axes (ii) re-scale the axes so that the length of the rotated version of a matches the length of b (iii) rotate this re-scaled vector until it sits on top of b. This suggests that it should be possible to decompose an arbitrary linear transformation by combining rotations and coordinate re-scaling. The SVD is nothing more than this representation of a general linear map as a rotation, a re-scaling, and a final rotation. An example is shown in Figure 3.2 (b), where the action of M is shown on the set {a} of all vectors with a certain length. The tips of these vectors trace out the surface of a sphere. The effect of M is to ‘squish’ this sphere into an ellipsoid. This has to be done in such a way that the black dot on the sphere ends up as the red dot on the ellipsoid. 3.1 The Singular Value Decomposition (a) 72 (b) Figure 3.2 Linear transformation. (a): a vector a is mapped onto a vector b by M , which may be seen as implementing a rotation of a onto the x-axis, followed by a coordinate re-scaling to increase the length of a until it matches the length of b, followed by a rotation onto b. (b): The action of M on the set of all initial vectors {a} (the grey sphere) produces a set of vectors {b} (the red ellipsoid). As an example, the red dot is the image of the black dot under under M . The final ellipsoid can be generated from the initial sphere by: rotating the sphere, re-scaling the axes and then rotating again. To produce the ellipsoid from the sphere, it is necessary to first rotate the sphere until the black dot is placed appropriately. Then the x, y and z axes are re-scaled to deform the rotated sphere into the ellipsoid ‘shape’. Finally, a second rotation puts the ellipsoid in the correct orientation, with the black dot sitting on top of the red dot. The above discussion is limited to the intuitive case of a real vector in three dimensions being mapped to another real vector in three dimensions. In fact, the SVD exists for all matrices M — all linear transformations. That is, M could be a complex rectangular matrix of any size, that maps a complex n-dimensional vector to a complex m-dimensional vector. In general, M can always be written in the form M = U DV † . (3.4) 3.1 The Singular Value Decomposition 73 Here U and V are both unitary matrices, and D is a real, positive, diagonal matrix. The properties of these types of matrices are reviewed in Appendix A. In terms of the discussion above, V † represents a rotation into a new coordinate system, D represents a re-scaling of this new coordinate system, and U represents a final coordinate rotation. The elements of D, lying along its diagonal, are the factors by which the coordinates are re-scaled in the second step. They are known as the singular values of M , and they contain a great deal of useful information about the transformation represented by M . Geometrically, the singular values correspond to the lengths of the semi-axes of the ellipsoid in Figure 3.2 (b). In this thesis, singular values are generally denoted by the symbol λ, the same as eigenvalues. It should always be clear from the context what is meant. By convention, the singular values, which are all positive, real numbers, are ordered in descending magnitude, so that D11 = λ1 is the largest singular value, D22 = λ2 is smaller, and so on. It is sometimes useful to visualize the structure of the matrices in the SVD, and so for reference we provide the following tableau, λ1 M = |u1 i . . . |um i | {z U }| .. . λm {z D }| hv1 | hvm | . {z } .. . V† (3.5) 3.1 The Singular Value Decomposition 74 Performing the matrix multiplications explicitly, we obtain M= X λ j Mj , (3.6) j where each of the matrices Mj = |uj ihvj | is an outer product of the j th columns of U and V (see Section A.2.1 in Appendix A). This last representation provides a natural way to interpret the action of M , as a sum of independent mappings from the column-space of V to the column-space of U . To understand why this is useful, recall that the sets of column vectors {|vj i} and {|uj i} of U and V are both orthonormal bases, because U and V are unitary (see Section A.4.5 in Appendix A). Therefore any vector |ai to which M is applied can be written in the coordinate system defined by the {|vj i}, and the result |bi can always be written in terms of the {|uj i}, a = |ai = a1 |v1 i+a2 |v2 i+. . .+an |vn i, and b = |bi = b1 |u1 i+b2 |u2 i+. . .+bm |um i. (3.7) Applying (3.6) to |ai, and making use of the orthonormality of the u’s and v’s, hui |uj i = hvi |vj i = δij , we find that b1 = λ1 a1 , b2 = λ2 a2 , etc . . . . (3.8) So M can be viewed as a set of mappings between two special coordinate systems, each with a different ‘fidelity’, given by the singular values. M may seem like 3.1 The Singular Value Decomposition 75 a complicated transformation, but as long as we use the {|vj i} to define our input coordinate system, and the {|uj i} to define our output coordinate system, the action of M is always extremely simple. It just maps each coordinate from the input onto the corresponding output coordinate, re-scaled by the corresponding singular value. 3.1.1 Unitary invariance If the input coordinate system is rotated, before performing the SVD, this cannot change the singular values, but only the input basis that we should use. The same is true if the output coordinate system is rotated. More generally, consider applying some unitary transformation W to M . The SVD of this compound operator is then e DV † , W M = W U DV † = U (3.9) e = W U is a unitary matrix, since both W and U are unitary. The output where U basis vectors are modified by W , but the singular values — the elements of D — are unchanged. By the same token, the product M W also has the same singular values as M , but V must be replaced by W † V . Sometimes it is only possible to find rotated versions of a matrix M , in which case this property of the SVD is useful. 3.1.2 Connection with Eigenvalues The SVD is connected with the eigenvalues of the ‘square’ of M . More precisely, consider the normally and antinormally ordered products KN = M † M and KA = M M † . † † These two products are both Hermitian, since KN = KN and KA = KA . Therefore, 3.1 The Singular Value Decomposition 76 they both have real eigenvalues, and their eigenvectors each form orthonormal bases (see Section A.4.3 in Appendix A for a derivation of this fact). And inserting the decomposition (3.4), we see that KN = V D2 V † , and KA = U D2 U † , (3.10) where we have used the relations U † U = V † V = I, which must hold for unitary matrices. Therefore, the eigenvalues of KN and KA are both given by the squares of the singular values. 3.1.3 Hermitian SVD Note also that if M is a Hermitian matrix, with M = M † , we must have that U = V . That is, M = U DU † , (3.11) which is precisely the spectral decomposition of M , with the eigenvalues of M lying along the diagonal of D, and the eigenvectors of M making up the columns of U . So the SVD and the spectral decomposition of M are identical for Hermitian matrices. 3.1.4 Persymmetry Another case, which we will encounter in our treatment of Raman storage, is that of persymmetry. Suppose that M is a real, square matrix, with n = m. Then M is persymmetric if it is symmetric under reflection in its anti-diagonal — the diagonal 3.1 The Singular Value Decomposition 77 running from bottom left to top right, as shown in Figure 3.3. M = M P , where (M P )ij = Mm−j+1 m−i+1 . This is rather an unusual type of symmetry, and it is Figure 3.3 Persymmetry. The upper left and lower right portions of a real persymmetric matrix are mirror images of eachother under reflection in the anti-diagonal, represented as a grey stripe. not often discussed in textbooks. But the action of a persymmetric matrix can be viewed as very similar to that of a real Hermitian matrix, the only difference being that the result is ‘flipped around’. To see the implications of persymmetry for the SVD, let us write M = XH, where H is a real Hermitian matrix (i.e. symmetric under reflection in its main diagonal), and where X is a ‘flip’ matrix, with ones along its anti-diagonal, and zeros everywhere else (left blank for clarity below), X= 1 . .. 1 1 . (3.12) The action of X when multiplying a matrix is to flip it around a horizontal axis, so that its last row becomes its first, and vice versa. For every persymmetric matrix M , there must always be some Hermitian matrix H such that M = XH. The property 3.2 Norm maximization 78 of persymmetry can also be easily written in terms of X. A little mental acrobatics will verify that M P = XM † X, so that persymmetry requires M = XM † X. Inserting the spectral decomposition of H, we obtain e DU † , M = XM † X = X(XH)† X = XH † X † X = XHX 2 = XU DU † = U (3.13) where we used the Hermiticity of H and X, along with the fact that X 2 = I (two e = XU is a unitary matrix whose horizontal flips cancel each other out), and where U columns have been flipped round. Therefore the SVD of a persymmetric matrix is such that the columns of U , the basis for the output coordinate system, are flipped versions of the columns of V , the input basis. This result will be of use to us in Chapter 5. 3.2 Norm maximization Suppose that we would like to know how to choose a in order to maximize the norm (the length) of b. Since M is a linear transformation we can always increase the norm of b just by increasing the norm of a: if we double the length of a, the length of b also doubles. But this is not interesting. Clearly the direction of a matters; some directions will result in a larger norm for b. If the norm of a is fixed, how should we choose its direction? This question is easily answered if we are able to compute the SVD of M . Without losing generality, let ||a|| = a = 1. That is, 3.3 Continuous maps 79 |a1 |2 + |a2 |2 + . . . + |an |2 =1. Using (3.8), the norm of b is then given by b2 = |b1 |2 + |b2 |2 + . . . + |bm |2 = λ21 |a1 |2 + λ22 |a2 |2 + . . . λ2n |an |2 . (3.14) But by definition, λ1 is the largest of the singular values, so that b2 < λ21 (|a1 |2 + |a2 |2 + . . . + |an |2 ) = λ21 , (3.15) That is, the largest possible norm of b is b = λ1 . This maximum norm is obtained by choosing a1 = 1, with all other components vanishing. So the ‘optimal’ vector a, with regard to maximizing b, is a = v1 (or |ai = |v1 i). From the perspective of Figure 3.2 (b), this amounts to choosing a so that the resulting b lies along the largest semi-axis of the ellipsoid generated by M . 3.3 Continuous maps The preceding discussion of matrices can be extended, without essential modification, to the continuous map (3.2) describing the storage interaction in a quantum memory. The Green’s function K(z, τ ) has the same basic structure as a matrix, except that it has two continuous arguments, instead of two discrete indices. And it is also amenable to the SVD. The expression (3.6) for a matrix becomes, in the 3.3 Continuous maps 80 continuous case K(z, τ ) = X λj ψj (z)φ∗j (τ ). (3.16) j As before, the λ’s are the singular values. The functions ψj and φj are the continuous analogues of the input and output basis vectors |uj i and |vj i. We will refer to them as modes; φ1 is the first input mode, ψ2 is the second output mode, and so on. Alternatively, since the signal field and spin wave both appear in (3.2), it may be physically more meaningful to talk of the φ’s as the signal modes, and the ψ’s as spin wave modes. The inner product between two vectors a, b takes the form of an overlap integral, for continuous functions a(x), b(x), † a b = ha|bi = X a∗i bi Z −→ a∗ (x)b(x) dx. (3.17) i The orthonormality conditions hui |uj i = hvi |vj i = δij for basis vectors are therefore replaced by the relations Z 0 L ψi∗ (z)ψj (z) dz = Z ∞ −∞ φ∗i (τ )φj (τ ) dτ = δij . (3.18) These conditions tend to produce sets of ‘wiggly’ functions, so that the product of two different modes alternates between positive and negative values, and integrates to zero. Sets of oscillating modes, satisfying orthonormality conditions like (3.18), are common in harmonic analysis and acoustics; they are also the bread and butter of much of quantum physics — quantum memories are no exception. In general, the 3.3 Continuous maps 81 first modes, associated with the largest singular value λ1 , are slowly varying. They represent the basic ‘shape’ of the Green’s function, with broad brushstrokes. Higher modes tend to oscillate faster, and they represent smaller corrections, at increasingly fine levels of detail. 3.3.1 Normally and Anti-normally ordered kernels. The continuous analogues of the matrices KN and KA in (3.10) are found by integrating the product of two K kernels over one of their arguments, 0 KN (τ, τ ) = KA (z, z 0 ) = Z L K ∗ (z, τ )K(z, τ 0 ) dz, Z0 ∞ K(z, τ )K ∗ (z 0 , τ ) dτ. (3.19) −∞ These two kernels share the same eigenvalues; they satisfy the eigenvalue equations Z ∞ −∞ Z L KN (τ, τ 0 )φj (τ 0 ) dτ 0 = λj φj (τ 0 ), KA (z, z 0 )ψj (z 0 ) dz 0 = λj ψj (z 0 ). (3.20) 0 Sometimes these kernels are more convenient to work with than the kernel K. 3.3.2 Memory Optimization. Now it is clear how to optimize the performance of a quantum memory. We want the largest efficiency η = NB /NA . But from the expressions in (3.1), we see that NB is just the continuous analogue of b2 , the squared norm of the output spin wave. 3.3 Continuous maps 82 Suppose that we fix NA = 1, just as we fixed a2 = 1 in Section 3.2. How should we choose the signal field A(τ ) to maximize η? The answer carries over directly from our considerations of matrices. We should choose A(τ ) = φ1 (τ ). With this choice, the resulting spin wave is B(z) = λ1 ψ1 (z), and the optimal storage efficiency is η = λ21 . In practice, the SVD of the Green’s function is almost always computed numerically, and to do this, the continuous function K is discretized on a fine grid: it is converted into a matrix. Therefore our treatment of matrices applies directly when calculating the singular values, and the mode functions, for a quantum memory. Note that, since we must of course have η ≤ 1, any physical Green’s function will have λ1 ≤ 1, and this condition is a useful check that K has been sampled with a sufficiently fine grid. 3.3.3 Unitary invariance When deriving a form for the Green’s function K, it will sometimes be easier to work in terms of a transformed coordinate system. Suppose that we introduce a new variable y = y(z), where y(z) is some monotonic single-valued (i.e. invertible) function. The Green’s function, expressed in terms of y, has the same singular values as in the original coordinates, provided that we make the transformation unitary by including a Jacobian factor: 1 K(y, τ ) = K [z(y), τ ] × p , J(y) (3.21) 3.3 Continuous maps 83 where z(y) is the inverse transformation, relating z to y, and where J(y) = ∂z y is the Jacobian, relating the line elements dy and dz. Including the factor involving J ensures that the norm of K is preserved in the new coordinate system, Z |K(y, τ )|2 dy = Z |K [z(y), τ ] |2 × Z |K(z, τ )|2 dz. = ∂z dy ∂y The output modes transform accordingly, ψj [z(y)] . ψj (y) = p J(y) (3.22) An identical procedure is used when transforming the time coordinate. Instead of re-parameterizing the kernel, we may wish to transform to a frequency representation, instead of a temporal one. That is, we might be able to work out an e expression for the Fourier transformed Green’s function K e ω) = √1 K(z, 2π Z K(z, τ )eiωτ dτ. (3.23) Inserting the SVD expansion (3.16) into (3.23), we obtain e ω) = K(z, X λj ψj (z)φe∗j (ω), (3.24) j where the modes φej are Fourier transforms of the φj . These Fourier transformed 3.3 Continuous maps 84 modes also form an orthonormal set, Z Z Z Z 1 ∗ −iωτ 0 iωτ 0 0 φi (τ )e dτ φj (τ )e dτ dω 2π Z Z = δ(τ − τ 0 )φ∗i (τ )φj (τ 0 ) dτ 0 dτ φe∗i (ω)φej (ω) dω = = δij , (3.25) where in the penultimate step we used the plane-wave expansion of the Dirac delta function (see §D.2 in Appendix D), 1 δ(x) = 2π Z e±ixy dy. (3.26) e The expression (3.24) is therefore precisely the SVD of the transformed kernel K. e are the same as those of K. In fact the temporal Fourier The singular values of K transform we applied is just an example of a unitary transformation; unitary because, by Parseval’s theorem, it is norm-preserving. That is to say, Z 2 |f (τ )| dτ = Z |fe(ω)|2 dω, (3.27) for any function f and its Fourier transform fe. As described in Section 3.1.1, the singular values of a mapping are never altered by a unitary transformation. Other useful possibilities include Fourier transforming over the spatial variable, 3.4 Optimizing storage followed by retrieval 85 to find the Green’s function in so-called k-space, 1 e K(k, τ) = √ 2π Z K(z, τ )eikz dz. (3.28) Again the singular values are the same as those for K, but the output modes are k-space versions of the ψj . Generally the integration limits in these Fourier transforms would be [−∞, ∞], but when boundary conditions are needed, it will sometimes be useful instead to implement a unilateral transform, where the integration runs from 0 to ∞. The utility of these techniques will become clear when we examine the memory interaction more closely in the next chapter. A brief review of both the bilateral and unilateral Fourier transforms can be found in Appendix D. 3.4 Optimizing storage followed by retrieval Suppose that we are not interested in the efficiency of storage alone, but rather the combined efficiency of storage into, followed by retrieval from the memory. In many situations it is this combined efficiency that is most experimentally relevant. The same techniques as outlined above are directly applicable. The entire memory interaction can be characterized as a map between the input and the output signal fields, Z ∞ Aout (τ ) = −∞ Ktotal (τ, τ 0 )Ain (τ 0 ) dτ 0 . (3.29) 3.5 A Simple Example 86 The total efficiency of the memory is given by the ratio of the norms of Aout and Ain . The input mode that maximizes this efficiency is therefore found from the SVD of the Green’s function Ktotal , and the resulting optimal efficiency is given by ηcombined = λ21 , where λ1 is the largest singular value of Ktotal . If we neglect any decoherence of the spin wave (as we do throughout this thesis), the kernel Ktotal can be constructed from the kernels describing the storage and retrieval processes individually. Under certain circumstances — when the retrieval process is precisely the time-reverse of the storage process [133] — the combined kernel Ktotal is equal to the normally ordered kernel KN defined in (3.19), and ηcombined = η 2 . This is the optimal situation, and in general a mismatch between the storage and retrieval processes reduces the memory efficiency so that ηcombined < η 2 . These issues are explored in more detail in Chapter 6. 3.5 A Simple Example Before embarking on a detailed derivation of the equations of motion for an ensemble quantum memory, we run through a simple example that contains many of the features that emerge from a more rigorous analysis. We consider a classical optical signal pulse propagating through an ensemble of identical atoms. We use a classical Lorentz model for the atoms [134] , in which the electric field of the light pulse ‘pulls’ on an electron in each atom, while a harmonic restoring force ‘pulls back’, keeping the electrons bound around their equilibrium positions. Let the average displacement at time τ , away from equilibrium, of an 3.5 A Simple Example 87 electron located at position z, be given by x(z, τ ). The average is taken over all the atoms with the same position coordinate z, all of which behave identically. The restoring force on each electron is given by Frestore = −kx, for some constant k. This is a good approximation for any kind of restoring force, provided the typical displacement x is small enough. The force on an electron due to the electric field E of the signal pulse is Flight = eE, where e is the electronic charge. The classical equation of motion for x is then given by Newton’s second law, Ftotal = ma, where m is the electronic mass and a = ∂τ2 x is the acceleration of the electron. Putting this together yields the equation m∂τ2 + k x = eE. (3.30) We can write the signal field as E = Aeiωs τ , where ωs is the optical carrier frequency of the signal, and where A is a slow modulation describing the temporal profile of the pulse. Irradiating the atoms will produce a response with a similar temporal structure, so we write x = Beiωs τ , where B is some slowly varying envelope. Substituting this into the equation of motion, and neglecting the term ∂τ2 B, gives the equation ∂τ B = −iαA, where α = e 2mωs . (3.31) Here for simplicity we have assumed that the signal field frequency is tuned perfectly into resonance with the atoms, so that ωs = p k/m. This equation describes the response of the atoms to the field; to complete the picture, we would 3.5 A Simple Example 88 like to understand how the field responds to the atoms. As a first step, we apply a (bilateral) Fourier transform from τ −→ ω, so that (3.31) becomes e = α A, e B ω (3.32) e where the tildes denote the transformed variables. The electronic displacement B acts as a source for the signal field. More precisely, each oscillating electron generates e an electric field E(r) = ae/4π0 c2 r in proportion to its acceleration1 . Here 0 as usual denotes the permittivity of free space; r is the distance from the electron. Summing the contributions from all the atoms in a thin slice of the ensemble with thickness δz, we find the total field generated by the dipoles is etot = −iδz ωs en x e, E 4π0 c (3.33) where n is the atomic number density [137] . If we consider the propagation of the signal through a thin slice of the ensemble of thickness δz, we therefore have e + δz, ω) = A(z, e ω) − iδz ωs en B(z, e ω) A(z 4π0 c βδz e = 1−i A(z, ω), ω (3.34) 1 This can be derived rather neatly from relativistic equivalence. The electrostatic potential energy V = e2 /4π0 r of two electrons separated by r produces a relativistic mass increase M = V /c2 of the pair of electrons. The extra weight 12 gM of one of the electrons, in a gravitational field g, must have its origin in a vertical force eE exerted by the electric field E of the other electron. The equivalence principle demands that we cannot tell if the field g is swapped for an acceleration a. Putting this all together, we find the field due to an accelerating charge is E ∼ ae/4π0 c2 r [135,136] . 3.5 A Simple Example where β = e2 n 8π0 mc . 89 Taking the limit as δz −→ 0, we derive Beer’s law of exponential absorption for the signal field, with absorption coefficient β, βδz z/δz e 1−i Ain (ω) δz→0 ω e ω) = A(z, lim ein (ω), = e−iβz/ω A (3.35) ein (ω) = A(z e = 0, ω) is the initial spectrum of the signal field, at the start of where A the ensemble [138] . Substituting this into (3.32) gives an expression for the average electronic displacement in the Fourier domain, −iβz/ω e e B(ω, z) = α ω ein (ω). ×A (3.36) Using the result (D.40) from §D.5 in Appendix D, along with the convolution theorem (D.21), we can take the inverse Fourier transform of this to obtain the map Z Bout (z) = −iα T h p i J0 2 βz(T − τ ) Ain (τ ) dτ 0 , (3.37) 0 where Bout (z) = B(τ = T, z) is the electronic displacement at time τ = T , with T some time chosen to define the end of the interaction, after the signal pulse has passed through the ensemble. Here J0 is a zero’th order ordinary Bessel function of the first kind — we will encounter this function, through similar inverse Fourier transforms, frequently in Chapter 5. We have arranged for the notation in (3.37) to appear suggestive of the storage map described at the start of this Chapter. If we 3.5 A Simple Example 90 identify the electronic displacement Bout as the amplitude of a spin wave excited by the signal field, then the storage kernel K in (3.2) can be identified with the Bessel function appearing in the integrand of (3.37). A numerical SVD of the Green’s function h p i K(z, τ ) = −iαJ0 2 βz(T − τ ) (3.38) would reveal the temporal shape of Ain that maximizes the degree of atomic excitation. We have not taken care to normalize the signal field A, or the electronic displacement B, and so (3.1) is not quite true, using the current definitions. But up to some constant normalization, the optimal efficiency of a quantum memory described by the above model would be provided by squaring the largest singular value of (3.38). Our purpose in the above exposition has not been to describe a real quantum memory in any quantitative detail. But this simple example contains many of the ingredients we will encounter in the following Chapters. We have used a one dimensional propagation model, combined with some atomic physics, to obtain two equations of motion — the first describing the atomic response to the field, and the second describing the influence on the field due to the induced atomic polarization. The solution was found using a Fourier transform, and the result took the form of the storage map (3.2), with the kernel given by a Bessel function. This story applies equally to the fuller treatment given shortly. Finally, note that we have made no use of the formalism of quantum mechanics so far. Of course, a correct description of the atoms at least requires that we quantize the electronic energies. But the 3.5 A Simple Example 91 propagation is simply Maxwellian electrodynamics. It may help to keep in mind that, although we will treat the signal field quantum mechanically for the sake of completeness, the quantum memories we consider behave essentially classically. Or, we might say that their efficiencies may be derived classically, since efficiencies do not depend on correlation functions of the optical fields, and it is only in the photon statistics revealed by these correlation functions that non-classicality is manifest. Chapter 4 Equations of motion In this chapter we derive the dynamical equations describing the interaction of light with the atoms of an ensemble quantum memory. We focus on EIT and Raman storage, which can be treated together; inhomogeneously broadened memories, such as CRIB and AFC, are covered in Chapter 7. We borrow techniques from the treatment of Stokes scattering by Mostowski et al. [139] , and some notation from the later treatment of quantum memories by Gorshkov et al. [140] . The derivation divides broadly into three parts. First, we write down the Hamiltonian describing the interaction of a single atom with the signal and control fields, and we obtain Heisenberg equations for the atomic evolution. Next, we introduce Maxwell’s equations, describing the coupling of the signal field to the macroscopic atomic polarization as it propagates through the ensemble. Finally, we add up the contributions from many atoms to form the macroscopic variables describing the atomic polarization and the spin wave. Having derived the dynamical equations in a convenient form, we investigate 4.1 Interaction 93 various methods of solution, in order to optimize the performance of the memory, in the next chapter. 4.1 Interaction We consider the propagation of a weak signal field through an ensemble of Λ-type atoms, in the presence of a bright control field, tuned into two-photon resonance with the signal, as depicted in Figure (4.1). Figure 4.1 The signal (blue) and control (green) fields involved in a Λ-type ensemble quantum memory. As shown in Section C.4 in Appendix C, the interaction of a light beam with an atom is well described by the electric dipole Hamiltonian HED = −E.d, where E is the electric field associated with the light at the atomic position, and where d is the dipole moment associated with an optically active electron in the atom. 4.2 Electric Field 4.2 94 Electric Field The electric field is composed of two parts, a bright classical control field, and a weak signal field that we wish to store, both propagating along the z-axis, E = Ec + Es . (4.1) The control field is sufficiently intense that it is not affected by its interaction with the atoms of the memory, so that we do not need to treat it as a dynamical variable in the Hamiltonian. We therefore represent it as a classical field, Ec (t, z, ρ) = vc Ec (t, z, ρ)eiωc (t−z/c) + c.c., (4.2) where vc is the control polarization vector1 , ωc is the central frequency of the control field, and Ec (t, z, ρ) is the slowly varying envelope of the control, describing the spatio-temporal profile of the control pulse. Here ρ = xx̂ + y ŷ is a transverse position vector, as shown in Figure 4.1. The signal field is much weaker than the control, and in general it may be in a non-classical state (for example, a Fock state — see Appendix C). Therefore we treat the signal field quantum mechanically. The signal field, when well-collimated, 1 Note that if the polarization is not linear (circular, for instance), the polarization vector is complex, satisfying v ∗ .v = v † v = 1. 4.2 Electric Field 95 can be written in the form Z Es (z, ρ) = ivs g(ω)a(ω, ρ)e−iωz/c dω + h.c., where vs is the signal polarization vector, g(ω) = (4.3) p ~ω/4π0 c is the mode amplitude, and a(ω, ρ) is an annihilation operator for a signal photon with frequency ω, and transverse position ρ, which satisfies the equal-time commutation relation [a(ω, ρ), a† (ω 0 , ρ0 )] = δ(ω − ω 0 )δ(ρ − ρ0 ). (4.4) The form of (4.3) is very similar to the expression (C.10) in Appendix C, the only difference being the inclusion of the transverse position ρ, which allows us to treat diffraction (this is covered in Chapter 6). In the next chapter we will drop the transverse coordinate and work with a one dimensional propagation model. Note that all these operators have no time-dependence in the Schrödinger picture (see Appendix B). In the Heisenberg picture, commutation with the optical free-field Hamiltonian (see (4.14) below) gives the annihilation operators the simple timedependence a(ω, ρ, t) = a(ω, ρ)eiωt . (4.5) It will be useful to factorize the signal field into a carrier wave, and a slowly varying envelope, as we did with the control field in (4.2). To do this, we make use of the approximation that the bandwidth of the signal field will be very small in com- 4.2 Electric Field 96 parison to its central frequency ωs . Therefore, only terms with frequencies rather close to ωs will be important in the integral in (4.3). Since the dependence of the mode amplitude g(ω) on frequency is quite weak (only ‘square-root’), we make the replacement g(ω) −→ g(ωs ). We are then able to perform the frequency integral explicitly, to obtain Es (t, z, ρ) = ivs gs A(t, z, ρ)eiωs (t−z/c) + h.c., where gs = √ (4.6) 2πg(ωs ), and where we have defined the slowly varying time-domain annihilation operator A according to the relation −iωs (t−z/c) A(t, z, ρ) = e 1 ×√ 2π Z a(ω, ρ, t)e−iωz/c dω. (4.7) That it retains the property of a photon annihilation operator, albeit for spatiotemporal, rather than spectral modes, can be seen from its commutator with its Hermitian adjoint. Inserting (4.5) into (4.7), we find [A(t, z, ρ), A† (t0 , z 0 , ρ0 )] = δ(t − t0 − (z − z 0 )/c)δ(ρ − ρ0 ). (4.8) Aside from the leading phase factor, which removes the rapid time-dependence due to the carrier frequency ωs , the action of A(t, z, ρ) in free space can be understood as the annihilation of a signal photon in a spatio-temporal mode centred at position ρ and retarded time τ = t − z/c. 4.3 Dipole Operator 4.3 97 Dipole Operator Having developed a convenient notation for the signal and control fields, we now consider the atomic variables. The Coulomb interaction between the atomic nucleus and the electrons is of course rather complex in general, but the formalism of quantum mechanics comes to the rescue. The energy levels labelled |1i, |2i, |3i are eigenstates of the atomic Hamiltonian, and therefore they form an orthonormal basis for the Hilbert space of quantum states of the atom (see Appendices B and C). That is, any state of the atom can be described in terms of these states. And any operator acting on the atomic states can be expressed using the coordinate system defined by these states. In particular, the electric dipole operator d can be written in the following way, d= X djk σjk , (4.9) j,k where the coefficients djk = hj|d|ki are the matrix elements of the dipole operator, the σjk = |jihk| are ‘flip operators’ (sometimes known as transition projection operators), and where the double summation runs over the three atomic states. To preempt possible confusion, we should clarify that d is a three-dimensional vector in space, whose elements are quantum mechanical operators acting on the three-dimensional Hilbert space of the Λ-level atom. The Dirac notation (|1i, |2i, etc...) refers to vectors and/or operators in/on this Hilbert space, while the bold font notation (d, E, v, etc...) refers to vectors in ordinary space (whether their elements are numbers, or operators). 4.3 Dipole Operator 4.3.1 98 Parity So far, we have not used any properties of the dipole operator specifically — (4.9) is an identity that holds for any atomic operator. But now we can use the parity of the dipole operator to remove some terms from the sum in (4.9). Parity refers to the way a quantity transforms under the operation of inversion, when all spatial coordinates are reflected in the origin. That is, r −→ −r, where r is any position vector. As discussed in Appendix C, the dipole operator is simply given by d = −er, where now r is the position, with respect to the atomic centre-of-mass, of the optically active electron. Therefore under inversion, we have d −→ −d. The atomic dipole operator has negative parity. This means that all the diagonal dipole matrix elements must vanish, djj = 0. To see this, we can express the matrix element djj in terms of the wavefunction ψj of the state |ji, djj = hj|d|ji Z = ψj∗ (r)dψj (r) d3 r. (4.10) The integral runs over all space, but we can divide it into a pair of integrals: the first over half of all space, with positive coordinates r (+), and the second over the remaining half of space, with negative coordinates −r (−). Since the dipole operator 4.3 Dipole Operator 99 changes sign under inversion, the second integral exactly cancels with the first, Z djj 2 3 Z |ψj (r)| d d r + = Z = |ψj (r)|2 d d3 r − + |ψj (r)|2 (d − d) d3 r + = 0. (4.11) Here we used the fact that |ψj |2 has positive parity (i.e. is unaffected by inversion). This must be true since |ψj |2 describes the electronic charge density associated with the state |ji, and this must be spherically symmetric: there is no interaction to break the spherical symmetry of the bare atom. The electric dipole interaction is therefore completely off-diagonal, meaning that it only couples different states together, never the same state to itself. Furthermore, we require that the state |3i is long-lived, in order that it can serve as an effective storage state. Therefore we neglect any dipole coupling between the states |1i and |3i, so that no direct transitions between these states are mediated by the dipole operator: our goal, instead, is to implement an indirect transition, mediated by the control field. In the light of these arguments, we arrive at the somewhat pared-down expression d = d12 σ12 + d23 σ23 + h.c.. (4.12) 4.4 Hamiltonian 4.4 100 Hamiltonian We are now in a position to write down the Hamiltonian for the atom-light system, H = HA + HL + HED . (4.13) Here HL is the free-field energy of the light field, Z HL = Z ω a† (ω, ρ)a(ω, ρ) d2 ρ dω, (4.14) A where the transverse integral runs over the transverse area A of the signal field. Here we have neglected the zero-point energy, and also the fixed energy associated with the control field. HA is the Hamiltonian for the bare atom. The atomic states |1i, |2i, |3i are by definition eigenstates of HA , and therefore when written in terms of these states, HA is purely diagonal (See Section A.4.3 in Appendix A), HA = X ωj σjj , (4.15) j where ωj is the resonant frequency of state |ji. We can now use the Heisenberg equation (see B.10 in Appendix B) to find the time evolution of the atomic flip operators. Since these operators always commute with the optical free-field Hamiltonian HL , we drop this from the Hamiltonian (it 4.4 Hamiltonian 101 has no effect on the atoms), and we work with the equation ∂t σjk = i[σjk , HA + HED ]. (4.16) The flip operators satisfy the following multiplicative identity, σij σkl = σil δjk , (4.17) † and under Hermitian conjugation we have σjk = σkj . Using these relations we obtain five independent atomic equations; two for the atomic populations, ∂t σ11 = −iE. (d12 σ12 − h.c.) , ∂t σ33 = iE. (d23 σ23 − h.c.) , (4.18) and three for the atomic coherences, ∂t σ12 = iω21 σ12 − iE. [d∗12 (σ11 − σ22 ) + d23 σ13 ] , ∂t σ13 = iω31 σ13 − iE. [d∗23 σ12 − d∗12 σ23 ] , ∂t σ23 = iω32 σ23 − iE. [d∗23 (σ22 − σ33 ) − d12 σ13 ] , (4.19) where we have defined ωjk = ωj − ωk as the frequency difference between the states |ji and |ki. Note that P j σjj = I, the identity, so that the sum of the populations commutes with the Hamiltonian, and therefore has no time-dependence. This simply 4.5 Linear approximation (1) 102 expresses the fact that the atom remains in one of the states |1i, |2i, |3i, at all times. Therefore ∂t σ22 = −∂t (σ11 + σ33 ). The coupled equations (4.18) and (4.19) constitute a rather complex system, and it is not possible to extract an analytic solution, in general. Fortunately, the equations simplify considerably in the linear regime. 4.5 Linear approximation (1) Provided that we store a small number of photons in the quantum memory, such that most of the atoms remain in their ground states, with only very few atoms excited, we can ignore the dynamics of the atomic populations. We replace the operators σjj by their expectation values on the atomic ground state, σ11 −→ 1, σ22 −→ 0, σ33 −→ 0. (4.20) This leaves us with just the three equations for the coherences, ∂t σ12 = iω21 σ12 − iE. [d∗12 + d23 σ13 ] , ∂t σ13 = iω31 σ13 − iE. [d∗23 σ12 − d∗12 σ23 ] , ∂t σ23 = −iω23 σ23 + iE.d12 σ13 . (4.21) 4.6 Rotating Wave Approximation 4.6 103 Rotating Wave Approximation The leading terms in each of the equations in (4.21), of the form iωkj σjk , simply describe rapid oscillations. The electric field E is also oscillating rapidly; the combined dynamics of the coherences will therefore contain components oscillating at both the sum and difference frequencies of these oscillations. Physically, these contributions give rise to different time-orderings of the two-photon Raman transition between states |1i and |3i, as shown in Figure 4.2. When the detuning is small compared to optical frequencies, the sum frequencies are many orders of magnitude higher than the difference frequencies, and so the sum frequencies average out to zero: they can be neglected. This is the content of the rotating wave approximation (RWA). rotating counter-rotating Figure 4.2 Time-ordering. The rotating wave approximation neglects counter-rotating terms, which correspond to strongly suppressed time-orderings for the Raman process. To implement the RWA, we define rotating coherences by the ansatz σ ejk = σjk eiωjk τ , (4.22) 4.6 Rotating Wave Approximation 104 where τ = t − z/c is the retarded time (see the text following (4.8)). Including the dependence on z here will be useful when we consider propagation. Inserting (4.22) into (4.21) yields the equations ∂t σ e12 = −iE. d∗12 e−iω21 τ + d23 σ e13 e−iω23 τ , ∂t σ e13 = −iE. d∗23 σ e12 eiω23 τ − d∗12 σ e23 e−iω21 τ , ∂t σ e23 = iE.d12 σ e13 eiω21 τ . (4.23) Inserting the expressions (4.2) and (4.6) for Ec and Es into E, and multiplying out the resulting expressions, we find terms multiplying rapidly varying exponentials, like e−i(ω21 +ωs )τ , for instance, and also terms multiplying slowly varying exponentials, like e−i(ω21 −ωs )τ = e−i∆τ . Neglecting the fast oscillating terms, we obtain the equations ∂t σ e12 = −id∗12 . vc Ec e−i∆+ τ + ivs gs Ae−i∆τ − id23 . vc Ec e−i∆τ + ivs gs Ae−i∆− τ σ e13 , i h ∂t σ e13 = −id∗23 . vc∗ Ec∗ ei∆τ − ivs∗ gs A† ei∆− τ σ e12 + id∗12 . vc Ec e−i∆+ τ + ivs gs Ae−i∆τ σ e23 , i h e13 , ∂t σ e23 = id12 . vc∗ Ec∗ ei∆+ τ − ivs∗ gs A† ei∆τ σ where we have defined ∆+ = ω21 − ωc and ∆− = ω23 − ωs (see Figure 4.3). (4.24) 4.7 Unwanted Coupling (a) 105 (b) (c) Figure 4.3 Useful and nuisance couplings. (a) The desired quantum memory coupling. (b) The control field couples to the ground state, initiating spontaneous Stokes scattering. (c) The signal field couples to the storage state: it is very weak, and there is no significant population in this state, so the effect of this coupling is negligible. 4.7 Unwanted Coupling We have already succeeded in dramatically simplifying the dynamical equations, but it is still not obvious how the behaviour described by this system of equations allows for the implementation of a quantum memory. Other physical processes obscure the useful features of the system. For instance, the term involving d∗12 .vc represents the coupling of the control field to the |1i ↔ |2i transition. If this term is strong, the control field can initiate spontaneous Stokes scattering, as shown in Figure 4.3 (b). Aside from complicating the equations, this process can be problematic, since it generates excited atoms in the state |3i that are not correlated with the signal field. When we attempt to retrieve the signal field, some of these uncorrelated atoms may contribute a noisy background emission. In practice, it is possible to distinguish the signal from the noise using phasematched retrieval, which we discuss in §6.3.2 in Chapter 6. In any case, a description of the memory interaction requires that we can eliminate this unwanted coupling. As can be seen from Figure 4.3, the Stokes scattering process is detuned further from resonance, with detuning ∆+ > ∆. If 4.8 Linear Approximation (2) 106 ∆+ is sufficiently large (this requires that the splitting ω31 is large compared to ∆), the term representing Stokes scattering will oscillate quickly enough to be neglected. Alternatively, it may be possible to choose the polarization of the control such that the product d12 .vc vanishes due to a selection rule, although strict selection rules usually require the application of an external magnetic field. Polarization selection rules are discussed in the context of cesium in §F.4 of Appendix F. Regardless of the justification, in the following analysis we set d12 .vc = 0. By the same token, we set d23 .vs = 0, thereby disregarding any coupling of the signal field to the |2i ↔ |3i transition, as depicted in Figure 4.3 (c). The equations of motion are now given by ∂t σ e12 = d∗12 .vs gs Ae−i∆τ − id23 .vc Ec e−i∆τ σ e13 , ∂t σ e13 = −id∗23 .vc∗ Ec∗ ei∆τ σ e12 − d∗12 .vs gs Ae−i∆τ σ e23 , ∂t σ e23 = d12 .vs∗ gs A† ei∆τ σ e13 . 4.8 (4.25) Linear Approximation (2) Now that we have arrived at the more transparent set of equations (4.25), we can identify some terms in these equations that are ‘small’, in the sense that ‘weakly excited’ operators are involved. To be more concrete, we analyse the size of each term perurbatively. We attach a parameter to each of the coherences σ ejk , and also to the signal field A. This parameter just labels these quantities as ‘small’; in the case of the coherences, they are initially vanishing, and in the case of the signal field, 4.8 Linear Approximation (2) 107 only a very few signal photons are sent into the memory. The equations become ∂t σ e12 = d∗12 .vs gs Ae−i∆τ − id23 .vc Ec e−i∆τ e σ13 , ∂t σ e13 = −id∗23 .vc∗ Ec∗ ei∆τ e σ12 − d∗12 .vs gs 2 Ae−i∆τ σ e23 , ∂t σ e23 = d12 .vs∗ gs 2 A† ei∆τ σ e13 . (4.26) There are two terms proportional to 2 . They correspond to ‘second-order’ perturbative corrections to the dynamics, and they can be neglected. Dropping the label , the linearized equations of motion for the atomic coherences are ∂t σ e12 = d∗12 .vs gs Ae−i∆τ − id23 .vc Ec e−i∆τ σ e13 , ∂t σ e13 = −id∗23 .vc∗ Ec∗ ei∆τ σ e12 . (4.27) These equations contain the important physics of the quantum memory interaction. They describe how the coherence σ e12 is directly excited by the signal field A, and how this excitation is then coupled to the coherence σ e13 through the control field Ec . Macroscopically, when many identical atoms are involved, we can identify σ e12 with the atomic polarization, and σ e13 with the spin-wave. Before considering the collective dynamics of the ensemble as a whole, we first consider the propagation of the signal field. 4.9 Propagation 4.9 108 Propagation Heisenberg’s equations describe evolution in time, but it is not obvious how to treat the spatial propagation of the signal field in this formalism. In fact, it is possible to do this, but a more transparent derivation utilizes Maxwell’s equations. The atomic ensemble behaves as a dielectric medium in the presence of the signal field, for which the appropriate formulation is as follows, ∇.D = ρfree , ∇ × E = −∂t B, ∇.B = 0, ∇ × H = Jfree + ∂t D. (4.28) Here D is the displacement field, H is the magnetic field, E is the electric field and B is the magnetic induction. The distinction between B and H will not be very important for us, since all the materials we are concerned with are non-magnetic. We simply take B = µ0 H, where µ0 is the permeability of free space, and refer to B as the magnetic field. The quantities ρfree and Jfree are the free charge density and current, respectively. The designation ‘free’ refers to the fact that they are not induced by the fields: they are charges and currents that are not associated with any material dielectric properties. In any case we only deal with materials in which there are no free charges or currents, and therefore we set ρfree = Jfree = 0. A wave equation for the electric field is found by differentiating the equation for H, and 4.9 Propagation 109 substituting in the equation for E, ∇ × ∂t H = ∂t2 D, ⇒ −∇ × (∇ × E) = µ0 ∂t2 D. (4.29) The double curl derivative on the left can be simplified using the vector calculus identity ∇ × (∇ × E) = ∇ (∇.E) − ∇2 E. (4.30) The signal field is a transverse propagating wave, which is divergence free2 . This is evident from the definition of the Coulomb gauge (see Appendix C), in which ∇.A = 0, with E = ∂t A, where A is the magnetic vector potential. We therefore drop the first term in (4.30), to obtain ∇2 E = µ0 ∂t2 D. (4.31) The displacement field is formed from the sum of the electric field, and the material polarization P , D = 0 E + P , (4.32) where P is the polarization density, defined as the dipole moment per unit volume. Substituting (4.32) into (4.31), and using the relation 0 µ0 = 1/c2 , we arrive at the 2 These arguments apply to a homogeneous and isotropic dielectric containing no source charges [141] 4.10 Paraxial and SVE approximations 110 wave equation 1 2 2 ∇ − 2 ∂t E = µ0 ∂t2 P . c (4.33) This equation relates the propagation of the optical fields to the atomic polarization in the memory. Since the control field is so intense, it is not significantly affected by its interaction with the ensemble, and so we do not consider its propagation further. For the signal field, we use the equation 1 2 ∇ − 2 ∂t Es = µ0 ∂t2 Ps , c 2 (4.34) where Ps is the component of the atomic polarization which acts as a source for the signal field. That is, Ps is the component of the polarization oscillating at the signal carrier frequency ωs . Further simplification is accomplished by making use of the fact that the signal and control fields are collimated beams. 4.10 Paraxial and SVE approximations The slowly varying envelope approximation (SVE), and the paraxial approximation, are both implicit in the decomposition (4.6). We assume that the amplitude A (its status as an operator is not important for this discussion) is a smooth, slowly varying function of time and space. The exponential factor then represents the optical carrier wave, oscillating in time with frequency ωs , and oscillatiing along the z-axis with wavevector ks = ωs /c. The paraxial approximation allows us to treat the signal field as a beam traveling 4.10 Paraxial and SVE approximations 111 along the z-axis, with negligible divergence. This approximation is satisfied as long as the transverse spatial profile of A is much larger than the signal wavelength λs = 2π/ks . The SVE approximation allows us to treat the propagation of the signal field purely in terms of the envelope of the signal pulse, represented by the temporal shape of A, without having to explicitly model the very fast time-dependence of the carrier wave, which oscillates much faster. To make this approximation successfully, we should have that the temporal duration of the signal field is much longer than the optical period 2π/ωs . To see how these approximations simplify the theoretical description, we insert the signal field (4.6) into the wave equation (4.34). We consider only the positive frequency component of the signal field, since only this component is coupled to the atoms through the system (4.27). We also define the slowly varying atomic polarization Pes by factorizing out the signal frequency, Ps = Pes eiωs τ . (4.35) i h i 1 2 h 2 ∇ − 2 ∂t ivs gs Aeiωs (t−z/c) = µ0 ∂t2 Pes eiωs (t−z/c) . c (4.36) The resulting wave equation is We take the scalar product of both sides with the polarization vector vs∗ (this is the same as taking the inner product with vs† , see Appendix A), and apply the chain 4.11 Continuum Approximation 112 rule for the derivatives to obtain 1 2 ωs 1 ωs 2 µ0 1 2 2 2 ∇⊥ + ∂z − 2 ∂t − 2i ∂z + ∂t − A = −i vs∗ . ∂t2 + 2iωs ∂t − ωs2 Pes , − 2 ωs c c c c c gs (4.37) where we have divided out the exponential factor eiωs τ . The transverse Laplacian ∇2⊥ = ∂x2 + ∂y2 describes diffraction of the signal field as it propagates. Note that the term in square brackets on the left hand side vanishes. Concerning the remaining terms, we observe that according to the SVE approximation ks ∂z + 1 ∂t A ∂z2 − 1 ∂t2 A . 2 c c (4.38) We therefore drop the smaller term. For the same reason, we drop all but the last term on the right hand side, and we arrive at the equation 1 i µ0 ωs2 ∗ e 2 ∇⊥ + ∂z + ∂t A = − v .Ps . 2ks c 2gs ks s (4.39) This equation describes the coupling of the atoms to the signal field amplitude. It now remains for us to connect the atomic evolution equations (4.27) with the macroscopic polarization Ps . 4.11 Continuum Approximation We have in mind an ensemble of sufficient density that the atoms form an effective continuum. Consider a small region within the ensemble at position r = (z, ρ), 4.11 Continuum Approximation 113 with volume δV . We’ll call this a voxel. To make the continuum approximation, we should have many atoms in each voxel; nδV 1, where n is the number density of atoms in the ensemble. Each voxel should be ‘pancake-shaped’, so that its thickness δz along the z-axis satisfies δz λs , while its transverse area δA satisfies δA λ2s . The condition on δz ensures that the longitudinal optical phase ks z is roughly constant throughout the voxel. The condition on δA ensures that the typical interatomic separation is much larger than the signal wavelength λs , so that dipole-dipole interactions between the atoms can be neglected, and we can treat the atoms as isolated from one another. The macroscopic polarization at position r is then found by adding up the dipole moments in the voxel located at r, P = 1 X β d , δV (4.40) β(r) where the index β runs over all the atoms in the voxel at position r. Using the expression (4.12) for each atom, and recalling that the coherence σ23 is negligible (see Section 4.8), we have P = 1 X β d12 σ12 + h.c. . δV (4.41) β(r) We now define macroscopic variables involving sums over the coherences σjk , in order to ‘tie up’ the system of equations (4.27) with the propagation equation (4.39). For 4.11 Continuum Approximation 114 the macroscopic polarization, we define the operator P =√ 1 X β i∆τ σ e12 e . nδV (4.42) β(r) Note that P is not simply the magnitude of the vector P . They are closely related (see (4.47) below), but some book-keeping is required to ensure that we keep track of all the relevant constants. In the same vein, for the spin wave we define the operator B=√ 1 X β σ e13 . nδV (4.43) β(r) These definitions are motivated by analogy with the slowly varying photon annihilation operator A. For instance, the equal-time commutator of B with its Hermitian adjoint is given by h i B(z, ρ, t), B † (z 0 , ρ0 , t) = = XXh β γi 1 e31 σ e13 , σ (δV )2 n β(r) γ(r 0 ) (δV1)2 n × nδV (σ11 − σ33 ) 0 if r and r 0 label the same voxel, (4.44) otherwise. Using the linear approximation (4.20), we find, in the continuum limit δV −→ 0, h i B(z, ρ, t), B † (z 0 , ρ0 , t) = δ(z − z 0 )δ(ρ − ρ0 ). (4.45) 4.11 Continuum Approximation 115 Identical arguments yield the commutator h i P (z, ρ, t), P † (z 0 , ρ0 , t) = δ(z − z 0 )δ(ρ − ρ0 ), (4.46) for the polarization. Therefore both the operators P and B satisfy bosonic commutation relations. We interpret them as annihilation operators for an atomic excitation at position r. That is, P (z, ρ) annihilates a distributed excitation of the excited state |2i at position (z, ρ). And B(z, ρ) annihilates a distributed excitation of the storage state |3i. Substituting the definition (4.42) into (4.41), and taking the positive frequency component (i.e. the component oscillating at a frequency +ωs ), we find Pes = √ nd12 P, (4.47) for the slowly varying macroscopic polarization. We can now write down equations governing both the propagation of the signal field, and the atomic dynamics. These equations are, 1 i 2 ∇ + ∂z + ∂t A = −κ∗ P, 2ks ⊥ c ∂t P = i∆P + κA − iΩB, ∂t B = −iΩ∗ P, (4.48) 4.12 Spontaneous Emission and Decoherence 116 where we have defined the control field Rabi frequency Ω= d23 .vc Ec , ~ (4.49) and the coupling constant d∗ .vs √ d∗ .vs κ = 12 × ngs = 12 ~ ~ r ~ωs n . 20 c (4.50) The equations (4.48) describe the propagation and diffraction of a weak signal field through an ensemble of ideal atomic Λ-systems, prepared initially in their ground states. We have not yet included any description of decay processes, such as spontaneous emission from the excited state, or collisional de-phasing of the spin wave. These two processes are expected to happen on very different time-scales, but they can be treated in exactly the same way, which we now introduce. 4.12 Spontaneous Emission and Decoherence In Section (C.5) in Appendix (C), we discuss the description of Markovian decay in the Heisenberg picture using Langevin equations. We used a model in which a bosonic system was coupled to a large reservoir, also composed of bosons. In the present discussion, all the operators we consider — A, B and P — are bosonic in character (see Eqs. (4.8), (4.45) and (4.46) for their commutators). In the case of spontaneous emission, these operators are coupled to the electromagnetic field, which is a large reservoir of bosons, and so the model of Appendix C is applicable. Using 4.12 Spontaneous Emission and Decoherence 117 this model, spontaneous emission at the rate γ is described by incorporating an appropriate decay term into the dynamical equation for P , along with a Langevin noise operator FP , which introduces fluctuations that preserve the bosonic commutation relations of P . The noise operator is delta-correlated, meaning that no correlations exist between its values at different times. And since all the atoms in the ensemble are subject to independent fluctuations, no correlations exist between the noise at different positions. These properties are summarized by the relations hFP† (t, z, ρ)FP (t0 , z 0 , ρ0 )i = 2γ n̄P × δ(t − t0 )δ(z − z 0 )δ(ρ − ρ0 ), hFP (t, z, ρ)FP† (t0 , z 0 , ρ0 )i = 2γ(n̄P + 1) × δ(t − t0 )δ(z − z 0 )δ(ρ − ρ0 ), (4.51) where the expectation value is taken on the initial state of both the ensemble and the reservoir (i.e. the electromagnetic field to which the atoms are coupled). Here the number n̄P is the initial number of atoms thermally excited into state |2i, on average. When dealing with optical transitions, we typically have n̄P = 0, of course. Similarly, to treat decoherence of the spin wave B at a rate γB , we add a decay term and a noise operator FB , which satisfies identical relations to (4.51), again with n̄B = 0. The dynamical equations including these dissipative processes are then i 1 2 ∇ + ∂z + ∂t A = −κ∗ P, 2ks ⊥ c ∂t P = −γP + i∆P + κA − iΩB + FP , ∂t B = −γB B − iΩ∗ P + FB . (4.52) 4.12 Spontaneous Emission and Decoherence 118 Note that these decay rates are for the atomic coherences. The atomic populations decay at twice these rates. For instance, the number of spin wave excitations is given by Z LZ NB = 0 B † (z, ρ)B(z, ρ) d2 ρ dz. (4.53) A In the absence of any optical fields (with Ω = 0), we find that ∂t hNB i = −2γB hNB i, (4.54) where the expectation value is taken on an arbitrary state. Any terms involving the noise operator FB vanish when taking the expectation value because its fluctuations average to zero. The relation (4.54) shows that the number of spin wave excitations decays at the rate 2γB . The same argument applied to P shows that the number of excited atoms, in the state |2i, decays at the rate 2γ, so that the spontaneous lifetime of the state |2i is 1/2γ. Now that we have introduced losses and decoherence with a degree of formal rigour, we see that in fact the noise operators FB , FP may be neglected when optimizing the performance of a quantum memory. The reason is that the efficiency of a quantum memory depends only on the ratio of stored to input excitations. That is, only the number operators for the signal field and spin wave are involved. These number operators involve only normally ordered products, of the form A† A or B † B, and therefore only normally ordered products of the F operators enter into the efficiency. Since we have n̄ ∼ 0, any products of the form F † F vanish when an 4.12 Spontaneous Emission and Decoherence 119 expectation value is taken, as is clear from (4.51). Therefore the Langevin operators can be safely dropped from the system of equations (4.52). Of course the decay terms are important! Chapter 5 Raman & EIT Storage Here we use the equations of motion derived in the last chapter to study the optimization of the EIT and Raman quantum memory protocols. 5.1 One Dimensional Approximation The analysis is greatly simplified if we use a one dimensional model, so that we only consider propagation along the z-axis. This can always be made a good approximation by using laser-beams with low divergence. The effects of diffraction are considered in Chapter 6. In the following, we will average over the transverse coordinate ρ, to produce a one dimensional propagation model. We re-define the 5.1 One Dimensional Approximation 121 variables A, P and B by integrating over the transverse area A of the signal field, A(t, z, ρ) −→ A(t, z) = P (t, z, ρ) −→ P (t, z) = B(t, z, ρ) −→ B(t, z) = Z 1 √ A(t, z, ρ) d2 ρ, A A Z 1 √ P (t, z, ρ) d2 ρ, A A Z 1 √ B(t, z, ρ) d2 ρ. A A (5.1) Having averaged the variables in this way, we now drop the transverse Laplacian ∇2⊥ from the propagation equation for A in (4.52). Further simplifications follow. For instance, in the absence of any transverse structure, the control field envelope Ec , propagating undisturbed at the speed of light along the z-axis, can be written as a function of the retarded time τ = t − z/c only. We therefore make a change of variables from (t, z) to (τ, z), which enables us to write the control field Rabi frequency as Ω = Ω(τ ). Furthermore, the mixed derivative in the propagation equation for A is simplified, since ∂z 1 + ∂t = ∂z , c t z τ ∂t z = ∂τ z , (5.2) where the subscripted parentheses indicate the variables held constant. In this new coordinate system, the one dimensional equations of motion for a Λ-type quantum 5.1 One Dimensional Approximation 122 memory are therefore given by ∂z A(z, τ ) = −κ∗ P (z, τ ), ∂τ P (z, τ ) = −ΓP (z, τ ) + κA(z, τ ) − iΩ(τ )B(z, τ ), ∂τ B(z, τ ) = −iΩ∗ (τ )P (z, τ ), (5.3) where Γ = γ−i∆ is the complex detuning. Note that we have dropped the decay term associated with the decoherence of the spin wave B: by assumption this is negligible on the time-scale of the storage process, which is what we seek to optimize. We have also dropped the Langevin noise operator FP associated with spontaneous decay of the polarization P , since this noise does not affect the efficiency (see the end of Section 4.12 in the previous chapter). The equations (5.3) are mercifully rather simple. Certainly they are easier on the eye than any of their previous incarnations in Chapter 4! The elimination of the Langevin noise operators means that these equations are now entirely classical in nature: we can treat (5.3) as a system of coupled partial differential equations in three complex-valued functions A, P and B. Since the equations are linear, the solutions will be linear, and we need not worry about issues involving commutators or operator ordering. Our aim is to find an expression for the Green’s function K(z, τ ), relating the input signal field Ain (τ ) = A(z = 0, τ ) to the final spin wave Bout (z) = B(z, τ −→ ∞). In the absence of decoherence of the spin wave, the limit τ −→ ∞ is simply a mathematical shorthand for “the end of the storage interaction”, when the control and signal fields 5.2 Solution in k-space 123 have fallen away to zero. As described in (3.2) in Chapter 3, taking the SVD of K will tell us how to optimize the memory efficiency. We now attempt a solution of the system (5.3). 5.2 Solution in k-space 5.2.1 Boundary Conditions We must solve three first order partial differential equations in three functions, and therefore there must be three boundary conditions. For the storage process, we begin with no excitations of the atomic polarization, and no spin wave excitations, so the boundary conditions for the functions P and B are simply Pin (z) = P (z, τ −→ −∞) = 0, Bin (z) = B(z, τ −→ −∞) = 0. (5.4) As mentioned above, the boundary condition for the signal field is set by the initial temporal profile of the signal envelope A, as it impinges on the front face of the ensemble at z = 0, Ain (τ ) = A(z = 0, τ ). (5.5) Our analysis will tell us the shape for Ain that maximizes the memory efficiency. These boundary conditions are represented by the tableau in Figure 5.1. 5.2 Solution in k-space 124 Figure 5.1 Quantum memory boundary conditions. Example solutions for the functions A, P and B are shown in each panel, with the z coordinate running from top to bottom and the τ coordinate running from left to right. The red lines indicate the boundary conditions that must be specified to generate the solutions. 5.2.2 Transformed Equations To proceed with solving the equations of motion, it will be useful to reduce them to a system of coupled ordinary differential equations. This can be done by applying a unilateral Fourier transform over the z coordinate (see Appendix D). We define Fourier transformed variables accordingly, Z ∞ 1 e A(z, τ )eikz dz, A(k, τ ) = √ 2π 0 Z ∞ 1 Pe(k, τ ) = √ P (z, τ )eikz dz, 2π 0 Z ∞ 1 e B(k, τ ) = √ B(z, τ )eikz dz. 2π 0 (5.6) 5.2 Solution in k-space 125 Using the result (D.27) for the transform of the spatial derivative ∂z , we obtain e − √1 Ain = −κ∗ Pe, −ik A 2π e − iΩB, e ∂τ Pe = −ΓPe + κA e = −iΩ∗ Pe. ∂τ B (5.7) We remark that the independence of Ω from z is critical to the usefulness of this transformation. The spatial propagation has now been reduced to an algebraic e equation, which we can solve for A. 5.2.3 Optimal efficiency Even given unlimited energy for the control pulse, the storage efficiency is limited by spontaneous emission from the excited state |2i. The storage into the dark state |3i, which is not affected by spontaneous emission, is always mediated via coupling to |2i. Even with perfect transfer between states |2i and |3i, we can never store more efficiently into |3i than we can couple to |2i. Therefore the storage efficiency is bounded by the efficiency with which we can transfer population into |2i. To evaluate this upper bound, we simply neglect the spin wave, and solve the equation e we obtain for Pe with Ω = 0. Including the solution for A, |κ|2 e iκ e ∂τ P = − Γ + i Ain . P+√ k 2πk (5.8) 5.2 Solution in k-space 126 This equation can be integrated directly to give iκ 2 Pe(k, τ ) = Pein (k)e−(Γ+i|κ| /k)τ + √ 2πk Z τ e−(Γ+i|κ| 2 /k)(τ −τ 0 ) Ain (τ 0 ) dτ 0 . (5.9) −∞ Since we are concerned with storage, we can set the initial polarization to zero. We next assume that we are somehow able to transfer all the excitations from Pe e with no loss, at some time τ = T , which marks the end of to the spin wave B, eout , and the the storage interaction. We can then make the substitution Peout → B optimal storage process is described by the map Z ∞ e K(k, T − τ )Ain (τ ) dτ, eout (k) = B (5.10) −∞ where the k-space storage kernel is given by iκ −(Γ+i|κ|2 /k)τ e e . K(k, τ) = √ 2πk (5.11) e = 0, so that no storage takes place after τ = T . Note that for times τ > T , we set K Now, some comments are warranted. First, the optimal storage efficiency does not depend on the detuning ∆. To see this, note that e−Γτ = e−γτ × ei∆τ . The latter factor, involving the detuning, represents a pure phase rotation. We could absorb it into the definition of Ain without altering its norm. Therefore we can drop it from the kernel — it has no effect on its singular values, and no effect on the optimal efficiency. Second, the optimal efficiency only depends on the resonant optical depth, 5.2 Solution in k-space 127 defined by d= |κ|2 L . γ (5.12) To see this, we normalize the time and space coordinates, along with the spin wave and signal field amplitudes, to make them dimensionless. The spontaneous decay rate γ and the ensemble length L provide natural time and distance scales for this normalization. We denote the normalized variables by an overbar, τ = γτ, e A e= √ A , γ k = kL, e e = √B . B L (5.13) This notation is rather clumsy, but it serves to clarify the re-scaling. With these definitions, the storage map becomes e out k = B Z ∞ e k, T − τ Ain (τ ) dτ , K (5.14) −∞ where the kernel has been converted into the dimensionless form √ i d −(1+id/k)τ e K k, τ = √ e . k 2π (5.15) Here we have assumed for simplicity that κ is real, (i.e. κ = κ∗ ), and we have dropped the detuning for the reason mentioned above. It is now clear that the optical depth d is the only parameter associated with the interaction that plays any role in determining the efficiency of the storage process. This was first shown by 5.2 Solution in k-space 128 Gorshkov et al [133] . Their explanation is that, regardless of the memory protocol used, the branching ratio between loss — spontaneous emission — and storage — absorption — is fixed by the optical depth. We have derived the result by arguing that we cannot do better than is possible through direct, linear absorption into the state |2i. Clearly detuning from resonance cannot improve matters, and so from this perspective it is unsurprising that the best possible efficiency is only limited by the resonant coupling, parameterized by d. Below we examine the quantitative behaviour of the optimal efficiency in more detail. e The optimal efficiency is given by the square of the largest singular value of K. Note that it makes no difference whether the argument τ or the ‘flipped’ argument T − τ is used; this just flips the input modes without affecting the singular values. e has a singularity at k = 0, and this means we cannot approxiUnfortunately, K mate it as a finite matrix, in order to compute the singular values numerically. To proceed further, we need to transform to a different coordinate system to remove the singularity. One possibility is to apply an inverse Fourier transform, and this works well. Before doing this, however, we first introduce another way to analyse the kernel (5.15), which gives a degree of insight into its structure. In what follows, we will drop the overbar notation for the normalized parameters, as a concession to legibility. 5.2 Solution in k-space 5.2.4 129 Solution in Wavelength Space e are unaffected by a unitary transformation. Recall that the singular values of K Consider the coordinate transformation k −→ λ = 2π/k, from k-space to ‘wavelength space’. The kernel (5.15) is no longer singular in this coordinate system. To guarantee unitarity, the transformed kernel must include a Jacobian factor of √ √ 1/ ∂k λ = −i 2π/λ (see §3.3.3 in Chapter 3). We obtain the result √ e K(λ, τ) = d −(1+idλ/2π)τ e . 2π (5.16) e A , as described in (3.19) in ChapWe now form the anti-normally ordered product K ter 3, which takes the form 1 e A (λ, λ0 ) = 1 × . K 4π 2πi λ − d i − λ0 (5.17) e A give the singular values of K, e so we should try to solve the The eigenvalues of K following eigenvalue problem, 1 2πi Z ∞ −∞ ψej (λ0 ) dλ0 = ηj ψej (λ). 0 λ − 4π i − λ d (5.18) The integrand has a singularity at λ0 = λ − 4π d i. We consider extending the integral into the complex plane, and integrating along a semicircle-shaped contour, closed in the lower half of the complex plane (as depicted in Figure D.4 in Appendix D), so that the contour encloses the singularity. We assume that the mode functions 5.2 Solution in k-space 130 ψej fall away to zero along the curved portion of the contour, in the limit that its radius is made infinitely large. Under this assumption, the only contribution to the integral comes from the straight portion of the contour, along the real line, which is precisely the integral in (5.18). We can now use Cauchy’s integral theorem (see §D.1.1 in Appendix D) to evaluate the left hand side of (5.18), ψej λ − 4π d i = ηj ψej (λ). (5.19) By inspection, a possible form for the modefunctions is ψej (λ) ∝ e±iαj λ . (5.20) That is, plane waves in λ-space, each with some eigenfrequency αj . We should choose the minus sign in (5.20), so that the modefunctions are exponentially damped in the lower half of the complex plane, as we assumed above. The storage efficiencies ηj are then given by ηj = e−4παj /d . (5.21) This form is reassuring, since the ηj −→ 1 in the limit d −→ ∞, which makes sense. However, it is not obvious that there is any constraint on the eigenfrequencies αj . We need a ‘quantization condition’ on the modes. One possibility is to transform back into ordinary z-space, and look for a physically reasonable condition. First, we 5.2 Solution in k-space 131 transform back in to k-space. Remembering to include the Jacobian factor, we get 1 ψej (k) ∝ × e−2πiαj /k . k (5.22) The proportionality symbol reflects the fact that the modes should be properly normalized; at the moment we are simply concerned with their functional form. Transforming back into z-space requires taking the inverse Fourier transform of (5.22). The method is described in §D.5.2 of Appendix D. The result is ψj (z) ∝ J0 2 p 2παj z , (5.23) where J0 is a zero’th order ordinary Bessel function of the first kind. In the limit of large d, we expect that the signal field is completely absorbed, so that there is no signal left at the end of the ensemble. We are working in normalized coordinates, so the end of the ensemble is located at z = 1. With no remaining signal field, no spin wave excitations can be excited, so we should expect ψj (1) ∼ 0. This is satisfied if we choose αj = .23, 1.21, 2.98, 5.53, etc..., as shown in Figure 5.2. This choice of quantization is consistent with the orthogonality requirement on the modes, Z 1 ψi (z)ψj∗ (z) dz = 0, if i 6= j, (5.24) 0 which follows from the orthogonality condition (D.34) of the Bessel functions (see §D.5 in Appendix D). 5.2 Solution in k-space 132 1 0 0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5 √ Figure 5.2 The first three zeros of the function J0 (2 2πx). The condition that the spin wave should vanish at z = 1 picks out these zeros as the eigenfrequencies αj . The optimal storage efficiency is then given by η1 = e−4πα1 /d ≈ 1 − 2.9/d, (5.25) where the approximation holds in the limit of large d. The above quantization procedure used to derive the optimal efficiency (5.25) was rather ad hoc, and we should check it against a numerical SVD. We therefore return to the k-space kernel (5.15). As noted previously the singularity at k = 0 makes this form of the kernel inconvenient. Fortunately it is easy to take the inverse Fourier transform from k-space back into ordinary space (again using the method described in §D.5.2 of Appendix D). The result is √ K (z, τ ) = √ de−τ J0 2 dτ z . (5.26) We could perform a numerical SVD on this kernel directly. The numerical problem 5.2 Solution in k-space 133 converges better, however, if we form the anti-normally ordered product kernel KA , which is given by √ d 0 KA z, z 0 = e−d(z+z )/2 I0 d zz 0 . 2 (5.27) Here I0 is a zero’th order modified Bessel function of the first kind (see §D.5.3 in Appendix D for a clue as to how to perform the required integral). The optimal efficiency η1 is the largest eigenvalue of this kernel. In Figure 5.3, we plot the analytic prediction (5.25) alongside the numerical result, over a range of optical depths. The analytic formula is an excellent approximation for optical depths larger than ∼ 50. This scaling of the optimal storage efficiency was first noted by Gorshkov et al. [133] . They also provided an elegant proof of the optimality of the kernel (5.27), that does not rely on the heuristic assertion that ‘we cannot map more efficiently to B than we can to P ’. Nonetheless, the result is the same. 1 10 0 10 −1 10 −2 10 −3 10 0 400 800 1200 1600 2000 Figure 5.3 Optimal storage efficiency. The plot shows the difference between the optimal efficiency η1 and unity, on a logarithmic scale, as a function of the optical depth d. The analytic formula 1 − η1 ≈ 2.9/d derived above (green), is in excellent agreement with the numerical result (blue), found by diagonalizing the kernel (5.27). Having identified the best possible storage efficiency, we now continue with our 5.2 Solution in k-space 134 analysis of the equations of motion, including the control field. This analysis will reveal how the temporal profile of the optimal input mode depends on the control pulse, and in what regimes it is possible to effectively shape the optimal input mode by shaping the control. 5.2.5 Including the Control e and substituting We return to the system (5.7). Solving the first equation for A, the result into the second equation for Pe, yields a pair of coupled linear differential equations in time only. For each spatial frequency k, we need to solve for the e To do this, we define a vector |ψi whose temporal dynamics of the system (Pe, B). e two elements are the functions Pe and B, e )| ↓i, |ψ(τ )i = Pe(τ )| ↑i + B(τ (5.28) where the basis kets | ↑i and | ↓i are given by 1 | ↑i = , 0 and 0 | ↓i = . 1 (5.29) We have suppressed the dependence on the spatial frequency k, which is no longer a dynamical variable, but which of course must not be forgotten! The equation of motion for |ψi is found to be ∂τ |ψi = −iM |ψi + |α0 i, (5.30) 5.2 Solution in k-space 135 where the time-dependent matrix M is given by M (τ ) = |κ|2 k − iΓ Ω(τ ) , Ω∗ (τ ) 0 (5.31) and where the time-dependent vector |α0 i includes the signal field boundary condtion, |α0 (τ )i = i √ κ Ain (τ )| ↑i. 2πk (5.32) Using the normalized variables introduced in the previous section, in which all lengths are scaled by L and all frequencies by γ, we simply replace κ with √ d. The structure of (5.30) is very similar to Schrödinger’s equation (B.6) for a two-level system, except that the evolution is not unitary, because M is not quite Hermitian, and because of the ‘driving term’ |α0 i. Nonetheless, techniques applied to the solution of Schrödinger’s equation remain useful. First, we write down the formal solution. Suppose that we are able to construct a propagation matrix V (τ ) such that ∂τ V = iV M . We then have that ∂τ (V |ψi) = (∂τ V ) |ψi + V ∂τ |ψi = iV M |ψi + V (−iM |ψi + |α0 i) = V |α0 i. (5.33) 5.2 Solution in k-space 136 Integrating this gives |ψ(τ )i = V −1 −1 (τ )Vin |ψin i + V Z τ (τ ) V (τ 0 )|α0 (τ 0 )i dτ 0 , (5.34) −∞ e and where Vin is the where |ψin i contains the boundary conditions for Pe and B, propagation matrix at the start of the interaction. Clearly we must have Vin = I, the identity operator. To find the spin wave at the end of the storage process, for which |ψin i = 0, we take the limit τ −→ ∞ to obtain eout = h↓ |V −1 B out Z ∞ V (τ )|α0 (τ )i dτ, (5.35) −∞ where Vout = V (τ −→ ∞). The k-space storage kernel is then given by i e K(k, τ) = k r d −1 h↓ |Vout (k)V (k, τ )| ↑i, 2π (5.36) where we have now included the dependence of the V matrices on k explicitly, lest we forget it. If we can find an expression for V , we can construct the storage kernel, and so find the optimal input mode by means of its SVD. Here we comment that if the matrix M (τ ) were replaced by an ordinary function, we would simply have Z V (τ ) = exp i τ 0 0 M (τ ) dτ . (5.37) −∞ And in fact this is still true whenever M is a diagonal matrix, since then [M (τ ), M (τ 0 )] = 0 (diagonal matrices always commute). Alternatively, if M is constant in time, it 5.2 Solution in k-space 137 can be pulled out of the integral in (5.37), and there is no issue with commutation at different times. However, in general, when M is non-diagonal and time-varying, as we have whenever Ω is not simply a constant, the solution (5.37) is not correct1 . This is an example of what is sometimes known as the great matrix tragedy: the simple fact that eA eB 6= eA+B when [A, B] 6= 0 is responsible for most of the difficulty arising in quantum mechanical calculations! 5.2.6 An Exact Solution: The Rosen-Zener case It is possible to find an exact solution for V , in the particular case that the shape of the control field envelope is given by a hyperbolic secant, Ω0 sech Ω(τ ) = Tc τ Tc , (5.38) where Tc sets the duration of the control pulse, and Ω0 is a dimensionless constant that sets the strength of the pulse (see Figure 5.4 (a)). The method of solution is due to Rosen and Zener [142–144] . We include details of the derivation here for completeness, but in §5.4 below we introduce a numerical approach that is faster, more accurate and more general. The analytic solutions presented here do provide useful points of comparison, of course. We first transform to the interaction picture, to remove the rapid oscillations generated by the diagonal elements of M . To do this, we define the diagonal and 1 Sometimes the formal solution is written like this, but in general it is understood that the exponential must be time ordered 5.2 Solution in k-space 138 off-diagonal matrices M0 = 2β Tc 0 0 , 0 0 Ω , MX = Ω∗ 0 (5.39) where for later convenience we have defined 2β/Tc = d/k − iΓ. Clearly we have that M = M0 + MX . The interaction picture evolution operator VI is then defined by VI = V V0−1 , where V0 satisfies ∂τ V0 = iV0 M0 . We can write out V0 explicitly as V0 (τ ) = eiM0 τ since M0 is a constant. The equation of motion for VI is found to be ∂τ VI = (∂τ V ) V0−1 + V ∂τ V0−1 = (iVI V0 M ) V0−1 + VI V0 −iM0 V0−1 = iVI MI , (5.40) where MI = V0 MX V0−1 is the operator that generates time evolution in the interaction picture, given by MI = 0 Ω∗ e−2iβτ /Tc Ωe2iβτ /Tc . 0 (5.41) As described by Pechukas and Light [143] , the matrix elements of VI each satisfy a second order differential equation, which we find by differentiating the equation of 5.2 Solution in k-space 139 motion, ∂τ (∂τ VI ) = i (∂τ VI ) MI + iVI (∂τ MI ) = −VI MI2 + iVI (∂τ MI ) . (5.42) The square of MI is simply given by MI2 = |Ω|2 I, where I is the identity matrix. We can express the derivative of MI as ∂τ MI = MI G, where G is the diagonal matrix G= ∂τ Ω∗ Ω∗ − 2i Tβc 0 ∂τ Ω Ω 0 + 2i Tβc . (5.43) Inserting this into (5.42) gives ∂τ2 VI − (∂τ VI ) G + |Ω|2 VI = 0. (5.44) The boundary conditions are VI τ →−∞ = I, ∂τ VI τ →−∞ = iMI τ →−∞ . (5.45) The equation (5.44) is solved by making a temporal coordinate transformation. We define the normalized integrated Rabi frequency by 1 ω(τ ) = W Z τ −∞ |Ω(τ 0 )|2 dτ 0 , (5.46) 5.2 Solution in k-space 140 where W is a normalization related to the total energy in the control pulse, Z ∞ W = |Ω(τ )|2 dτ. (5.47) −∞ The coordinate ω runs from 0 to 1, as τ runs from −∞ to ∞ (see Figure 5.4 (b)). ω can be thought of as the time coordinate marked out by a clock that is powered by the control field. Using the control field profile (5.38), we can evaluate the integral in (5.46) explicitly, to get ω(τ ) = 12 tanh (a) 1 0 -6 -4 -2 0 2 4 6 τ T + 12 . (5.48) (b) -6 -4 -2 0 2 4 6 Figure 5.4 The Rosen-Zener model. The control field, shown in (a) with Ω0 = Tc = 1, takes the form of hyperbolic secant. (b): the integrated Rabi frequency ω marks out time at a rate given by |Ω|2 . Under the transformation τ −→ ω, the temporal derivative ∂τ transforms as 5.2 Solution in k-space 141 follows, ∂τ = (∂τ ω) ∂ω = = 1 τ 2 sech ∂ω 2Tc Tc ΩTc 2 1 ∂ω , 2Tc Ω0 (5.49) where we used the control shape (5.38). The second derivative is then given by ∂τ2 = = ∂τ2 ω ∂ω + (∂τ ω)2 ∂ω2 " # 1 ΩTc 2 1 ΩTc 4 2 (1 − 2ω) ∂ω + ∂ω . Tc2 Ω0 4 Ω0 (5.50) Putting all this together, and using ΩTc Ω0 2 = 4ω(1 − ω), (5.51) we find the equation ω(1 − ω)∂ω2 VI 1 + (∂ω VI ) + iβZ − ω + Ω20 VI = 0, 2 (5.52) where Z is the Pauli matrix 1 0 . Z= 0 −1 (5.53) 5.2 Solution in k-space 142 The boundary conditions, in terms of the new variable ω, are given by VI ω→0 = I, where θ± = 1 2 ∂ω VI ω→0 = iΩ0 0 ω −θ− ω −θ+ 0 , (5.54) ± iβ. The equation (5.52) is known as a hypergeometric differential equation. The solutions are known as hypergeometric functions, denoted2 by the symbol F , and parameterized by the coefficients appearing in the equation. The hypergeometric functions are special functions that can be evaluated using a mathematics application such as Matlab or Mathematica. Matching the general solutions to the boundary conditions (5.54), the solution for VI is given by VI (ω) = iΩ0 θ+ θ+ ω F (θ+ F (Ω0 , −Ω0 , θ+ , ω) iΩ0 θ− θ− ω F (θ− + Ω0 , θ+ − Ω0 , 1 + θ+ , ω) . F (Ω0 , −Ω0 , θ− , ω) + Ω0 , θ− − Ω0 , 1 + θ− , ω) (5.55) The properties of these special functions can be used to show that at the end of the storage process we have VI ω→1 = Γ2 (θ+ ) Γ(θ+ +Ω0 )Γ(θ+ −Ω0 ) i sin(πΩ0 ) cosh(πβ) i sin(πΩ0 ) cosh(πβ) Γ2 (θ− ) Γ(θ− +Ω0 )Γ(θ− −Ω0 ) , (5.56) where Γ(x) is the Euler Gamma function. We substitute these matrices into (5.36) to obtain an expression for the storage kernel in terms of k and ω. Some manipulations 2 Sometimes the symbol 2 F1 is used, and the designation Gauss hypergeometric function then distinguishes this from the generalized hypergeometric functions p Fq . 5.2 Solution in k-space 143 reveal that the determinant of the matrix (5.56) is 1, which simplifies forming the −1 inverse Vout . The exponential factor ei2βτ /Tc that enters when transforming from VI back to V = VI V0 translates into the factor e i2βτ /Tc = ω 1−ω iβ . (5.57) After some algebra, we arrive at the result √ iβ i d ω iQ(k, ω) e ×p , K(k, ω) = √ k 2π 1 − ω 2ω(1 − ω)/Tc (5.58) where we’ve defined the function Q in the following way Q(k, ω) = Ω0 θ− Γ2 (θ+ ) ω F (θ− + Ω0 , θ− − Ω0 , 1 + θ− , ω) Γ(θ+ + Ω0 )Γ(θ+ − Ω0 ) θ− sin(πΩ0 ) − F (Ω0 , −Ω0 , θ+ , ω). (5.59) cosh(πβ) It helps to keep in mind that θ± is a function of k through β. Note we have included the Jacobian factor p √ 2ω(1 − ω)/Tc = ∂τ ω in the denominator of (5.58), to make the transformation from τ to ω a unitary one (see §3.3.3 in Chapter 3). We would now like to extract the optimal input mode and its associated optimal efficiency by applying the SVD. However, just as in the case of (5.15), the kernel in (5.58) has a singularity at k = 0. As before, we remove this by transforming form k-space into λ-space, where λ = 2π/k is the wavelength of the spin wave excitation. Including 5.2 Solution in k-space 144 e in terms of λ and ω is, the Jacobian factor, the expression for K i e K(λ, ω) = 2π r Tc d −θ− (1 − ω)−θ+ Q(λ, ω), ω 2 (5.60) where Q is given, as before, by (5.59), the only difference being that in λ-space, the parameter β takes the form 2β/Tc = dλ 2π − iΓ. There is a minor pathology associated with the points ω = 0 and ω = 1, which blow up, but in practice this is easily addressed by introducing a small regularization that shifts the singularities into the complex plane, ω → ω + i. (5.61) After these steps, we have a non-singular kernel that is amenable to a numerical SVD. From this we obtain the singular values {λj }, and a set of input modes {φej (ω)}. The optimal storage efficiency is already given by η1 = λ21 . To find the temporal mode of the signal field that is stored with this optimal efficiency, we need to transform the mode φe1 (ω) back into the temporal domain. Including the Jacobian factor, we have 1/2 2 ω(τ ) [1 − ω(τ )] φe1 [ω(τ )] Tc sech Tτ e 1 τ 1 √ φ1 2 tanh +2 . Tc 2Tc φ1 (τ ) = = (5.62) e in (5.60) provides a check on the numerical optimizaThe analytic solution for K tions we present in §5.4 (see Figures 5.8 and 5.9). In fact, evaluating the function 5.2 Solution in k-space 145 F (a, b, c, ω) with complex a, b or c can be time-consuming, since these values are generated by analytic continuation of F into the complex plane. This procedure is not always accurate, and so the direct numerical optimizations presented at the end of this chapter are both faster and more reliable. Finally, the analytic solution (5.60) is, of course, only valid for the particular control (5.38). It would be more convenient if we could derive an expression that holds for a range of control field profiles. We now show how to construct an approximation to V that holds in the adiabatic limit, which is essentially the limit of a slowly varying control. 5.2.7 Adiabatic Limit The idea behind adiabatic evolution is to adjust Ω sufficiently slowly that at each moment we can neglect the time dependence of M , and treat the problem as if it were time-stationary. In this limit, the state |ψi remains in an instantaneous eigenstate of M at all times. As M changes, the eigenstates of M slowly evolve, and we arrange for the populated eigenstate at the end of the storage interaction to overlap with the | ↓i state; in this way excitations are transferred into the spin wave. To see how this works, we re-cast the equation of motion for V in terms of the adiabatic basis, which is the basis formed by the instantaneous eigenstates of M . Suppose M has the following eigenvalue decomposition (see §A.4.4 in Appendix A), M = RDR−1 . (5.63) 5.2 Solution in k-space 146 We define the operator Vad = V R, and differentiate it to obtain its equation of motion, ∂τ Vad = (∂τ V ) R + V ∂τ R = iV M R + V ∂τ R = iV RDR−1 R + V R R−1 ∂τ R = iVad Mad , (5.64) where Mad = D − iR−1 ∂τ R generates the time evolution in the adiabatic basis. The content of the adiabatic approximation is to neglect the term R−1 ∂τ R in Mad , so that Mad is a purely diagonal matrix. This allows us to solve the equation of motion for Vad , using the result (5.37). That is, Z Vad (τ ) = exp i τ 0 0 Mad (τ ) dτ . (5.65) −∞ Armed with this solution, we can construct the propagation matrix V = Vad R−1 , and therefore the storage kernel (5.36). We now implement this programme explicitly, after which the conditions under which the adiabatic approximation is justifiable will become clearer. Here we remark that corrections to the adiabatic approximation can be generated, in the current formalism, by making use of the Magnus expansion [143,145] , or Salzman’s expansion [146] , which provide approximations to the propagator V 5.2 Solution in k-space 147 when Mad is non-diagonal and time-dependent. These corrections quickly become unwieldy however, and so we do not present them here: the numerical approach presented in §5.4 obviates the need for them. To find the adiabatic kernel, we start by finding the instantaneous eigenvalues of M (τ ), by solving the equation |M − λI| = 0, where the vertical bars denote the determinant (see §A.4.2 in Appendix A). The resulting eigenvalues are λ± = b ± p b2 + |Ω|2 , (5.66) where we have defined 2b = d/k − iΓ (thus replacing the notation 2β/Tc defined in the previous section). Solving the equation M |±i = λ± |±i for the elements of the eigenvectors |±i, we find λ+ , |+i ∝ ∗ Ω λ− . |−i ∝ ∗ Ω (5.67) The diagonalizing transformation R is the matrix with |+i as its first column and |−i as its second. But there is some freedom as to the normalization of the vectors |±i. This is fixed by requiring that limτ →−∞ R = I, the identity. This just codifies our knowledge that Ω = 0 at the start of the storage interaction, so that M is initially diagonal, which means that no transformation need be applied to diagonalize it at 5.2 Solution in k-space 148 τ −→ −∞. A suitable form for R, that satisfies this boundary condition, is 1 R= Ω∗ λ+ λ− Ω∗ 1 . (5.68) Note that limτ →−∞ λ− = 0. The inverse transformation is then given by R−1 = 1 1− λ− λ+ 1 −Ω∗ λ+ −λ− Ω∗ 1 . (5.69) We have Mad = diag(λ+ , λ− ), so that Vad is given by i h R τ 0 0 0 exp i −∞ λ+ (τ ) dτ Vad (τ ) = i . h R τ 0 exp i −∞ λ− (τ 0 ) dτ 0 (5.70) Combining these results together and substituting them into (5.36), we find there is only a single non-vanishing term contributing to the storage kernel, i e K(k, τ) = − k r R∞ Ω∗ (τ ) exp −i τ λ− (τ 0 ) dτ 0 d × . 2π λ+ (τ ) − λ− (τ ) (5.71) We can further simplify this result if we make the assumption that |b| |Ω| (5.72) at all times. This does not have anything to do with the rate at which we change the control field, but it is usually considered as part of the adiabatic approximation, 5.2 Solution in k-space 149 as discussed in §5.2.9 below. With this approximation, we can write λ+ ≈ 2b + |Ω|2 , 2b and λ− ≈ − |Ω|2 . 2b (5.73) Inserting these expressions into (5.71), and making the replacement 2b = d/k − iΓ, we find, after a little algebra, " R∞ # √ Z ∞ id 0 )|2 dτ 0 |Ω(τ 1 1 d 2 e K(k, τ) = √ × |Ω(τ 0 )|2 dτ 0 × exp Γ τ . Ω∗ (τ ) exp − Γ1 k + i Γd k + i Γd 2π Γ τ (5.74) The first exponential factor represents the accumulation of phase due to the dynamic Stark effect, in which the strong control field ‘dresses’ the atoms and causes a time dependent shift in the |2i ↔ |3i transition frequency. The second exponential factor represents the ‘meat’ of the interaction: the response of the atoms to the incident signal field. The kernel (5.74) is already amenable to a numerical SVD. However, to make a connection with previous work, we take the inverse Fourier transform. We use the method detailed in §D.5.2 in Appendix D, and apply the shift theorem, to get the adiabatic z-space kernel √ Z ∞ d ∗ 1 0 2 0 |Ω(τ )| dτ + dz ×J0 K(z, τ ) = −i Ω (τ ) exp − Γ Γ τ ! √ sZ ∞ d 0 2 0 2i |Ω(τ )| dτ z . Γ τ (5.75) Note the additional contribution dz to the exponential factor. This contribution represents the change in refractive index experienced by the signal field as it propagates through the ensemble. On resonance (with Γ = 1 in normalized units), this 5.2 Solution in k-space 150 term describes exponential attenuation i.e. absorption, with an absorption coefficient of d. This explains why the quantity d is known as the optical depth: it directly quantifies the optical thickness of the ensemble on resonance. A numerical SVD could also be applied to the kernel (5.75) to obtain the optimal temporal input mode φ1 (τ ), and its associated optimal efficiency η1 = λ21 . This adiabatic solution appears in the work of Gorshkov et al. [133] , although the method of derivation differs slightly. They provided a method to find the optimal input mode, in the limit of large control field energy, but they did not use the SVD: The SVD is a more direct method, and it does not require the assumption of large control energy. We will discuss the differences between the SVD method and the method of Gorshkov et al. shortly. First, we make a coordinate transformation that removes the explicit dependence of K on the shape of the control field. As in the Rosen-Zener solution above, we transform from τ to ω, where ω is the normalized integrated Rabi frequency, defined by (5.46). We do not assume a particular shape for the control: its profile can be arbitrary in the present case (within the limits of the adiabatic approximation, to be discussed below). By removing the dependence on the control field with this coordinate transformation, we only need to perform the SVD once, in the transformed coordinate system, in order to obtain the optimal input mode for any control field shape. To make the coordinate transformation unitary, we include a Jacobian factor of √ W /Ω∗ (τ ) in the transformed kernel (see §3.3.3). This rather conveniently cancels with the factor of Ω∗ (τ ) in (5.75), so the transformed kernel 5.2 Solution in k-space 151 can be written as √ K(z, ω) = −i dW −(W −W ω+dz)/Γ √d p e J0 2i Γ (W − W ω)z , Γ (5.76) A final, cosmetic simplification is achieved by flipping the ω coordinate, ω −→ 1 − ω. This has no effect on the singular values of the kernel, but simply flips the input modes around. The optimal mode for adiabatic storage in a Λ-type quantum memory, with an arbitrary control field profile, is now found by taking the SVD of √ K(z, ω) = −i dW −(W ω+dz)/Γ √dW √ e J0 2i Γ ωz . Γ (5.77) We note that both ω and z run from 0 to 1. The optimal temporal input mode, is then found from the mode φ1 (ω) by the relation Ω(τ ) φ1 (τ ) = √ φ1 [1 − ω(τ )] . W (5.78) √ The factor of Ω/ W arises from the Jacobian relating the coordinates ω and τ . We now have a simple prescription for finding the optimal input mode for a quantum memory. Given a fixed value for W , which essentially quantifies the total energy in the control pulse, and given values for the detuning and the optical depth, we form the kernel (5.77) and take the SVD. Then, for any arbitrary shape of the control, we can construct the optimal input mode φ1 , using the transformation (5.78) above. Some examples of the optimal input modes predicted using this approach can be 5.2 Solution in k-space 152 found in Figures 5.8, 5.9, 5.10 and 5.11 in §5.4 at the end of this chapter. In general, the storage efficiency depends on the optical depth, the detuning (through Γ), and the control pulse energy, through W , since the kernel (5.77) depends on all of these quantities. We also know however, from the discussion in §5.2.3, that the best possible storage efficiency only depends on the optical depth. Below we connect these two results together. 5.2.8 Reaching the optimal efficiency In §5.2.3 we derived an expression for the optimal storage efficiency possible in a Λ-type ensemble memory. In fact, it is possible to reach this optimal efficiency in the adiabatic limit. As was first shown by Gorshkov et al. [133] , the anti-normally ordered kernel formed from the adiabatic storage kernel (5.77) is equal to the optimal kernel (5.27), in the limit of large control pulse energy. To see this, we substitute (5.77) into the expression (3.19) for the anti-normally ordered kernel, and perform the integral over ω. In the limit W −→ ∞, we can evaluate the integral analytically. After some leg work — see §D.5.3 in Appendix D — we find that the result is exactly (5.27). That is, Z lim W →∞ 0 1 ∗ 0 Z K(z, ω)K (z , ω) dω = 0 ∞ K(z, ω)K ∗ (z 0 , ω) dω = √ d −d(z+z 0 )/2 e I0 (d zz 0 ). 2 (5.79) This shows that it is possible to saturate the upper bound on the storage efficiency, even in the adiabatic limit. The adiabatic limit is not only useful because we can 5.2 Solution in k-space 153 construct the optimal input mode explicitly. It is also useful because it is possible to shape the input mode by shaping the control. From the form of (5.78), it is clear that, by an appropriate choice of Ω(τ ), we can choose the shape of φ1 (τ ). In particular, we can choose the control so that φ1 (τ ) matches the temporal profile of some ‘given’ input field. This is of considerable practical importance, since it may be experimentally much easier to shape the bright control field, than to shape the weak input field (this is discussed in Chapter 8). The combination of these two facts — that adiabatic storage can be optimal, and also that it enables one to shape the input mode — makes the adiabatic limit an important regime for the operation of a quantum memory. Under what circumstances is the equality in (5.79) achieved? The limit of W −→ ∞ is not really required. Examining the form of (5.77), it is clear that we just need to make W large enough that the exponential factor e−W ω/Γ , evaluated at the limit ω = 1, is sufficiently small that extending the integral further would make no difference. We should certainly have that W |Γ| then. In addition, we should ensure that any contribution from the Bessel function is negligible at ω = 1. Using √ the approximation J0 (ix) = I0 (x) ∼ ex / x for x 1, we see that we should also have that W d. To summarize, adiabatic storage is optimal if the control pulse is sufficiently energetic that the conditions W max (|∆|, d) (5.80) 5.2 Solution in k-space 154 are satisfied. Recall that ∆ is the common detuning of the signal and control fields from resonance. We can connect W with the total energy Ec in the control pulse. The total energy is Ec = A Z ∞ Ic (τ ) dτ, (5.81) −∞ where Ic = 20 c|Ec |2 is the cycle-averaged control pulse intensity. From the definition of the Rabi frequency (4.49), we then find 20 cA Ec = γW. d23 .vc 2 ~ (5.82) Here the presence of the factor of γ indicates that we have converted back into ordinary units (rather than normalized units). From the definition of W (5.47), we find it has the dimensions of frequency, and so in ordinary units it is accompanied by the factor γ. Let us define the number of photons in the control pulse by Nc = Ec /~ωc . We also define Na as the number of atoms in the ensemble addressed by the optical fields, Na = nLA. Using (4.50) and (5.12), we can express the optical depth as ∗ d12 .vs 2 ~ωs × Na . d= ~ 20 cAγ (5.83) With these definitions, we see that the condition W d in (5.80) amounts, essentially, to the condition Nc Na (we have used ωs ≈ ωc and |d∗12 .vs | ≈ |d23 .vc |). We might therefore describe this condition as describing a ‘light-biased’ interaction, where the number of control photons dominates over the number of atoms; this is 5.2 Solution in k-space 155 why it is the latter quantity that limits the efficiency. In the work of Gorshkov et al. [133] , a method is presented for optimizing the adiabatic storage efficiency in this light-biased limit. The method works by combining the optimal anti-normally ordered kernel (5.27), whose eigenfunctions are the optimal spin waves, with the adiabatic storage kernel (5.75), which connects the control field profile to the signal field profile. Combining these two kernels together is possible only when Nc Na , and also W ∆. In this limit their method works extremely well. One advantage of the SVD method however, is that we can apply it directly to the kernel (5.77), without making this approximation, and we can therefore find the optimal input modes for arbitrary values of W . 5.2.9 Adiabatic Approximation We have employed several approximations in the name of adiabaticity. We now examine the physical content of these approximations. The first assumption we made, in the text following (5.64), was to neglect the term R−1 ∂τ R in the adiabatic generator Mad . Clearly this term vanishes if the control field is held constant, and so the size of this term is set by the rate of variation of the control field. To make the adiabatic approximation, we must therefore limit the bandwidth of the control pulse. To find this limit, we introduce the second approximation we made in (5.72), namely that |b| |Ω|. (5.84) 5.2 Solution in k-space 156 With this approximation, we can write the diagonalizing transformations R, R−1 in the form 1 R= Ω∗ 2b Ω − 2b 1 ; R−1 = 1 ∗ − Ω2b Ω 2b 1 . (5.85) For adiabatic evolution, we should have that ||R−1 ∂τ R|| ||D||, where D = diag(2b + |Ω|2 /2b, −|Ω|2 /2b). Here the double bars represent the Frobenius norm, which is found by adding in quadrature the magnitudes of all the elements in a matrix. We neglect terms of order |Ω/b|, which are small by assumption. We then arrive at the condition ∂τ Ω 2 2 2b |Ω| . (5.86) Recall that b = 21 (d/k − iΓ) depends on the wavevector k, and that Ω varies with time, so we should be careful to satisfy this condition for all the values of these two parameters that play a role in the storage process. We can distinguish two regimes. First, the EIT regime (see §2.3.1 in Chapter 2), in which the optical fields are tuned into resonance with the excited state. In this case, ∆ = 0 and so Γ = 1 (in normalized units). Since d 1 for reasonably efficient storage, the contribution to b from the complex detuning Γ is small, and we have |b| ∼ d/2k. Therefore the adiabatic conditions (5.84) and (5.86) vary strongly with k, and we must take some care to identify the range of wavevectors that are important for the storage process. The second regime is the Raman regime (see §2.3.2 in Chapter 2), in which the detuning is large compared to the excited state linewidth; ∆ 1 in normalized 5.2 Solution in k-space 157 units. In this regime |b| ∼ 21 (d/k − ∆). In this case the large contribution to b from the detuning makes the adiabatic conditions less dependent on k. What range of wavevectors are important in the storage process? One answer is provided by inspection of the k-space kernel (5.74). Using the expansion (k+i Γd )−1 = −i Γd + Γ2 k d2 3 + i Γd3 k 2 + . . ., we can re-write the k-space map in the form i e K(k, ω) ≈ − √ 2π r ΓW W 2 W × ei d ωk × e− d2 ωk , d (5.87) where we have introduced the integrated Rabi frequency ω, and ‘flipped’ the kernel, as we did in deriving (5.77). This expression is a good approximation if |d/Γ| is large, which is not guaranteed in the Raman regime, but which is generally true in the EIT regime. The final exponential factor describes a Gaussian profile in k-space, with a characteristic width given by δk = δk(ω) = √ d , Wω (5.88) in the EIT regime with Γ = 1. In this case the most restrictive form of the adiabatic condition (5.84) can be written as (dropping an unimportant factor of two) Ωmax d , δk(ωmax ) (5.89) where Ωmax = max(Ω) is the peak Rabi frequency of the control, and where ωmax is the value of the integrated Rabi frequency when this peak occurs. For a symmetric 5.2 Solution in k-space 158 control pulse, we would have ωmax = 21 ; in general ωmax will be some fraction that for our purposes we may approximate as ∼ 1. Substituting (5.88) into (5.89), we find that this condition in fact restricts the bandwidth δc of the control pulse: δc 1. (5.90) Here we made the approximation W ∼ Ω2max /δc . The condition (5.90), in ordinary units, is δc γ. That is, the bandwidth of the control should not exceed the natural linewidth of the |2i ↔ |3i transition in the ensemble, to achieve adiabatic storage on resonance. This suggests that EIT is a memory protocol best suited for the storage of narrowband fields. However, the analysis by Gorshkov et al. [133] reaches a different conclusion: that the adiabatic restriction on the bandwidth of the control is δc dγ (in ordinary units). This much less stringent condition is derived by evaluating the adiabatic condition (5.86) using b ∼ d/max(δk) with max(δk) ∼ 1. Identifying the maximum of the quantity |∂τ Ω/Ω| with the bandwidth δc , one arrives at their result. One justification for using δk ∼ 1 is that for optimal storage, we only need to access the optimal spin wave mode; higher modes are irrelevant. Since the optimal mode is that mode which is most slowly varying in space, its width in k-space is limited to a relatively small region, and the adiabatic approximation need only be satisfied within this range. It is difficult to argue rigorously about these approximations, but as we will see, numerics reveal that the reality lies somewhere between these two cases: the adiabatic approximation breaks down rather quickly in the EIT regime 5.3 Raman Storage 159 as the bandwidth approaches the natural linewidth, although it is true that this is mitigated somewhat by increasing the optical depth. In the Raman regime, the analysis is simpler. The adiabatic condition is simply that Ωmax |∆|, (5.91) independent of k. We might comment that there may be some particular value of k such that there is a cancellation, and b becomes small, but the effect of this isolated point is generally negligible. The limitation on the control bandwidth comes from (5.86), which yields the condition δc |∆|. (5.92) Therefore adiabatic evolution is guaranteed in the Raman case whenever the detuning ∆ is the dominant frequency involved in the interaction, with both the Rabi frequency and the bandwidth of the control field small by comparison. 5.3 Raman Storage So far we have studied the properties of storage in a Λ-type ensemble for arbitrary values of the detuning. We now specialize to the case of large detunings, ∆ γ (or ∆ 1 in normalized units). In this Raman regime the storage kernel simplifies further, and an interesting connection between the input and spin wave modes emerges. The following treatment forms the basis of our Rapid Communication on 5.3 Raman Storage 160 Raman storage [77] . Taking the limit ∆ 1, we can write Γ ≈ −i∆, and the storage kernel (5.77) becomes √ K(z, ω) = C × e−i(W ω+dz)/∆ × J0 (2C ωz), (5.93) where we have defined the Raman memory coupling by √ C= Wd ≈ |Γ| √ Wd . ∆ (5.94) The exponential factor in (5.93) represents only phase rotations applied to the ω and z coordinates. We can absorb these phases into the signal field and spin wave, and therefore we can drop them from the kernel (we will be careful to ‘put them back’ when we write down the optimal modes). We can now write down a very simple recipe for constructing the optimal input mode in the Raman regime. First, we form the kernel √ K(z, ω) = CJ0 (2C ωz). (5.95) This depends only on the memory coupling C, therefore this parameter uniquely determines the efficiency of the memory in the Raman limit [77] . We note that the kernel K is real, and symmetric under the exchange of ω and z. That is, K is Hermitian (see §A.4.3 in Appendix A). Therefore, its SVD is the same as its spectral decomposition (see §3.1.3 in Chapter 3). The singular values of K are also its eigenvalues, and the input modes {φj }, as functions of ω, have the same form as 5.3 Raman Storage 161 the spin wave modes {ψj }, as functions of z. Their phases differ though, because of the phase rotations we ‘absorbed’. To be precise, let us define the functions {ϕj } as the eigenfunctions of the kernel (5.95). That is to say, ϕj satisfies Z 1 √ CJ0 (2C xy)ϕj (x) dx = λj ϕj (y). (5.96) 0 The optimal input mode for the signal field, including the correct phase rotation and transforming back from ω to τ , is given by 1 φ1 (τ ) = √ Ω(τ ) × exp W iW [1 − ω(τ )] ∆ × ϕ1 [1 − ω(τ )]. (5.97) The optimal output mode for the spin wave, to which this optimal input mode is mapped by the storage process, is given by ψ1 (z) = e−idz/∆ ϕ1 (z). (5.98) The optimal storage efficiency is given by the square of the largest eigenvalue in (5.96); η1 = λ21 . Figure 5.5 shows the variation of this optimal efficiency with C. In taking the Raman limit, we have neglected spontaneous emission. This means that the predicted efficiency can approach unity, as long as C is large enough, even if the optical depth is low. That is, a smaller d can be ‘compensated’ by a larger W — a more energetic control. Of course the optimal efficiency derived in §5.2.3 remains correct, even in the Raman limit: the best achievable efficiency is always 5.3 Raman Storage 162 limited, through spontaneous emission, by the optical depth. If the efficiency predicted by (5.96) is larger than the upper limit (5.25), then we have reached a regime where spontaneous emission dominates over other losses. However, generally a Raman memory requires a large optical depth to operate efficiently, as we discuss below. Therefore the dominant loss mechanism in a Raman memory is not spontaneous emission, but insufficient coupling. That is, the large detuning from resonance makes the interaction weak, so that the biggest problem in getting a Raman memory to work is to make the coupling strong enough that the signal field is completely absorbed. The utility of the kernel (5.95), is that it provides a simple way to analyze the Raman limit before spontaneous emission becomes a limitation. Examples of the optimal modes predicted by the Raman kernel are shown in Figures 5.8 and 5.10 in §5.4 at the end of this chapter. What is notable about the behaviour of the (a) (b) 1 10 0.8 10 0.6 10 0.4 10 0.2 10 0 0 0.5 1 1.5 2 2.5 3 3.5 4 0 −1 −2 −3 −4 −5 10 0 0.5 1 1.5 2 2.5 3 3.5 4 Figure 5.5 Raman efficiency. (a) the optimal storage efficiency η1 = λ21 predicted by the kernel √ (5.95) in the Raman limit ∆ 1, versus the memory coupling C = W d/|Γ|. The efficiency ‘saturates’ at around C ∼ 2. (b) A logarithmic plot of the difference 1 − η1 between the predicted efficiency and unity. Raman efficiency is that it rises steeply for small values of C, before ‘saturating’ 5.3 Raman Storage 163 at C ∼ 2. Physically, this saturation point coincides with the stimulated scattering regime. For ordinary Stokes scattering, the scattering process becomes efficient as the coupling is increased beyond this point (see Figure 10.6 in Chapter 10). In the case of a Raman quantum memory, the transmission of the signal field through the ensemble drops sharply, and the efficiency of the memory becomes limited by spontaneous emission, rather than insufficient coupling. Therefore in designing a Raman memory, one need only ensure that C & 2, in order that the scheme is viable (see §9.8 in Chapter 9 and §10.9 in Chapter 10). 5.3.1 Validity As mentioned above, in deriving the Raman kernel we neglected spontaneous emission. We did this tacitly when we dropped the real part of Γ in the exponential factor appearing in the storage kernel. This is valid so long as neither W nor d is too large. To see this, we define the balance R according to the relation r R= W . d (5.99) Note that the balance has no relation to the matrix R introduced in §5.2.7 above. Using the arguments employed in §5.2.8, we see that the balance really expresses the extent to which the interaction is dominated by light, or matter: R2 ∼ Nc . Na (5.100) 5.3 Raman Storage 164 That is, the case R 1 corresponds to a light-biased interaction — as described earlier, this is the limit in which the adiabatic kernel saturates the upper bound on the storage efficiency — and the case R 1 describes a matter-biased interaction, with a weak control, but a large/dense ensemble. Using the balance, and the definition C= √ W d/|Γ|, we can re-express the adiabatic storage kernel (5.77) as follows, √ K(z, ω) = Ce−iθ × e−iC cos θ(Rω+z/R) × eC sin θ(Rω+z/R) × J0 (2Ce−iθ ωz), (5.101) where we have defined the phase angle θ such that tan θ = 1/∆; the complex detuning being given by Γ = −i|Γ|eiθ . (5.102) The Raman limit is the limit of large detuning, which corresponds to the limit θ 1. The exponential factor involving cos θ in (5.101) corresponds to the phase rotations we dealt with previously. On the other hand, the exponential factor involving sin θ comes from the real part of Γ, and represents spontaneous emission. It cannot be removed by a unitary transformation of the modes, and it reduces the storage efficiency. This is the term we neglect in the Raman limit. For this approximation to hold for both the optical and spin wave modes together, both terms in the exponent should be small. This is true provided we satisfy the conditions C sin θR 1, and C sin θ × 1 1. R (5.103) 5.3 Raman Storage 165 Using sin θ ≈ tan θ, setting C ∼ 1 (as we should have for a reasonably efficient memory), and using (5.100), we find that the Raman kernel is a good approximation whenever ∆ γ, and γ2 Nc ∆2 , ∆2 Na γ2 (5.104) using ordinary units. That is to say, the interaction should be roughly ‘balanced’, with broadly equal contributions to the coupling originating from atoms and light. The larger the detuning, the more ‘leeway’ there is to bias the interaction one way or the other. Suppose that we set R ∼ 1, so that d ∼ W . For reasonable efficiency, we should have C ∼ 1, or thereabouts. Squaring this, and using R ∼ 1 then, we find d ∼ ∆. In a Raman memory, ∆ 1, and therefore d 1. That is, a Raman memory described by (5.95) generally requires a large optical depth. This explains why spontaneous emission is not as important a limiting factor as is the issue of sufficiently strong coupling. Finally, we comment that since ∆ ∼ d, the adiabatic condition on the bandwidth of the control field can be written as δc d, or in ordinary units δc dγ. This is the same as the condition derived by Gorshkov et al. for the limitation on the control bandwidth for resonant EIT storage. It is therefore arguable that a Raman memory does not allow for more broadband storage than an EIT memory does. But the adiabatic condition in the Raman case is rather more robust that it is in the EIT case, since it does not depend on identifying an ‘important’ region in kspace. Numerics show that the adiabatic storage kernel is a better approximation 5.3 Raman Storage 166 in the Raman case, for more broadband control pulses, than it is in the EIT case. Broadband storage provided the motivation for studying the Raman memory, but its advantages in this respect are not clear cut. Other considerations that may favour detuning from resonance include the possibility of dealing with a more complex excited state manifold. Suppose that, instead of a single excited state |2i, there are a host of states, perhaps resulting from spinorbit or hyperfine splitting (see, for example, Figure 10.3 in Chapter 10). Tuning into resonance with one of these states could make the dynamics rather complicated. There may be some direct absorption of the signal field into the nearby states, followed by spontaneous emission and loss. Detuning away from all of the states puts the contribution from each state on an equal footing, so that the dynamics can be treated just as we did the simple three level system, where we need only swap the coupling C to a single state for an equivalent coupling that includes the scattering amplitudes for all the states (see §F.4 in Appendix F). And by detuning we eliminate the possibility of absorption losses. In addition, a Raman memory is tunable, since if we tune further from resonance, we can maintain strong coupling by increasing the control pulse energy. The Raman memory is also affected less adversely by inhomogeneous broadening of the excited state than an EIT memory might be. Again, this is because the coupling to the ensemble is not dominated by resonance with a single frequency. 5.3 Raman Storage 5.3.2 167 Matter Biased Limit We have already shown how the anti-normally ordered product formed from the adiabatic storage kernel tends to the optimal kernel (5.27) in the limit of large control pulse energy. This is the light-biased limit, with R 1. To reach this limit, the balance R should exceed the upper limit in (5.104), so we should have R > ∆ (in normalized units). Now, there is a degree of symmetry to the structure of (5.101): The kernel is unchanged when we swap z and ω, if at the same time we send R −→ 1/R. It therefore follows that in the matter-biased limit R 1 — that is Na > ∆Nc /γ in ordinary units — the normally ordered kernel tends to a limit defined not by the optical depth, but by the control pulse energy: Z lim d→∞ 0 1 K ∗ (z, ω)K(z, ω 0 ) dz = √ W −W (ω+ω0 )/2 e I0 (W ωω 0 ). 2 (5.105) The kernel on the right hand side has precisely the same form as (5.27), except that the efficiency is limited by W rather than by d. Therefore, if we are limited by the energy of the control, the optimal efficiency achievable is given by η1 = 1 − 2.9/W . Of course, extremely high energy lasers are readily available, whereas the size of the ensemble is generally not easily varied. Nonetheless, in cases where the ensemble may be damaged by a high energy laser pulse — as may be the case for a solid state memory — it may be that the control energy becomes a limitation. 5.3 Raman Storage 5.3.3 168 Transmitted Modes. In this section we describe a connection between the modes that are stored in the memory, and the modes that are transmitted through it when the efficiency is not perfect. This connection holds in the Raman limit, and it is clearest when we form the equations of motion for the signal field and the spin wave in the adiabatic approximation. This is the way the adiabatic approximation is most commonly introduced, and so it is informative to run through the procedure. We start with the equations of motion (5.3), which we reproduce below in normalized units, √ ∂z A(z, τ ) = − dP (z, τ ), ∂τ P (z, τ ) = −ΓP (z, τ ) + √ dA(z, τ ) − iΩ(τ )B(z, τ ), ∂τ B(z, τ ) = −iΩ∗ (τ )P (z, τ ). (5.106) The adiabatic approximation is made by setting ∂τ P = 0 on the left hand side of the second equation. This is reasonable when the natural dynamics of the optical polarization P are overwhelmed by the motion driven by the signal and control fields — known as adiabatic following. In this situation, 1/|Γ| is the shortest timescale in the problem, so that when far-detuned we must have (δc , Ωmax ) |∆|, precisely the adiabatic conditions (5.91), (5.92) derived above. We can then solve the second equation for P algebraically, and substitute the result into the first and third 5.3 Raman Storage 169 equations. The result is √ Ω d A = i B, Γ √ |Ω|2 Ω∗ d ∂τ + B = −i A. Γ Γ d ∂z + Γ (5.107) We then switch coordinates from (τ, z) to (ω, z), and we define new variables α and β for the signal field and spin wave as follows α(z, ω)e−(W ω+dz)/Γ = √ W A(z, τ ) , Ω(τ ) β(z, ω)e−(W ω+dz)/Γ = B(z, τ ). (5.108) These new variables incorporate the Jacobian factor associated with the coordinate transformation, and also the phase rotations associated with the dynamic Stark shift and the ensemble refractive index. In general, the transformation linking A and B to α and β is not quite unitary, because Γ is not strictly imaginary, so the exponential factors on the left hand side of (5.108) make the norms of {A, B} different to the norms of {α, β} — they are not pure phase rotations. However, taking the Raman limit ∆ 1, we find Γ ≈ −i∆, and in this case, the transformation is unitary. This will be important shortly. For now, observe that the equations of motion simplify greatly when they are cast in terms of the transformed variables {α, β}. We have removed the control field, and also the homogeneous terms on the left hand side, so 5.3 Raman Storage 170 that we obtain the system ∂z α = −Cβ, ∂ω β = Cα. (5.109) Note the symmetry of this system of equations. If we swap α for β, and β for −α, and then swap z and ω, the system of equations is unchanged. As we will see below, this symmetry simplifies the form of the solution, so that there are only two different Green’s functions, instead of a potential four. We solve these equations by applying a unilateral Fourier transform. Having eliminated the control field, we can choose either to apply the transform over the ω coordinate or the z coordinate. We will apply the transform over the z coordinate, as we have done previously, and solve the equations in k-space. Using tildes to denote the transformed variables, we have α0 −ik α e− √ 2π e = −C β, ∂ω βe = C α e, (5.110) where α0 = α(z = 0, ω) represents the boundary condition for the signal field. Solving the first equation for α e yields α e = −i Ce i β+√ α0 . k 2πk (5.111) 5.3 Raman Storage 171 Substituting this result into the second equation, and integrating, we obtain the solution for β, e ω) = e−iC 2 ω/k βe0 + i √C β(k, 2πk Z ω e−iC 2 (ω−ω 0 )/k α0 (ω 0 ) dω 0 , (5.112) 0 where βe0 is the Fourier transform of the spin wave boundary condition β0 = β(z, ω = 0). Finally, substituting this back into (5.111) gives the solution for α, C i C2 2 α e(k, ω) = −i e−iC ω/k βe0 + √ α0 (ω) + √ k 2πk 2πk 2 Z ω e−iC 2 (ω−ω 0 )/k α0 (ω 0 ) dω 0 . 0 (5.113) These k-space solutions are singular at k = 0, but fortunately it is possible to take the inverse Fourier transform analytically. We use the results described in §D.5 in Appendix D, along with the convolution theorem for the unilateral Fourier transform (see §D.4.2 in Appendix D), to obtain the rather formidable-looking solution Z ω r h p i z 0 ) α (ω 0 ) dω 0 J 2C z(ω − ω α(z, ω) = α0 (ω) − C 1 0 ω − ω0 0 Z z h p i −C J0 2C (z − z 0 )ω β0 (z 0 ) dz 0 , 0 Z zr i h p ω 0 )ω β (z 0 ) dz 0 β(z, ω) = β0 (z) − C J 2C (z − z 1 0 z − z0 0 Z ω h p i J0 2C z(ω − ω 0 ) α0 (ω 0 ) dω 0 . +C (5.114) 0 We re-write this in terms of Green’s functions, or propagators, in order to bring out 5.3 Raman Storage 172 its structure. Z 1 0 0 0 L(ω, ω )αin (ω ) dω − 0 Z Z 1 L(z, z 0 )βin (z 0 ) dz 0 + βout (z) = αout (ω) = 0 1 Z K(ω, z)βin (z) dz, 0 1 K(z, ω)αin (ω) dω, (5.115) 0 where αout (ω) = α(z = 1, ω) is the transmitted signal field, and βout = β(z, ω = 1) is the spin wave at the end of the interaction. We have introduced the Green’s functions L and K, defined as follows, L(x, y) = δ(x − y) − CΘ(x − y) × √ h p i K(x, y) = CJ0 2C x(1 − y) . √ 1 J1 (2C x − y), x−y (5.116) Note that only two distinct Green’s functions are required, because of the symmetry between α and β in the adiabatic equations of motion (5.109). The Green’s function K(x, 1−y) is precisely the adiabatic Raman storage kernel (5.95), and this is why we have used the same notation. The kernel L describes the relation of the transmitted signal field to the input field, or equivalently it relates the final to the initial spin wave. We are considering storage, so that βin = 0, but note that the same solutions can be used to describe retrieval from the memory; retrieval is dicussed in Chapter 6. The Heaviside step function Θ(x) in L makes the kernel causal, so that the transmitted signal field is never influenced by future values of the input field. We now show that there is a connection between the SVD of K, which tells us about the optimal input mode and its associated efficiency, and the SVD of L, 5.3 Raman Storage 173 which tells us about the transmitted fields. Note that it is only correct to associate the singular values of the Green’s function K with storage efficiencies when the transformation connecting {A, B} to {α, β} is unitary, and this is only true in the Raman limit ∆ 1. Therefore the following analysis applies only in this limit, and we will assume that we are tuned far from resonance in the remainder of this section. To see the connection between the SVDs of K and L, it will help us to use matrix notation, since it is much more compact. As discussed in Chapter 3, the solutions 5.115 may be considered as the infinite-dimensional limit of the following matrix equations, |αout i = L|αin i − K|βin i, |βout i = L|βin i + K|αin i. (5.117) To be more precise, we define the vector |αout i as a discretized version of the continuous function αout , αout (0) α ( 1 ) out N −1 |αout i = αout ( N 2−1 ) .. . αout (1) , (5.118) where N is the number of discretization points. The other vectors are defined 5.3 Raman Storage 174 similarly. The matrices K and L are given in terms of the continuous kernels by Kjk = K j−1 k−1 , N −1 N −1 1 × , N −1 Ljk = L j−1 k−1 , N −1 N −1 × 1 . N −1 (5.119) The continuous equations are recovered by taking the limit N −→ ∞, when matrix multiplication is replaced by integration, and the factors 1/(N − 1) become the increments dz, dω. It is useful to be able to swap between the continuous and discretized representations, in order to use both the machinery of calculus and linear algebra in their ‘natural habitats’. The relationship between the SVDs of the matrices K and L is fixed by the equations of motion (5.109), which are unitary. To see this, observe that these equations imply the identity ∂z (α† α) + ∂ω (β † β) = 0. (5.120) We have used the † notation for Hermitian conjugation, rather than the ∗ notation for complex conjugation, since this equation holds whether or not we treat α, β as ordinary functions, or as quantum mechanical annihilation operators3 . Integrating this identity over all z and ω gives the condition Nα,out + Nβ,out = Nα,in + Nβ,in , 3 (5.121) Strictly, the Langevin noise operators introduced in §4.12 of Chapter 4 contribute unless an expectation value is taken, but in the Raman limit we neglect spontaneous emission, and these operators along with it. 5.3 Raman Storage 175 where Nα,out , Nα,in are the numbers of transmitted and incident signal photons, respectively, and where Nβ,out , Nβ,in are the numbers of final and initial spin wave excitations: 1 Z Nα,out = 0 Z Nβ,out = 0 † αout (ω)αout (ω) dω, 1 † βout (z)βout (z) dz, Z 1 Nα,in = Z01 Nβ,in = 0 † αin (ω)αin (ω) dω, † βin (z)βin (z) dz. (5.122) The condition (5.121) fixes the combined transformation of signal field and spin wave, implemented by the memory interaction, as unitary, meaning that the total number of ‘particles’ is conserved. This unitarity holds in the Raman limit approximately, since we have neglected spontaneous emission, which process would scatter particles into modes other than the signal field or the spin wave, thus violating the conservation law (5.121). As discussed in §5.3.1, this approximation is generally a good one for a Raman memory, since large optical depths are required, making spontaneous emission losses negligible. The conservation condition (5.121) has the same form as that for a beamsplitter, which mixes a pair of input modes without loss, to produce a pair of output modes. Indeed, this is a helpful perspective from which to view the action of the Raman memory (see Figure 5.6 below). The difference with a conventional beamsplitter, as we will see, is that the Raman interaction couples multiple modes together in a pairwise fashion, with each pair of modes ‘seeing’ a different reflectivity. We derive this fact by combining the solutions for α and β into 5.3 Raman Storage 176 a single transformation, |xout i = U |xin i, (5.123) where the vectors |xout i and |xin i are defined by |αout i |xout i = , |βout i |αin i |xin i = , |βin i (5.124) and where the matrix U is given by L −K . U = K L (5.125) In this notation, the conservation condition (5.121) is written as hxout |xout i = hxin |xin i. (5.126) Substituting the transformation (5.123) into the above condition, we find that U † U = I, where I is the identity matrix. Using this result, we multiply (5.123) by U † from 5.3 Raman Storage 177 the left, to obtain the inverse transformation |xin i = U † |xout i. (5.127) Substituting this into the conservation condition, we then find U U † = I. So we see that U is indeed a unitary transformation (see §A.4.5 in Appendix A). To see the implications of this for the matrices K and L, we substitute the form (5.125) for U into the conditions U † U = U U † = I, and perform the matrix multiplications. This yields the conditions L† L + K † K = LL† + KK † = I. (5.128) In terms of the normally and antinormally ordered products, these conditions are written as LN + KN = I; LA + KA = I. (5.129) Therefore we must have that LN and KN commute: [LN , KN ] = [LN , I − LN ] = [LN , I] − [LN , LN ] = 0. (5.130) And similarly for the antinormally ordered products, [LA , KA ] = 0. As discussed in §3.1.2 of Chapter 3, the eigenvectors of KA are the output modes of K, and the eigenvectors of KN are its input modes. That is, if we write the SVD of K as † 2 U † , and K 2 K = UK DK VK† , then KA = UK DK N = VK DK VK . The fact that KA K commutes with LA implies that the eigenvectors of LA are the same as those of KA 5.3 Raman Storage 178 (see §A.3.1 in A). So we can write LA as 2 † LA = UK DL UK , (5.131) 2 is a diagonal matrix of positive eigenvalues. The same argument applies where DL for the normally ordered products, so that we have 2 † LN = VK DL VK . (5.132) Note that the eigenvalues of LN are necessarily the same as those of LA . To summarize, the unitarity of the memory interaction constrains the matrices K and L such that their SVDs are built from a common set of input and output modes: L = UK DL VK† K = UK DK VK† . (5.133) And substituting these expressions into the conditions 5.129, we find the following relationship between the singular values, 2 2 DL + DK = I. (5.134) Our analysis is nearly complete. Let us review what we have discovered. A general transformation from |xin i to |xout i could involve four Green’s functions — that is, there could have been four matrices to deal with, one for each of the elements of U in (5.125). But the symmetry of the equations of motion (5.109) allows us to 5.3 Raman Storage 179 express the transformation in terms of just two Green’s functions, L and K. Each of these has an SVD, so there are potentially four sets of orthogonal modes to deal with, associated with the input and output modes of the two Green’s functions L and K. The further condition of unitarity on U , however, allows us to reduce the number of orthogonal sets from four down to two, since L and K must share the same input modes, and also the same output modes, as one another. We now show that in fact the output modes are just flipped versions of the input modes, so that the entire interaction can be fully described using just a single set of orthogonal modes. The final step is accomplished by noticing that the matrices L and K are persymmetric. That is, they are symmetric under reflection in their anti-diagonal (see §3.1.4 in Chapter 3). This can be seen by examining the functional forms (5.116) of the Green’s functions. For example, the value of L(x, y) only depends on the difference x − y. Therefore the contours of L all lie parallel to the line y = x, which corresponds to the main diagonal of the matrix L. L is accordingly unchanged by a reflection in the main anti-diagonal, that is to say, L is persymmetric. That K is also persymmetric follows from the fact that K(x, 1 − y) is Hermitian. As shown in §3.1.4 in Chapter 3, the input and output modes associated with the SVD of a persymmetric matrix are simply ‘flipped’ versions of one another. Putting this last result together with our previous analysis, we can express the Green’s functions 5.3 Raman Storage 180 entirely in terms of a single set of modefunctions {ϕj }, K(x, y) = X ϕj (x)λj ϕj (1 − y), j L(x, y) = X ϕj (x)µj ϕj (1 − y), (5.135) j where the singular values add in quadrature to unity, µ2j + λ2j = 1. (5.136) The modes ϕj can be found from a numerical SVD of the kernel K. Or, equivalently, they are given by the eigenvalue equation (5.96) introduced previously in §5.3. The procedure used to connect the SVDs of K and L through the unitarity of U is known as the Bloch-Messiah Reduction [147–149] . The resulting decomposition makes the assertion that Raman storage may be understood by analogy with a beamsplitter into a rigorous correspondence. If we define aj , (bj ) as the annihilation operator for a photon (spin wave excitation) in the j th input mode ϕj , and if Aj (Bj ) annihilates a photon (spin wave excitation) in the j th output mode, then for each mode, the memory interaction can be written as Aj = µj aj − λj bj , Bj = µj bj + λj aj . (5.137) These relations are precisely those arising in the quantum mechanical description of 5.3 Raman Storage 181 an optical beamsplitter, coupling input modes aj , bj , to output modes Aj , Bj , with reflectivity R = ηj = λ2j (see Figure 5.6). Optimal storage corresponds to the case R = 1, so that the incident signal field is entirely ‘reflected’ into a spin wave mode. Figure 5.6 Raman storage as a beamsplitter. Optical and spin wave modes are mixed pairwise by the storage interaction, as light beams are on a beamsplitter. The ideal quantum memory, with unit storage efficiency, would see the beamsplitter replaced by a perfect mirror. A possible use of a Raman quantum memory with non-unit storage efficiency is as part of the modified DLCZ quantum repeater protocol described in §1.7 in Chapter 1. Instead of using a 50 : 50 beamsplitter in combination with an ideal quantum memory to generate number state entanglement, a single Raman quantum memory with η1 = 50% can be used, as shown in Figure 5.7. The quality of the entanglement generated relies on good overlap of the transmitted field modes on the final beamsplitter, so the preceding theoretical characterization of the temporal structure of these modes simplifies the analysis of this type of protocol. 5.4 Numerical Solution 182 D1 QML BS SL SR D2 QMR Figure 5.7 Modified DLCZ protocol with partial storage. When a single detector fires behind the final beamsplitter, number state entanglement is generated between the memories. The protocol is explained in §1.7 in Chapter 1. The only difference is that here, a single memory with 50% efficiency replaces the combination of a beamsplitter and an ideal memory. 5.4 Numerical Solution So far we have succeeded in deriving a form for the storage kernel in the adiabatic limit, or for arbitrary bandwidths in the case of a hyperbolic secant control. The general problem of finding the storage kernel for both arbitrary bandwidths and arbitrary control pulse profiles has not been solved analytically. But it is possible to construct the kernel numerically. This can be done by integrating the system of coupled equations (5.106) multiple times, each time with a different boundary condition. Provided the boundary conditions form a set of orthogonal functions, the Green’s function can be reconstructed. The easiest set of boundary conditions to implement is the set of ‘impulses’ — delta functions. To see why this works, recall 5.4 Numerical Solution 183 the definition of the storage kernel, Z ∞ Bout (z) = K(z, τ )Ain (τ ) dτ, (5.138) −∞ where we used normalized units for z. Now, if we insert a delta function δ(τ − τj ) as the signal field boundary condition Ain (τ ), where τj is some particular time slot, the resulting spin wave Bout,j is Bout,j (z) = K(z, τj ). (5.139) We can therefore reconstruct an approximation to the entire Green’s function by numerically solving for Bout,j repeatedly, with the times τj chosen from a finite grid. The grid should range over a sufficient range of times that all of the Green’s function is sampled, and we should make the grid sufficiently fine that no features of the Green’s function are missed. So long as these requirements can be met while keeping the computation reasonably fast, this is a convenient way to find the storage kernel without requiring the adiabatic approximation, and without imposing any restriction on the temporal profile of the control field. This method was previously used to reconstruct the adiabatic Green’s functions describing stimulated Stokes scattering in a dispersive ensemble by Wasilewski and Raymer [148] . In this thesis we solve the equations of motion numerically using Chebyshev spectral collocation for the spatial derivatives, and a second order Runge-Kutta (RK2) method for the time stepping. The method of solution is explained in detail in Appendix E. 5.4 Numerical Solution 184 In the figures below we compare the optimal input modes predicted by the various methods presented in this chapter. In Figures 5.8 and 5.9 we plot, side by side, the optimal input modes found from numerically constructed Green’s functions, the Rosen-Zener kernel (5.60), the adiabatic kernel (5.77) and the Raman kernel (5.95). In Figure 5.8 the adiabatic approximation is well satisfied; the Raman approximation only poorly so, and therefore there is good agreement between all predictions, save for the Raman prediction, which deviates slightly. But note that the adiabatic and Raman kernels do slightly overestimate the phase due to the dynamic Stark- 1 Rosen Zener Numerical Adiabatic Raman 6 4 0.5 2 0 −5 0 5 −5 0 −5 5 Time 0 5 −5 0 Figure 5.8 Comparison of predictions for the optimal input modes in the adiabatic limit. Here we used a hyperbolic secant control, given by (5.38) with Tc = 1 and Ω0 = Ωmax = 3, an optical depth of d = 300 and a detuning ∆ = 15, in normalized units. The control intensity profile |Ω(τ )|2 is indicated by the dotted lines, scaled for clarity. The blue lines show the predicted optimal intensity profiles |Ain (τ )|2 = |φ1 (τ )|2 , and the red lines show the variation of the temporal phase of the mode φ1 in radians, referred to the axes on the right hand side. The storage efficiency is ∼ 90% in all cases, and the predictions are in good agreement generally. But both the adiabatic and Raman kernels overestimate the phase shift due to the dynamic Stark shift. And the detuning is not quite large enough to render the Raman kernel correct. In Figure 5.9 the fields are tuned into resonance, the optical depth is reduced, 5 0 Phase Intensity (arb. units) shift. 5.4 Numerical Solution 185 and the control intensity is increased, so that the adiabatic approximation is no longer satisfied. There is rough agreement between the numerical and Rosen-Zener predictions, which do not rely on the adiabatic approximation, although there is a slight discrepancy that we attribute to the accumulation of numerical errors in the evaluation of the Hypergeometric functions in the Rosen-Zener kernel (5.60). The optimal modes predicted by these two methods exhibit oscillations, as can be seen from the phase jumps indicating sign changes. These oscillations are missing from the predictions of the adiabatic kernel; the prediction of the Raman kernel is 1 Rosen Zener Numerical Adiabatic Raman 6 4 0.5 2 0 −5 0 5 −5 0 5 −5 Time 0 5 −5 0 Figure 5.9 Comparison of predictions for the optimal input modes outside the adiabatic limit. Again, the control takes the form of a hyperbolic secant, this time with Tc = 1 and Ω0 = Ωmax = 5. We reduce the optical depth down to d = 10, and we tune into resonance, putting ∆ = 0. The numerical and Rosen-Zener predictions roughly coincide, and both the predictions exhibit oscillations characteristic of non-adiabatic ‘ringing’. Numerical errors in the Rosen-Zener kernel amplify the size of these oscillations slightly. The adiabatic kernel does not correctly reproduce the oscillations, while the Raman kernel is totally inappropriate for modelling this resonant interaction. The optimized storage efficiency predicted by the numerics is around 82%. The numerically constructed Green’s functions yield the most reliable predictions. In addition, they take less time to calculate then the Rosen-Zener predictions 5 0 Phase Intensity (arb. units) catastrophically wrong, as we would expect on resonance. 5.4 Numerical Solution 186 — the Matlab code for the numerical method runs in ∼ 40 s on a 3 GHz machine; the Rosen-Zener kernel takes several hours to construct, simply because the hypergeometric functions are so difficult to evaluate efficiently. Also, the numerical method is more flexible, since arbitrary control profiles can be used. In general, then, the direct numerical approach is the method of choice for optimizing a Λ-type quantum memory. In the adiabatic and Raman limits, however, it is faster to use the analytic kernels derived using these approximations. And of course, the optimization can then be trivially generalized to arbitrary control pulse shapes using the transformations (5.78) and (5.97). In Figures 5.10 and 5.11 below we compare the numerical predictions with those of the adiabatic and Raman kernels in the resonant and Raman limits, using much more broadband control pulses. As expected, there is good agreement between all three predictions in the Raman limit. In the resonant case, describing EIT storage, the Raman kernel of course fails completely, but the adiabatic kernel remains reasonably reliable. That said, its agreement with the numerical prediction is not as good as in the Raman limit, which is symptomatic of the fragility of the adiabatic approximation on resonance. When far into the Raman limit, the numerical, adiabatic and Raman methods all predict a significant temporal phase variation due to the dynamic Stark shift. Whereas in the EIT case, the optimal input mode has a flat phase. Depending on the flexibility of the technology used to shape the optical pulses before storage, it may be easier to implement EIT storage for this reason. On the other hand, Raman 5.4 Numerical Solution Numerical 1 Raman Adiabatic 6 4 0.5 2 0 −0.2 0 0.2 −0.2 0 Time 0.2 −0.2 0 0.2 Phase Intensity (arb. units) 187 0 Figure 5.10 Broadband Raman storage. Here the control is a −(τ /Tc )2 Gaussian pulse Ω(τ ) = Ωmax ep , where Tc = 0.1 and W = 302.5, so that Ωmax = (2/π)1/4 W/Tc = 49.1. The optical depth is d = 300, and the detuning is ∆ = 150. As usual all these quantities are in normalized units. These parameters give a Raman memory coupling of C = 2.0 and a balanced interaction with R = 1.0. Note that the control pulse duration is roughly one tenth of the spontaneous emission lifetime of the excited state |2i. Nonetheless, due to the large detuning, the adiabatic and Raman approximations are well satisfied, and the agreement between the numerical and analytic predictions is clear. The optimized storage efficiency is predicted by all the methods shown to be ∼ 98%. storage allows for more freedom in the carrier frequencies of the signal and control fields. 5.4.1 Dispersion All of the optimal modes shown in the figures above are slightly delayed in time with respect to the control pulse. This is due to the characteristic dispersion associated with an absorption process, which produces a superluminal group velocity. Of course, there is no question of violating causality, it is simply that as the trailing edge of the signal field is absorbed, the ‘centre of mass’ of the signal pulse advances, giving the appearance of superluminal propagation. The efficiency of the memory 5.5 Summary Numerical Adiabatic Raman 6 4 0.5 2 0 −0.2 0 0.2 −0.2 0 Time 0.2 −0.2 0 0.2 Phase Intensity (arb. units) 1 188 0 Figure 5.11 Broadband EIT storage. All parameters are the same as above in Figure 5.10, except that the storage is performed on resonance, with ∆ = 0. The Raman kernel fails utterly, but the adiabatic prediction fairs better, comparing well with the numerical prediction. That the agreement is poorer on resonance is a signature that the adiabatic approximation is less robust on resonance. The optimized storage efficiency predicted by the numerical method is ∼ 99%. is maximized by ‘pre-compensating’ for this effect, and this explains the time shift common to the optimal input modes. In the case of resonant storage, this view conflicts with our characterization of EIT in §2.3.1 of Chapter 1 as working by slowing the group velocity of the signal. Perhaps a better perspective is therefore that on resonance the control must precede the signal in order to ‘prepare’ the transparency window. 5.5 Summary We have covered rather a lot of material in this chapter. Here we review the main results. 1. The analysis of the storage process is greatly simplified if we use a one di- 5.5 Summary 189 mensional propagation model. All the results listed below make use of this model. 2. The best possible storage efficiency is limited by the optical depth, being given by η = 1 − 2.9/d. 3. The method of Rosen and Zener yields an analytic expression for the storage kernel for a hyperbolic secant control. It is, however, difficult to evaluate this efficiently and accurately (at least using Matlab). 4. In the adiabatic approximation, an analytic expression for the storage kernel can be derived that holds for all control pulse profiles. This is quick to evaluate. 5. When far detuned, the expression further simplifies, yielding the Raman kernel, which only depends on the Raman memory coupling C. The Raman memory can be decomposed as a set of beamsplitter transformations between light and matter, using just a single set of modefunctions. 6. If none of the above methods are appropriate, the storage kernel can be directly constructed by repeated numerical integration of the equations of motion for the memory. This method yields the correct optimal input mode for arbitrary control profiles and detunings, for all values of the optical depth. In the next chapter, we consider retrieval from a Λ-type memory. Chapter 6 Retrieval So far we have considered the optimization of storage in a Λ-type quantum memory. In some circumstances this optimization is sufficient to maximize the combined efficiency of storage into, followed by retrieval from the memory. Specifically, this is true when the retrieval process is the time reverse of the storage process [133] . When this is not the case, the optimization of this combined efficiency is a distinct problem. In this Chapter, we discuss various strategies for retrieval, and we present the results of a numerical analysis of their effectiveness. 6.1 Collinear Retrieval In order to convert the stored excitation back into a propagating optical signal, we send a second control pulse — a read out pulse — into the ensemble. Since the control mediates coupling between the signal field and the spin wave, the rationale is that the spin wave excitations will transfer back to the optical field, emerging 6.1 Collinear Retrieval 191 as a collimated pulse traveling in a well-defined direction. We have met with some success in analyzing the storage process using a one dimensional propagation model, and so it is natural to consider read out from this perspective. We will see later that certain advantages accrue if a small angle is introduced between the signal and control fields. For the moment, suppose that we have stored a signal photon collinearly, as described in the previous chapter. Confining ourselves to the same one dimensional model, there are two possible read out geometries: forward, and backward retrieval. 6.1.1 Forward Retrieval We can describe forward retrieval using the same equations of motion as we used for storage — (5.3), or (5.106) in normalized units. For the retrieval process, the signal field boundary condition is set to zero, and the spin wave boundary condition is set by the coherence generated in the medium during the storage process. Let us denote quantities associated with the retrieval process with a superscript r. In this notation the boundary conditions at read out are Arin (τ r ) = 0, r Bin (z) = Bout (z). (6.1) The second condition assumes that there is no loss of coherence during the time that the excitations are stored. This kind of loss is homogeneous in space, and so it would r . The optimization appear only as a constant factor reducing the amplitude of Bin of the memory is therefore unaffected by the neglect of decoherence. 6.1 Collinear Retrieval 192 The first step in analyzing the efficiency of the retrieval process is to examine the form of the map from the stored excitations to the retrieved signal field, Arout (τ r ) = Z 1 r K r (τ r , z)Bin (z) dz. (6.2) 0 In fact, in the adiabatic limit, the retrieval kernel K r is identical in form to the storage kernel K, when expressed in terms of the integrated Rabi frequency ω r , rather than the time τ r . This is guaranteed by the symmetrical form of the equations of motion in the adiabatic limit, which is discussed in §5.3.3 in Chapter 5. For completeness, we verify this by explicit calculation. In the notation of §5.2.5 in Chapter 5, the retrieved signal field is expressed in k-space as √ d r r r e ein A (k, τ ) = −i √ h↑ |V r−1 (k, τ r )Vinr (k)| ↓iB (k). 2πk (6.3) Using Vinr = I, along with the adiabatic approximations (5.70) and (5.73), we obtain √ Z τr 1 d r r 1 r 0 2 0 A (k, τ ) = − √ Ω (τ ) exp − |Ω (τ )| dτ Γr −∞ 2π Γr " id R τ r # r (τ 0 )|2 dτ 0 |Ω 1 r2 e r (k). × exp Γ −∞ B in k + i Γdr k + i Γdr r r (6.4) Here Ωr is the Rabi frequency describing the temporal profile of the read out control pulse, which may differ from that of the storage pulse, and Γr = γ − i∆r allows for the possibility that the detuning is changed for the readout process. We transform the time coordinate from τ r to ω r , and take the inverse Fourier transform from k 6.1 Collinear Retrieval 193 to z-space. Using the result (D.40), along with the shift and convolution theorems (D.18), (D.26) in Appendix D, we find that the retrieval kernel is given by Ωr∗ (τ r ) K r (τ r , z) = √ × K r [ω r (τ r ), z] , r W (6.5) where √ i dW r −[W r ωr +d(1−z)]/Γr h √dW r p r K (ω , z) = i e J0 2i Γr ω (1 − z) . Γr r r (6.6) This kernel has precisely the same form as (5.76), except that ω has been switched for z, and z has been switched for ω r . Now we can write down the adiabatic map describing the combined processes of storage followed by forward retrieval, Arout (τ r ) = Z ∞ Ktotal (τ r , τ )Ain (τ ) dτ. (6.7) −∞ The kernel Ktotal describes the entire memory interaction (see §3.4 in Chapter 3), and is given by Ktotal (τ r , τ ) = Ωr∗ (τ r )Ω(τ ) √ × Ktotal [ω r (τ r ), ω(τ )] , r W W (6.8) where r Z Ktotal (ω , ω) = 1 K r (ω r , z)K(z, ω)dz. (6.9) 0 Here the kernel K is the adiabatic storage kernel given in (5.76) in Chapter 5, and K r , given in (6.6), is related to K by the symmetry connecting the storage and 6.1 Collinear Retrieval 194 retrieval kernels just derived above. The input mode that optimizes the total memory efficiency is found from the SVD of Ktotal in (6.9). The kernel looks superficially similar in form to the normally ordered kernel KN formed from the storage kernel K. But Ktotal 6= KN because K r 6= K ∗ . If these two kernels were equal to one another, then the optimal input mode found from the SVD of Ktotal would be equal to the mode found from the SVD of K; this follows from the properties of KN (see §3.3.1 in Chapter 3). Since these two kernels are not equal to each other, the optimization of storage followed by forward retrieval is different to the optimization of storage alone. Note that the optimal input mode does not depend on the shape Ωr of the read out control; it only depends on the total energy in the read out pulse, parameterized by W r . But changing the shape of Ωr changes the temporal profile of the signal field retrieved from the memory. This is useful: one can optimize the memory efficiency once, for a fixed energy W r , and then changing the shape of Ωr allows one to produce an output signal pulse of any shape, within the limits of the adiabatic approximation. Novikova et al. have already demonstrated optimal storage, followed by shaped retrieval, experimentally, on resonance [74,150,151] , using the theory of Gorshkov et al. [133] , which applies in the light-biased limit (see the end of §5.2.8 in Chapter 5). In Figure 6.1, we show an example of how the optimal input mode for storage, followed by retrieval, differs from the optimal storage mode. We used the adiabatic kernel (5.76) to model the storage interaction, and we evaluated the expression (6.9) for Ktotal numerically. This is easily done by discretizing the coordinates so that K 6.1 Collinear Retrieval 195 becomes a matrix. We use the same control pulse and detuning for both storage and readout, Ω = Ωr and ∆ = ∆r , and then Ktotal is given by the product of two copies the matrix K. Taking the SVD of the result yields a radically different optimal input mode to that found from the SVD of K alone. The difference can be understood by considering the shape, in space, of the spin wave Bout (z) produced by the storage interaction. When optimizing storage alone, this is given by the optimal output mode ψ1 (z) found from the SVD of the storage kernel K. The shape of this mode generally takes the form of a decaying exponential, as shown in part (b) of Figure 6.1. This shape is consistent with Beer’s law absorption: as the signal pulse propagates through the ensemble it is increasingly likely to be absorbed, so that at the exit face, at z = 1, there is very little probability for the signal pulse to excite an atom. Therefore the spin wave decays in magnitude steadily from the input to the exit face. But forward retrieval from a spin wave of this shape is problematic: if the amplitude of the spin wave is concentrated close to the entrance face of the ensemble, a retrieved excitation, that has been converted into a signal photon, must propagate a large distance through the ensemble before reaching the exit face. There is therefore a high probability that the retrieved photon will be re-absorbed, so that it never emerges from the ensemble, and this greatly reduces the efficiency of the read out process. For forward retrieval, it is much better that the bulk of the spin wave is concentrated towards the exit face of the ensemble, so that re-absorption losses are minimized. In fact, the optimal spin wave mode for retrieval is precisely the space-reverse of the mode generated by the storage process; this can be derived 6.1 Collinear Retrieval 196 from the symmetry relating the storage and retrieval kernels. It is the fact that the optimal shape of the spin wave for retrieval is improperly matched to the spin wave mode generated by optimal storage that complicates the combined optimization of storage followed by retrieval. A compromise between these two shapes must be found. Such a compromise is shown in part (d) of Figure 6.1: the amplitude of the spin wave at z = 1 is much larger than for the spin wave that results from only optimizing storage. And there is a marked suppression of the spin wave amplitude at z = 0, since even though the storage interaction naturally excites atoms close to the entrance face, it is extremely deleterious to the retrieval efficiency to concentrate the stored excitations there. If we could run the storage process backwards, so that the retrieval process was precisely the time-reverse of the storage process, we should be able to achieve a retrieval efficiency equal to the storage efficiency, ηretrieve = ηstorage . It should then be possible to achieve a combined efficiency of storage followed be 2 . However, even after optimizing retrieval of ηcombined = ηstorage × ηretrieval = ηstorage 2 the combined efficiency, it still falls far short of this optimum, ηcombined ηstorage . 2 In the example shown in Figure 6.1, ηcombined = 28%, whereas ηstorage = 80%. The tension between Beer’s law during storage and re-absorption during retrieval makes forward retrieval generally inefficient. In the next section we consider backward retrieval, which performs better. 197 Input mode 1 Spin wave 6 (a) (b) 4 0.5 2 0 6 (c) (d) 4 0.5 2 0 −2 0 2 0 0.5 1 Phase Intensity (arb. units) 0 1 Phase Intensity (arb. units) 6.2 Backward Retrieval 0 Figure 6.1 Forward retrieval. We consider collinear storage, followed by forward retrieval, using a Gaussian control √ pulse Ω(τ ) = 2 Ωmax e−τ , with W = 9, so that Ωmax = (2/π)1/4 W = 2.68. The optical depth is d = 300, and the detuning is ∆ = 15, in normalized units. (a) shows the input mode that optimizes just the storage efficiency, and (c) shows the input mode that optimizes the combined efficiency of storage followed by retrieval. The shape of the control intensity profile is indicated by the black dotted lines. The former optimization gives a storage efficiency of ∼ 90%, whereas the latter optimization gives a combined efficiency of only 28%. In parts (b) and (d) we show the ‘intensity’ |Bout (z)|2 and spatial phase of the spin wave generated in the ensemble after the storage process is complete, for the storage and combined optimizations, respectively. The combined optimization produces a spin wave that reduces re-absorption losses by shifting the ‘centre of mass’ of the spin wave away from the entrance face at z = 0. 6.2 Backward Retrieval The arguments given in the previous section suggest that retrieving the signal field backwards should be much more efficient than forward retrieval. If the storage interaction produces a spin wave with its amplitude concentrated toward the entrance face of the ensemble, re-absorption losses are minimized if retrieved signal photons propagate backwards to re-emerge from the entrance face. This is indeed the case, but issues of momentum conservation arise if the energy splitting between the ground 6.2 Backward Retrieval 198 and storage states |1i and |3i is non-zero. To see why, consider the diagram shown in part (a) of Figure 6.2. Momentum conservation requires that the wavevectors ks , kc and κ associated with the signal, control and spin wave, respectively, sum to zero. When this is the case, the storage interaction is said to be phasematched, since the spatial phases accumulated by the optical fields as they propagate through the ensemble are ‘matched’ to the spatial phase of the spin wave. This means that the slowly varying envelopes of these fields are strongly coupled to the spin wave, as described by the equations of motion for the quantum memory (see (5.106) in Chapter 5, for instance). As we will see below, when momentum is not conserved, the interaction is not phasematched, and destructive interference greatly reduces the strength of the coupling, and with it the memory efficiency. Fortunately, the storage process is always phasematched, because there is no fixed dynamical relationship between the spatial phase of the spin wave and its energy. This is because the atoms comprising the spin wave do not interact with another, so that there is no coupling at all between the spatial shape of the spin wave and the frequency splitting between the ground and storage states. In the storage process, a signal photon is absorbed, and a control photon is emitted. The spin wave therefore acquires a wavevector that ‘takes up the slack’, given by the difference of the signal and control wavevectors, κ = ks − kc = ωs − ωc z, c (6.10) 6.2 Backward Retrieval 199 where z is a unit vector pointing along the positive z axis. In the last equality we used the optical dispersion relation for a plane wave, k = ω/c, (6.11) where k is the magnitude of the wavevector and ω is the angular frequency of the wave (not the integrated Rabi frequency!). This relation fixes the momentum associated with the spin wave as given by the difference between the signal and control field frequencies, which is in turn fixed by the energy splitting that separates the ground and storage states, in order to satisfy two-photon resonance. Even though there is no intrinsic connection between the spatial phase of the spin wave and its energy, the kinematics of the scattering process that generates the spin wave does, in fact, tie these two quantities together. The efficiency of the retrieval process depends critically on whether it is possible to phasematch the retrieval interaction. When we attempt to retrieve the spin wave excitations by sending the readout control pulse in the backward direction, momentum conservation requires that the wavevector of the retrieved signal field is formed from the sum of the spin wave and the read out control wavevectors, ksr = κ + kcr . (6.12) Substituting (6.10) into (6.12), we obtain ksr − kcr = ks − kc . (6.13) 6.2 Backward Retrieval 200 For collinear storage, we have ks = ks z and kc = kc z. For backward retrieval, we have ksr = −ksr z and kcr = −kcr z. Inserting these expressions into (6.13), and using the optical dispersion relation (6.11) along with the two-photon resonance condition, we get to the condition ωc − ωs ωs − ωc = . c c (6.14) As is clear from part (b) of Figure 6.2, this condition can only be satisfied when ωs = ωc , and κ = 0, which requires that the ground and storage states are degenerate in energy. Experimentally, it is important to be able to distinguish the weak signal field from the bright control pulse, and the ability to spectrally filter these two fields should not be surrendered lightly. Therefore it is very useful to consider storage in memories where ωs 6= ωc , and then the issue of momentum conservation becomes important when considering backward retrieval. (a) (b) Figure 6.2 Phasematching considerations for backward retrieval. (a) The momenta of the control field and spin wave must sum to that of the signal field, from which they are ‘scattered’. The storage process is automatically phasematched, since the magnitude of κ is initially a free parameter, that is determined by ks and kc during storage. (b) The spin wave momentum is pointing in the ‘wrong direction’, when backward retrieval is attempted: it is not possible to simultaneously satisfy the two-photon resonance condition ksr − kcr = (ωs − ωc )/c, and the phasematching condition ksr = κ + kcr . In order to optimize the combined efficiency of storage, followed by backward 6.2 Backward Retrieval 201 retrieval, we need to find an expression for Ktotal that describes the entire memory interaction in this case. The SVD of this kernel will then provide us with the optimal input mode. The equations of motion for the retrieval process have the same form as those describing storage, but they must describe propagation in the backward direction, with the z coordinate is reversed. The kernel Ktotal is simply constructed using the adiabatic solutions for storage and retrieval separately, provided we are careful about how the boundary conditions are ‘stitched together’. Let us denote the flipped z coordinate for the retrieval process by z r , so that z r = L − z, in ordinary units, where z is the coordinate describing propagation during storage. The atomic coherence at the start of the retrieval process is identical — as always assuming no decoherence — to that generated by the storage process. We can write this as r σ13,in (z r ) = σ13,out (z), (6.15) where the σ13 (x) denotes the Raman coherence associated with an atom located at the longitudinal position x. The spin wave B, defined in (4.43), makes use of the slowly varying operators σ e13 , introduced in (4.22) (see Chapter 4), which incorporate an exponential factor eiω13 τ . Recall that τ is in fact the retarded time, τ = t − z/c, so there is a spatial phase built into the spin wave; this represents the momentum imparted to the spin wave by the optical fields that create it during the storage 6.2 Backward Retrieval 202 process. Equating the boundary conditions, as in (6.15), gives the relation r r Bin (z r ) = Bout (z) × e−iω13 τout × eiω13 τin , r = Bout (z) × e−iω13 (tin −tout ) × e−iω13 (z−z = Bout (L − z r )e2iω13 z r /c r )/c . , (6.16) Here we used the definition τ r = tr − z r /c = tr − L/c + z/c, and in the last line we dropped some unimportant constant phase factors. The spatial phase factor in (6.16) represents the phase mismatch shown in Figure 6.2 (b). Note that it vanishes if ω13 = 0. The phase mismatch causes oscillations of the spin wave in space that can dramatically reduce the retrieval efficiency. This can be seen by considering the form of the retrieval map Arout (τ r ) Z 1 = Z0 1 = r K r (τ, z r )Bin (z r ) dz r r K r (τ r , z r )eiδkz Bout (1 − z r ) dz r . (6.17) 0 Here we have switched to normalized units, and defined the dimensionless phase mismatch δk = 2ω13 L . c (6.18) Note that δk is a negative quantity if the storage state |3i is more energetic than the ground state |1i. It is clear that, regardless of the form of the retrieval kernel K r , or of the spin wave Bout , the integral can be made to vanish if δk is made large enough. 6.2 Backward Retrieval 203 This explains why a phase mismatch can be very detrimental to the efficiency of backward retrieval. More precisely, if kmax represents the width in k-space of the integrand K r × Bout , the value of the integral becomes strongly suppressed when δk kmax . It is therefore possible to mitigate the effect of the phase mismatch, to some extent, by localizing the spin wave strongly in space, so that kmax becomes large. Below, we show explicitly how to optimize storage, followed by backward retrieval in the presence of a phase mismatch; the result essentially generates a more strongly localized spin wave, for exactly this reason. To perform the optimization, we construct the kernel Ktotal using the solution (6.4) for the retrieval process, making the replacement z −→ z r . The result is Arout (τ r ) Z ∞ Ktotal (τ r , τ )Ain (τ ) dτ, (6.19) Ωr∗ (τ r )Ω(τ ) √ × Ktotal [ω r (τ r ), ω(τ )] , W rW (6.20) = −∞ with Ktotal (τ r , τ ) = and where r Z Ktotal (ω , ω) = 1 r K r (ω r , z r )K(1 − z r , ω)eiδkz dz r . (6.21) 0 Suppose that the readout control pulse is identical in shape and frequency to the storage control pulse, with Ωr = Ω and ∆r = ∆. Suppose also that the storage state is degenerate with the ground state, so δk = 0. Then consider the kernel Ktotal in 6.2 Backward Retrieval 204 (6.21) with its output time argument flipped around, Z r 1 K r (1 − ω r , z r )K(1 − z r , ω) dz r Ktotal (1 − ω , ω) = 0 Z = 1 K(1 − z r , ω r )K(1 − z r , ω) dz r , (6.22) 0 The second line follows from the symmetry K r (1 − x, y) = K(1 − y, x) that connects the retrieval kernel (6.6) to the storage kernel (5.76). This shows that Ktotal is, in the case under consideration, very similar in structure to the normally ordered kernel KN that would be formed from K(1 − z r , ω). The only difference is that the first instance of K in (6.22) is replaced by its complex conjugate in KN . If K is real, then this complex conjugation is not important, and Ktotal = KN , which means that the optimal input modes derived from the SVD of Ktotal are equal to the optimal input modes derived from the SVD of K alone. This observation shows that the optimal input mode for storage alone is close to the optimal input mode for storage followed by backward retrieval, if identical storage and readout control pulses are used, and if there is no phase mismatch. The two optimizations are the same whenever K is real, and more generally whenever ψ1 (z), the output spin wave mode generated by optimal storage, is real. In Figure 6.3 below we show an example of how the optimal input mode for storage alone can differ from the optimal mode for storage followed by backward retrieval. Without a phase mismatch (parts (c) and (d)), the difference between the two optimizations is rather small, though non-negligible, for the example shown. 6.2 Backward Retrieval 205 When a significant phase mismatch is introduced (parts (e) and (f)) the optimizations differ more markedly. It is a general feature of the optimization of storage with backward retrieval, that as |δk| is increased, the shape of the optimal input mode approaches that of the control pulse. The reason is as follows. Recall that the dispersion experienced by the signal field as it propagates causes its group velocity to differ from that of the control (see the end of §5.4 in Chapter 5). If the signal initially overlaps with the control, then the coupling between these two fields is initially high, but as the signal walks off from the control, the coupling decays away. The bulk of the storage interaction therefore occurs near the entrance face of the ensemble at z = 0. As is clear from part (f) of Figure 6.3, the spin wave generated by the optimal input mode is indeed concentrated more closely towards the entrance face. This makes the spin wave more strongly localized, which reduces the ‘wash-out’ caused by the phase mismatch, as discussed above, while at the same time reducing re-absorption losses. A discussion of similar optimizations can be found in the work of Gorshkov et al. [133] . In this work time-reversal arguments are employed to show that for 2 optimized storage with backward retrieval, ηcombined = ηstorage if ψ1 (z) is real. This condition is always satisfied in the light-biased limit (see the end of §5.2.8 in Chapter 5), when the anti-normally ordered kernel KA becomes real, independent of whether or not the adiabatic approximation is satisfied (see (5.27 in Chapter 5)). As shown above in the adiabatic limit, when ψ1 is not real, implementing ‘true’ time reversal requires that the phase of the spin wave is conjugated. Without a practical method 1 206 Input mode Spin wave 10 (a) 6 (b) 4 5 0 1 0 0 6 6 (c) −2 0 2 Intensity (arb. units) 4 2 2 0 0 8 6 4 2 0 6 (f) 4 2 0 0.5 1 Phase (e) 0.5 0 (d) 4 0.5 0 1 2 Phase Intensity (arb. units) 0.5 Phase Intensity (arb. units) 6.2 Backward Retrieval 0 Figure 6.3 Backward Retrieval. We consider collinear storage followed by backward retrieval, using a Gaussian control pulse Ω(τ ) = /Tc )2 Ωmax e−(τp , with Tc = 0.1 and W = 100, so that Ωmax = (2/π)1/4 W/Tc = 28.25. The optical depth is d = 300, and the detuning is ∆ = 150, in normalized units. Parts (a) and (b) show the optimal input mode for storage alone, alongside the spin wave mode ψ1 (z) generated in the medium. We use the adiabatic form for the storage kernel; comparison with the prediction derived from the numerically constructed kernel verifies that the adiabatic approximation is well satisfied. The optimal efficiency for storage alone is ηstorage ∼ 96%. In parts (c) and (d) we show the optimal input mode for storage followed by backward retrieval, and the generated spin wave, with degenerate ground and storage states, δk = 0. The opti2 mized total efficiency is ∼ 88%, which is about 5% less than ηstorage . The most notable difference between these two optimizations is the appearance of a ‘wiggle’ in the phase of the input mode for the combined optimization. But this occurs while the signal intensity is rather low, so it is actually less important than the more subtle re-shaping that occurs. In parts (e) and (f), we show the optimal input mode for storage and backward retrieval, along with the generated spin wave, with a phase mismatch δk = −5. Note that this optimization results in a spin wave with its ‘centre of mass’ concentrated more closely towards the entrance face of the ensemble at z = 0 than the other optimizations. This reduces the effective length of the ensemble — the length over which there is significant excitation — which diminishes the effect of the phase mismatch. The optimized combined efficiency of storage and retrieval is around 75% in this case. 6.3 Phasematched Retrieval 207 2 for doing this, ηcombined < ηstorage , and the optimal mode for storage alone differs from the mode that optimizes the combination of storage and backward retrieval. In this and the previous section we have shown that retrieving the stored excitations from a quantum memory efficiently is a non-trivial problem. Forward retrieval is plagued by re-absorption losses — a modematching issue. Backward retrieval is beset by momentum conservation problems — a phasematching issue. In the next section we present a solution to both of these problems, which requires a departure from the one dimensional treatment we have worked with thus far. 6.3 Phasematched Retrieval The best memory efficiency possible is achieved by optimizing collinear storage, followed by backward retrieval, with δk = 0. But to spectrally filter the signal field from the control, we should have δk 6= 0. By introducing a small angle between the signal and control beams, both at storage and at read out, we can maintain proper phasematching, even when δk 1, which allows for efficient retrieval. This idea was proposed by Karl Surmacz, and the following treatment is adapted from our paper [152] ; the numerical simulations were performed by me. The principle of the scheme is easily understood by considering the phasematching diagrams in Figure 6.4. Here we assume for simplicity that all the beams used, both at storage and at read out, are confined to a common plane, so that we need only consider two space dimensions. The spin wave momentum is determined by the signal and control field wavevectors according to (6.10). If we fix the detuning, so that ∆r = ∆, then 6.3 Phasematched Retrieval 208 the magnitudes of ks and kc are unchanged at read out, and then the condition (6.12) uniquely defines the direction of ksr that is phasematched. By symmetry, the angle θ between the signal and control beams is the same at read in as at read out. Provided we adhere to the geometry shown in Figure 6.4 (a), the retrieval process remains correctly phasematched for all choices of θ. To maximize the efficiency of the memory, we should reduce θ as far as is possible, so that 1. we approach a collinear geometry, which maximizes the overlap of the signal and control pulses, and 2. the spin wave generated by the storage process overlaps well with the optimal spin wave mode for retrieval. An heuristic choice of θ that satisfies these desiderata is that shown in part (b) of Figure 6.4, with ks cos θ = kc . The signal and control wavevectors are close to parallel, but they are arranged so that the spin wave momentum κ is orthogonal to kc . This way, when the control field direction is reversed for backward retrieval, no phase mismatch is introduced and the signal field emerges at an angle θ with respect to the read out control pulse, as shown in part (a) of Figure 6.5. This choice for θ assumes that the storage state is more energetic that the ground state, so that δk is negative, and ks > kc . We also consider the possibility that the storage state lies energetically below the ground state. In many systems the ground state is prepared artificially by optical pumping (see §10.12 in Chapter 10), and it is then quite feasible to select the ground state so that ω1 > ω3 . In this case, δk > 0 and 6.3 Phasematched Retrieval 209 ks < kc . We then choose θ so that kc cos θ = ks ; the geometry is shown in part (b) of Figure 6.5. (b) (a) Figure 6.4 Non-collinear phasematching. (a) the general geometry required to phasematch storage and retrieval, when the signal beam makes an angle θ with the control. (b) a geometry that closely approaches the collinear one, while preserving correct phasematching. In the next section we include dispersion, which modifies the length of ks and ksr . Storage Retrieval (a) (b) Figure 6.5 Efficient, phasematched memory for positive and negative phase mismatches. We show the beam directions of the control (green) and signal (blue) fields at storage and retrieval, with (a) δk < 0, and (b) δk > 0. Off-resonance, the angles used depend on both δk and on the material dispersion (see §6.3.1 below). 6.3 Phasematched Retrieval 6.3.1 210 Dispersion So far we have used the dispersion relation (6.11), which applies to light propagation in vacuo. The phase mismatch δk quantifies the spatial phase imparted to the spin wave with collinear storage calculated using only this vacuum dispersion. But even in the case of degenerate ground and storage states, the spin wave is not always entirely real, as discussed in the previous section (see part (b) of Figure 6.3). This spatial phase arises dynamically, and has its origin in the material dispersion experienced by the signal field as it propagates through the ensemble. The control field is unaffected, since it couples states |2i and |3i, both of which are unpopulated. But the signal field couples the populated ground state |1i to state |2i, and therefore it propagates subject to an augmented refractive index. This effect can be described by incorporating a dispersive term into our definition for the magnitude of the signal field wavevector, ks −→ kd = ks − kdisp , (6.23) where, working in ordinary units, we define kdisp = dγ∆ |κ|2 ∆ = 2 . |Γ|2 (γ + ∆2 )L (6.24) This definition is motivated by inspection of the adiabatic storage kernel (5.77) derived in Chapter 5, which includes an exponential factor with an exponent whose imaginary part contains the spatial phase kdisp z. The form of kdisp is consistent with the refractive index associated with an absorptive resonance at ∆ = 0. Note that on 6.3 Phasematched Retrieval 211 resonance, kdisp vanishes, and the signal field wavevector assumes its vacuum value. At large detunings, kdisp ∝ 1/∆, so it is generally small in the Raman limit, but not negligible. Incorporating the dispersive phase into the signal wavevector allows us to use the phasematching scheme outlined in the previous section to compensate both for non-degeneracy of the ground and storage states and for material dispersion in the ensemble. By eliminating even this latter dynamical phase, we essentially render the spin wave real, so that the efficiency of backward retrieval (albeit at a small angle to the z-axis) is equal to the storage efficiency, which is optimal. And with all spatial phases removed in this way, the optimization of storage alone is the same as the optimization of storage with backward retrieval, and so it is only necessary to perform the former, simpler optimization. 6.3.2 Scheme Including material dispersion, then, the phasematching scheme is summarized by the following choice for θ, −1 θ = cos r; r = max kd kc , kc kd , (6.25) where kc = ωc /c, and kd is given by (6.23). This formula holds for arbitrary values of δk, and for all detunings. Provided that θ is sufficiently small, we expect that many of the results pertaining to the optimization of collinear storage remain approximately valid. Our general 6.3 Phasematched Retrieval 212 strategy is to use the one dimensional theory of Chapter 5 — in particular the adiabatic storage kernel (5.77) — to find the optimal temporal input mode, and then implement the above phasematching scheme by introducing the small angle θ between the signal and control fields. Note that this scheme is experimentally appealing for three reasons. First, it allows for storage with non-degenerate ground and storage states, so the signal and control fields can be spectrally filtered. Second, introducing an angle between the signal and control beams makes it possible to spatially filter the signal from the control, with a pinhole or an optical fibre tip, for example. Third, the effect of unwanted Stokes scattering can be eliminated. This is because any spin wave excitations produced by Stokes scattering — the process shown in part (b) of Figure 4.3 in Chapter 4 — will not have the same momentum as those produced by absorption of the signal field. At read out, we choose the direction of the control pulse so that retrieval of the stored signal field is correctly phasematched. This same control pulse could potentially drive anti-Stokes scattering, converting the unwanted excitations into noise photons with the same frequency as the retrieved signal. But this process will generally not be correctly phasematched, and is therefore greatly suppressed. All these features of the scheme constitute a compelling case for the utility of a non-collinear phasematched memory protocol. Here we should remark that this type of phasematching scheme produces a spin wave with a larger momentum than arises from collinear storage, since |κ| ≥ |δk|. If the atoms in the ensemble are free to move — for instance if the storage medium is 6.4 Full Propagation Model 213 a warm atomic vapour — then any large spatial frequencies in the spin wave may be washed out as atoms diffuse over the phase fronts of the spin wave. Therefore our phasematching scheme may be more susceptible to decoherence, if D > λκ = 2π/|κ|, where D is the distance over which the atoms diffuse during the storage time (see §10.5 in Chapter 10). This problem does not affect solid state memories, which are appealing for this and many other reasons. To investigate the efficacy of our phasematching scheme, we should consider the effect of walk-off between the signal and control fields as they propagate, since when θ 6= 0 their paths inevitably diverge. In the next section we present the results of numerical simulations that account both for walk-off and diffraction. These vindicate the scheme (6.25), and they show, as might be expected, that the beams should be loosely focussed, so as to maximize the overlap of the pulses. 6.4 Full Propagation Model We model the propagation of the signal and control fields in three dimensions: two spatial dimensions, and time. The signal field propagates along the z-axis, and the control field is launched at an angle θ to the z-axis, in the (x, z)-plane, intersecting the signal beam in the centre of the ensemble. We consider that both fields are Gaussian beams — defined below — focussed at the centre of the ensemble, as depicted in Figure 6.6. The equations of motion for the interaction are given in (4.52) at the end of Chapter 4. For numerical convenience, we work in normalized units that render all quantities dimensionless, and as close as possible to ∼ 1. We summarize 6.4 Full Propagation Model 214 the normalizations we employ below; they differ slightly from the normalized units used in the preceeding analytic treatments. 1. The z coordinate is measured in units of L, and the magnitudes ks , kc , are measured in units of 1/L. We define the z coordinate so that it runs from − 12 to 12 , so that the centre of the ensemble lies at z = 0. This definition is useful for parameterizing the control field. 2. The x coordinate is measured in units ws , where ws is the transverse beamwaist of the signal field. That is, at z = 0, the transverse signal field amplitude 2 2 is a Gaussian e−(x/ws ) in ordinary units, and by definition e−x in normalized units. To capture all relevant dynamics, x runs from −3 to 3. 3. The time coordinate τ , which as usual is the retarded time τ = t − z/c, is measured in units of Tc , the duration of the control field. This is a departure from the time normalization used previously, in which times were measured in units of 1/γ. With this new normalization, the control field has roughly unit duration, as does the signal field (since the optimal signal pulse is generally comparable in duration to the control). All relevant dynamics are then captured by considering times from τ = −3 up to τ = 3, or thereabouts. 6.4 Full Propagation Model 215 The equations of motion, written using these normalized units, are given by p iα2 2 ∂x + ∂z A = − dγP, 2ks p ∂τ P = −ΓP + dγA − iΩB, ∂τ B = −iΩ∗ P. (6.26) As in all of our previous analyses, we have dropped the Langevin noise operators associated with spontaneous emission and decoherence, since they have no effect on the efficiency of the memory. The normalization factor α = L/ws is the aspect ratio of the ensemble, which appears to correctly account for the difference in units used for the x and z coordinates. Note that the amplitudes A, P and B are not the averaged quantities defined in (5.1) at the start of Chapter 5. They are closer to those quantities given in (4.7), (4.42) and (4.43) in Chapter 4, which depend on the transverse spatial coordinates ρ, as well as τ and z. In fact, since we have dropped any dependence on the y-coordinate, we have implicitly averaged over this dimension. But this has no effect on the equations. 6.4.1 Diffraction The transverse derivative ∂x2 in (6.26) describes the diffraction of the signal field as it propagates. Its importance in the dynamics is set by the ratio of α2 to ks . Although both of these are generally large quantities, their ratio can be small, or large, depending on how tightly the signal is focussed: if ws is small, diffraction 6.4 Full Propagation Model 216 Figure 6.6 Focussed beams. The signal (blue) and control (green) beams are cylindrically symmetric Gaussian beams focussed at the centre of the ensemble, with a small angle between them. can become significant. In fact, the signal field Rayleigh range, which quantifies the length of the region over which the focussed signal beam is well-collimated, is given in our normalized units by zs = ks /(2α2 ). As expected, the contribution from the diffractive term becomes significant when the Rayleigh range falls below 1, which is the length of the ensemble. 6.4.2 Control Field Walk-off between the signal and control field appears through the spatial dependence of the control field Ω. To construct the correct expression, consider first the representation of a pulse with a Gaussian transverse profile, and also a Gaussian 6.4 Full Propagation Model 217 temporal profile, propagating along the z-axis, Ωmax Ω(x, τ, z) = exp W (z/zc ) ( i 1 − 2 R(z/zc ) W (z/zc ) x wc 2 − i tan−1 z zc ) 2 ×e−τ . (6.27) Here zc = kc wc2 /(2α2 ) is the Rayleigh range of the control field, with wc the beamwaist of the control, measured in units of ws . The structure of (6.27) is well known in optics, since the Gaussian transverse profile arises naturally as the lowest order mode supported by the confocal cavity in a laser. When a laser beam of this form is directed through a lens, it retains its Gaussian profile, but its width narrows as it approaches the focal plane, and then widens out aftwerwards. The functions W and R parameterize the beam size and the radius of curvature of the field phase fronts, and are given by p 1 + z2, 1 R(z) = z 1 + 2 . z W (z) = (6.28) The focus lies at z = 0, at which point R −→ ∞ and W = 1. The term involving tan−1 is known as the Gouy phase; its effect on the memory efficiency is negligible, but we include it for completeness. Direct substitution shows that (6.27) indeed satisfies the paraxial wave equation 2 wc 2 i ∂ + ∂z Ω = 0. 4zc x (6.29) 6.4 Full Propagation Model 218 When the control propagates at an angle θ to the z-axis, the control amplitude is described by a similar expression to (6.27), except that we rotate the (x, z) coordinates by the angle θ. We should be careful to apply this rotation with t held constant, not τ , and we must take account of the different units used for z and x. The correct transformation is found by making the replacements x −→ x0 = cos(θ)x + α sin(θ)z, z −→ z 0 = cos(θ)z − and τ −→ τ 0 = τ + sin(θ) x α z − z0 . c (6.30) The factors of α ensure that the units of x and z are interconverted as they should be, and the modification to τ represents the change in the apparent velocity of the control pulse in the reference frame of the signal field. 6.4.3 Boundary Conditions The boundary conditions must be specified in much the same way as in the previous one dimensional analyses. One additional feature is the dependence on the x coordinate. The boundary conditions associated with this new degree of freedom are simply that all the quantities A, P and B vanish as |x| −→ ∞. In practice, this is achieved by fixing the value of the signal amplitude A to zero at x = ±3 (i.e. at the edges of the region covered by the numerics). We use Chebyshev spectral collocation to treat the spatial derivatives when solving the system (6.26), and we 6.4 Full Propagation Model 219 build this boundary condition directly into the differentiation matrices; the method is explained in Appendix E. Fixing A at these boundaries also fixes P and B, since the interaction is local: there can be no atomic excitation where there is no input light. To model the storage process, we launch the signal field along the z-axis with a Gaussian transverse profile at the entrance face of the ensemble, (" Ain (x, τ ) = A z = − 21 , x, τ = exp # ) i 1 − x2 φ1 (τ ), R − 12 /zs W 2 − 21 /zs (6.31) where φ1 (τ ) is the optimal input mode for storage alone, found using the one dimensional theory of the previous Chapter — we use the adiabatic kernel (5.77), so that φ1 is given by (5.78). As described above, the phasematching scheme should eliminate the spatial phase of the spin wave, so that the optimal mode for collinear storage alone should be close to optimal for phasematched storage and retrieval. Note that the absolute magnitude of the signal field is irrelevant, because the memory interaction is linear. This boundary condition is also incorporated into the differentiation matrices we use to solve the dynamics numerically. For the storage process, both the spin wave B and the polarization P are initially zero, Bin (x, z) = B(τ −→ −∞, x, z) = 0, Pin (x, z) = P (τ −→ −∞, x, z) = 0. (6.32) We use an RK2 method for the time-stepping, and these boundary conditions are trivially implemented by zeroing the vectors representing B and P at the collocation 6.4 Full Propagation Model 220 points on the first iteration. See Appendix E for a description of the RK2 method, and how it is used in combination with spectral collocation to arrive at a solution. 6.4.4 Read out To model the read out process, we solve the equations of motion once more, this time with no incident signal field, and a pre-existing Raman coherence determined by the storage process. We are only able to model the build up of the signal field along a specific direction, the z-axis, and so we must re-orient our coordinate system so that the z-axis points along the direction in which the retrieval process is phasematched. This is the direction of ksr in Figure (6.4). The angle through which we rotate our coordinate system is therefore the angle between ks and ksr in Figure (6.4), which is given by kc sin(θ) π = θ0 = −2 sin−1 q kd2 + kc2 − 2kc kd cos(θ) 2θ − π if kc > kd . (6.33) if kd > kc The first result holds for general values of θ, while the second result applies to the case where θ is chosen according to the scheme (6.25). At read out, the signal field is initially zero, Arin (xr , τ r ) = 0, (6.34) 6.4 Full Propagation Model 221 as is the polarization, r Pin (xr , z r ) = 0. (6.35) Here xr , z r are the rotated coordinates describing the retrieval process, xr = cos(θ0 )x + α sin(θ0 )z, z r = cos(θ0 )z − sin(θ0 ) x. α (6.36) As in (6.30), the factors of α appear to interconvert the units of x and z. The initial spin wave at read out is set by the spin wave generated by the storage process, but we must be careful to include any spatial phase factors arising from the above coordinate transformation. When the signal and control fields are not collinear, the definition of the slowly varying coherence σ e13 (see (4.22) in Chapter 4) must be modified accordingly, σ e13 (r) = σ13 (r)eiω13 t+i(ks −kc ).r . (6.37) Here the argument r indicates that the coherence refers to an atom at that position in the ensemble. The boundary condition for B is found by equating the coherences at the end of storage and the start of read out, r σ13,in (r r ) = σ13,out (r). (6.38) The spin wave amplitudes are built from the slowly varying coherences. Dropping 6.4 Full Propagation Model 222 some unimportant temporal phase factors, we therefore obtain r r r Bin (r r ) = Bout (r)ei[(ks −kc ).r r −(k s −kc ).r] . (6.39) Writing this boundary condition out explicitly in terms of the retrieval coordinate system gives the relation r Bin (xr , z r ) = Bout (xr c0 − s0 αz r , z r c0 + s0 xr /α) × (6.40) exp i (ks − kc c)(1 − c0 ) − kc ss0 z r + i (ks − kc c)s0 − kc s(1 + c0 ) xr /α , where c and s denote cos(θ) and sin(θ), respectively, and c0 , s0 are equal to cos(θ0 ) and sin(θ0 ). Note that in the absence of dispersion, with ks = kd , the phase factor identically vanishes; it is removed by the phasematching scheme. When dispersion is significant, the slowly varying envelope Bout itself contains a spatial phase variation. The phasematching scheme is then designed so that the exponential factor in the second line of (6.40) cancels this spatial phase, rendering B r as smooth as possible, for efficient retrieval. 6.4.5 Efficiency The efficiency of the memory is calculated in two stages. First, we simulate storage and evaluate the storage efficiency, which is given by R ∞ R 1/2 ηstorage = 2 −∞ −1/2 |Bout (x, z)| dz dx R∞ R∞ . 2 −∞ −∞ |Ain (x, τ )| dτ dx (6.41) 6.5 Results 223 We then simulate retrieval, and we calculate the retrieval efficiency, R∞ R∞ −∞ ηretrieval = R−∞ ∞ R 1/2 |Arout (xr , τ r )|2 dτ r dxr r r r 2 r r −∞ −1/2 |Bin (x , z )| dz dx . (6.42) The total memory efficiency is then given by ηcombined = ηstorage × ηretrieval . 6.5 Results In Figures 6.7 and 6.8, we present the results of simulations performed to examine the effectiveness of our phasematching scheme. We simulated both resonant EIT storage, and also off-resonant Raman storage. Each simulation was run twice, once with a tightly focussed control field, with wc = 1, and once with a loosely focussed control, with wc = 2. In the latter case the energy in the control pulse was quadrupled, so that the intensity of the control was the same in both cases. Figure 6.7 shows the angle θ at which the combined efficiency of storage and retrieval is largest, over a range of values of the phase mismatch δk. The numerical solutions generally bear out the analytic prediction (6.25) of our phasematching scheme. But the agreement with our prediction is much better when the control is loosely focussed. This is to be expected, since transverse walk-off is less important when the control is wider than the signal, and since a wider control diffracts less, so that its intensity is more homogeneous over the ensemble. For EIT, collinear storage with θ = 0 is only optimal when δk = 0, since there is no dispersion on resonance. But for the Raman memory, collinear storage is optimal when δk ∼ −2, when the 6.5 Results 224 (a) −3 θ (radians) 8 x 10 EIT 6 4 2 0 −4 8 θ (radians) −3 −2 −1 0 1 2 3 4 −2 −1 0 1 2 3 4 −3 (b) x 10 6 Raman 4 2 0 −4 −3 Figure 6.7 Effectiveness of our phasematching scheme. We plot the angle θ at which the combined efficiency of storage followed by retrieval is optimized. The blue solid line shows the prediction of our phasematching scheme (6.25). The black filled circles correspond to tight control focussing, with wc = 1; the red open diamonds correspond to loose control focussing with wc = 2. We plot the results for a typical EIT protocol in (a): we set γ = 1, d = 30 and ∆ = 0, in units normalized as described in §6.4 above. In (b) we present equivalent results for a Raman protocol d = 300, γ = 0.1 and ∆ = 15. The bandwidth of the control is 10 times larger, with respect to the natural atomic linewidth γ, in this latter protocol. To make up for the increased detuning in the Raman protocol, the optical depth is also much larger. In both protocols, the control field amplitude was given by Ωmax = 5. storage state lies significantly above the ground state in energy. This is because the material dispersion alters the signal field wavevector, kd 6= ks , so that even with degenerate ground and storage states, the spin wave acquires a spatial phase. In Figure 6.8 we plot the optimal memory efficiencies obtained from the numerical simulations alongside the analytic predictions for the best collinear efficiencies, again over a range of values of δk. The efficiencies obtained using the phasematching scheme (6.25) are generally very close to the best efficiencies achieved in the sim- 6.5 Results 225 efficiency (a) 1 0.8 0.8 0.6 0.6 0.4 0.4 0.2 0 −4 (c) efficiency (b) 1 Narrow EIT −2 0 0.2 2 0 −4 4 (d) 1 0.8 0.6 0.6 0.2 0 −4 0.4 Narrow Raman −2 0 0.2 2 4 −2 0 2 4 0 2 4 1 0.8 0.4 Wide EIT analytic collinear backwards collinear forwards phasematched 0 −4 Wide Raman −2 Figure 6.8 Comparing phasematched and collinear efficiencies. Plots (a) and (b) contain the results for the EIT protocol, with tight and loose focussing, respectively. Plots (c) and (d) show equivalent results for the Raman protocol. The parameters used are the same as those used in produciing Figure 6.7. The solid lines represent the efficiencies obtained if the phasematching scheme (6.25) is used. The black filled circles are the best efficiencies achievable. The green dotted lines show the optimal efficiencies predicted for collinear storage with backward retrieval, calculated using the one dimensional adiabatic kernel (6.21). The dashed red lines show the efficiency achieved by collinear storage with forward retrieval, where the input signal pulses were shaped using the adiabatic kernel (6.9), where this improved the efficiency. ulations. These efficiencies are greatest when θ = 0, since at these points walk-off between the signal and control is eliminated. As δk is increased, so does the angle θ required for optimal efficiency, and walk-off therefore reduces the memory efficiency. But the efficiency falls only slowly with increasing δk. With loose focussing it is possible to exceed the optimal collinear efficiency when δk is large enough, using either EIT or Raman storage. This demonstrates what we sought to show with our simulations, that even including the effects of diffraction and walk-off, non-collinear storage and retrieval, with proper phasematching, is preferable over collinear stor- 6.5 Results 226 age, with either forward or backward retrieval. The lower efficiencies observed for tight control focussing confirms that diffraction and walk-off can be detrimental, but provided sufficient laser energy is available, it is possible to reduce their effects by simply widening the control beam waist. The red dashed lines in Figure 6.8 are the result of simulating collinear storage with forward retrieval, using the full numerical propagation model. For the EIT protocol, we used the one dimensional adiabatic kernel (6.9) to predict the optimal input profiles, and these input pulse shapes performed well — better than any other pulse profiles we tried. This at least confirms the utility of the one dimensional theory in this context. However, for the Raman protocol, this optimization did not produce the best pulse profiles. In fact, optimal pulse profiles for backward retrieval worked better, even though the excitation was retrieved in the forward direction. So these input profiles were used to produce the red dashed lines in parts (c) and (d). The reason for the failure of the forward retrieval optimization is that the Raman protocol is more sensitive to the control intensity than is EIT. The control intensity is highest in the centre of the ensemble, where it is focussed; its intensity falls off towards the exit face of the ensemble. But as noted in §6.1.1, optimizing for forward retrieval tends to shift the bulk of the spin wave towards the exit face. When diffraction is introduced in the full model, this optimization is no longer beneficial, since it shifts more of the spin wave to where the control intensity is lower. This explains why the pulses optimized for backward retrieval performed better in the simulations of forward retrieval for the Raman protocol. If the control beam waist 6.5 Results 227 is increased above 2.4, diffraction effects become small enough that the optimization for forward retrieval does indeed produce better results than the backward retrieval optimization, for both the EIT and Raman protocols. Comparison of Figures 6.7 and 6.8 reveals that the efficiency falls off sharply in the region kd > kc , to the left of the point at which θ = 0. That is, the EIT efficiency falls sharply for δk < 0, and the Raman efficiency falls quickly as δk is reduced below around −2. The efficiency remains high even for large angles when kd < kc . This is easily explained by considering the geometry in part (b) of Figure 6.4. When kd < kc , the retrieved signal field propagates exactly backwards with respect to the direction of the input signal beam, since θ0 = π (see (6.33)). Therefore the region over which the retrieved signal experiences gain overlaps precisely with the region in which the spin wave is deposited during the storage process. We could say that the ‘retrieval overlap’ is high, and the efficiency is correspondingly large. On the other hand, when kc < kd , θ0 6= π, and the retrieved signal field propagates at an angle to the direction defined by the input signal field. The retrieval overlap suffers as θ is increased, and this explains the sharp drop in memory efficiency in this regime: although the memory is correctly phasematched, the atomic excitations are not distributed favourably for efficient retrieval. The foregoing discussion suggests that, as a rule, positive phase mismatches are preferable to negative ones, although dispersion complicates things slightly. That is to say, for efficient phasematched retrieval, it is generally better to arrange for the ground state |1i to have a higher energy than the storage state |3i. Although this 6.6 Angular Multiplexing 228 may be rather non-standard, when the ground state manifold is prepared by optical pumping, it is not problematic. We have succeeded in showing that proper phasematching can indeed boost the efficiency of retrieval from a quantum memory. In the next section we describe how the angular selectivity of phasematching offers the possibility of storing multiple modes in a single ensemble. 6.6 Angular Multiplexing There is a sense in which an ensemble memory represents a degree of ‘overkill’ if just a single photon is to be stored. The number of atoms is many times larger than the number of input photons, so there is an enormous redundancy built into the memory. A more efficient use of resources would aim to store multiple input pulses; in this way a single physical memory allows for highly parallel information processing. Several schemes that make use of the idea of multimode memories have already been proposed [17,38,153,154] , in connection with both quantum computing and quantum repeaters. A general difficulty with multimode storage in an ensemble is the possibility that the control field used for storing one signal pulse, may also retrieve another signal pulse that was previously stored. Phasematching provides a way to suppress this kind of unwanted coupling. We consider storing multiple pulses in a single ensemble by changing the angle θ between the signal and control for each input pulse. After a pulse has been stored, momentum conservation picks out a single, unique direction, in which the retrieval 6.6 Angular Multiplexing 229 process is phasematched, as shown in part (a) of Figure 6.4. As long as the control field used to store a subsequent pulse propagates in a different direction to this, the previously stored pulse will not be retrieved. By successively scanning the angle of either the signal or control field, it is possible to store many pulses in a single ensemble. And it is also possible to selectively retrieve just one of the stored pulses, by aligning a retrieval control pulse with the appropriate direction for the chosen mode. The directional selectivity afforded by phasematching is demonstrated using our numerical model in part (a) of Figure 6.9. Here we simulated EIT storage, with θ chosen using the phasematching scheme (6.25). Upon retrieval, however, we rotated away from the optimal angle by θd . In the case δk = 0, we have θ = 0 and the spin wave momentum vanishes, so that all retrieval directions are phasematched. The efficiency is therefore only limited by walk-off, so it remains high over a large range of deviation angles θd . But when we set δk = 1, phasematching selects a unique retrieval direction, and the efficiency falls off quickly as θd is increased. This shows how it is possible to ‘switch off’ retrieval from a particular spin wave by tuning the angle of the control. 6.6.1 Optimizing the carrier frequencies Angular multiplexing requires that we implement storage with a range of angles θ between the signal and control fields. Now, for a given pair of frequencies ωc , ωs for the control and signal fields, there is only one angle that satisfies the phasematching 6.6 Angular Multiplexing (b) efficiency (a) 1 0.8 0.6 0.4 0.2 0 0 230 0.02 0.04 0.06 0.08 0.1 θd (radians) Figure 6.9 Angular multiplexing. In (a) we plot the dependence of the memory efficiency on the deviation angle θd , for EIT storage with δk = 0 (dotted line) and δk = 1 (solid line). We used a wide control beam, with wc = 9; all other parameters are the same as those used for the EIT protocol in Figures 6.7 and 6.8. The efficiency is only limited by walk-off when δk = 0, but when δk = 1 phasematching restricts the range of angles over which efficient retrieval is possible. In part (b) we illustrate the principle of multimode storage by angular multiplexing. The number of pulses that can be stored is given by N = ∆θ/δθ, where ∆θ is the largest angle, permitted by walk-off, for which efficient retrieval is possible, and where δθ is the smallest angle separating two modes, such that storage of one mode does not result in the accidental retrieval of another previously stored mode. Inspecting part (a), we see that ∆θ ∼ 0.04, and δθ ∼ 0.02, so N ∼ 2. scheme (6.25). As our simulations showed, this angle provides the optimal memory efficiency. To retain this optimum as θ is scanned, we should change the signal and control frequencies accordingly. For example, in the case that ωs < ωc (i.e. δk > 0), the phasematching scheme imposes the relationship ωs = ωc cos(θ) + ckdisp between θ, ωs and ωc . In the Raman protocol, it is possible to perform this kind of fine tuning on the frequencies, since the efficiency changes slowly with the detuning. Of course, EIT is defined by the condition that ∆ = 0, and so for this protocol there is no freedom to tune the frequencies in this way. 6.6 Angular Multiplexing 6.6.2 231 Capacity How many modes can be stored in such a multiplexed memory? One constraint is that it must be possible to resolve one retrieved signal field from another. Therefore the minimum angle δθ by which the direction of phasematched retrieval must change between two modes must be greater than the angular divergence of the retrieved signal field. This angular divergence is set by Fraunhofer diffraction, which yields the condition δθ & λs /ws — the angle between consecutive modes must exceed the ratio of the wavelength and beam waist of the signal field. Equally important is the requirement that there is no ‘cross-talk’ between neighbouring modes — that δθ is large enough that retrieval of an undesired mode is well-suppressed by its momentum mismatch. An estimate of the magnitude of the momentum mismatch is given by δθks (see Figure 6.10). This momentum mismatch is directed approximately perpendicular to the z-axis, so we can approximate the phase accumulated by the mismatch as δθks ws (working in ordinary units). For good suppression of unwanted retrieval, this phase should exceed 2π, roughly, and from this we again derive the condition δθ & λs /ws . Therefore the multimode capacity of a multiplexed memory is bounded by the number of diffraction limited modes that can be addressed. For the example in Figure 6.9 (a) this condition implies δθ ∼ 0.01, which appears to be broadly correct. The number of modes that can be stored is given by N = ∆θ/δθ, where ∆θ is the largest angle at which efficient retrieval is possible, as limited by walk-off between the signal and control fields. We approximate ∆θ as the angle subtended by the 6.6 Angular Multiplexing 232 Figure 6.10 Minimum momentum mismatch. Two signal fields with momenta k1 and ks differ in their directions by a small angle δθ. When retrieval for one field is phasematched, retrieval for the other is not, with the momentum difference shown in red, which points essentially across the ensemble. It’s magnitude is roughly δθks (ignoring any dispersion that may slightly alter the lengths of k1 , ks away from ks ). The total phase accumulated through propagation across the ensemble, of width ws , is then roughly δθks ws . When this phase is larger than 2π, phasematching is sufficiently selective to effectively suppress any accidental retrieval of stored excitations. control waist across the length of the cell, ∆θ ∼ wc /L (in ordinary units). The multimode capacity is then given by N ∼ wc ws /λs L. This is essentially equal to the geometric mean of the Fresnel numbers Fs , Fc associated with the signal and control fields, N∼ p Fs Fc , where Fs,c = 2 ws,c , Lλs,c (6.43) where we used the fact that the control wavelength λc is generally rather close to the signal wavelength. The Fresnel number of a beam quantifies the number of diffraction limited modes it supports, so this result has the interpretation that the number of modes that may be stored in a multiplexed memory is given by the average (in the sense of the geometric mean) number of modes supported by the optical fields used. For the example used in Figure 6.9 (a), the formula (6.43) predicts N ∼ 3, and this agrees reasonably well with what might be estimated ‘by eye’ (see caption of Figure 6.10). 6.6 Angular Multiplexing 233 A high capacity multiplexed memory requires wide beams, with Fs,c 1. In addition, our numerics showed that wc should be at least 2 (better 3) times larger than the signal beam waist. The feasibility of this kind of memory depends on the availability of energetic lasers that can maintain a high intensity, even when very loosely focussed. But laser technology has advanced sufficiently that this is certainly possible for beams with a width on the order of centimeters, giving multimode capacities on the order of 100. In the next chapter we continue to investigate multimode storage, both in the Raman and EIT protocols covered so far, and in a number of alternative memory protocols. Chapter 7 Multimode Storage At the end of the last Chapter we described a way to store multiple pulses in a single ensemble memory. We considered scanning the propagation direction of either the control or the input signal beams: each independently stored pulse represented a separate spatial mode, characterized by the direction of its wavevector. However, in many practical situations, it is desirable to fix the alignment of all components in an optical system. It may therefore be more useful to consider using a different degree of freedom to define a basis of modes. As discussed at the start of Appendix C, we might also consider polarization or spectral modes. And in fact since polarization is only two dimensional, any such consideration of multimode storage must focus on spectral modes. Equivalently, we might talk about temporal modes, since time and frequency are Fourier conjugates; they are two interchangeable representations of the same underlying space. This space can be thought of as the space of all possible input profiles for the signal field. The method used in the previous Chapters to 7.1 Multimode Capacity from the SVD 235 optimize the storage efficiency of a quantum memory is immediately useful in this connection. The SVD of the storage kernel provides a ‘natural’ basis for the space of signal profiles: the input modes {φk }. In this Chapter we apply the SVD to the study of multimode storage: we show how to optimize and quantify the multimode capacity of the Raman, EIT, CRIB and AFC memory protocols. The following treatment expands upon the account published in Physical Review Letters [155] . 7.1 Multimode Capacity from the SVD Suppose that we would like to store a ‘message’ in a quantum memory. We must select an alphabet in which to write the message. If we encode the message in an optical signal, each letter in the alphabet is represented by a mode of the optical field. The multimode capacity of the quantum memory is the size of the largest alphabet we can use for our message compatible with efficient storage. In general, this capacity will depend on the particular alphabet chosen: some modes of the optical field cannot be stored at all. To take an absurd example, we certainly could not encode any letters as X-rays! The multimode capacity also depends on the brightness of the signals used to encode the message. For instance, when two similar modes are bright, we can distinguish them and encode two letters, whereas when they are dim we can encode only a single letter (see Figure 7.1). In a quantum communication protocol, security requires that we encode each letter using just a single photon (see §1.5 in Chapter 1). This means that there can be no overlap whatsoever between the modes comprising the alphabet. That is, we 7.1 Multimode Capacity from the SVD A 236 B Bright Dim Figure 7.1 Bright overlapping modes are distinct. Two “letters” A and B are encoded with similar optical pulses. When they are bright, the letters can be distinguished; not so when their intensity is reduced. must find an orthonormal basis of modes. The multimode capacity of a quantum memory in this context is therefore the number of orthonormal modes that can be stored. It remains true that this capacity depends on the choice of basis. For example, it may not be the same for time bin modes as it is for frequency modes (see §C.1 in Appendix C). What, then, is the optimal basis? Which encoding maximizes the capacity? It is the basis formed by the input modes {φk } determined from the SVD of the storage kernel K. That is, for any number N , storage of the first N input modes φ1 , . . . , φN is more efficient than the storage of any other set of N orthonormal modes. Each input mode φk is stored with efficiency ηk = λ2k . In the optimal situation of phasematched backward retrieval without dispersion, the retrieval efficiency is also equal to ηk (see the discussion towards the end of §6.2 in 6). The total memory efficiency for the k th mode is then ηk2 = λ4k . The multimode capacity is given by the width of the ‘distribution’ of efficiencies (see Figure 7.2). Below we consider two ways of quantifying this width. 7.1 Multimode Capacity from the SVD 237 Figure 7.2 Visualizing the multimode capacity. The total memory efficiency — the efficiency of storage followed by retrieval — for the k th mode is ηk2 = λ4k . These efficiencies define a decaying distribution, and the multimode capacity is related to the width, or ‘variance’ of this distribution. The Schmidt number quantifies this width, but takes no account of the absolute efficiencies. If we introduce a threshold efficiency ηthreshold , we can define an operational measure N for the multimode capacity by considering the number of modes with memory efficiency greater than the threshold. 7.1.1 Schmidt Number The Schmidt number S is defined as follows [156,157] , 1 S = P∞ , ek4 k=1 η where η2 ηek2 = P∞k 2. j=1 ηk (7.1) The ηek2 are defined so that they sum to unity, turning them into effective probabilities. In fact, if we consider that these probabilities are the eigenvalues of a quantum mechanical density matrix, describing a statistically mixed state, then S is the inverse of the purity of this state [158] . From the definition (7.1) it’s clear that S = 1 if there is just one non-zero singular value. In general, if there are N equal singular values, with all others vanishing, then S = N , so the Schmidt number does indeed count the number of modes that may be stored with equal efficiency. When the singular values are not equal, but instead decay in magnitude smoothly as the index 7.1 Multimode Capacity from the SVD 238 k increases, S still provides a sensible measure of the number of coupled modes. The normalization of the η’s makes S independent of the overall efficiency of the memory: it measures how many modes are involved in the memory interaction with comparable strength, independent of what that strength is. This makes S an independent figure of merit for characterizing the performance of a quantum memory. One might choose to optimize the efficiency (i.e. the optimal efficiency) of a quantum memory, or the Schmidt number, or both. In most situations, however, it is more useful to quantify the ability of a memory to store multiple modes efficiently, so that a metric that combines these two measures is desirable. We introduce such a metric in the next section. But before abandoning the Schmidt number entirely, we observe that it is possible to develop an intuition for the value of S — at least whether or not it is large — by looking at the form of the Green’s function describing the storage interaction (see Figure 7.3). The reason is that the contribution of a single mode to the Green’s function necessarily takes the form of a product of two functions; each with contours orthogonal to the other’s: one function represents the input mode; the other the corresponding output mode. Now, the Schmidt number is independent of the basis of modes we consider, because the singular values are. We are therefore always free to choose a basis of modes comprised of localized pulses. Each mode is orthogonal to the others if the pulses do not overlap. The product of two such modes produces a contribution to the Green’s function in the form of a rectangular ‘blob’, with sides parallel to the coordinate axes, and rounded edges if the pulses have smooth edges. Counting the number of such blobs required to make up the 7.1 Multimode Capacity from the SVD 239 Green’s function provides an estimate of the number of modes required to construct it. Any feature of the Green’s function that traces out a curve is ‘multimode’, as is any feature that runs at an angle to the coordinate axes. The Schmidt number is sometimes useful because it is amenable to this kind of dead reckoning, as we will see in §7.3.2 below. (a) Single mode (b) Multimode Figure 7.3 The appearance of a multimode Green’s function. Inspection of the contours of the Green’s function describing a memory interaction offers insight into the degree to which it is multimode, as quantified by the Schmidt number S. By reconstructing the approximate shape of these contours using non-overlapping rectangles, one can estimate the number of modes of which it is composed. In (a) we show an archetypal single mode Green’s function, which looks like an isolated ‘hill’; a single elliptical contour is drawn. The contour is roughly approximated by just a single rectangle, indicating that only one mode contributes to its structure. In (b) we illustrate how curved or angled contours admit a decomposition involving more modes. 7.1.2 Threshold multimode capacity To be more quantitative, we introduce a threshold efficiency ηthreshold that allows us to delineate those modes that are stored and retrieved with acceptable efficiency from those that are not. We then define the multimode capacity as the largest number of modes N for which the mean memory efficiency, averaged over all N 7.1 Multimode Capacity from the SVD 240 modes, exceeds ηthreshold . The average memory efficiency of the first k input modes is given by Pk Λk = 2 j=1 ηj k . (7.2) Then the multimode capacity can be written as N= ∞ X k=1 βk , where βk = 1 if Λk > ηthreshold 0 otherwise . (7.3) A more conservative estimate could replace the average efficiency (7.2) with the minimum efficiency ηk2 . The results presented below are not qualitatively altered by making this alternative definition, but as N becomes large, this definition represents an increasingly pessimistic estimate of the performance of a memory for storing random ‘words’ drawn from the alphabet defined by the {φk }. The capacity defined in (7.3) is a convenient measure, with a clear interpretation. In the next section, we investigate the scaling of N for various memory protocols. We set ηthreshold = 70%; this value makes the form of the scaling apparent for parameters that are tractable numerically, and indeed experimentally. We return to the one dimensional model used in Chapter 5, since we are able to write down explicit expressions for the storage kernels in this case. The simulations presented in the previous Chapter confirm that the one dimensional approximation works well, provided that the control field is more loosely focussed than the signal field (see §6.5 in Chapter 7). 7.2 Multimode scaling for EIT and Raman memories 7.2 241 Multimode scaling for EIT and Raman memories In Chapter 5 we showed that the best possible storage efficiency for a Λ-type quantum memory is found from the SVD of the anti-normally ordered kernel (5.27) [133] . We repeat it here for convenience, √ d 0 KA z, z 0 = e−d(z+z )/2 I0 d zz 0 . 2 (7.4) Here we are using normalized units for the z coordinates. The kernel (7.4) is valid for both EIT and Raman protocols, in the light-biased limit where the control pulse is sufficiently energetic (see §5.2.8 in Chapter 5). The eigenvalues of this optimal kernel are equal to the largest possible values of the storage efficiencies ηk = λ2k , so we can calculate the multimode capacity N by diagonalizing KA . This can be done numerically without difficulty, and we present the results of such a numerical diagonalization in part (a) of Figure 7.4 below. In part (b) we plot the capacity predicted by the Raman kernel (5.95) introduced in §5.3 of Chapter 5. In both cases the capacity rises only slowly with the optical depth d. To explain these results, we briefly return to the treatment given in §5.2.3 in Chapter 5. There we derived an approximate expression (5.21) for the storage efficiencies, valid in the limit d 1: ηk = e−4παk /d , (7.5) 7.2 Multimode scaling for EIT and Raman memories 242 √ where αk is the k th zero of the function J0 (2 2πx). The zeros of J0 are distributed approximately linearly along the real line (see §15.4 in Chapter XV in G.N. Watson’s famous treatise [159] ). Therefore the αk increase, roughly, with the square of the index k. Setting αk ∼ qk 2 for some constant q, we find that the index of the mode whose total efficiency ηk2 falls below ηthreshold is given approximately by r k∼ ηthreshold √ × d. 8πq (7.6) The multimode capacity N should scale in the same way, so that we expect N∼ √ d, (7.7) for large optical depths, regardless of the threshold ηthreshold . Inspection of Figure 7.4, which was generated by setting ηthreshold = 70%, indeed reveals a scaling of N≈ √ d/3. 7.2.1 A spectral perspective The kernel (7.4) is derived by considering the optimal efficiency for absorption into the excited state |2i, and the square root scaling of N with d can be understood by considering the bandwidth of the absorption line associated with an ensemble of twolevel atoms. If each atom has a Lorentzian lineshape with natural linewidth 2γ, the absorption profile of the entire ensemble is given by exponentiating the single-atom 7.2 Multimode scaling for EIT and Raman memories 243 profile, 2dγ 2 . F (ω) = exp − 2 γ + ω2 (7.8) The full width at half maximum (FWHM) of F is given by r ∆ω = 2 2dγ 2 − γ 2 ≈ 2γ ln 2 r 2d , ln 2 (7.9) where the approximation holds for d 1. An estimate of the multimode capacity is provided by counting the number of spectral modes with FWHM 2γ that ‘fit’ inside this absorption profile, which procedure yields N∼ ∆ω √ ∼ d. 2γ (7.10) We can therefore identify the origin of the square root scaling of N with d as the scaling of the bandwidth of the absorption of an ensemble with its optical thickness. This scaling is rather poor: in order to store two modes in such a memory, we must quadruple the optical depth required for storing a single mode. If this optical depth were divided up into four separate ensemble memories, we could store four modes — one mode in each memory — and so the multimode scaling for EIT and Raman memories is decidedly sub-optimal. The reason is that when the optical depth is increased, new atoms are added ‘behind’ the old: they absorb at the same frequencies as the other atoms, and so they improve the absorption at these frequencies, but they do not provide much coupling at other frequencies — other modes. In the next 7.2 Multimode scaling for EIT and Raman memories 244 section we will show that adding an inhomogeneous broadening to the ensemble can improve the multimode scaling, by redistributing optical depth over a range of frequencies. This turns the square root scaling with d into a linear one, consistent with what one might expect to achieve from operating separate memories in parallel. And this is essentially the way such broadened protocols work, albeit within a single physical ensemble. (a) (b) Optimal kernel Raman kernel 8 6 4 2 0 0 100 200 300 400 500 0 20 40 60 80 100 Figure 7.4 Multimode scaling for Raman and EIT memories. (a): the multimode capacity found by numerical diagonalization of the optimal kernel KA in (7.4). This represents the best possible scaling for both EIT and Raman memories; the capacity scales only with the square root of the optical depth. (b): we also show the multimode capacity found by diagonalizing the Raman kernel (5.95), which is parameterized by the Raman memory coupling C (see §5.3 in Chapter 5). This kernel is valid in the far off-resonant limit — though if the control pulse intensity is increased sufficiently, the optimal kernel KA should be used instead. We plot the Raman capacity against C 2 , since this quantity is proportional to d. The same square root scaling with ensemble density is evident. Equivalently, if d is held constant, the Raman capacity scales with the square root of the control pulse energy. We used ηthreshold = 70% in both plots. 7.3 CRIB 7.3 245 CRIB The CRIB memory protocol is introduced in §2.3.3 in Chapter 2. Storage is achieved by direct absorption into the excited state |2i, which is artificially broadened by application of an external field. Once storage is complete, the excitation is ‘shelved’ by application of a short, bright control pulse, which transfers the excitation to the storage state |3i. To implement retrieval, another control pulse transfers the excitation back to |2i, and the inhomogeneous broadening is ‘flipped’, so that the atomic dipoles re-phase and the signal field is re-emitted. The same considerations regarding retrieval discussed in the previous Chapter apply to this protocol, just as they do to Raman and EIT memories: retrieval in the forward direction is inefficient due to re-absorption losses [160] , while backward retrieval is vulnerable to phasematching problems. However, in this protocol, the control pulse is applied after the signal has been absorbed, so it is possible to distinguish the two fields temporally. Spectral and spatial filtering is therefore less important, and so it is feasible to use an ensemble where the states |1i and |3i are degenerate. This allows for efficient collinear storage, followed by phasematched backward retrieval. Since there is no dispersion on resonance, the stored excitation has no spatial phase, and the retrieval efficiency is equal to the storage efficiency, 2 with ηcombined = ηstorage . In the following, we restrict our attention to this situation, which is optimal. 7.3 CRIB 7.3.1 246 lCRIB We first consider lCRIB, in which the broadening is applied longitudinally [84,161,162] . The resonant frequency of the |1i ↔ |2i transition varies linearly along the z-axis. Since the control field is only applied after the signal field has been resonantly absorbed, the equations of motion for lCRIB are given by the system (5.106), with Ω = 0, and with the spatial variation of the detuning included. √ ∂z A(z, τ ) = − dP (z, τ ), ∂τ P (z, τ ) = −Γ(z)P (z, τ ) + where Γ(z) = 1 − i∆(z), with ∆(z) = ∆0 z − 1 2 √ dA(z, τ ), (7.11) . Here we have returned to the normalized units of Chapter 5, with all frequencies measured in units of γ, and with z running from 0, at the entrance face of the ensemble, up to 1, at the exit face. The width of the applied spectral broadening is ∆0 . As usual, we solve these equations by applying a unilateral Fourier transform over the z-coordinate. Using the formula (D.27) from Appendix D, we obtain the transformed system √ e − √1 Ain = − dPe, −ik A 2π √ e ∂τ Pe = − 1 − ∆0 ∂k + i ∆20 Pe + dA, (7.12) 7.3 CRIB 247 where Ain is the temporal profile of the incident signal field. Solving the first equation e and substituting the result into the second equation yields for A (∂τ − ∆0 ∂k ) Pe = − 1 + i ∆20 + i kd √ d Pe + i √ Ain . 2πk (7.13) Now if we temporarily define the composite variable s = k + ∆0 τ , we can replace the derivatives on the left hand side with a single time derivative, √ ∂τ Pe = −f (s, τ )Pe + i √ d Ain , 2π(s − ∆0 τ ) (7.14) where the partial derivative ∂τ is taken with s held constant, and where we have defined the function d . f (s, τ ) = 1 + i ∆20 + i s−∆ 0τ (7.15) Integrating (7.14) gives the solution Peout (s) = Pein (s)e− RT −∞ f (s,τ ) dτ √ Z T RT d 0 0 + i√ Ain (τ )e− τ f (s,τ ) dτ dτ, 2π −∞ (7.16) where Pout (s) = P (s, T ) is the atomic excitation at the moment τ = T that the short control pulse is applied, marking the end of the storage interaction. We do not explicitly model the atomic dynamics induced by the control; we simply assume that it is sufficiently intense to transfer — effectively instantaneously — all the atomic excitations into the storage state |3i, so that Pout −→ Bout . We set Pin = 0, since no atoms are excited at the start of the storage process. We then perform the integral 7.3 CRIB 248 in the exponent of the integrand in the second term, and convert back from s to k, to arrive at the following expression for the lCRIB storage map in k-space, Z ∞ e K(k, T − τ )Ain (τ ) dτ, eout (k) = B (7.17) −∞ e defined by with the kernel K √ d e K(k, τ ) = i √ e−(1+i∆0 /2)τ × k id/∆0 (k + ∆0 τ )−id/∆0 −1 2π (7.18) e = 0 when τ < 0. The multimode scaling of lCRIB is determined for τ ≥ 0, with K by the singular values of this kernel. Unfortunatey we cannot extract these singular values directly because there is a singularity at k = −∆0 τ . A number of alternatives are open to us. First, we can return to the equations of motion, and construct the Green’s function in (z, τ )-space directly by numerical integration (see §5.4 in Chapter 5 and §E.5 in Appendix E). This kernel is not singular, and is therefore amenable to a numerical SVD. We’ll refer to this as simply the ‘numerical method’. Second, we can remove the singularity in (7.18) by applying a small regularization to k. That is, we replace k by k − i, where is some small real number. This shifts the singularity off the real axis, making the kernel well-behaved. But it also changes the singular values, since it is not a unitary transformation. We fix this with the following procedure. We apply a numerical inverse Fourier transform from k-space back to zspace. This can be performed very efficiently with a fast Fourier transform (FFT), an implementation of which is standard in Matlab. The Fourier transform is unitary, so 7.3 CRIB 249 it does not affect the singular values. Next, we multiply the result by the exponential factor e−z . By the shift theorem (see §D.3.4 in Appendix D), this compensates for the regularization, so that the result is equal to the Fourier transform of (7.18) without the regularization applied. Because the Fourier transform is unitary, we can extract the singular values of (7.18) by taking a numerical SVD of the kernel generated by this procedure. We will refer to this as the ‘Fourier method’. We use both of these methods below. But to gain some insight into the form of the multimode scaling of lCRIB, we now introduce a third approach. The approach is only valid for very large broadenings, but it clarifies the scaling behaviour exhibited by the numerical techniques just described. 7.3.2 Simplified Kernel We perform a series of unitary transformations that simplify the form of the kernel (7.18), while leaving the singular values unchanged. The first of these transformations is to trop the exponential factor iei∆0 τ /2 , since it is just a phase rotation. Next we drop the factor k id/∆0 . This is, again, a pure phase rotation, which is well defined as long as ∆0 6= 0 and k 6= 0. The second of these conditions arises because there is a logarithmic singularity in the phase of k id/∆0 at k = 0. The effect of this singularity is small when d/∆0 is small, so in the following we assume a large broadening, with ∆0 d. The resulting kernel is √ d e K(k, τ ) = √ e−τ (k + ∆0 τ )−id/∆0 −1 . 2π (7.19) 7.3 CRIB 250 We now take the inverse Fourier transform, from k-space back to z-space. I have to confess to not knowing how to perform this transform, but Mathematica provides an answer! Combining this with the shift theorem gives √ K(z, τ ) = α (d/∆0 ) de−τ z id/∆0 ei∆0 zτ , (7.20) where α is given by 1 α(x) = − e−πx/2 sinh(πx)Γ(ix), π (7.21) with Γ denoting the Euler Gamma function (not the complex detuning!). Finally, we note that the factor z id/∆0 is another pure phase rotation, well-defined if ∆0 6= 0 and z 6= 0. We can drop it without affecting the singular values, again in the limit ∆0 d, which gives the simple kernel, √ K(z, τ ) = α(d/∆0 ) de(i∆0 z−1)τ . (7.22) Now we form the anti-normally ordered kernel KA , by integrating the product of two of these kernels from τ = 0 to τ = ∞ (see §3.3.1 in Chapter 3), KA (z, z 0 ) = d 1 |α(d/∆0 )|2 × . ∆0 2/∆0 − i(z − z 0 ) (7.23) This kernel has the simple structure we have been seeking. It takes the form of a Lorentzian peak centred on the line z = z 0 , with a width set by ∆0 , and a height determined by the ratio d/∆0 . We can draw two conclusions from the form of this 7.3 CRIB 251 kernel. First, in the limit of large broadening ∆0 d, the Schmidt number becomes independent of the optical depth, being determined only by ∆0 . This follows from the fact that the functional form of (7.23) does not depend on d, which only affects its overall magnitude. The Schmidt number does not depend on this overall magnitude, because of the normalization of the η’s in (7.1), and so S is independent of d. The contours of (7.23) form a strip along the diagonal line z = z 0 . Application of the estimation technique described in Figure 7.3 suggests that the multimode capacity is proportional to the ratio of the length to the width of the strip, so that S ∝ (2/∆0 )−1 ∼ ∆0 (see Figure 7.5). Numerics confirm this to be the case: the Schmidt number of an lCRIB memory rises linearly with the applied broadening (see Figure 7.6). The second conclusion we can draw from the structure of (7.23) is that the threshold multimode capacity N rises linearly with the optical depth d — a significant improvement over the square root scaling derived for EIT and Raman memories in §7.2. To see why, consider a situation where the optimal memory efficiency η12 exceeds the threshold efficiency ηthreshold by a reasonable margin, so that N ≈ S (see Figure 7.7). If we double the applied broadening, the Schmidt number doubles. But the ratio d/∆0 is then halved, so that the overall efficiency falls below ηthreshold , and N ≈ 0. In fact, the function α(x) ≈ 1 for x 1, so that halving d/∆0 approximately halves the ηk , and divides the total memory efficiencies ηk2 by four. To bring the overall efficiency back to its previous value, above the threshold, we must 7.3 CRIB 252 (a) Many modes (b) Few modes 1 1 0.5 0.5 00 0.5 1 00 0.5 1 Figure 7.5 Scaling of Schmidt number with broadening. In (a) and (b) we plot |KA | for large and small broadenings of ∆0 = 100 and ∆0 = 10, respectively, with d = 30 in both cases. Below these plots we illustrate the mode-counting procedure described in Figure 7.3, which provides a way to understand why the Schmidt number increases linearly with the applied broadening (see Figure 7.6). double the optical depth. The multimode capacity N is then doubled, because we have both doubled S and maintained the correspondence N ≈ S by keeping the overall efficiency above ηthreshold . This argument is quite general: we can increase the number of modes contributing to the storage interaction by increasing the applied broadening, but the coupling is ‘shared’ between all these modes, so we must increase the optical depth at the same time in order to maintain efficient storage over all the modes. If our aim is to maximize N , there is an optimal value for the ratio d/∆0 , which depends on ηthreshold . If the ratio is too large, we can afford to increase the broadening and introduce more modes without bringing the efficiency below the threshold. Converseley, if the ratio is too small, we should sacrifice some 7.3 CRIB 253 (a) (b) 1 30 20 0.5 10 0 0 100 200 300 0 0 100 200 300 Figure 7.6 Comparison of the predictions of the kernels (7.23) and (7.18). (a): the optimal total memory efficiency η12 predicted by the simplified kernel (7.23) (green dotted line) is plotted alongside the efficiency predicted by the kernel (7.18), using the ‘Fourier method’ (blue solid line), as a function of the applied broadening. The optical depth is set at d = 30; the simplified kernel compares well with the Fourier method for ∆0 & 3d. (b): the Schmidt number predicted by the two kernels. The agreement between the two, even at the boundaries of the regime of validity of the simplified kernel, is excellent. The linear scaling of S with ∆0 expected from the form of (7.23) is clear. modes and reduce the broadening to boost the efficiency above the threshold. This analysis shows that the multimode capacity N of an lCRIB memory scales linearly with d, provided that the broadening ∆0 is increased linearly with d at the same time. The preceding discussion applies in the limit ∆0 d. We confirm the persistence of linear multimode scaling outside this regime using the direct numerical method described earlier, in which we construct the Green’s function by integrating the equations of motion. The result is plotted in Figure 7.8 at the end of §7.3.3 below, alongside the results for EIT and Raman shown earlier, and the results for tCRIB, which we now turn to. 7.3 CRIB 254 (a) (b) (c) Figure 7.7 Understanding the linear multimode scaling of lCRIB. In the limit of large broadening, the Schmidt number of the lCRIB storage interaction depends linearly on the width of the applied spectral broadening, but the overall storage efficiency is determined by the ratio d/∆0 . To double the multimode capacity N , given a fixed threshold efficiency ηthreshold , we need to double the Schmidt number, while keeping the overall efficiency the same. The process is shown in parts (a) to (c). We begin with η12 > ηthreshold , so that S ≈ N . In (b) we double the spectral broadening, which doubles S, but reduces the efficiency (in fact, by roughly a quarter). To return to the same efficiency as in (a), in part (c) we double the optical depth. This has no effect on S, but it returns the ratio d/∆0 to the same value as in part (a). By doubling the optical depth, we have suceeded in doubling the multimode capacity. These arguments explain why the threshold capacity N depends linearly on d for lCRIB. 7.3.3 tCRIB In a tCRIB memory, the direction of the broadening is perpendicular to the z-axis. That is, the resonant frequency of the atoms varies across the ensemble (see part (a) of Figure 2.10 in Chapter 2). The theoretical treatment requires that we divide the ensemble into frequency classes, where all the atoms in one frequency class have the same detuning from the signal field carrier frequency. Thorough treatments are given by Gorshkov et al. [163] , and also by Sangouard et al. [160] , and we adapt their techniques to our purpose, which is to derive the Green’s function for the tCRIB memory interaction. Suppose that the applied broadening produces an inhomogeneous line with spec- 7.3 CRIB 255 tral absorption profile p(∆). This means that a proportion p(∆)d∆ of the the atoms have their resonant frequencies shifted by ∆ away from their nominal frequency. The profile is normalized, so that Z p(∆) d∆ = 1. (7.24) The total optical depth d is divided amongst all the frequency classes, so that the optical depth contributed by atoms detuned by ∆ is d(∆)d∆ = dp(∆)d∆. The equations of motion for the storage process are found by a straightforward generalization of the system (7.11) to the case of multiple frequency classes (and with no variation of ∆ with z in this case, of course), Z p ∂z A = − d(∆)P (∆) d∆, ∂τ P (∆) = −Γ(∆)P (∆) + p d(∆)A. (7.25) Here we have emphasized the functional dependence of the complex detuning Γ(∆) = 1 − i∆, and we have defined P (∆) as the slowly varying polarization associated with the frequency class of atoms detuned by ∆. We quickly encounter difficulties if we attempt to solve this system of equations with our usual trick of Fourier transforming over z. Instead, we get to the solution by applying a unilateral Fourier transform over time, from τ to ω. Note that here ω is the frequency conjugate to the retarded time τ ; it certainly has no relation to the integrated Rabi frequency used in Chapter 7.3 CRIB 256 5! The transformed equations are given by e = − ∂z A Pin (∆) −iω Pe(∆) − √ 2π Z p d(∆)Pe(∆) d∆, = −Γ(∆)Pe(∆) + p e d(∆)A. (7.26) Solving the second equation, and substituting the result into the first yields the following equation for the signal field, r e z) = − [∂z + df (ω)] A(ω, d 2π Z p p(∆)Pin (z; ∆) d∆, Γ(∆) − iω (7.27) where we have defined the lineshape function Z f (ω) = p(∆) d∆, 1 − i(∆ + ω) (7.28) which is essentially the convolution of the Lorentzian spectral response of the atoms with the inhomogeneous profile p(∆). Recall that in our normalized units the natural atomic linewidth is defined to be equal to 1. We can immediately solve for the signal field, r e z) = e−df (ω)z A ein (ω) − A(ω, d 2π Z p Z p(∆) 0 z 0 e−df (ω)(z−z ) Pin (z 0 ; ∆) 0 dz d∆. Γ(∆) − iω (7.29) ein is the spectrum of the incident signal field. Now that we are in possession of Here A a solution for the signal, it is a matter of algebra to construct the Green’s function. 7.3 CRIB 257 But it is not enough to find the storage map alone. We can write down an expression for the atomic excitation at the end of the interaction, but since it is distributed over all frequency classes, it is not clear what we should optimize for efficient storage. There is no single spin wave whose norm represents the storage efficiency. Therefore, we construct the kernel Ktotal that relates the input signal to the retrieved signal field (see §3.4 in Chapter 3). The modes found from the SVD of this kernel have a clear interpretation as those modes that are eventually retrieved from the memory. To find this kernel, we first solve for the polarization Pout at the end of the storage process, setting Pin = 0. This requires taking an inverse Fourier transform from ω back to τ = T , where T marks the end of the storage interaction, 1 Pout (z, ∆) = √ 2π Z ∞ e−iωT Pe(ω, z; ∆) dω, (7.30) −∞ where, using the second line of (7.26), Pe(ω, z; ∆) = = p e z) d(∆)A(ω, Γ(∆) − iω p d(∆)e−df (ω)z ein (ω), ×A Γ(∆) − iω (7.31) where in the second line we used (7.29). This completes the description of the storage process. At time τ = T , the control pulse shelves all the excited atoms. To describe retrieval of the excitations, we again use (7.29). As usual we use a superscript r to identify quantities associated with the retrieval process. There is no input field for er = 0. The retrieved signal field, at the exit face of the ensemble with retrieval, so A in 7.3 CRIB 258 z r = 1, is given by r erout (ω r ) = − A d 2π 1 Z p Z r r p (∆ ) e−df r (ω r )(1−z r ) 0 r (∆r , z r ) Pin dz r d∆r . Γ(∆r ) − iω r (7.32) We ‘stitch’ together the storage and retrieval interactions by making the identification r Pin (∆r , z r ) = Pout (∆, z). (7.33) This says that the atomic polarization is the same at the end of the storage process as it is at the start of the retrieval process. But in between storage and retrieval, the inhomogeneous profile is flipped, so that red-detuned atoms become blue-detuned and vice-versa. This is crucial for reversing the atomic dynamics so that they re-emit the signal field. We model this flipping of the detunings by setting ∆ = −∆r , (7.34) so that frequency classes with positive and negative detunings swap places. For backward retrieval with no phase mismatch — the optimal situation — we also swap the z-coordinate: z = 1 − zr. (7.35) Note that in this case the frequency ω r is conjugate to the retarded time τ r , which is the time coordinate in a frame moving at the speed of light backwards with respect to the initial z-axis used for the storage process. We now combine the relations (7.35) 7.3 CRIB 259 and (7.34) with (7.33), and substitute this into (7.32), using (7.30) and (7.31). The result is er (ω r ) = A out Z ∞ e total (ω r , ω)A ein (ω) dω, K (7.36) −∞ where the Green’s function for the memory interaction is given by r −iωT e total (ω , ω) = −e K Z p Z 1 pr (∆r )p(−∆r ) d r r r r e−d(1−z )[f (ω )+f (ω)] dz r . d∆ × r r r [Γ(∆ ) − iω ][Γ(−∆ ) − iω] 2π 0 (7.37) Now we assume that the inhomogeneous profile is the same for both the storage and retrieval processes, and symmetric about the unbroadened resonance. then pr (∆r ) = p(∆r ) = p(−∆r ), and the lineshape is also unchanged, f r = f . Performing the integrals over ∆r and z r , we find r −d[f (ω )+f (ω)] − 1 e total (ω r , ω) = 1 e K , 2π 2 + i(ω r + ω) (7.38) where we have dropped the exponential factor of e−iωT , since it represents only an unimportant phase rotation. Here we note that a nearly identical treatment was first given by Gorshkov et al. [163] . The crucial difference for us is that we used a unilateral Fourier transform, rather than a Laplace transform, to solve the ein and A er have a natural equations. This means that the transformed amplitudes A out interpretation as the spectra of the input and retrieved signal fields. The norm of the signal spectrum is the same as the norm of the temporal profile, by Parseval’s theorem (or, alternatively, by energy conservation), and so the SVD of the kernel 7.3 CRIB 260 Ktotal in (7.38) tells us about the storage efficiency of the memory, and indeed, its multimode capacity. Now that we have found an explicit form for the Green’s function describing tCRIB, the multimode capacity can be found by taking its SVD: the singular values of Ktotal are equal to the ηk . In Figure 7.8 we plot the resulting prediction for the multimode scaling of tCRIB. We used a rectangular broadening profile with total width ∆0 , p(∆) = 1/∆0 if |∆| ≤ ∆0 /2, 0 otherwise. (7.39) And we optimized N with respect to the width ∆0 , using a threshold ηthreshold = 70%. It is clear that the scaling of N with d is linear. Furthermore, it is the same as the scaling of lCRIB, whose multimode capacity is also plotted. This suggests that the multimode capacity of both CRIB protocols is identical. The scaling of unbroadened EIT and Raman protocols, as derived from (7.4), is shown for comparison. Clearly CRIB dramatically outperforms equivalent unbroadened protocols: given the same optical depth — the same total number of atoms — more modes of the optical field can be stored by adding a controlled broadening. We found that for both tCRIB and lCRIB, the multimode capacity scaled roughly as N ∼ d/25, and the optimal width for the spectral broadening scaled as ∆opt 0 ∼ 9d/5. This confirms the validity of the arguments given at the end of §7.3.2 for lCRIB. Just as the poor multimode scaling of unbroadened protocols can be explained by considering the absorption bandwidth of a homogeneous ensemble (see §7.2.1), so 7.3 CRIB 261 we can understand the improved scaling of CRIB by considering the inhomogeneous linewidth. Consider the storage of a spectral mode with bandwidth 2γ. An optical depth of order 10 is required to efficiently absorb the incident light. To store N such spectral modes ‘side by side’ in frequency, we should have a total optical depth d ∼ 10N , spread over a spectral width ∆0 ∼ 2γN . That is, the multimode capacity scales linearly with d, and so does the broadening ∆0 , which is precisely what we have found to be the case for CRIB. And it is clear why there is an optimal broadening: if ∆0 is too large, there is insufficient optical depth over the bandwidth of a mode for efficient absorption. In the next section, we consider a modification to the Raman protocol which takes advantage of spectral broadening to improve its multimode scaling. 20 10 0 0 100 200 300 400 500 Figure 7.8 Multimode scaling for CRIB memories. The multimode capacity N is shown as a function of the total optical depth d for tCRIB (green dashed line), lCRIB (red dotted line) and, for comparison, unbroadened EIT and Raman protocols (solid blue line). For the tCRIB calculation, a numerical SVD was applied directly to the kernel (7.38). For lCRIB, we constructed the storage kernel by solving the equations of motion (7.11) numerically — we found this to be the most reliable method, and of course no approximations are required. Nonetheless some numerical error is apparent at large broadenings, since the numerical problem becomes increasingly stiff as ∆0 increases. For both the CRIB protocols, we optimized N over ∆0 , and as expected, the optimal broadening width was found to scale linearly with d. We set ηthreshold = 70% for all calculations. 7.4 Broadened Raman 7.4 262 Broadened Raman Is it possible to improve the multimode capacity of the EIT and Raman protocols by incorporating a spectral broadening? Here we consider a simple modification to the Raman protocol, in which a longitudinal broadening is applied to the storage state. That is, the energy of the state |3i varies linearly along the z-axis, covering a range of frequencies ∆0 . The treatment is very similar to that of lCRIB given earlier, and we will see that it is indeed possible to recover the linear multimode scaling characteristic of CRIB. As usual, we start with the equations of motion describing one dimensional propagation. Adapting the system (5.106) to the present scenario, with an added broadening, we have √ ∂z A(z, τ ) = − dP (z, τ ), ∂τ P (z, τ ) = −ΓP (z, τ ) + ∂τ B(z, τ ) = i∆0 z − 1 2 √ dA(z, τ ) − iΩ(τ )B(z, τ ), B(z, τ ) − iΩ∗ (τ )P (z, τ ). (7.40) The first term on the right hand side of the Heisenberg equation for B just describes a position-dependent energy shift of the storage state. We make no attempt to solve these equations exactly. Rather, we proceed directly to study the adiabatic limit, in which any driven dynamics are much slower than the timescale 1/∆ set by the detuning. In this limit, we can eliminate the polarization P , by setting the left hand side of the second equation to zero (see §5.3.3 in Chapter 5, and also the papers by Gorshkov et al. [133] ). Solving the resulting algebraic equation for P , and substituting 7.4 Broadened Raman 263 the result into the other two equations yields the system √ d Ω(τ ) d ∂z + A(z, τ ) = i B(z, τ ), Γ Γ √ |Ω(τ )|2 Ω∗ (τ ) d 1 ∂τ + − i∆0 z − 2 B(z, τ ) = −i A(z, τ ). Γ Γ (7.41) As we did for lCRIB, we now apply a unilateral Fourier transform over z. The adiabatic equations of motion in k-space are then e τ) = A(k, √Γ Ain (τ ) 2π √ e τ) + iΩ(τ ) dB(k, d − ikΓ √ 2 ∗ ∆0 |Ω(τ )| e Ω (τ ) d e ∂τ − ∆0 ∂k + i + B(k, τ ) = −i A(k, τ ). 2 Γ Γ , (7.42) Again we encounter the combined derivatives ∂τ − ∆0 ∂k , which we deal with by transforming from the coordinates (k, τ ) to (s, τ ), where s = k+∆0 τ . The derivatives can then be replaced with ∂τ , where now s is held constant. Substituting the first equation of (7.42) into the second, and integrating, we arrive at the solution − eout (s) = B ein (s)e B R∞ −∞ f (s,τ ) dτ r −i d 2π Z ∞ −∞ R∞ Ω∗ (τ ) 0 0 e− τ f (s,τ ) dτ Ain (τ ) dτ, d − iΓ(s − ∆0 τ ) (7.43) where the function f is given by ∆0 |Ω(τ )|2 d + 1− . f (s, τ ) = i 2 Γ d − iΓ(s − ∆0 τ ) (7.44) 7.4 Broadened Raman 264 ein = 0. At the end of the storage interaction, To model the storage process, we set B at time τout −→ ∞, the coordinate s is given by s = k + ∆0 τout . That is, k is related eout , so it does not affect to s by a constant offset. This does not affect the norm of B the efficiency of the memory. In the following we therefore make the replacement s −→ k, since k has the clearer physical meaning: it is the spatial frequency of the spin wave. We also drop the first term i∆0 /2 from f , since it represents only a phase rotation, which also does not affect the memory efficiency. The storage map for a broadened Raman memory can then be written as Z ∞ eout (k) = B e K(k, τ )Ain (τ ) dτ, (7.45) −∞ where the k-space storage kernel is given by r e K(k, τ ) = −i R 1 ∞ d ∗ 0 2 0 0 Ω (τ )g(k, τ )e− Γ τ |Ω(τ )| [1−dg(k,τ )] dτ , 2π (7.46) with g defined by g(k, τ ) = 1 . d − iΓ(k − ∆0 τ ) (7.47) Note that the broadening introduces a timescale into the dynamics that is not set by the control field, so we cannot conveniently write this kernel in terms of the integrated Rabi frequency used in Chapter 5 (see (5.46)). But if we set ∆0 = 0, (7.46) does indeed reduce to the standard adiabatic storage kernel (5.74) for a Λ-type memory. 7.4 Broadened Raman 265 We can extract the singular values from (7.46) with a numerical SVD. This can be a little problematic, since for large detunings the kernel becomes nearly singular. But it is easy to resolve this issue by introducing a regularization, Fourier transforming, and then compensating — this ‘Fourier method’ is described at the end of §7.3.1 above, where we applied it to the k-space storage kernel for lCRIB, which is also singular. In Figure 7.9 we show how the square root scaling of the unbroadened protocol is transformed into linear scaling upon application of a broadening. To see how this scaling arises from the structure of the storage kernel, we repeat the arguments used for lCRIB which allow us to simplify the kernel. We consider the special case of control field with a rectangular profile, since this allows us to perform e the integral in the exponent of K, Ω(τ ) = Ωmax for 0 ≤ τ ≤ T , 0 otherwise. (7.48) The kernel evaluates to r e K(k, τ ) = −i i dΩ22max Γ ∆0 d g(k, τ ) 2 Ωmax g(k, τ )e−Ωmax T /Γ . 2π g(k, T ) (7.49) In the Raman limit ∆ 1, we can write Γ ≈ −i∆, to obtain 2 2 C∆ −iΩ2max T /∆ i C 1−i C e K(k, τ ) = −i √ e × [g(k, T )] ∆0 T × [g(k, τ )] ∆0 T , 2πT (7.50) 7.4 Broadened Raman 266 where C is the Raman coupling (this is defined in (5.94) in §5.3 of Chapter 5). In this far-detuned limit, the exponential factor is a pure phase rotation that we are free to remove, and so is the factor involving g(k, T ), since g is purely real and it is here raised to a purely imaginary power. As in our treatment of lCRIB, there is a logarithmic singularity at g(k, T ) = 0, but its effect becomes negligible in the limit ∆0 C 2 /T of large broadening. Dropping these terms, and some other spurious phase factors, we find C e K(k, τ) = √ 2πT d − k − ∆0 τ ∆ i C2 ∆0 T −1 . (7.51) Taking the inverse Fourier transform, with the help of Mathematica, and the shift theorem, and dropping a further phase factor, we find the z-space kernel K(z, τ ) = α C 2 /∆0 T p C 2 /T e−i∆0 zτ , (7.52) where the function α is defined in (7.21) in §7.3.2 above. This kernel is nearly identical in form to (7.22). Instead of being damped in time by the factor e−τ due to the spontaneous lifetime of the excited state, the kernel is instead truncated at τ = T , when the control field is switched off. But the functional form is the same: an exponential parameterized only by the spectral broadening ∆0 . The magnitude of the kernel is also the same as for (7.22), except the optical depth d has been swapped for the quantity C 2 /T . By analogy with the facts we know for lCRIB, we therefore conclude the following. First, in the limit of large broadening, the Schmidt 7.4 Broadened Raman 267 number of the Raman memory depends only on ∆0 . Second, the multimode capacity N , given some threshold efficiency, scales linearly with C 2 /T , provided that as this quantity is increased, the applied broadening ∆0 is also increased proportionately. Note finally that C 2 ∝ d, so the multimode capacity scales linearly with the optical depth, as it does for CRIB. These assertions are confirmed by the results of a numerical SVD performed on the kernel (7.46) using the ‘Fourier method’ described above. We used a Gaussian control pulse, but the shape of the control makes no difference to the multimode scaling in the adiabatic limit. This is to be expected, since in this regime the dynamics adiabatically follow the control, and so its temporal profile only affects the shapes of the input modes, not the efficiency of the memory. For the parameters shown, we found N ∼ d/300, with an optimal broadening width of ∆0 ∼ d/77. We should remark that the multimode capacity is much smaller than that found for CRIB, or even than is predicted by the optimal kernel KA in (7.4). This is because the memory operates far off resonance, so that a much higher optical depth is required to achieve a strong interaction. Of course the other advantages of Raman storage — for instance, broadband capability, tunability, and insensitivity to unwanted inhomogeneous broadening — are retained in the present scheme. Since investigating this protocol, a demonstration has been implemented by Hétet et al. in Canberra [164] , although only a single optical mode was stored and retrieved: this is quite difficult enough to start with! An analysis of this protocol was also recently conducted by Moiseev and Tittel [165] . 7.5 AFC 268 10 5 0 0 1000 2000 3000 Figure 7.9 The multimode scaling of a broadened Raman protocol. The blue solid line shows the multimode capacity calculated from the kernel (7.46) using the ‘Fourier method’, optimized over the width ∆0 of the applied broadening. The green dashed line shows the square-root scaling obtained if ∆0√is set equal to zero. We used a 2 Gaussian√control pulse, with Ω(τ ) = 10de−(τ /0.1) , and a detuning of ∆ = 90d. These parameters maintain the adiabatic condition that ∆ Ωmax , so that we remain in the Raman limit as the coupling is increased. We used a threshold efficiency of 70% for both calculations. The last memory protocol we consider is the AFC memory protocol proposed by Afzelius et al. in Geneva [91] , which is introduced in §2.3.4 of Chapter 2. 7.5 AFC In the AFC protocol, an ensemble with a naturally broad inhomogeneous absorption line is prepared by optical pumping, producing an atomic frequency comb. That is, atoms are removed (or pumped into a ‘shelf’ state) that have resonant frequencies lying between the ‘teeth’ of a spectral comb, so that the ensemble only absorbs light at the evenly spaced frequencies of the comb teeth. The great advantage of this approach, is that the broad spectral bandwidth covered by the ensemble absorption arises naturally. Adding more teeth to the comb to increase the absorption 7.5 AFC 269 bandwidth only requires that fewer atoms are removed, the density or size of the ensemble need not be increased. This should be contrasted with CRIB, in which broadening the absorption bandwidth requires an increase in the total optical depth, if the same level of absorption is to be maintained. As a result, the multimode capacity of AFC does not depend on the density of the ensemble, making it by far the most ‘multimode’ protocol yet proposed. We model AFC using precisely the same approach as we used for tCRIB. We assume that we have succeeded in preparing an ensemble so that it has an inhomogeneous absorption profile that takes the form of a series of M equally spaced resonances, covering a total spectral width ∆0 , each with optical depth d, p(∆) = M X δ(∆ − δj ), where δj = ∆0 j=1 Note that we have defined p so that R 1 j−1 − . M −1 2 (7.53) p(∆) d∆ = M . The total optical depth associ- ated with the entire frequency comb is then dtotal = M d. We reserve the designation d for the optical depth associated with a single tooth of the frequency comb, since this is set by the density of the ensemble. This definition allows direct comparison of AFC with the other protocols studied in this Chapter, where d quantifies the physical resources — density, length — required to build the memory. The lineshape function for AFC is a sum of Lorentzian lines, f (ω) = M X j=1 1 1 − i(δj + ω) (7.54) 7.5 AFC 270 (again, recall that the ‘1’ in the denominator represents the natural atomic linewidth in our normalized units). We construct the Green’s function for the AFC memory in the same way as we did for tCRIB. The only difference is that the inhomogeneous profile is not flipped around for retrieval: the discrete structure of the comb means that the phase of the atomic dipoles undergoes periodic revivals, without requiring any external meddling. We still consider phasematched retrieval in the backward direction, however, since this is the optimal situation. The equivalent expression to (7.37) for AFC is therefore e total (ω r , ω) = −e−iωT K Z d p(∆) d∆× r [Γ(∆) − iω ][Γ(∆) − iω] 2π Z 1 e−d(1−z r )[f (ω r )+f (ω)] dz r , 0 (7.55) where we used ∆r = ∆, pr = p and f r = f . Performing the integrals, and dropping the unnecessary phase factor, we find the Green’s function for AFC to be r r −d[f (ω )+f (ω)] − 1 e total (ω r , ω) = 1 f (ω ) − f (ω) × e K . 2π ωr − ω f (ω r ) + f (ω) (7.56) The first quotient on the right hand side exhibits a removable singularity at ω r = ω, but this can be dealt with using L’Hôpital’s rule: −2df (ω) − 1 e total (ω, ω) = 1 f 0 (ω) e , K 2π 2f (ω) where f 0 (ω) = M X j=1 iω . [1 − i(δj + ω)]2 (7.57) In Figure 7.10 we show the multimode scaling for AFC derived from the SVD of this kernel. In part (a) we show the multimode capacity N as function of the tooth optical 7.5 AFC 271 depth d, for various numbers of comb teeth. For each number of teeth M , the square root scaling characteristic of an unbroadened memory is apparent. But it is possible to increase the multimode capacity arbitrarily, just by adding more teeth to the comb. In principle, the multimode capacity of this memory protocol is infinite! Of course, there is a limitation in practice. First, the number of modes stored can never exceed the number of atoms in the ensemble. A more important restriction however comes from the spectral width of the initial inhomogeneous line. To achieve efficient retrieval, the teeth comprising the comb must be ‘well-separated’, so that they do not overlap in frequency. If they overlap, the re-phasing of the atomic dipoles will not be complete, and there will be only partial retrieval of the signal field. Because the lineshape function f is a sum of Lorentzians, which are functions that do not have compact support, a rather large frequency separation between the comb teeth is desirable. Therefore, as more teeth are added to the comb, the total width ∆0 of the comb must be increased, in order to accomodate the increased number of teeth with the same separation between them. Eventually, the width of the comb will approach the width of the initial inhomogeneous absorption profile from which the comb was prepared. More teeth cannot then be added without compromising the efficiency of the retrieval process. We found that the memory efficiency begins to suffer seriously if the finesse F falls below around 30, with F defined by F = ∆0 . 2(M − 1) (7.58) 7.5 AFC 272 The finesse is just the ratio of the tooth separation δj+1 − δj and the natural atomic linewidth 2 (the FWHM in normalized units). For our threshold of ηthreshold = 70%, a finesse of greater than 100 was required. In part (b) of Figure 7.10 we show the multimode scaling of AFC as a function of the total optical depth dtotal , alongside the scaling for CRIB and unbroadened protocols. For these protocols, dtotal = d; the plots are simply reproduced from Figure 7.8. This plot shows that, as we might expect, the multimode capacity of AFC does indeed scale linearly with dtotal . That is, N remains proportional to the total number of atoms available for the protocol. The advantage of AFC lies in the use of a ‘natural resource’ of atoms, so that increasing the number of atoms used in the protocol does not require a higher ensemble density. Note however, that even as a function of dtotal , the AFC protocol still outperforms CRIB, with a scaling of N ∼ 2d/25, and an optimal comb width of ∆opt 0 ∼ 5d. This concludes our investigation of multimode storage. The techniques used are quite general, and are readily applicable to new quantum memory protocols as they are invented. Experimental implementations of multimode storage will likely by challenging, but the technical advantages for applications of memory to quantum communication makes the research a worthwhile endeavour. Practically, it is hard to imagine how one might encode or detect photons in the optimal input modes φk (τ ). These modes have non-trivial shapes, which do not necessarily overlap well with the ‘natural’ modes of photon detectors. Generally these detectors, be they avalanche photodiodes (APDs), photomultiplier tubes 7.5 AFC 273 (PMTs) or superconducting bolometers (SSPDs), are engineered to have a broad, flat spectral response. This means they couple to a temporal mode that looks like a short, sharp, spike. The natural modes for single photon detection are therefore time-bin modes, where information is encoded in the time of arrival of a photon. These time-bin modes are not usually the same as the φk , so time-bin encoding is sub-optimal, and the multimode capacity predicted by the SVD cannot be reached. But the discrepancy between the optimal capacity N and the achieved capacity depends on the overlap between the subspace of signal profiles spanned by the timebin modes with the subspace spanned by the first N optimal modes. As N becomes large, this overlap grows, and the capacity for time-bin modes rapidly approaches the optimal capacity. Without attempting a detailed calculation, the plausibility of this claim can be appreciated by comparing parts (a) and (b) in Figure 7.5 in §7.3.2. Here, we estimated the Schmidt number by decomposing the Green’s function using non-ovelapping rectangular pulses as spatial modes. If we consider a Green’s function defined in the temporal domain, such pulses are precisely time-bin modes. And it is clear that in part (a), the multimode Green’s function is more faithfully reconstructed than its few-mode counterpart in part (b). This is quite general. As a Green’s function becomes multimode, it admits an increasingly fine-grained decomposition in terms of time-bin modes. This shows that the time-bin basis quickly becomes ‘just as good’ as the φk . Therefore the multimode capacity for time-bin modes fast approaches that calculated from the SVD. A similar argument applies if a frequency-bin encoding is used, or, for that matter, if any other orthonormal 7.5 AFC 274 basis is chosen for the signal alphabet. The multimode capacities calculated in this chapter can accordingly be characterized as a tight upper bound on the performance of the protocol concerned. In the next chapter, we wrap up our analysis of quantum memory by addressing the optimization of storage when the signal field is given, and cannot be ‘shaped’. 7.5 AFC 275 (a) (b) 40 150 20 30 100 14 20 8 50 10 2 0 0 0 100 200 300 400 500 0 100 200 300 400 500 Figure 7.10 The multimode scaling of the AFC memory protocol. (a): We show the multimode capacity of AFC for various numbers of comb teeth — indicated by the numbers in the plot — as a function of the optical depth d associated with a single comb tooth. For each number of comb teeth M , the capacity scales with the square root of d, just like an unbroadened protocol. But adding more teeth allows to increase N arbitrarily, so the multimode capacity is not limited by the optical depth. (b): Here we show the multimode capacity of AFC (green dashed line) as a function of the total optical depth dtotal = M d associated with the entire frequency comb. This shows that N scales linearly with the total number of atoms involved in the protocol, just as it does for the CRIB protocols. For comparison, the capacities of tCRIB, lCRIB and EIT/Raman protocols are also shown (red dotted line, black dot-dashed line and blue solid line, respectively). For these latter protocols dtotal = d, so these capacities are just taken from Figure 7.8. It is interesting that the capacity of AFC is larger than that of CRIB, even when evaluated as a function of dtotal . Of course, d is the more relevant physical resource for AFC, since this sets the ensemble density. We optimized the capacities plotted for AFC in both (a) and (b) over the spectral width ∆0 of the frequency comb. The threshold efficiency used was ηthreshold = 70%, as usual. Chapter 8 Optimizing the Control The SVD has proved an extremely useful tool for the analysis of quantum storage. Given a set of parameters, including the ensemble geometry and density, as well as the detuning and the control pulse profile, it is possible to construct a Green’s function that contains all the information we might want to know about the memory interaction. In particular, taking its SVD provides us with the optimal input mode, so that we can achieve efficient storage by shaping the signal field. But suppose that we are simply given a signal field, and asked to store it. This is probably the more likely situation in practical applications, where we have control over our memory, but not over the source that generates the signal field to be stored. In this case we are not able to shape the signal so that it matches the optimal mode φ1 . How do we achieve efficient storage in this situation? We must try to deform the optimal mode φ1 into the shape of our signal! This can be done by shaping the control field in order to sculpt the Green’s function; in this chapter we will explore how the correct 8.1 Adiabatic shaping 277 shape for the control may be found. 8.1 Adiabatic shaping The adiabatic limit is precisely the limit in which the natural atomic response is sufficiently fast that the atoms can ‘follow’ the driving fields. In this limit, the atomic response function is completely determined by the control field profile. Therefore the adiabatic limit is the regime we should work in, if our aim is to be able to affect the Green’s function by shaping the control. Fortunately, another consequence of adiabatic following is that the Green’s function may be completely parameterized by the integrated Rabi frequency ω, defined in (5.46) in §5.2.6 of Chapter 5. Later in Chapter 5 we derived an analytic solution for the adiabatic storage kernel of a Λ-type ensemble memory (5.77). It is extremely convenient that the optimal input mode found from the SVD of this kernel applies universally for all control profiles (within the adiabatic limit, of course). Given a control profile, it is easy to find the temporal shape of the optimal input mode, simply by converting back from the ω to τ . The conversion is given by (5.78). The optimization problem in this case is then simple. We must adjust the control field until the shape of the optimal temporal mode is the same as the shape of the signal field we intend to store. Changing the shape of the control pulse has no effect on the singular values, since the storage kernel depends only the control pulse energy through W . Therefore, when the shaping is complete, the same storage efficiency is achieved, as would have been if we had shaped the signal field and left the control fixed. 8.1 Adiabatic shaping 278 In Figure 8.1 we show some examples of this kind of optimization. The control field is parameterized by a vector Ω of 2N real numbers, giving the real and imaginary parts of Ω(τ ) at a set {τj } of N discrete points in time. We choose the points τj to lie on a Chebyshev grid, since this makes polynomial interpolation of the control field to other times numerically stable (see §E.1.2 in Appendix E). Starting with the initial guess Ω(τj ) = Ain (τj ), we optimize Ω using a simplex search algorithm — ‘fminsearch’ in Matlab — to minimize the norm of the difference Ain (τ ) − φ1 (τ ). At each iteration the optimal mode φ1 (τ ) is determined from Ω by interpolating to find the envelope Ω(τ ), and then using (5.78). Convergence does not take longer than 1 minute on a 3 GHz machine. After each optimization, we use the optimized control profile along with the numerical method described in §5.4 of Chapter 5 to construct the storage kernel without the adiabatic approximation. The resulting optimal mode is then shown for comparison with the signal profile to be stored. This provides a direct way to examine the accuracy of the optimizations. The method used relies on the validity of the adiabatic kernel (5.77), and it is discernible that the optimization performs better in the Raman limit than for EIT, since the adiabatic approximation is less robust on resonance (this is discussed in §5.2.9 of Chapter 5). As with so much of this thesis, an excellent account of closely related work can be found in the papers of Gorshkov et al. [133] , who studied this problem in the lightbiased limit of large control pulse energy, where the optimal kernel (5.27) is valid (see §5.2.8 in Chapter 5). They did not use the SVD, or the integrated Rabi frequency ω, so their approach is a little more convoluted, but nonetheless it is accurate and 8.2 Non-adiabatic shaping 279 ingenious. 8.2 Non-adiabatic shaping The opposite to the adiabatic limit is the limit of a very short, broadband control pulse. If the pulse is sufficiently short, there is no time for the coherence induced on the |2i ↔ |3i transition to couple to the optical polarization P connecting states |1i and |2i. In this limit, the shape of the control field makes no difference to the form of Green’s function describing absorption of the signal. Instead, the control simply induces Rabi oscillations between |2i and |3i that transfer population from the excited state to the storage state. The efficiency of this population transfer is given by sin2 (θ/2), where θ = 2 R Ω(τ ) dτ is the pulse area (note the difference between this dimensionless quantity and the integrated Rabi frequency ω, which is relevant in the adiabatic limit). Perfect transfer is achieved by an infinitely short π-pulse: this kind of instantaneous map is assumed in the CRIB and AFC protocols treated in the last Chapter, and in the derivation of the optimal kernel (5.27) in Chapter 5. If we use such a control pulse, the optimal signal input mode is fixed by the homogeneous and inhomogeneous lineshape of the ensemble. If the signal field mode does not coincide with this optimal mode, there is nothing we can do to optimize the storage efficiency. But there is a large middle-ground between the adiabatic regime and the extreme case of a π-pulse control. In this middle-ground, the analytic solution for the storage kernel (5.77) is of limited use, but the shape of the control field still has an effect, at 8.2 Non-adiabatic shaping 280 least to some degree, on the form of the storage Green’s function. Since the analytic solution breaks down, we must resort to numerical solutions of the equations of motion to optimize the control. The method we use is the most direct method one could imagine. Given a vector Ω parameterizing the control, we construct the full control profile Ω(τ ) by interpolation, and then we integrate the equations of motion numerically, and we extract the storage efficiency η. We then use a simplex search algorithm to optimize the elements of Ω so as to maximize η — using fminsearch in Matlab, we minimize −η. This approach works well, but relies on the ability to repeatedly solve the equations of motion quickly and accurately. Fortunately, this is possible using the method described in Appendix E, where we use Chebyshev spectral collocation for the spatial propagation, and RK2 for the time-stepping. To improve the convergence of the optimization, it helps to start with an initial guess that is already close to optimal. We can make use of the SVD in this connection. We start with an initial guess Ω, determined by setting Ω(τ ) = Ain (τ ). This is motivated by the fact that φ1 (τ ) = Ω(τ ) in the adiabatic limit, when the coupling is small (i.e. when d or C is small), so it is a reasonable opening gambit. Next we use the numerical method described in §5.2.9 of Chapter 5 to construct the resulting Green’s function, and we take its SVD to extract the optimal input mode φ1 (τ ). In general, since the coupling is not small, φ1 (τ ) 6= Ain (τ ) — this is why the optimization is necessary. By fitting a cubic spline interpolant to φ1 , and also to (k) Ain , we are able to build a sequence of n functions {φ1 } that represent a smooth (1) (n) deformation of this optimal mode into Ain . That is, φ1 = φ1 , and φ1 = Ain , with 8.2 Non-adiabatic shaping 281 the intervening functions lying ‘between’ these two. This is easily done in Matlab by (k) building the φ1 using splines with coefficients found by interpolating between the coefficients for φ1 and Ain . Now, Ω already describes the optimal control for storing (1) the mode φ1 , by construction. We now use the optimization method described (2) in the previous paragraph to optimize Ω for storing the mode φ1 . The hope is (1) that this is sufficiently ‘close’ to φ1 that the optimization will converge quickly. We then repeat the optimization, this time using the newly optimized vector Ω as (3) our initial guess, and using the mode φ1 as the target input mode. The pattern is now clear. We iterate these steps, each time switching the target input mode (k) from φ1 (k+1) to φ1 , and using the previous result for Ω as an initial guess. As we procede, the target modes approach Ain , and Ω approaches the vector describing the optimal control for storing Ain . The rationale behind this approach is that the control remains near-optimal at all times. We begin with an optimal control for the ‘wrong’ target, and by degrees we deform this target into Ain , all the time ‘bringing along’ the control. We found that this method is helpful in the most non-adiabatic situations, where the initial guess Ω(τ ) = Ain (τ ) is particularly poor. When the dynamics are more adiabatic, optimizing for Ain directly in a single step is often sufficient. In Figure 8.2 we show some examples of this type of optimization. Clearly the numerical optimization works well where the analytic optimization described in the previous section fails. The method is not particularly time consuming on a modern computer, although it is of course slower than the analytic optimization. It is quite 8.2 Non-adiabatic shaping 282 general though, and it provides an easy way to optimize the storage efficiency for a given input, and also to check the final proximity of the optimal input mode to the desired shape. Gorshkov et al. have also studied the optimization of the control field outside the adiabatic limit, in the fourth of their series of papers on the subject [166] . They use an interesting approach involving gradient ascent, in which they derive an explicit formula for the incremental change in the control profile that will improve the storage efficiency. Numerical solution of the equations of motion provides them with this incremental change, and by iterating they are able to generate the optimal control for storage. This is a robust and efficient optimization that produces very similar results to the method we describe here. They comment that they are not able to verify the optimality of their results, and this is true for our optimizations too. But in both cases, the SVD of the numerically constructed Green’s function does make it possible to check the resemblance between the desired input mode, and the optimal mode resulting from the control. When these two modes match up, as they do in the cases shown in Figure 8.2, it is clear that the optimization has converged. When these modes do not match, of course, it is not obvious whether the result is a global or simply a local optimum. This completes our theoretical treatment of optimal storage and retrieval from ensemble memories. In the next Chapter we present a derivation of the equations of motion for the memory interaction in diamond, since this is one of the media we have used in our experiments, and the theory of Chapter 4 is not directly applicable. 8.2 Non-adiabatic shaping 283 Finally, in Chapter 10 we review the status of our experiments, before concluding the thesis in Chapter 11. 1 Broadband Raman (a) Broadband EIT 8 6 (b) 6 0.5 4 4 2 2 0 0 −0.2 0 0 −0.2 0.2 1 0 0.2 Non-adiabatic Narrowband EIT 8 (c) (d) 6 6 0.5 0 −2 0 2 4 4 2 2 0 Phase Intensity (arb. units) 284 Phase Intensity (arb. units) 8.2 Non-adiabatic shaping 0 −2 0 2 Figure 8.1 Adiabatic control shaping. Given an input signal profile, as well as the available control pulse energy W , optical depth d and detuning ∆ we construct the adiabatic kernel (5.77), and extract the optimal input mode, as a function of the integrated Rabi frequency ω. We then optimize the control profile Ω(τ ) until the temporal profile of the optimal mode matches the given signal. In parts (a) 2 to (d), the signal is assumed to be a Gaussian, Ain (τ ) = e−[(τ −τs )/Ts ] . Its intensity profile |Ain (τ )|2 is shown as a dashed black line. The timing of the signal pulse τs is arbitrary, and in each case we choose it so as to aid the convergence of the optimization. The temporal intensity profile |Ω(τ )|2 of the optimized control is represented by the blue solid line, with its temporal phase shown by the red solid line (referred to the axes on the right-hand side). We used N = 21 Chebyshev points to parameterize the control in all cases. The green dotted lines show the temporal intensity profiles of the optimal input mode determined from the optimized control using numerical integration to construct the Green’s function. These should coincide with the signal mode, if the optimization has succeeded. Part (a) shows the result for an optimization with W = 2.5, ∆ = 150 and d = 300, and a signal duration of Ts = 0.1, all in normalized units. This describes off-resonant Raman storage. The adiabatic approximation is well satisfied, and the optimization performs well. In part (b) we show the result for the same optimization when the detuning is set to zero, which describes broadband EIT. Here the adiabatic approximation is not well satisfied, and the optimization performs rather poorly. In part (c) we optimize the control for a narrowband signal with Ts = 1 (note the difference in time scale on the horizontal axis). The optimization performs much better. Finally in part (d) we reduce the optical depth to d = 10, and we increase the control energy to W = 49. The detuning is set to 0, and the signal duration remains Ts = 1. Despite the narrow signal bandwidth, the adiabatic approximation that worked well for part (c) is now ‘broken’, because of the large Rabi frequency, and the optimization fails entirely. The storage efficiencies achieved in parts (a) to (d) were 94%, 74%, 94% and 54%, respectively. For comparison, the efficiencies that would have been achieved if the signal field were exactly equal to φ1 (τ ) were 97%, 99%, 96% and 82%. 8.2 Non-adiabatic shaping 285 Broadband EIT 10 (a) 6 (b) 4 0.5 5 2 0 −0.2 0 0.2 0 −0.2 1 0 0.2 0 Non-adiabatic Narrowband EIT 6 (c) (d) 10 4 0.5 5 2 0 −2 0 2 0 −2 0 2 Phase Intensity (arb. units) Phase Intensity (arb. units) Broadband Raman 1 0 Figure 8.2 Non-adiabatic control shaping. Parts (a) to (d) show the results of the direct numerical optimization described above, where n = 10 steps were used in deforming the initial target Ain (τ ) = 2 (1) φ1 (τ ) into the final target Ain (τ ) = e−[(τ −τs )/Ts ] . The optimizations each ran in around 2 minutes on a 3 GHz machine. The parameters are identical to those used in parts (a) to (d) of Figure 8.1. The numerical optimization copes well with non-adiabatic dynamics, and in all cases comparison of the target signal mode with the optimized input mode shows that the optimizations have met with some success. In part (b), the broad bandwidth of the signal makes the adiabatic approximation poor, and it is noticeable that the optimized control profile features a small oscillation which is not present in the control in part (c). The adiabatic approximation fails entirely in part (d), and here the control involves a large oscillation, with its energy distributed into two ‘pulses’. These oscillations are typical of non-adiabatic shaping, since the ‘ringing’ of the atomic dynamics must be compensated by the control to produce a smooth input mode. The storage efficiencies achieved in parts (a) to (d) were 97%, 99.5%, 97% and 81%, respectively. For comparison, the efficiencies that would have been achieved if the signal field were exactly equal to φ1 (τ ) were 97%, 99.6%, 97% and 82%. Chapter 9 Diamond In this chapter we build a theoretical description of quantum memory in diamond. The end result is a set of equations with precisely the same form as (5.107) describing Raman storage in a vapour, so that our analysis of optimal storage and retrieval applies in diamond just as it does to atomic vapours (see §5.3.3 in Chapter 5). But some justification of this claim is required, and so we present a derivation below. 9.1 Diamond Scheme Diamond is a singular material, both physically and aesthetically. It is the hardest and most transparent mineral, with the highest thermal conductivity of any material, and a very large refractive index (around 2.4). These properties arise in part from its extremely simple structure. It is comprised entirely of carbon atoms; each is connected to four others by strong covalent bonds. The bonds are all equivalent, and this symmetry produces a tetrahedral arrangement of atoms that is exceedingly 9.1 Diamond Scheme 287 robust. The diamond structure can be visualized by convolving a basis of two carbon atoms with a face-centred cubic bravais lattice [167] , as shown in Figure 9.1. Figure 9.1 The crystal structure of diamond. The diamond lattice is FCC (face-centred cubic); here we show just a single unit cell, outlined for clarity with thin ‘rods’. Each basis is shown as a pair of atoms connected by a thick ‘tube’, one atom is coloured red; the other grey. The grey atom of each basis is located at the sites of the FCC lattice. Each atom is connected by four covalent bonds (not shown) to its neighbours, forming a tetrahedral pattern. If the basis is deformed, the bonds produce a large restoring force, and so in response to an impulse the atoms of the basis can undergo harmonic oscillations relative to one another. The frequency of these ‘internal’ vibrations is rather high, because the interatomic bonds are very ‘stiff’. It is these high frequency vibrational modes we seek to excite when implementing a quantum memory in diamond. Eventually any relative motion within a basis couples to collective motion of the basis with respect to its neighbouring bases, and the energy is dissipated as 9.2 Quantization 288 waves of motion of the lattice sites with respect to one another: sound waves. This process limits the lifetime of a diamond quantum memory, and in fact the lifetime — on the order of picoseconds — is much too short to make diamond a useful medium for quantum storage. Nonetheless, it is possible to store very broadband pulses in diamond, because the oscillation frequency is so high, and a solid-state, room-temperature, broadband quantum memory is interesting in its own right. 9.2 Quantization Just as the electromagnetic field can be quantized, revealing photons as the constituents of light, so the harmonic oscillations of a crystal can be quantized. The quanta of crystal vibrations are known as phonons. Our aim in the present context is to describe the coherent mapping of a single photon to a single phonon in the diamond crystal. The crystal vibrations may be quantized by imposing periodic boundary conditions on the atomic displacements within a cubic crystal of side length L. To illustrate this procedure we consider a one dimensional wave b(z) = eikz describing the atomic displacement at position z. Periodic boundary conditions require that b(0) = b(L), so that we must have kL = 2πm for some integer m. Therefore the momenta associated with crystal vibrations are quantized, with k = 2πm/L. The quanta are phonons. The smallest non-zero wavevector allowed is found by setting m = 1, whence we obtain the mode separation δk = 2π/L. There is also an upper limit to the allowed 9.2 Quantization 289 wavevectors. The wavelength of any vibration cannot be meaningfully defined if it falls below 2a, where a is the lattice constant — the distance separating neighbouring atoms. The wave is only ‘sampled’ at the atomic positions, and the spatial frequency of the waves cannot exceed the sampling rate. The maximum wavevector, set by this coarseness of the crystal structure, is ∆k = π/a. As shown in Figure 9.2 below, any wave with a larger wavevector k is physically indistinguishable from a wave with momentum k − 2∆k, and so we only consider phonons with wavevectors lying within the range [−∆k, ∆k]. This region in k-space is known as the first Brillouin zone, or just the Brillouin zone. That all physically distinct phonon modes are contained within the Brillouin zone can be confirmed by counting them: the number of modes in the Brillouin zone is 2∆k/δk = L/a = N , where N is the number of atoms in the crystal. So the number of phonons is equal to the number of atoms. This must be true in one dimension, since each atom has one vibrational degree of freedom. The generalization to three dimensional vibrations turns the Brillouin zone into a three dimensional volume in k-space, often with a non-trivial shape. But this is not important for us. As in most of the rest of this thesis, we will use a one dimensional model to describe the quantum memory. The lattice constant in diamond is around 3.6 Å, so the Brillouin zone boundary lies at ∆k ∼ 1010 m−1 . By comparison, the wavevector associated with visible light at around 500 nm is about 107 m−1 . Any momentum imparted to the crystal by interaction with optical fields is therefore very small, on the scale set by the crystal lattice. The excitations produced by a quantum memory may safely be considered to 9.3 Acoustic and Optical Phonons 290 lie at the zone centre, with δk ∼ 0. This, of course, greatly simplifies our theoretical displacement description. Figure 9.2 Phonon aliasing. A crystal vibration with a wavelength smaller than 2a is physically equivalent to one with a longer wavelength. In the example shown, the blue solid line represents the profile of a short wavelength vibration, and the dashed blue line shows the profile of the equivalent longer wavelength vibration. Note that the atomic positions, indicated by the red circles, are identical for both waves. The black circles show the equilibrium positions of the atoms, equally spaced by the lattice constant a. 9.3 Acoustic and Optical Phonons Phonons come in two varieties, as we have already hinted. Acoustic phonons are the quanta of sound waves in a crystal. They represent compression and rarefaction within the crystal lattice, and the energy associated with this lattice distortion clearly vanishes as its wavelength becomes large, since in the limit of infinite wavelength there is no distortion, and therefore no restoring force. As mentioned above, optical wavelengths are already much larger than a unit cell, so the energies of optically accessible acoustic phonons are very small. A typical dispersion relation for acoustic phonons is shown in part (a) of Figure 9.3. The energy of acoustic phonons rises linearly with their wavevector k for k ∆k. Scattering from low-energy zone- 9.3 Acoustic and Optical Phonons 291 centre acoustic phonons is conventionally known as Brillouin scattering, but it is not of interest to us here. We will focus instead on Raman scattering, which in crystals refers to the scattering of light from so-called optical phonons. These represent the second variety of phonon; they are the quanta associated with the high-frequency internal vibrations of the crystal bases. The bases oscillate essentially independently of their neighbours, provided the wavelength of the oscillation is not so small that their neighbours are ‘pulling’ on them. In the limit of infinite wavelength, the optical phonon energy does not vanish, but is set by the frequency associated with the normal modes describing the natural internal oscillations of the basis. A typical dispersion relation for optical phonons is shown in part (a) of Figure 9.3. The non-vanishing energy of optical phonons at the zone-centre distinguishes them from acoustic phonons. The Raman interaction in diamond couples an incident signal field to these zone-centre optical phonons, and it is these phonons that play the role of the storage state in a diamond quantum memory. Broadly speaking, these phonons are like the metastable state |3i used for storage in the atomic systems discussed previously. 9.3.1 Decay In many crystals, the basis is composed of unlike atoms or ions, and so the optical phonons are associated with an electric dipole moment. This means they couple strongly to the electromagnetic field — that is why they are given the designation ‘optical’. However in diamond, which is homopolar, with both atoms in its basis 9.3 Acoustic and Optical Phonons (a) 292 (b) Optical branch Acoustic branch Figure 9.3 Phonon dispersion. (a): typical dispersion curves for acoustic and optical phonons. The former have negligible energy near the zone-centre, which is the region to which we have access with optical fields. The latter have large energies, and a flat dispersion relation close to the zone-centre. It is the zone-centre optical phonons that we use to store an incident photon. (b): The decay of optical phonons is dominated by the Klemens channel, in which anharmonic coupling allows a zone-centre optical phonon (black dot) to decay into a pair of acoustic phonons with large, opposite momenta (red dots). We only show the dispersion in the first Brillouin zone. The dotted lines on the left and right hand sides indicate the boundaries of this zone, which are identified, meaning that any we could wrap the plots around a cylinder and stitch these two lines together: they represent physically equivalent momenta. identical, there is no separation of charge associated with internal oscillations. The optical phonons in diamond are therefore not, in fact, optically active. That is, they do not directly radiate or absorb electromagnetic radiation. This is advantageous for quantum memory, since the optical phonons in diamond are accordingly longer-lived than in many other materials. As touched upon above, the dominant decoherence process for these phonons is the decay into acoustic phonons via anharmonic couplings: the covalent bonds do not behave like perfect springs, and their deviation from Hooke’s law allows the optical and acoustic phonons to exchange energy. To conserve momentum the acoustic phonons are produced in pairs with approximately opposite momenta, as illustrated in part (b) of Figure 9.3. This process is known 9.3 Acoustic and Optical Phonons 293 as the Klemens channel [168,169] . The anharmonicities that give rise to the Klemens channel are largely geometrical in origin, and the lifetime of optical phonons is only weakly affected by temperature [170] . The purity and quality of the crystal also contribute, but all diamonds have an optical phonon lifetime1 τp . 10 ps. 9.3.2 Energy There are three optical phonons in diamond, corresponding to basis oscillations in three orthogonal directions. The three phonons are degenerate at the zone centre, because diamond is symmetric with respect to the interchange of these three directions. With a one dimensional interaction we will excite just a single phonon mode; the degeneracy of the modes means we need not worry about which mode this is. The optical phonon energy at zone centre is Ep = 0.17 eV, which corresponds to a wavenumber of νp = 1332 cm−1 , or an angular frequency of ωp = 2.5 × 1014 s−1 . This would correspond to an infra-red wavelength of λp = 7.5 µm. The large phonon energy in diamond is advantageous for the following reasons: First, the energy scale kB T associated with room temperature (T ∼ 300 K) is around 1/40 eV, which is much smaller than the phonon energy. Therefore there are very few thermally excited optical phonons at room temperature: using the Boltzmann formula pthermal = e−Ep /kB T we predict a population of around 1.7×10−3 thermal phonons per mode. Therefore demonstrating quantum memory at room 1 These phonon lifetimes were studied in our research group by Felix Waldermann, and later by K.C. Lee and Ben Sussman, using a technique they named TCUPS [171] . In these experiments, a pair of delayed pump pulses directed through a diamond crystal produces a corresponding pair of Stokes pulses. The phase coherence, as measured from the visibility of spectral interference, between the two Stokes pulses directly measures the coherence of the optical phonons. 9.4 Raman interaction 294 temperature in diamond is feasible. Second, the bandwidth of the stored signal field cannot exceed the phonon frequency, since ωp sets the frequency difference between the signal and control fields, and these should not overlap spectrally. Since the phonon frequency is so large, the signal bandwidth can be large, meaning that a short pulse can be stored. Taking τs ∼ 1/ωp as a rough estimate of the shortest signal pulse duration that can be stored, we find τs ∼ 1000τp . If the ratio of the shortest storable pulse duration to the maximum storage time τs /τp is taken as a figure of merit for a memory, a diamond quantum memory is actually rather impressive! 9.4 9.4.1 Raman interaction Excitons The Raman interaction in diamond involves an intermediate state, just as it does in the atomic case considered in earlier chapters. The optical fields are detuned from resonance with this state, but the interaction nonetheless requires strong coupling to this state to mediate the storage of the signal field. The relevant intermediate state in diamond is an exciton. To understand what this is, recall that the electronic orbitals in an extended crystal arrange themselves into disjoint bands. Electronic band structure arises from Bragg scattering of electrons from the periodic potential associated with the regular lattice of atomic nuclei in the crystal. As the De Broglie wavelength of an electron approaches 2a, the reflected and transmitted components 9.4 Raman interaction 295 of the electronic wavefunctions interfere destructively, leading to the appearance of forbidden energy bands, containing no allowed electronic states, at the edges of the Brillouin zone. This is illustrated in Figure 9.4. (a) Free (b) Periodic (c) Bands Figure 9.4 Band structure. (a): the dispersion relation of a free electron is parabolic, since its kinetic energy E = p2 /2m is quadratic in k. (b): electrons in a periodic lattice must have a periodic dispersion relation. To a first approximation, this is found by simply adding ‘copies’ of the free electron dispersion relation at intervals of 2∆k = 2π/a in k-space. (c): scattering of the electrons from the periodic potential produced by the atomic cores results in anti-crossings in the dispersion relations, leaving energy gaps, shown shaded in gray. Here we have removed any electronic states lying outside the first Brillouin zone (red shaded area), since they are not physically meaningful. An exciton is produced when an electron in the lower energy band, the valence band, is promoted into the upper energy band, the conduction band, leaving behind a hole. The hole in the valence band simply represents the absence of an electron, but it is convenient to think of it as a particle in its own right, with positive charge, negative energy and negative mass. An exciton is the combined system of an electron in the conduction band and a hole in the valence band (see Figure 9.5). In fact, since these two particles have opposite charge, they attract one another, and it is possible for them to form a bound system, similar to a hydrogen atom, or the positronium 9.4 Raman interaction 296 of particle physics. The binding energy is rather small, however, and here we will treat the electron and hole as if they were free particles2 . Conduction band Valence band Figure 9.5 An exciton. A photon (not shown) promotes an electron (blue dot) from the valence band into the conduction band, leaving behind a positively charged hole (red dot). Note that the curvature of the valence band for the hole has been reversed, since the hole has negative energy. In this picture, the electron and hole have approximately equal and opposite momenta, so that the total momentum of the exciton is small, consistent with the small momentum of the photon. 9.4.2 Deformation Potential An incident photon can produce an exciton if it has sufficient energy to breach the band gap. This process provides the coupling between the optical fields and the diamond. To excite an optical phonon, there should be some coupling between excitons and phonons. In polar crystals, there are direct couplings between the dipole fields of excitons and phonons, in the form of the Fröhlich and piezoelectric interactions [173–176] . In diamond, these long-range electric couplings are absent, but 2 A precise characterization of the Raman cross-section in diamond does require an account of bound excitons [172] , but our aim is simply to study the feasibility of using this interaction for quantum storage, so we sacrifice rigour for simplicity 9.4 Raman interaction 297 there remains a short range coupling known as the deformation potential. The origin of this interaction can be understood as follows. The diamond structure takes the form of two interpenetrating FCC lattices, offset from one another. A zone centre optical phonon — with infinite wavelength — may then be interpreted as the rigid displacement of one sub-lattice with respect to the other. At any instant, the crystal structure is accordingly deformed, much as if it were subject to an external strain, and the electronic band structure, which depends on electron scattering from the crystal potential, is altered. Therefore electronic energy levels are coupled to crystal vibrations. More specifically, excitons are coupled to optical phonons. Optical phonons are not energetic enough to create or destroy an exciton outright, but an exciton with some momentum k can scatter from the deformation potential to produce an exciton with a momentum k 0 , and an optical phonon with momentum κ. Phasematching of this process over the length of the crystal ensures that momentum is conserved, with κ = k − k 0 . We now have to hand all the ingredients necessary to unpick the Raman interaction in diamond, which is shown in Figure 9.6, 1. A signal photon produces a virtual exciton — ‘virtual’ because the energy of the signal photon is smaller than the band gap. This is analogous to the virtual, or dressed state to which the signal couples in atomic systems, when it is detuned from the excited state |2i. 2. The virtual exciton scatters from the deformation potential to produce an exciton with a different momentum, and also a zone-centre optical phonon. 9.4 Raman interaction 298 3. The remaining virtual exciton recombines — the electron decays back into the valence band, filling the hole — and emits a control photon. From this description, it’s clear that the Raman interaction in diamond is third order, rather than second order as it is in the atomic systems considered in earlier Chapters. In principle this makes it weaker than in atomic systems, but the extremely high density of electrons in the solid state more than makes up for the extra perturbative order, and the Raman cross section in diamond is, in fact, extremely large. (a) (b) Figure 9.6 The Raman interaction in diamond. (a): An energy level diagram for the Raman quantum memory interaction. The relevant states are written in the form |n, mi, where n is the number of excitons involved, and m is the number of optical phonons. A signal photon (blue wavy arrow) produces a virtual exciton (indicated by the dotted line, detuned from the real exciton state). The deformation potential interaction then produces another virtual exciton, and an optical phonon (orange arrow). Finally, the virtual exciton recombines, emitting a control photon (green arrow) and leaving a single optical phonon behind. (b): A Feynman diagram for the same process. Here the dotted lines indicate the world lines of the virtual electron and hole comprising the intermediate exciton. The optical phonon is indicated by the orange spiral. 9.5 Propagation in Diamond 9.5 299 Propagation in Diamond It is advantageous to describe both the optical fields and the excitations of the diamond crystal in the Heisenberg picture, in order to treat propagation. This was done for the atomic case in Chapter 4 by studying the dynamics of the flip operators σjk describing the atomic evolution. The analysis was greatly simplified by considering just three discrete atomic states. In diamond however the electronic states form a quasi-continuum in each energy band. It is therefore not immediately obvious how the approach of Chapter 4 carries over to the present case. In addition to this issue, there is the problem of how to describe the local dynamics of the crystal excitations. To examine the spatial distribution of these excitations, we would like to obtain equations for the phonon or exciton amplitudes at some position z within the crystal. In the atomic case the interaction was entirely local, since each atom scattered light at a point, and independently of all other atoms. The local dynamics was therefore determined by the Hamiltonian of a single atom, and propagation was treated by summing these local contributions. The situation in diamond is conceptually different. First, phonons are global excitations of the crystal. Second, the electrons in a crystal are not localized around the atomic cores; rather they form a quasi-free Fermi gas distributed over the volume of the crystal. The above problems are essentially cosmetic, as we’ll see below. Our general strategy is as follows. To work in the Heisenberg picture, we write the crystal 9.5 Propagation in Diamond 300 Hamiltonian H in second-quantized form, H= X Hαβ |αihβ|, (9.1) α,β where in the sum, both α and β run over a complete set of states, and where the Hαβ = hα|H|βi are the matrix elements of the Hamiltonian connecting these states. The dynamics are then determined from the Heisenberg equations of motion for the flip operators |αihβ|. This approach broadly mirrors that used in Chapter 4 (see §4.3). To extract the spatial variation of the crystal excitations we seek to express the crystal Hamiltonian in the form 1 H= L Z L H(z) dz, (9.2) 0 where H is an effective local Hamiltonian. It turns out that this representation of the Hamiltonian emerges naturally from the periodic structure of the crystal lattice. 9.5.1 Hamiltonian The Hamiltonian for a diamond quantum memory is comprised of three parts, H = HER + HEL + H0 . (9.3) 9.5 Propagation in Diamond 301 The first two contributions represent the interaction of the electrons in the diamond with the radiation field and with the lattice respectively. The last part accounts for the energy of the excitations in the diamond. We neglect the Hamiltonian HL of the free radiation field, as we did in §4.4 in Chapter 4, since it plays no part in the equations of motion. 9.5.2 Electron-radiation interaction The Hamiltonian HER is simply the A.p interaction introduced in §C.4 in Appendix C. This form of the light-matter interaction is more appropriate than the E.d electric dipole Hamiltonian, because the electrons in diamond are not localized around the atomic cores. Instead, they are spread over the entire volume of the crystal, in so called Bloch waves3 . Signal and control fields We divide the vector potential into two parts; the weak signal field and the strong classical control, A = As + Ac . (9.4) Ac (t, z) = vc Ac (t, z)eiωc (t−nz/c) + c.c., (9.5) The control field is written as 3 The vector potential is not actually a physical field, and strictly we should apply the PZW transformation (C.20) to the Bloch states, in order that our treatment is gauge invariant [177,178] . However, this transformation simply shifts the electron momentum (c.f. (C.23)), and has no effect on the transition matrix elements. In any case, in the Heisenberg picture we are free to choose a gauge such that A(t = 0) = 0, whence the PZW transformation becomes trivial [177] . 9.5 Propagation in Diamond 302 where we have included the factor n = 2.417 in the exponent, which is the refractive index of diamond. We treat the signal field quantum mechanically, so it is written in second-quantized notation as Z As (z) = vs g(ω) a(ω)e−iωnz/c dω + h.c., ω (9.6) where we have used the one dimensional formula (C.10) along with (C.13) from §C.4 in Appendix C. The mode amplitude g(ω) = p ~ω/4π0 nAc includes the refractive index. Just as we did in (4.6) in Chapter 4, we anticipate the compact spectral support of the signal pulse about its carrier frequency ωs by pulling the mode amplitude g(ω)/ω out of the integral and defining a slowly varying envelope operator A, As (t, z) = Here gs = √ vs gs A(t, z)eiωs (t−nz/c) + h.c.. ωs (9.7) 2πg(ωs ). We have also introduced the time dependence of the operators arising in the Heisenberg picture (see Appendix B). Electron wavefunctions It is sufficient to consider the interaction of the optical fields with just a single active electron. This is because the complicated many-body physics governing the behaviour of all the electrons in the crystal can be swept under the rug of Fermi-Dirac statistics: the Pauli-exclusion principle prevents more than one electron from occupying each state, and since all valence band states in the crystal are initially occupied, any electron-electron scattering in this band is ‘frozen 9.5 Propagation in Diamond 303 out’ because no electron can change its state without occupying a previously filled orbital. The upshot of this is that we may consider p to be the momentum operator for a single electron. The wavefunction ψk,n (r) of an electron with wavevector k in the nth energy band is given by the product of a spatial phase factor with a periodic Bloch function, ψk,n (r) = eikz uk,n (r). (9.8) The Bloch functions have the same translational symmetry as the crystal lattice, uk,n (z + a) = uk,n (z), which is a consequence of Floquet’s theorem. Note that such states are not exactly eigenstates of the momentum operator; this is why the A.p interaction can induce electronic transitions. Matrix elements We are interested in the matrix element hα|HER |βi connecting two quantum states. Neglect for the moment the state of the signal mode. This matrix element is then given by the spatial overlap of the electronic orbitals describing the initial and final states, with the operator A.p inserted between the two orbitals: e hα|HER |βi = − m Z ψα∗ (r)A(z).pψβ (r) d3 r. (9.9) crystal Here the indices α, β are standing in for the wavevectors and band indices of the initial and final orbitals. Now, the coordinate representation of the momentum operator is p = −i∇. Applying this to ψβ (r), and using (9.8), we can write the 9.5 Propagation in Diamond 304 matrix element as e hα|HER |βi = − m Z ei(kβ −kα )z A(z). [u∗α (r)(p + kβ )uβ (r)] d3 r. (9.10) crystal Since all the momenta are very close to the zone-centre, the spatial variation of the exponential factors in the integrand is very slow, and this is also true for any variation of the optical field A(z) (which contains only slowly varying envelopes or similarly long-wavelength exponential factors). On the other hand, any rapid spatial variation of the Bloch functions uα,β is periodic, repeated in every unit cell. We therefore factorize the integral into two parts, as follows, e X i(kβ −kα )zj hα|HER |βi = − A(zj ). e m Z j u∗α (r)(p 3 + kβ )uβ (r) d r , (9.11) unit cell where zj is the position of the j th unit cell. We define the matrix element pαβ as N times the overlap integral inside the square brackets, where N = AL/a3 is the total number of unit cells in the crystal. Taking the continuum limit for the sum over the zj , we can write e hα|HER |βi = − pαβ . mL Z L A(z)ei(kβ −kα )z dz. (9.12) 0 We have now succeeded in separating out the local from the bulk dynamics. It only remains for us to introduce the flip operators |αihβ| in a convenient form. 9.5 Propagation in Diamond Excitons 305 The momentum operator has negative parity, just as the atomic dipole operator djk does (see §4.3.1 of Chapter 4). Therefore pαα = 0, so there is no coupling of any state to itself. As mentioned above, if all electrons are in the valence band, an electron has ‘nowhere to go’, since all the valence band states are occupied. The only possibility is to promote an electron into the empty conduction band, creating an exciton. This process, and its time-reverse — exciton recombination — are the only important scattering processes involved in HER . Once an exciton has been created, it is of course possible for either the conduction band electron or one of the valence electrons to undergo scattering (this latter process is the same as scattering of the hole) within their respective bands via the A.p interaction. But the energies involved are much smaller than the photon energies in the signal or control fields, so these processes do not conserve energy, and may be neglected. The Hamiltonian can therefore be written as HER (t) = − e mL Z 0 L A(t, z). X ν,k pνk,0 s†νk eikz + h.c. dz, (9.13) where the subscript 0 on p denotes the crystal ground state, and where s†νk creates an exciton with momentum k and energy ων . In the notation of (9.12), k = kβ −kα . The Hermitian conjugate component destroys an exciton, representing recombination. The energy of an exciton does not depend on its wavevector k, because the electron and hole comprising the exciton might have large but opposite momenta, giving a small total momentum, but a large total energy. Therefore the exciton state 9.5 Propagation in Diamond 306 is independently parameterized by the two quantities ν and k. Excitons, being composed of pairs of fermions (a hole is a quasiparticle obeying fermionic statistics), are bosons. The annihilation operators sνk therefore satisfy the same commutation relation as photon mode annihilation operators (see Appendix C), h i sνk , s†µk0 = δν,µ δk,k0 . (9.14) A merciful simplification is achieved by neglecting any dependence of the matrix elements pνk,0 on ν or k, since the dependence of the Bloch functions describing the electron and hole on wavevector is rather weak. We thus write pνk,0 = ip, where p is the constant magnitude of the matrix element, and where the factor of i appears because the momentum operator is purely imaginary. With these simplifications, the Hamiltonian can now be written in the form of (9.2), with the effective local Hamiltonian given by X † ie sνk eikz − h.c. . HER (t, z) = − p.A(t, z) m (9.15) ν,k As a final step, we can perform the sum over momenta in (9.15) explicitly. We define the local exciton operator 1 X Sν (z) = √ sνk e−ikz , L k (9.16) 9.5 Propagation in Diamond 307 which satisfies the commutation relation h i Sν (z), Sµ† (z 0 ) = δν,µ δ(z − z 0 ). (9.17) The local Hamiltonian then takes the form " # √ X ie L † p.A(t, z) HER (t, z) = − Sν (z) − h.c. . m ν 9.5.3 (9.18) Electron-lattice interaction The Hamiltonian HEL for the electron lattice interaction is just given by HEL = V |with phonon − V |no phonon , (9.19) where V is the potential experienced by an electron, generated by all the atomic cores and other electrons. Let the displacement between the sub-lattices of diamond caused by a zone-centre optical phonon by given by u. If this displacement is small, a Taylor expansion to first order gives HEL = ∂V .u. ∂u (9.20) When we quantize the lattice vibrations, the leading factor becomes the deformation potential matrix elements, or just the ‘deformation potentials’ [174,179,180] , and the second factor becomes the operator for the optical phonon amplitude. For vibrations 9.5 Propagation in Diamond 308 along the direction vp with wavevector κ the phonon amplitude operator can be written [180,181] uκ = gκ vp eiκz (b†κ + b−κ ). (9.21) Here b†κ creates an optical phonon with momentum κ. Phonons are bosons, so that we have (see (C.8) in Appendix C) [bκ , b†κ0 ] = δκ,κ0 . (9.22) The phonon mode amplitude gκ , which has the dimensions of length (it is the lattice displacement due to a single phonon), is given by r gκ = ~ , N M ωκ (9.23) where M is the mass of a carbon atom and ωκ is the phonon frequency. For reasons discussed in the previous section, the only states where electronic scattering can occur are the exciton states, with an electron in the conduction band and hole in the valence band. The action of the deformation potential is therefore to destroy an exciton with energy and momentum (ν, k), and to produce a new exciton with modified parameters (ν 0 , k 0 ). Summing over all possibilities gives the expression HEL = X κ,µ,ν,k,k0 Z 0 L 1 0 vp .Dνµkk0 κ s†µk0 sνk ei(k −k)z × gκ eiκz (b†κ + b−κ ) aL dz. (9.24) The factor of 1/a appears to give the deformation potentials D the dimensions of 9.5 Propagation in Diamond 309 energy. The integral over space, along with the factor of 1/L, arises in precisely the same way as it did in (9.12) above. A dramatic simplification of (9.24) is possible, since the deformation potentials D depend only very weakly on the phonon and exciton momenta so close to the zone centre. By the same token, the phonon frequency ωκ = ωp is independent of κ close to the zone centre (see Figure 9.3). Dropping these dependencies, we can perform the summations over k, k 0 and κ to obtain the following expression for the effective local Hamiltonian, HEL (z) = h i L3/2 g X Dµν Sµ† (z)Sν (z) B † (z) + B(z) , a µ,ν (9.25) where Dµν = Dνµ = vp .Dµν is the real magnitude of the deformation potential connecting excitons with energies ωµ and ων , and where we have defined the local phonon operator 1 X B(z) = √ bκ e−iκz , L κ (9.26) i B(z), B † (z 0 ) = δ(z − z 0 ). (9.27) which has the commutator h 9.5.4 Crystal energy The energy H0 of the excited crystal is simply found by counting the number of excitons and phonons. Using the number operators for these particles (see (C.9) in 9.6 Heisenberg equations 310 Appendix C), we find H0 = X ων s†νk sνk + ν,k X ωκ b†κ bκ . (9.28) κ Or, in terms of the operators S and B, the local energy takes the form H0 X = ων Sν† (z)Sν (z) + ωp B † (z)B(z). L ν 9.6 (9.29) Heisenberg equations Now that we have constructed the Hamiltonians describing the Raman interaction, we can write down the Heisenberg equations governing time evolution of the operators Sν and B. The dynamics of A are derived in the next section using Maxwell’s equations, as was done for the atomic case in Chapter 4. Commutation of Sν and B with H, using the relations (9.17) and (9.26), yields the equations √ e Lg X Dµν Sµ (B † + B), ∂t Sν = iων Sν + √ p.A + i a m L µ √ D Lg X ∂t B = iωp B + i Dµν Sµ† Sν . a µ,ν (9.30) 9.6 Heisenberg equations 9.6.1 311 Adiabatic perturbative solution e in a rotating frame, so that Let us define local operators Seν and B Seν = Sν e−iων (t−nz/c) , e = Be−iωp (t−nz/c) . B (9.31) For notational convenience, let us also define B(t, z) = B † (t, z) + B(t, z) e † (t, z)e−iωp (t−nz/c) + B(t, e z)eiωp (t−nz/c) . = B (9.32) e are then given The equations of motion for the slowly varying operators Seν and B by √ e Lg X −iων (t−nz/c) √ p.Ae = Dµν Seµ eiωµν (t−nz/c) B, +i a m L µ √ X e = i Lg e−iωp (t−nz/c) ∂t B Dµν Seµ† Seν e−iωµν (t−nz/c) , a µ,ν ∂t Seν (9.33) where ωµν = ωµ − ων . The spatial phase factors are included for convenience when considering the exponentials within A. e the local phonon amplitude, in terms Our aim is to obtain an equation for B, of the signal and control fields. We achieve this by eliminating the intermediate excitons Seν adiabatically. The procedure is related to that used in §5.3.3 in Chapter 9.6 Heisenberg equations 312 5. We start by formally integrating the equation for Seν in (9.33), Seν (t) = Z t e 0 √ p. A(t0 )e−iων (t −nz/c) dt0 m L 0 √ Z t X Lg 0 +i Dµν Seµ (t0 )B(t0 )eiωµν (t −nz/c) dt0 . a 0 µ (9.34) e = 0) = 0, since there are no excitons in the crystal initially. We have set S(t Unfortunately (9.34) does not provide a direct solution for Seν , because it is coupled to all the other excitons through the summation on the right hand side. We settle instead for a perturbative solution. Substituting the first term on the right hand side of (9.34) into the second term yields the approximate result Seν (t) = Z t e 0 √ p. A(t0 )e−iων (t −nz/c) dt0 (9.35) m L 0 # √ Z t" Z t0 Lg X e 0 00 −iωµ (t00 −nz/c) 00 √ p. +i Dµν A(t )e dt B(t0 )eiωµν (t −nz/c) dt0 , a m L 0 0 µ This solution is correct to first order in the deformation potential. To perform the integrals in (9.35), we use the fact that the time variation of the exponential factors is much faster than the temporal dynamics of the optical fields, and also faster than the dynamics of the crystal excitations produced by these fields. This adiabatic approximation requires that the detunings ∆ν = ων − ωs of the excitons from the signal frequency are all much larger than the bandwidths of the signal or control fields. In diamond, the bandgap is in the ultraviolet, so this condition is always very well satisfied if optical frequencies are used. We proceed by pulling all slowly varying 9.6 Heisenberg equations 313 amplitudes out of the integrals and integrating only the exponentials. The resulting e expression for Seν contains 12 terms. Inserting this into the equation (9.33) for B provides us with the dynamical description we have been looking for, although now there are 144 terms! Fortunately, of the first order terms in the product Seµ† Seν , only very few contribute significantly to the memory dynamics: most terms are oscillating at high frequencies, so that they average to zero. After some legwork, we obtain e = iKΩ∗ A, ∂t B (9.36) where we have defined the control field Rabi frequency Ω(t, z) = evc .pAc (t, z) , m~ (9.37) and where the coupling constant K, with the dimensions of (length)−1/2 × (time)1/2 , is given by K= Dµν Dµν g ep.vs gs 1 X × ×√ + . (9.38) ~a ~ωs m L µν (ωµ + ωc )(ων + ωs ) (ωµ − ωc )(ων − ωs ) Let us suppose that we are sufficiently close to resonance that ων + ωs ων − ωs . This need not be the case, because the bandgap in diamond is very large, but it is a convenient simplification to neglect the ‘counter-rotating’ terms with summed frequencies in their denominators (c.f. §4.6 in Chapter 4). The coupling now takes 9.7 Signal propagation 314 the approximate form K= Dµν g ep.vs gs 1 X × ×√ , ~a ~ωs m L µ,ν (∆µ + ωp )∆ν (9.39) where we have used the Raman resonance condition ωs = ωc + ωp . The double appearance of the detuning in the denominator is characteristic of third-order scattering. Equation (9.36) is very similar in form to the equation for the spin wave (5.107) derived in Chapter 5. Thus encouraged, we proceed in the next section to derive the equation describing the propagation of the signal field. 9.7 Signal propagation The electric field of the signal is the solution to Maxwell’s wave equation, with the polarization in the diamond acting as a driving term (see (4.34) in Chapter 4). The polarization is the dipole moment per unit volume. That is, er = R P dV , where r is the one-electron position operator. We can relate the matrix elements of r to those of the momentum operator p, as follows (see (F.6) in §F.2 of Appendix F), pαβ = −imωαβ rαβ , (9.40) 9.7 Signal propagation 315 where ωαβ is the frequency splitting between the energy eigenstates |αi, |βi. Using this relation, a second-quantized form for the polarization operator can be found: P pνk,0 † ikz p0,νk sνk e − sνk e−ikz ων ων ν,k ep X 1 e† −iων (t−nz/c) √ = − Sν e + h.c. . mA L ν ων = ei X mAL (9.41) Making the SVE approximation (see (4.39) in Chapter 4), we find the propagation equation ∂z + n µ0 ωs2 ∗ e ∂t A = − v .Ps , c 2gs ks s (9.42) where Pes is the component of the polarization oscillating at the signal frequency ωs . Substituting the solution (9.35) into (9.41), performing the integrals with the help of the adiabatic approximation, and retaining only those terms with the appropriate time dependence, we arrive at the final equation for the signal, ∂z + n e ∗ ΩB, e ∂t + iχ A = iK c (9.43) e is very similar to K, where K X ωs Dµν e = g × ep.vs gs × √1 K , ~a ~ωs m L µ,ν ων (∆µ + ωp )∆ν (9.44) 9.7 Signal propagation 316 and where χ represents a spatial phase picked up by the signal due to the crystal dispersion, evs .pgs 2 1 X ωs × χ= . ~mωs L ν ων ∆ν (9.45) e is probably spurious, and may arise from the The discrepancy between K and K secular approximation we use in eliminating terms oscillating at the ‘wrong’ frequencies [177,178] . In any case, the difference between the two expressions is small close to resonance, and never more than an order of magnitude. In what follows, we set e −→ K. K The coupling constants are admittedly rather dense combinations of various material and optical parameters. In the next section we will try to extract a prediction for the coupling strength of a diamond quantum memory. But for the moment, what is important is the form of the equations (9.36) and (9.43). They are essentially identical to the Raman memory equations of §5.3.3 in Chapter 5. We can make the similarity explicit with a number of coordinate transformations and re-normalizations. First, we introduce the retarded time τ = t−nz/c (note the presence of the refractive index n). When the equations are written in terms of τ and z, the derivative ∂z + nc ∂t becomes simply ∂z , while the time derivative ∂t becomes ∂τ . Next, we remove the dispersive factor χ with a phase rotation, by making the transformation A −→ A = Aeiχz , e −→ B = Be e iχz . B (9.46) 9.7 Signal propagation 317 To remove the dependence on the control field profile, we introduce the normalized integrated Rabi frequency ω = ω(τ ) — defined in (5.46) in Chapter 5 — and the dimensionless transformed variables √ A(z, τ ) α(z, ω) = i W , Ω(τ ) √ β(z, ω) = LB(z, τ ). (9.47) Here we have also re-scaled the longitudinal coordinate by L so that z runs from e with K (as mentioned above), 0 up to 1. As a final simplification, we replace K and we assume that K is real. The equations of motion for the diamond quantum memory then take the form ∂z α = −Cβ, ∂ω β = Cα, (9.48) where the dimensionless Raman memory coupling is C= √ LW K. (9.49) This tells us that a diamond memory will behave in precisely the same way as a Raman memory based on an atomic vapour (see (5.109) in §5.3.3 of Chapter 5). We know that efficient storage is possible if the Raman memory coupling C is of the order of unity (we should have C & 2). And all of the results pertaining to optimal shaping carry over. So given the coupling C, we can directly find the optimal input mode φ1 for the signal, using (5.97). 9.8 Coupling 9.8 318 Coupling Here we show that efficient storage is possible in a small sample of diamond, so that diamond quantum memory need not be a luxury enjoyed only by the super rich. To estimate the memory coupling C, we must evaluate the summations in (9.39). To do this we first write Dµν = Dδµ,ν , which holds approximately, since the wavefunctions of distinct excitons overlap poorly [172] . Next we assume that both conduction and valence bands are parabolic around the zone centre (as shown in Figure 9.5), so that we may write the energy of an exciton as [182] ων −→ ω(k) = ω0 + ~k 2 , 2m? (9.50) where ~k is the momentum of the electron relative to the hole, and where m? is an effective reduced mass for the exciton that describes the local band curvature. Parameterizing the exciton energies in this way, the summations in K may be rewritten approximately as an integral over a sphere in k-space, X µ,ν Dµν (∆µ + ωp )∆ν −→ ≈ D (9.51) (∆ν + ωp )∆ν ν Z kmax AL 4πk 2 D dk, (2π)3 0 (ω0 − ωs + ωp + ~k 2 /2m? ) (ω0 − ωs + ~k 2 /2m? ) X where kmax is an arbitrary, suitably large cut-off. The integral can be performed 9.8 Coupling 319 with a partial fraction expansion and a trigonometric substitution, Z κ 0 b−3/2 a−c k2 dk = (a + bk 2 )(c + bk 2 ) b−3/2 (a − c) = # Z √bκ a c dx − dx (9.52) a + x2 c + x2 0 0 p √ p i h√ a tan−1 b/aκ − c tan−1 b/cκ . "Z √ bκ Choosing kmax sufficiently large, we get to the result [182] X µν Dµν ALD ≈ (∆µ + ωp )∆ν 4πωp 2m? ~ 3/2 √ p ∆ + ωp − ∆ , (9.53) where ∆ = ω0 − ωs is the detuning of the signal field from the conduction band minimum. We can express W in terms of the energy Ec in the control pulse as follows, Z ∞ W = −∞ |Ω(τ )|2 dτ = 2πα|vc .p|2 × Ec , ~m2 ωc2 An (9.54) where here α = e2 /4π0 ~c = 1/137 is the fine structure constant. We estimate the momentum matrix elements as vc .p ≈ vs .p = ~ × 2π/a. This is justified by noting that the band gap at the zone centre is produced by Bragg scattering that mixes electrons with k = 0 and k = 2∆k = 2π/a, and it is this latter component that is responsible for interband transitions. Other parameters are as follows. The bandgap4 in diamond is ~ω0 = 13 eV, and 4 This energy corresponds to the direct gap in diamond (at the zone centre, or ‘Γ-point’). The conduction band minima occur elsewhere in the Brillouin zone (~ω0 ∼ 5 eV at the ‘L-point’), but transitions to these states are mediated by phonons. They are therefore suppressed somewhat. Although they may not be insignificant, we neglect these processes here. 9.8 Coupling 320 the deformation potential has a value of around D ≈ 7~ω0 ≈ 90 eV. As mentioned before, the phonons have wavenumber νp = 1332 cm−1 . The lattice constant in diamond is a = 3.6 Å. The refractive index is n = 2.417, and the mass of a carbon atom is M = 12 a.m.u. (indeed, by definition). For simplicity we assume a reduced exciton mass m? = m. We then consider a crystal of length L = 1 mm, illuminated by beams with waists 100 µm in diameter. If the control pulse is taken from a modelocked pulse train with an 80 MHz repetition rate and an average power of 10 mW, it has an energy of 0.13 nJ. Using a signal field with central wavelength λs = 800 nm, we estimate a memory coupling of C = 2.4. (9.55) The optimal storage efficiency is therefore around 99.9% (see Figure 5.5 in §5.3 of Chapter 5). This demonstrates that efficient storage in a small sample of diamond is extremely feasible. Of course, the estimate we have made is very crude, but since neither the laser energy nor the sample size required are close to any practical limitations, a downward revision of C by a factor as large as 100 could still be accommodated. In fact (9.55) is something of an underestimate, since we have neglected the counter-rotating terms in K. When the signal is so far detuned from the conduction band edge, these terms still contribute significantly. 9.9 Selection Rules 9.9 321 Selection Rules The large Stokes shift (i.e. the large phonon energy) in diamond makes it easy to distinguish the signal and control fields spectrally. For instance, if the signal field wavelength is 800 nm, the control wavelength is around 894 nm. In addition, however, it turns out that the crystal symmetry requires the signal and control fields to be orthogonally polarized. The Raman interaction that couples the input and output fields is constrained to be proportional to an irreducible representation of the symmetry group associated with the optical phonons. The zone-centre phonons are always described by a subgroup of the crystal point group — the group of reflections and rotations that leaves the crystal unchanged. In the case of diamond, the crystal point group is the cubic group m3m (sometimes written Oh ), and the optical phonons transform as the subgroup Γ+ 5 (sometimes written T2g ). The optical fields are 3 dimensional vectors, and so the Raman interaction must be proportional [180] to the 3 × 3 irreducible representation of the group Γ+ 5 , which takes the form + Γ5 (x) = 1 + 1 , Γ5 (y) = 1 1 , Γ+ (z) = 1 5 1 where zero elements have been left blank for clarity, and we have assumed that the z-axis is aligned with the [001] direction (that is, parallel to a vertical edge of the cubic unit cell; see Figure 9.1). The Raman interaction requires that vc = Γ+ 5 (z)vs , from which it is clear that vc should be perpendicular to vs . This polarization , 9.10 Noise 322 selection rule adds to the experimental attractiveness of a diamond memory. 9.10 Noise The foregoing analysis of storage in diamond implicitly ignored the possibility of unwanted couplings. For instance, the intense control pulse can stimulate strong Stokes scattering, creating optical phonons and Stokes photons in pairs. This is the same problem as that shown for the atomic case in part (b) of Figure 4.3 in §4.7 of Chapter 4. Here it is exacerbated because the detuning from resonance is typically so large that both processes — storage of the signal and Stokes scattering — occur with roughly equal probabilities. In §6.3.2 of Chapter 6, a solution to this problem is described that involves introducing a small angle between the signal and control beams, and this solution certainly carries over to the diamond memory. Another interesting possibility is that of modifying the optical dispersion so that the unwanted Stokes light cannot propagate. It would be possible to do this by building a Bragg grating — alternating layers of diamond and air — with a spatial frequency equal to that of the unwanted Stokes light. Interference within the Bragg structure would then suppress any Stokes generation. But we won’t consider this further; a proofof-principle demonstration of broadband storage in room temperature diamond is challenging enough, without engaging in micro-fabrication [183] . In the next chapter, we review the experimental progress made in our group towards the goal of demonstrating a Raman quantum memory. Chapter 10 Experiments Although most of this thesis is theoretical, I had initially intended to build a working quantum memory. I have not succeeded as yet, but the ‘Memories’ subgroup is continuing its efforts in this direction. In this chapter we discuss some of the ongoing experimental work; its goals and future prospects. 10.1 Systems The experimental programme divides into three projects. 1. Diamond, 2. Quantum dots, 3. Atomic vapour. My experimental research has been focussed on the last of these; the theoretical analysis in the preceeding chapter constitutes my contribution to the first. We will 10.1 Systems 324 not describe the quantum dot project here; suffice it to say that quantum dots may be thought of as artificial atoms, so that a sample containing many dots behaves like an atomic ensemble, in which Raman storage may be implemented. Light storage in atomic vapour is becoming standard in quantum optics. In almost all cases, resonant EIT is used, and narrowband diode lasers provide the signal and control fields [74,150,151] . Our research group has some considerable experience with ultrafast lasers, however, and it was decided that a broadband Raman memory would be interesting. Atomic vapour is an ideal system for demonstrating such a memory; indeed this is the system that is considered when deriving the memory equations in Chapter 4. A common feature of all the experiments is the requirement of strong Raman coupling between a laser field and a material sample. The easiest way to verify the existence of strong coupling is to observe strong Stokes scattering. This is therefore the first step in all our experiments — the general strategy is shown schematically in Figure 10.1. Unfortunately, we have not yet been able to achieve this first prerequisite in the atomic vapour experiments. This is rather depressing, but we are persevering, since there are many improvements to be made. In the rest of this chapter, we will introduce the atomic species used in our experiments with atomic vapour. We then discuss the theory of Stokes scattering, and we estimate the strength of the coupling we expect to achieve. Finally, we describe the experimental techniques we have developed in our attempts to see strong Stokes scattering. We finish with a discussion of the planned realization of a Raman quantum memory. 10.2 Thallium Laser 325 Raman pump Stokes Vapour cell Filter Figure 10.1 Observing Stokes scattering as a first step. Strong Raman coupling between a bright laser and the atomic ensemble is required for a Raman quantum memory. The ability to produce, and detect, strong, stimulated Stokes scattering is a sine qua non for implementing the memory. 10.2 Thallium The first incarnation of the atomic vapour experiment used thallium (Tl) as the storage medium. Thallium is an extremely toxic poor metal, with a history of use in rat poison, and homicide generally. However the atomic structure of thallium exhibits a well-defined Λ-system, with a large Stokes shift (see Figure 10.2). For this reason, attempts were made to build a Raman quantum memory with thallium vapour, provided by means of a heated glass cell containing solid thallium. After around a year of unsuccessful attempts to observe strong Raman scattering from thallium vapour, it was realized that the vapour pressure of thallium is too low for a strong Raman interaction to be engineered (see the discussion in §10.9.1, and part (a) of Figure F.1 in Appendix F). 10.3 Cesium 326 7S1/ 2 378 nm F=1 F=0 6P3/ 2 F=2 F=1 1283 nm 6P1/ 2 F=1 F=0 Figure 10.2 Thallium atomic structure. The three levels of the Λsystem are discernible, marked by the thickened line segments. The hyperfine structure (indicated by the fine branches) is ignored, being negligible in the ground P -wave manifolds. The large Stokes splitting between the J = 1/2, 3/2 states in the P -state manifold makes it ideal for broadband storage. Unfortunately Thallium has a low vapour pressure, making an efficient thallium quantum memory impractical. 10.3 Cesium To increase the Raman coupling, an atomic species with a much higher vapour pressure was required. Our current experiment uses cesium1 (Cs), which is many orders of magnitude denser than thallium at room temperature (see part (b) of Figure F.1 in Appendix F). Cesium is a soft, gold-coloured alkali metal that is, thankfully, non-toxic, although it reacts explosively on contact with water, even the water vapour in air! Our cesium is sealed in an evacuated glass cell, along with a small amount of neon (10 Ne), which acts as a buffer gas (see §§10.4 and 10.5 below). There is only one stable isotope of cesium, namely 133 Cs. Needless to say we refer only to this isotope in what follows; our sample is naturally isotopically pure. The 1 Cesium is the American spelling; the British spelling Caesium retains some of the latinate flavour if its etymology. The word derives from the latin for ‘sky’, because the 7P ↔ 6S1/2 doublet lines are a brilliant blue colour. However American spell-checkers and journal styles have worn me down, and now the British spelling seems odd. 10.3 Cesium 327 Λ-system is implemented in the so-called cesium D2 line at 852 nm. This is the second of the strong ‘doublet lines’ (the D1 line is at 894 nm) characteristic of alkali metals — the same doublet in Sodium illuminates the night-time activities of most of this planet. The relevant atomic structure is shown in Figure 10.3. 852 nm 9.2 GHz Figure 10.3 Cesium atomic structure. The F = 3 and F = 4 hyperfine levels in the 6S1/2 ground state — the famous ‘clock states’ — provide the ground and metastable states for the Λ-type quantum memory. The 6P3/2 manifold collectively provides the excited state. The upper 6P3/2 state is split by the hyperfine interaction: The nuclear spin of I = 7/2 combines with the total electronic spin of J = 3/2 to produce four hyperfine levels with total angular momentum quantum numbers F = 2, 3, 4 and 5. The interaction is weak however, and the splitting between each state is around 200 MHz. For the purposes of the memory, we therefore treat the excited state manifold as a single state, which plays the role of |2i in the Λ-system. The hyperfine interaction is much stronger in the 6S1/2 electronic ground state. The reason is that the ground state is an S-wave, meaning that it has no orbital angular momentum. It therefore has no azimuthal phase, and so it remains well- 10.4 Cell 328 defined at the origin without vanishing — a wavefunction with such a phase is multivalued at the origin unless it is zero, so higher orbitals must disappear at the origin. In the ground state, then, the electron penetrates into the cesium nucleus. There is therefore a strong magnetic dipole coupling between the nuclear and electronic spins, known as the Fermi contact interaction, which acts to separate the two hyperfine states with F = 3 and F = 4 by an enormous 9.2 GHz. These two states are sometimes known as the clock states because coherent oscillations between them are used in cesium atomic clocks. In fact the hyperfine splitting is now defined to be exactly 9, 192, 631, 770 Hz, with the duration of the second being a derived quantity. The clock states form the two lower states in the quantum memory Λ-system. 10.4 Cell We contain the cesium vapour in a glass cell (the TT-CS-75-V-Q-CW from Triad Technology in Colorado, USA). The cell is 10 cm in length, and is made of glass that is resistant to heating up to temperatures of 500 ◦ C. The cell windows are 25 mm across, and are made of optically polished quartz, to which an anti-reflection coating has been applied which reduces reflection losses for light at the D2 wavelength of 852 nm down to around 0.1%. The cell is evacuated, and then a small sample of solid cesium is introduced. Finally, the cell is backfilled up to a pressure of 20 torr (∼ 2700 Pa) with neon gas. The reason for introducing this buffer gas is explained in §10.5. The cell is then hermetically sealed, by pinching shut the glass tube through which the cell contents are delivered. 10.4 Cell 10.4.1 329 Temperature control The cell is wrapped in heating tape: a thin weave of high resistance wires surrounded by thermal insulation. A thermocouple connected to an electronic temperature controller allows one to set and maintain the cell temperature as desired. The glass protuberance remaining after the cell is sealed provides a convenient ‘cold finger’ — an unheated region protruding from the cell where the cesium preferentially condenses. As long as the cell windows remain hotter than this cold finger, cesium does not collect on the cell windows, and the cell remains transparent to our laser beams. 10.4.2 Magnetic shielding The hyperfine levels in cesium are not pure quantum states. A hyperfine state with total angular momentum quantum number F is (2F +1)-degenerate, being comprised of Zeeman sublevels with quantum numbers m = −F, −F + 1, . . . , F − 1, F . These quantum numbers represent the projection of the atomic angular momentum along some axis, known as the quantization axis (see §F.4 in Appendix F). In the presence of a magnetic field, the atomic angular momenta precess around the direction of the field, and it is natural to define the quantization axis as aligned with this direction. With this definition, the quantum numbers m remain good quantum numbers, but the degeneracy of the Zeeman sublevels is lifted — this is known as the Zeeman shift. Optical transitions between the Zeeman sublevels occur subject to selection rules, which determine whether or not a transition conserves angular momentum, 10.4 Cell 330 based on the polarization of the incident light. With a judicious choice of polarized lasers, one can prepare the ensemble in just one Zeeman sublevel. Such a ‘spin polarized’ ensemble is used by Julsgaard et al. when implementing their continuousvariables memory in cesium [122] (see §2.4 of Chapter 2). This type of ensemble state is exquisitely sensitive to external magnetic fields, and it is common to build a magnetic shield of so-called µ-metal (an alloy with a high magnetic permeability that deflects magnetic field lines) around the vapour cell. In our system, the spectral bandwidth of the laser pulses comprising the signal and control fields is much larger than any feasible Zeeman shift one could produce (see §10.6). Therefore the Zeeman sublevels cannot be resolved in the quantum memory interaction, and so we neglect the Zeeman substructure of the hyperfine states. There is no need to spin-polarize the ensemble, and no need for magnetic shielding. In §F.4 of Appendix F, we show that orthogonal circular polarizations are not coupled by the Raman interaction. This is a further reason why it is not useful to polarize the ensemble: it might have been possible to use the Zeeman selection rules to our advantage (as is explained in the Appendix), but this result obviates this possibility. An unpolarized ensemble is in a mixed quantum state. It can be thought of as several independent sub-ensembles, each with a different spin polarization. Each subensemble interacts coherently with the optical fields, however, and the theoretical description of the memory interaction given in Chapter 4, 5 remains valid for each 10.5 Buffer gas 331 sub-ensemble. The lack of any magnetic shielding means that there may be stray magnetic fields that introduce a distribution of frequencies into the evolution of the Raman coherence through the Zeeman shift. This may cause the spin wave to dephase, so that the coherence is lost. However, as long as the magnetic fields remain constant over the memory lifetime, the spin wave will re-phase, because the number of Zeeman components is finite (this periodic re-phasing of a discrete set of oscillators is the principle behind retrieval from the AFC quantum memory [91] ; see §2.3.4 in Chapter 2). The periodic beating between the Zeeman sublevels restricts the times at which efficient retrieval is possible to those times where the spin wave is ‘in phase’, but the efficiency of the memory is not adversely affected. The above discussion justifies our decision not to build a magnetic shield for our cesium cell, and indeed not to attempt to polarize the vapour. However, since we have not yet been able to observe strong Stokes scattering, we cannot be sure that a magnetic shield would not help. We are investigating the construction of such a shield. 10.5 Buffer gas The 20 torr of neon buffer gas is added in order to extend the time that the cesium atoms spend in the interaction region. The cesium atoms are deflected by collisions with the neon atoms, so that their motion is diffusive rather than ballistic. The mean time between collisions is given by 1/γp , as can be verified by computing 10.5 Buffer gas hτ i = Rτ 0 332 τ ps (τ ) dτ using the ‘survival’ distribution (F.15) in Appendix F. If we follow the trajectory of a cesium atom, it will trace out a random walk with an average step length l = vth /γp , where vth = p 2kB T /M is the average thermal velocity of the atom. After N such steps, the root-mean-square displacement of the atom from its starting point is D = √ N l. The beam diameter is roughly √ A, so an atom escapes the beam after a time tescape , where D(tescape ) ∼ ⇒ tescape ∼ √ A, AM γp . 2kB T (10.1) A typical diffusion time for a beam with diameter 100 µm is around 10 µs at room temperature, which compares with around 0.5 µs in the absence of a buffer gas. During the course of this Brownian motion, the atom undergoes around 105 collisions. As described in Appendix F, these collisions randomize the phase of the optical polarization, causing pressure broadening. But the hyperfine coherence — the spin wave — is unaffected. The ground state hyperfine levels represent different orientations of the cesium nuclear spin, but neon is spinless, so there is no magnetic dipole interaction between the atomic nuclei in a collision. The Raman coherence is therefore maintained, despite the frequent collisions of cesium atoms with the buffer gas. The natural lifetime of the hyperfine coherence is several thousand years, so the memory lifetime is set by tescape . As well as slowing the escape of the cesium atoms, collisions with the buffer gas 10.6 Control pulse 333 change their velocities, so that the cesium atoms diffuse spectrally: as the cesium atoms’ velocities change, so do their Doppler shifted resonant frequencies. Before an atom leaves the interaction region, its resonance will ‘wander’ across the entire Doppler profile (see §10.10 below). This means that the atoms can be optically pumped (see §10.12 below) using a narrowband laser tuned into resonance with stationary atoms, since all atoms will eventually drift into resonance with the pump laser. Without a buffer gas, it is necessary to dither the pump laser frequency in order to address all the atoms, which is inconvenient (although certainly possible). 10.6 Control pulse The control pulse, or alternatively the Raman pump light, is sourced from a bespoke Ti:Sapphire laser oscillator: a Spectra-Physics Tsunami. The oscillator is actively modelocked using an acousto-optic modulator installed in the laser cavity, and the laser produces a pulse train with a repetition rate of 80 MHz, which is set by the cavity round-trip time. 10.6.1 Pulse duration The bandwidth of the control pulse cannot exceed the 9.2 GHz Stokes splitting between the two lower hyperfine states, if a quantum memory is to be implemented. However the gain bandwidth of Ti:Sapphire is very wide (several hundred nanometers!), so some care was taken to limit the bandwidth of the laser. A Gires-Tournois Interferometer (GTI) is installed inside the laser cavity. This is essentially a Fabry- 10.6 Control pulse 334 Perot etalon, with the rear of the two plates having a reflectivity of 100%. All incident energy is therefore reflected, but the spectral phase inside the cavity undergoes periodic jumps, with the free spectral range determined by the plate separation. Any pulse in the laser cavity with a bandwidth spanning one of these phase jumps experience catastrophic dispersion, which greatly reduces the efficiency with which the pulse — being heavily distorted — can extract energy from the Ti:Sapphire crystal. The bandwidth of the laser is therefore limited to bandwidths smaller than the free spectral range of the GTI. Use of a GTI with an unusually large plate separation produces a modelocked pulse train with a spectral bandwidth or around 1.5 GHz (∼ 3.6 pm at 852 nm). x Mirror Beamsplitter Mirror Retroreflector intensity (arb. units) Photodiode 1000 500 0 150 200 250 x (mm) 300 Figure 10.4 First order autocorrelation. The set up is essentially what is known as a Fourier transform spectrometer. An incident pulse is split into two components, and one is delayed with respect to the other. The arrangement with the retroreflector and facing mirror makes the movable arm insensitive to misalignment as the retroflector is translated over a large distance. To characterize the pulse train, measurements of the first and second order correlation functions of the output light were made. A Michelson interferometer was 10.6 Control pulse 335 built, so that interference of the optical field with a delayed copy of itself could be observed. The first order field autocorrelation, shown in Figure 10.4, produces a fringe pattern whose envelope is equal to the Fourier transform of the spectral intensity I(ω) of the laser output (see (F.16) in Appendix F for an example of this relationship). The spectral intensity is found to be approximately Gaussian, with a FWHM bandwidth of 1.5 GHz, as mentioned above. This is consistent with a Fourier-transform-limited pulse duration of 300 ps, but it may be that distortions to the spectral phase of the pulse train ‘smear out’ the pulses, producing longer durations. To investigate this possibility, we performed a second order interferometric autocorrelation. The experimental set up is shown in Figure 10.5. Two delayed copies of an incident pulse are focussed into a non-linear crystal — a small piece of β-Barium Borate (BBO). Blue light scattered in the forward direction results from the frequency upconversion of one pulse with its delayed counterpart. The envelope of the resulting fringe pattern can be related to the pulse duration. Although the shape of the autocorrelation envelope is not uniquely related to the pulse envelope (it is not possible to ‘invert’ the autocorrelation to retrieve the pulse profile), it is possible to infer the presence of spectral phase distortions by inspection of the lower portion of the interferogram. In the presence of ‘chirp’ (a drift in the carrier frequency of the pulse as a function of time), interference between the leading edge of one pulse, and the trailing edge of the other, is suppressed, since the carrier frequencies are no longer commensurate. This causes a distinctive narrowing of the lower portion of the interferogram that is not present in our measurement. On the basis of 10.6 Control pulse 336 this result, we conclude that the pulses produced by our laser source are close to being transform-limited, with approximately Gaussian spectral and temporal profiles. The FWHM duration of each pulse is around 300 ps, and the spectral bandwidth is approximately 1.5 GHz. On the one hand, this is much narrower than the Stokes splitting, as required. On the other hand, the pulses are much more broadband than have been used to date in quantum memory experiments (pulse durations of 100’s of microseconds are the shortest that are employed for EIT [74] ). Filter BBO Lens x Mirror Beamsplitter Mirror Retroreflector intensity (arb. units) Photodiode 600 400 200 0 150 200 250 x (mm) 300 Figure 10.5 Second order interferometric autocorrelation. A nonlinear crystal is inserted, and the frequency upconversion of one pulse with its delayed counterpart is detected. The inset shows the measured envelope of the interferogram. Again, the red lines are Gaussian fits. These data, along with the data in Figure 10.4, are the only experimental data in this thesis that were taken by me! 10.6.2 Tuning The plate separation of the GTI can be temperature tuned using a knob on top of the laser, and this also provides a convenient way to smoothly tune the laser frequency over small frequency shifts, of order 10 or 20 GHz (it appears that the spectral phase 10.7 Pulse picker 337 introduced by the GTI introduces differential gain over its free spectral range, which allows small changes to the FSR to sweep the laser frequency). Larger frequency shifts can be ‘dialled in’ using a birefringent filter (this is actually an intra-cavity Lyot filter — see §10.13.2 below). 10.6.3 Shaping Although a great deal of the theoretical work in this thesis deals with the optimization of quantum storage by appropriate shaping of the control or signal pulse profiles, it is not feasible to shape the pulses used in this experiment. They are too short to be shaped electronically, and two narrowband to be shaped spectrally! However, since we are still struggling to demonstrate storage in the first place, we are content to walk before attempting a steeplechase. The excellent work of Novikova et al. [74] on resonant shaped storage is a vindication of the theory of Gorshkov et al. [133] , and by extension the theory in this thesis, since they are so closely related (at least at optical depths less than 50 or so [151] ). 10.7 Pulse picker The time between consecutive pulses in the output of our laser is 12.5 ns. This is much shorter than the lifetime of the Raman coherence (see §10.5 above), so if the pulse train from the laser is sent ‘as is’ into the cell, the ensemble does not recover after each Raman interaction. If one’s aim is to generate strong Stokes scattering, this may be beneficial. The Stokes process can be stimulated both the 10.8 Stokes scattering 338 presence of Stokes photons, and by the presence of spin wave excitations, so the Raman coherence produced by a previous pulse may stimulate the Raman interaction in subsequent pulses. However it is not possible to investigate the efficiency of a quantum memory if different realizations of the memory interaction are coupled. To demonstrate the quantum memory, it is necessary to reduce the repetition rate of the laser so that each implementation of the memory is independent of all others. This is done with a ‘pulse picker’, which is a fast optical switch based on the electro-optic effect, known as a Pockels cell. The device is synchronized with the laser output, and the optical switch is set to transmit every 80, 000th pulse, reducing the laser repetition rate to 1 kHz. The ensemble now has 1 ms to ‘reset’ between pulses, which is plenty of time. In the next section we introduce the theory of Stokes scattering. 10.8 Stokes scattering The requirements for observing strong Stokes scattering are very similar to the requirements of a Raman quantum memory. A medium must be prepared in the ground state of a Λ-system. There should be strong Raman coupling between an incident laser pulse and the excitations of the medium. And it should be possible to observe scattering at the Stokes-shifted wavelength by filtering out the strong Raman pump. A description of Stokes scattering runs along very similar lines to that of a Raman quantum memory: the signal and control fields simply trade places, becoming pump 10.8 Stokes scattering 339 and Stokes fields (see part (a) of Figure 1.6, or part (b) of Figure 4.3). However, instead of absorption, the Stokes field experiences gain, with energy being transferred from the pump field into the Stokes field, while at the same time excitations are generated in the Raman medium. We are interested in transient Stokes scattering, which refers to the regime in which the duration of the Raman pump pulse is much shorter than the lifetime of the excitations in the medium. This is consistent with the requirement that the same medium should be useful as a quantum memory, which requires that the spin wave excitations far outlive the optical pulses. The equations describing transient Stokes scattering in the adiabatic limit can be written in the following simple form, ∂z α = Cβ † , ∂ω β = Cα† . (10.2) Here all the notation has the same meaning as in (5.109) from §5.3.3 in Chapter 5, except that α is a dimensionless annihilation operator for the Stokes mode. Recall that ω is the integrated Rabi frequency (defined in (5.46) in §5.2.6 of Chapter 5), where now the relevant Rabi frequency is that of the Raman pump pulse. The coupling constant C is the same as the Raman memory coupling, so the efficiency of both Stokes scattering and quantum memory are characterized by the same number. The equations (10.2) describe a squeezing, or Bogoliubov transformation as opposed to a beamsplitter interaction: optical and material excitations are produced in cor- 10.8 Stokes scattering 340 related pairs. This interaction is the basis of the DLCZ quantum repeater protocol, since the Stokes mode becomes entangled with the ensemble (see §1.6.4 in Chapter 1). The equations can be solved in precisely the same manner as the system (5.109). The solution for the Stokes field is r h √ i C2 0 α (ω 0 ) dω 0 αout (ω) = α0 (ω) + I 2C ω − ω 1 0 ω − ω0 0 Z 1 h p i I0 2C (1 − z)ω β0 (z) dz. +C Z ω (10.3) 0 Here the I’s denote modified Bessel functions, which describe exponential gain. We are interested in the intensity of spontaneously initiated Stokes scattering, when there are initially no spin wave excitations and no photons in the Stokes mode. The average number of Stokes photons produced is given by the normally ordered product Z hNout i = 0 1 † hαout (ω)αout (ω)i dω. (10.4) On substitution of (10.3) into (10.4), a number of terms vanish. The cross terms involving the products hα0† (ω)β0† (z)i and hβ0 (z)α0 (ω)i are both zero, since in the first case the inner product between singly-excited and vacuum states is taken, and in the second case the vacuum state is annihilated. The same is true of the term involving hα0† (ω)α0 (ω)i. The only non-vanishing term involves the anti-normally ordered combination hβ0 (z)β0† (z)i. To evaluate this term, we observe that β is a bosonic annihilation operator which satisfies the same commutation relation as B 10.8 Stokes scattering 341 (see (4.45) in Chapter 4, or (9.27) in Chapter 9), h i β(z), β † (z 0 ) = δ(z − z 0 ). (10.5) We therefore have that hβ0 (z)β0† (z 0 )i = hδ(z −z 0 )+β0† (z 0 )β0 (z)i = δ(z −z 0 ). Inserting this result into the expression for hNout i, we obtain2 hNout i = 2C 2 I02 (2C) − I12 (2C) − CI0 (2C)I1 (2C). (10.6) Figure 10.6 shows the average number of scattered Stokes photons as a function of the coupling C. C = 1 corresponds to hNout i ∼ 1, and this may be thought of as marking the onset of the stimulated scattering regime, when previously scattered Stokes photons stimulate further scattering, so that the scattering efficiency begins to grow exponentially with C. For efficient Raman storage, a quantum memory requires C ∼ 1 also, so the possibility of producing stimulated Raman scattering is a necessary condition for implementing a Raman quantum memory in any system. In the next section we explain how to calculate a prediction for the coupling constant C, when an atomic vapour is used as the memory medium (a calculation of C for the case of a diamond quantum memory was undertaken at the end of the previous Chapter). 2 This result appears as equation (38) in the seminal paper on Stokes scattering by Michael Raymer and Jan Mostowski [139] , who specialized to the case of a square pump pulse. 10.9 Coupling 342 10 10 10 10 10 5 0 −5 0 2 4 6 8 Figure 10.6 Stokes scattering efficiency. As the coupling C increases, the number of Stokes photons scattered rises, growing exponentially for C & 1. The model neglects depletion of the energy in the pump pulse: of course the number of Stokes photons scattered cannot exceed the number of photons in the pump pulse, or indeed the number of atoms in the ensemble. 10.9 Coupling Throughout this thesis we have made reference to the quantities d, Ω, C etc... Fortunately these quantities are not difficult to calculate in the context of an atomic vapour. 10.9.1 Optical depth Combining the definition (5.12) in §5.2.3 of Chapter 5 with (4.50) in §4.11 of Chapter 4, the optical depth can be written as d= |d∗12 .vs |2 ωs nL . 2γ0 ~c (10.7) The spontaneous emission rate 2γ can be expressed in terms of the dipole moment, using Fermi’s golden rule to derive the stimulated emission rate, and then Einstein’s 10.9 Coupling 343 relations to connect this to the spontaneous rate [107] . The result is 2γ = 3 |d |2 ω21 12 . 3π0 ~c3 (10.8) Here we have neglected any factors arising from degeneracy of the states involved. The factor of 1/3 represents an average over all spatial direcitons. Substituting (10.8) into (10.7), and assuming that the signal polarization vs is aligned with the dipole moment d∗12 , gives the result d = 3 × nλ2 L 4π ∼ nλ2 L, (10.9) where λ = 2πc/ω21 is the wavelength associated with the |1i ↔ |2i transition, and where we have made the approximation ωs ≈ ω21 (any detuning is much smaller than an optical frequency). (10.9) is consistent with the notion that the scattering crosssection of an atomic transition is roughly λ2 . The approximation in the second line of (10.9) is not generally accurate, but it is extremely useful as an ‘order of magnitude’ estimate for the optical depth. The number density n can be found from the vapour pressure, given the temperature of the atomic ensemble (see §F.1 in Appendix F). For an optical transition with λ ∼ 1 µm and a typical ensemble length of L ∼ 1 cm, we have d ∼ n[m−3 ] × 10−14 . As a rule of thumb, this allows one to easily estimate the atomic number density required for an efficient memory (with d & 100). In general, a more accurate value for the optical depth should be calculated by 10.9 Coupling 344 using empirical values for the rate γ (recall that 2γ is the spontaneous emission rate) and for the dipole moment d12 . Sometimes data tables list the values of oscillator strengths associated with atomic transitions. These are dimensionless numbers that quantify the dominance of a transition over other possibilities within the atom. The connection between the oscillator strength and the dipole moment for a transition is derived in §F.2 of Appendix F. The optical depth can be expressed in terms of the oscillator strength f12 for the |1i ↔ |2i transition as follows, d= π~α m × f12 nL , γ (10.10) where m is the electron mass and α = 1/137 is the fine structure constant. Figure 10.7 below shows the variation of the optical depth of our cesium cell, with a length of 10 cm, as a function of temperature. 10.9.2 Rabi frequency The Rabi frequency is given by Ω= d23 .vc Ec , ~ (10.11) where Ec is the electric field amplitude of the control associated with its positive frequency component. The instantaneous intensity of the control field is given by 1 Ic = 0 c × |2Ec |2 . 2 (10.12) 10.9 Coupling 345 8 Optical depth 10 6 10 4 10 2 10 300 350 400 450 Temperature, K 500 Figure 10.7 Cesium optical depth. This is the optical depth associated with the complete 6P3/2 manifold. The optical depth is the same whether we consider transitions from the F = 3 ground state level, or the F = 4 level. The dipole moment for the cesium D2 line is around 10−29 Cm, according to the data provided by Daniel Steck [184] (this figure has been divided by a factor of 3, because we consider linearly polarized incident fields). The length of our cell is 10 cm. The number density is found from the vapour pressure curve in Figure F.1 in Appendix F. An estimate of the energy in the control pulse is then given by Ec = Ic ATc . Neglecting any complex phase in Ec , we can express the peak Rabi frequency of the control in terms of this energy, Ωmax d23 .vc × ≈ ~ s Ec . 20 cATc (10.13) This can, of course, be expressed in terms of the oscillator strength for the |2i ↔ |3i transition using the conversion formula given in Appendix F. If the control pulse is taken from a pulse train — for example the output of a modelocked laser — we can write Ec = P/R, where P is the average power in the beam, and where R is the pulse repetition rate. The pulse duration Tc needs to be measured (see §10.6 above). 10.9 Coupling 10.9.3 346 Raman memory coupling The Raman memory coupling is defined in ordinary units as C = √ dγW /∆. An approximate expression, assuming a top-hat shape for the control pulse, is 2 C ≈ Tc dγ × Ωmax ∆ 2 . (10.14) If (10.13) is used in this expression, the factors of Tc cancel, and the assumption of a top-hat profile can be dropped. Since adiabatic evolution requires that ∆ Ωmax , the above form of the Raman coupling makes it easy to see why an efficient, adiabatic Raman memory should have Tc dγ 1 [133] . That is to say, the bandwidth δc ∼ 1/Tc of the pulses stored (the optimal signal bandwidth is close to the control bandwidth) should be small compared with dγ. When estimating the magnitude of the Raman coupling, the following form can be useful3 , C2 ≈ ~ mA∆ 2 × (πα)2 × (f12 f23 ) × Na Nc , (10.15) where Na is the number of atoms in the ensemble that are illuminated by the optical fields, and where Nc is the number of photons in the control pulse. Putting f12 ≈ f23 ≈ 1 provides an upper limit to the Raman coupling that is achievable with a given number of atoms and control photons. As an example, an optical depth of d ∼ 3.8 × 104 is predicted for our cesium cell, of length 10 cm, when heated to 90 ◦ C. Setting the control field beam waist to 330 3 This form of the coupling appears in our Rapid Communication on Raman storage [77] , but there is a typo! The form given here is correct. 10.9 Coupling 347 µm gives a Rayleigh range (see §6.4 of Chapter 6) of 5 cm, so that the beam remains collimated over the full length of the cell. Setting the average power from our laser oscillator to 600 mW gives a pulse energy of 7.5 nJ. One such pulse, focussed as just described and detuned by ∆ = 20 GHz from the D2 line is sufficient to produce a Raman memory coupling of C ∼ 2.4, corresponding to a storage efficiency of ∼ 99.9%. The memory balance is then R ∼ 0.5, so the memory is well within the Raman regime (i.e. the interaction is balanced; see §5.3.1 in Chapter 5). The detuning is around 50 times larger than the Doppler linewidth (see §10.10 below), so the inhomogeneous broadening may be neglected. Adiabaticity is guaranteed because the detuning is by far the largest frequency involved in the interaction, ∆/γ ∼ 104 , ∆Tc ∼ 40, ∆/Ωmax ∼ 6. 10.9.4 Focussing Substitution of (10.10) and (10.13) into (10.14) reveals that the Raman coupling depends on the geometry of the ensemble in the following way, C2 ∝ L . A (10.16) This is because the optical depth only depends on the length, and the squared Rabi frequency depends only on the inverse of the beam area. The ensemble is addressed by collimated laser beams, which in the ideal case (when the laser is working well) have a Gaussian transverse profile. The Rayleigh range zR of a focussed Gaussian 10.9 Coupling 348 beam of wavelength λ is related to its cross-sectional area at the focus by diffraction, zR = A . λ (10.17) The beam remains well collimated over a region of length zR either side of the focus, but after this is quickly diverges, and the intensity drops rapidly. Therefore we may consider that the length of the ensemble over which the memory interaction is reasonably strong is limited by the Rayleigh range, L ∼ zR . Making this identification in (10.16), and using the relation (10.17), we find C2 ∝ 1 . λ (10.18) That is, the geometrical dependence of C drops out. The Raman coupling is independent of how the beams are focussed. Loosely focussed beams are better described by the one-dimensional theory, and indeed numerical simulations show that a loosely focussed control beam improves the memory efficiency (see 6.5 of Chapter 6). But if the beams are too loosely focussed, their Rayleigh range will extend beyond the cell length, and the coupling will be limited by the dimensions of the cell. Therefore the optimal situation obtains when the Rayleigh range of the beams is matched to the cell length, zR = L/2. 10.10 Line shape 10.10 349 Line shape Figure 10.8 shows an absorption spectrum for the D2 line, measured using a laser with a spectral bandwidth of 1.5 GHz, with the cell heated to 70 ◦ C. A number of factors contribute to the shape of the D2 absorption lines in a warm cesium vapour. The absorption linewidth of each transition in the D2 line is around 250 MHz at room temperature for a single atom. But the large optical depth widens the absorption lines of the ensemble (see §7.2.1 in Chapter 7, and §10.11 below). The dominant line broadening mechanism is Doppler broadening, with pressure broadening contributing at the level of ∼ 10 MHz. These mechanisms are explained in §F.3 of Appendix F. The Doppler width is calculated using (F.12) by setting M = 133 a.m.u. — the mass of a cesium atom — and ω0 = 2πc/λ0 , with λ0 = 852 nm the cesium D2 line wavelength. The pressure-broadened linewidth is calculated from (F.18), but the parameters used are those of the buffer gas. The cesium cell contains neon, which was back-filled up to a nominal pressure of pbuffer = 20 torr. The buffer gas reduces the mean free path of the cesium atoms, so that they stay in the interaction region — the volume illuminated by the light beams — for a longer time than they otherwise would (see §10.5 below). The number density of neon is much greater than the number density of cesium, at all reasonable temperatures, so a cesium atom is much more likely to collide with a neon atom than with another cesium atom. The relevant number density in estimating the pressure-broadened linewidth is therefore n = pbuffer /kB T0 , 10.10 Line shape 350 1000 Signal, mV 800 Absorption spectrum No cell 600 400 200 0 −10 0 10 Detuning, GHz 20 Figure 10.8 Cesium D2 absorption spectrum. This plot appears thanks to Klaus Reim, a D.Phil student currently dividing his time between the cesium and quantum dot projects. A weak laser is scanned across the cesium D2 line at 852 nm. The blue line shows the signal from a photodiode placed after the cesium cell. The red dotted line indicates the signal detected if the cell is removed. The two dips correspond to the two ground state hyperfine levels: on the left is the F = 3 state; the F = 4 state is on the right. The 9.2 GHz splitting between these states is evident. The hyperfine structure of the upper 6P3/2 state is not resolved. The laser used has a Gaussian spectrum with a FWHM bandwidth of ∼ 1.5 GHz (see §10.6), and this contributes to the wide absorption linewidth. Doppler broadening, along with the ‘smearing’ effect of the hyperfine splitting in the upper state, and the large optical depth (d ∼ 104 at 70 ◦ C), accounts for the remainder of the linewidth, with pressure broadening contributing negligibly. where T0 ∼ 300 K is the temperature at which the cell was filled. The cell is sealed, so the number density of the buffer gas is fixed. The relevant collision velocity is the thermal velocity of Neon atoms, calculated using M = 10 a.m.u.. A reasonable figure for the collision cross-section σ is found by choosing datom = (300 + 50)/2 pm, which represents an average of the atomic diameters of a cesium and neon atom. Figure 10.9 shows the variation in the Doppler and pressure-broadened linewidths with temperature for our cesium cell. The temperature dependence is very weak, 10.11 Effective depth 351 since both linewidths scale linearly with the thermal velocity of the atoms in the vapour cell, which in turn scales with √ T . It’s clear that Doppler broadening dom- inates over pressure broadening. The linewidth is similar to the hyperfine splitting in the 6P3/2 excited state manifold, so these hyperfine states are not resolved in our sample. Fortunately it is not important, either for optical pumping (see §10.12 below), or for the quantum memory interaction, to distinguish the hyperfine structure of the excited state. Provided that Raman storage is implemented with a detuning that is large compared to both the Doppler linewidth and the hyperfine splitting, a theoretical description that ignores these complications remains appropriate. In fact, Gorshkov et al. have shown that even resonant storage is unaffected by Doppler broadening provided the ensemble is sufficiently optically thick [163] . However, characterizing the optical depth of the ensemble is not entirely straightforward, in the presence of line-broadening. 10.11 Effective depth If a weak probe beam is tuned into resonance with one of the D2 line transitions, the measured attenuation is much less than would be predicted on the basis of the optical depth estimated using (10.9). The reason for this is the redistribution of optical depth over a wide spectral range, because of Doppler broadening. Suppose one measures the transmitted power Pout , and the input power Pin . The effective optical depth is defined by Pout = Pin × e−2deff . (10.19) 10.11 Effective depth 352 9 10 Linewidth, Hz Doppler 8 10 Pressure 7 10 6 10 300 350 400 450 Temperature, K 500 Figure 10.9 Absorption linewidth. The pressure (green) and Doppler (blue) linewidths γ/2π for our cesium sample, calculated using the formulae (F.12) and (F.18) derived in Appendix F. The red dotted line shows the half-width at half-maximum of the resulting Voigt profile, which is the line profile resulting from the convolution of the pressure-broadened Lorentzian and Doppler-broadened Gaussian lineshapes. Doppler broadening dominates however, and the lineshape is essentially Gaussian. The spectrum of the transmitted beam can be found from (7.29) in Chapter 7, setting z = 1 and Pin = 0. In the presence of pressure-broadening, the polarization decay rate is modified, γ −→ γ 0 = γ + γp /2, (10.20) and the pressure broadened optical depth dp = d×γ/γ 0 should be used. The effective optical depth is then given by deff 1 = ln 2 R R Iin (ω) dω Iin (ω)e−2dp <[f (ω)] dω , (10.21) where Iin (ω) is the spectral intensity profile of the probe beam, and where the lineshape function f (ω) is defined in (7.28) in Chapter 7, the only difference being 10.12 Optical pumping 353 that all frequencies are normalized by γ 0 instead of γ. Using a beam with an average power of 1 µW, taken from the output of a modelocked Ti:Sapphire oscillator with a spectral bandwidth of 1.5 GHz (see §10.6), we measure deff ∼ 3 at a cell temperature of 90 ◦ C. If we invert the formula (10.21), using γ = 16.4 MHz (the spontaneous lifetime of the cesium D2 line is τ = 1/2γ = 30 ns), γ 0 ≈ 70 MHz and γd ≈ 260 MHz, we infer a ‘real’ optical depth of d ≈ 3,500. This is consistent with the prediction d = 3,700, found from (10.7) using the number density plotted in part (b) of Figure F.1. 10.12 Optical pumping Even though the hyperfine splitting is extremely large, it represents a very small energy gap at room temperature, with Esplitting ≈ kB T /680. Therefore the populations of both the clock states are equal. From the perspective of quantum storage, this means there is a large thermal background of incoherent material excitations that would swamp any stored signal. But the situation is worse than this: the memory interaction is totally destroyed if the populations of the two lower states are equal. As signal photons are absorbed, they are also produced by Stokes scattering from the thermal population in the storage state. The gain from thermally seeded Stokes scattering exactly balances the quantum memory absorption, and the memory is rendered useless. This effect is demonstrated in Figure 10.10. For an efficient quantum memory, therefore, one of the lower two states must be emptied. This is done in the laboratory by by optical pumping. 10.12 Optical pumping (b) Intensity (arb. units) (a) 354 Figure 10.10 Equal populations destroy quantum memory. (a): Absorption of a Gaussian signal pulse with the storage state empty. The memory efficiency is 73% (b): With equal populations in both ground and storage states, the absorption vanishes. The memory efficiency here is ∼ 0.01%. These plots are produced by numerically integrating the full system of Maxwell-Bloch equations, without making the linear approximation introduced in §4.5 of Chapter 4. There are 8 equations to solve, one for each optical field (i.e. the signal and the control pulses), one for the population of each state in the Λsystem, and three equations describing the coherences between pairs of states. The numerical methods used are described in Appendix E. We consider a Raman quantum memory, with ∆ = 150, d = 300, and W = 122.5, and a Gaussian control pulse with duration Tc = 0.1 (working in normalized units with frequencies expressed in terms of γ, the natural linewidth of the excited state — pressure and Doppler broadening are not modelled). The Raman memory coupling is then C = 1.28, and the balance is R = 0.64 (see §5.3.1 in Chapter 5). The signal field profile is identical to the control profile (no optimization is applied), but its amplitude is 1000 times smaller than that of the control (the relative amplitudes of the signal and control are important when the equations are not linearized). Suppose we wish to use the F = 4 state as the ‘ground state’ for the memory — such a choice allows for efficient phasematched backward retrieval, as discussed in §6.3 of Chapter 6. We then need to pump all the atoms into the F = 4 state, leaving the F = 3 state completely empty. We accomplish this by tuning a laser into resonance with the |F = 3i ↔ |2i transition, where |2i stands for any of the states in the 6P3/2 manifold. The laser field excites atoms into |2i, which then decay 10.12 Optical pumping 355 by spontaneous emission back into both the F = 3 and F = 4 ground states (the branching ratio is roughly equal). The laser field then re-excites the atoms, and the process repeats (see Figure 10.11). Consider the effect of consecutive cycles of excitation and decay. With each cycle, only around 50% of the initial population of the F = 3 state ends up back in F = 3. The rest ends up in the F = 4 state. Since there is no laser field exciting atoms out of this state, population builds up in F = 4, while the F = 3 state is eventually emptied entirely. This is the principle behind optical pumping. Figure 10.11 Optical pumping. A CW diode laser is tuned into resonance with the |F = 3i ↔ |2i transition. Spontaneous emission redistributes the excited population more-or-less equally between the two ground hyperfine levels. After several cycles, population builds up in the un-pumped F = 4 state, while the F = 3 state is emptied. 10.12.1 Pumping efficiency To characterize the optical pumping efficiency a simple measurement was made (this was done by Virginia Lorenz and Klaus Reim), based on an experiment performed by Jean-Louis Picqué in 1974 [185] . The set-up is illustrated in Figure 10.12. We use an external cavity diode laser to provide the pumping light. This is a single-mode 10.13 Filtering 356 diode laser with a diffraction grating placed in front of it in the Littrow orientation: the first order of diffraction is directed back into the laser. By fine-tuning the grating angle, it is possible to select those frequencies that experience the greatest optical gain (other frequencies are misaligned by the grating and are lost). The diode laser frequency can therefore by stabilized and tuned using the grating. The specular reflection from the grating provides the laser output, which is a continuous wave (CW) beam with a linewidth of around 100 MHz. Using the measurement scheme shown in Figure 10.12, we detect the fluorescence signal from a weak probe beam. The suppression of this signal with the pump beam set at 30 mW over the signal with the pump blocked indicates that we achieve a pumping efficiency of around 95%, with the cell temperature set at T = 50 ◦ C. At higher temperatures, the optical pumping efficiency drops as the energy in the pump is absorbed by the larger number of atoms. We are currently installing an ECDL with an output power of 100 mW, which should improve our pumping efficiency for higher temperatures. 10.13 Filtering The greatest experimental challenge posed by the implementation of a Raman quantum memory is the ability to filter out the very weak — even single photon — signal field from the strong classical control pulse. The problem is particularly difficult in cesium, because the Stokes shift of 9.2 GHz corresponds to a very small spectral shift. At 852 nm — the D2 resonance wavelength — the signal and control frequencies differ by around 20 pm. The required filter contrast is of order 108 or better, 10.13 Filtering 357 Photodiode Lens Lock-in Pump Cs cell Heaters Beamsplitter Probe Chopper Laser diode Grating Figure 10.12 Verifying efficient optical pumping. The beam from an external cavity diode laser (ECDL) tuned into resonance with the |F = 3i ↔ |2i transition is split into two parts at an asymmetric beamsplitter (just a microscope slide). 96% of the diode beam is used to pump the cesium atoms in the heated cell. 4% of the beam is sent through an optical chopper, which applies a 1 kHz modulation, before being directed through the cell at a small angle to the pump beam. Fluorescence is collected by a lens and focussed onto a photodiode. A lock-in amplifier is used to isolate the component of the fluorescence due to the modulated probe beam. This signal quantifies the optical pumping efficiency. The inset shows the reduction in the fluorescence signal as the pump power is increased. A pumping efficiency of 95% is achieved. These data appear thanks to Virginia Lorenz and Klaus Reim. so that no single filter provides sufficient rejection of the pump: several filters must be used in tandem. In this section we give details of the various filtering techniques we have at our disposal. 10.13 Filtering 10.13.1 358 Polarization filtering The Raman interaction couples orthogonal linear polarizations: a horizontally polarized control pulse will store a vertically polarized signal pulse, and vice versa, provided the detuning is sufficiently large. And any Stokes scattering induced by a linearly polarized pump pulse is polarized orthogonally to the pump. These facts are derived in §F.4 in Appendix F. Therefore the pump can be effectively rejected after the cell using a polarizer aligned orthogonally to the pump polarization. We use Glan-Laser polarizers from Foctek, which use total internal reflection at the boundary of two calcite crystals with perpendicular optic axes to preferentially transmit one linear polarization. The extinction ratio of these polarizers is of order 10−6 . However, stress-induced birefringence of the cell windows can distort the pump polarization, causing leakage of the pump light through the polarizer. The effective polarization extinction is therefore of order 10−4 . 10.13.2 Lyot filter A Lyot filter is a polarization-based spectral filter, popular among astronomers because of its wide working numerical aperture. In general a Lyot filter consists of a number of concatenated filter stages, with each stage improving the finesse of the filter (that is, reducing the spectral width of the pass-band, while increasing the spectral width of the region over which frequencies are blocked). We built the simplest possible type of Lyot filter, which involves a single stage. The filter then has a sinusoidal transmission as a function of frequency. Figure 10.13 shows the 10.13 Filtering 359 structure of the filter. It is essentially a Mach-Zender interferometer for polarization: an incident pulse is split into fast and slow polarization components inside a birefringent retarder. The polarizations are then combined on a polarizer, where they interfere to produce spectral fringes. The free spectral range (i.e. the period of the fringes in frequency) depends inversely on the length and birefringence of the material used for the retarder. In order to be useful as a spectral filter for Stokes scattering on the cesium D2 line, a free spectral range of ∆f = 2 × 9.2 GHz is required, so that transmission of the Stokes light is accompanied by rejection of the Raman pump light (see Figure 10.14). Calcite has the largest birefringence available, with |no − ne | = 0.17, but 10 cm of calcite are still required to produce the required free spectral range! We use three pieces of calcite, each around 3 cm long, mounted in series, along with a pair of Foctek Glan-laser polarizers (see previous section). The filter contrast is limited to 99%, however, because of phase distortions arising from the surface roughness of the calcite faces. The Lyot filter can be tuned in frequency over one free spectral range by tilting one of the calcite crystals slightly, which alters the optical path length. 10.13.3 Etalons Since the Lyot filter does not have sufficient contrast for filtering the Stokes light, we ordered some custom Fabry-Perot etalons. These are fixed air-gap etalons — pairs of optically flat glass plates that form a planar cavity with a discrete transmission spectrum. The free spectral range of the etalons depends on the separation L of 10.13 Filtering 360 (a) (b) optic axis Fringes B1 P2 P1 Retarder B2 Figure 10.13 Lyot filter. (a): The filter consists of a piece of birefringent material of length L — a retarder — placed between polarizers P1 and P2. The optic axis of the retarder is aligned at 45◦ to P1, so that an incident pulse of wavelength λ (purple) is split into ordinary and extraordinary polarization components (red and blue pulses). One component is delayed with respect to the other, because the refractive indices no , ne associated with each component are different. The phase retardance is given by φ = 2π|no − ne |L/λ. When the components recombine at P2, they interfere, producing sinusoidal spectral fringes, with a free spectral range ∆f = c/L|no − ne |. (b): Analogy with a Mach-Zender interferometer, in which two delayed pulses are mixed on a beamsplitter, producing fringes. the plates as ∆f = c/2L, since 2L is the distance an optical wave must traverse to make a round-trip of the cavity. Requiring a free spectral range of 18.4 GHz, as in the case of the Lyot filter, sets the etalon plate separation to be 8.2 mm. As mentioned in §10.6 below, the laser bandwidth is around 1.5 GHz, so the width of the pass band should not be smaller than this, in order to transmit the full Stokes spectrum (which follows the Raman pump spectrum). The maximum finesse is then F = 18.4/1.5 ∼ 12. The finesse of a Fabry-Perot etalon is fixed by the reflectivity R of the plates comprising the cavity, according to the formula √ π R . F= 1−R (10.22) 10.13 Filtering 361 Pump Stokes 9.2 GHz Figure 10.14 Stokes filtering. A Lyot filter with a free spectral range of 18.4 GHz can be used to suppress the Raman pump, while transmitting the weaker Stokes field. The finesse F = ∆f /δf , where δf is the FWHM of the pass band, is only 2, so this filter is far from ideal. The above considerations therefore fix the plate reflectivity to be 78%. This limits the out-of-band extinction to around 95%, so one of these etalons is not quite as good a filter as the Lyot filter. However, they are considerably more convenient, being very easy to align, and much smaller! Six custom etalons with these specifications, and a large clear aperture of 15 mm, were ordered from CVI Melles Griot. Initially it was intended to concatenate several etalons together, in series, so as to combine their extinction. However, when the etalons are placed one after another, their behaviour becomes complicated by interference effects arising from reflections between the etalons. This is mitigated somewhat because the etalons we use have anti-reflection coatings on their outside faces, which are also deliberately wedged. Nonetheless it can be problematic to use the etalons consecutively. 10.13 Filtering 10.13.4 362 Spectrometer Most recently, a grating spectrometer was built to filter the Stokes light (this was done by Virginia Lorenz and Klaus Reim). The resolution required is R = λ/δλ > 852/0.02 ≈ 106 . The resolution of a grating spectrometer is limited by the number of lines ruled on the grating, R ≈ N . The spacing between consecutive lines cannot fall below half the optical wavelength, or the first order of diffraction does not exist. For light at λ = 852 nm, this limits the maximum groove frequency to less than 2300 lines mm−1 . Using this groove frequency, a grating roughly 40 cm across is required. Such a large grating is not available, but a spectrometer with a grating around 6 cm across has been built, using a large off-axis parabolic mirror to focus the diffracted light onto a photon-counting CCD camera. Although the resolution of this spectrometer is sub-optimal, the ability to visualize the spectrum of the light downstream from the cesium cell makes it an extremely useful apparatus. All the above techniques have been used in an attempt to observe strong Stokes scattering from the cesium cell. So far we have not been successful, but our combined filter contrast is still improving. 10.13.5 Spatial filtering The utility of spatial filtering for a quantum memory is discussed in §6.3.2 of Chapter 6. In a quantum memory, where both signal and control beams are directed by the experimenter, it is possible to introduce a small angle between the beams so that the strong control pulse can be blocked after the cell. When looking for stimulated 10.13 Filtering 363 Stokes scattering however, it is necessary to look along the axis of the Raman pump, since the Raman gain is restricted to this direction. Therefore it is not generally possible to spatially filter Stokes scattering, although attempting to detect scattered light at very small angles from the pump direction may be sensible. A pinhole, or even an optical fibre — fibre in-coupling is very spatially selective — can be used; we are investigating these possibilities. It is possible, however, to look for Stokes scattering in the backward direction. The duration of our pulses is sufficiently long that they extend, in space, over the length of the cell: cTc ∼ L. This means that a Stokes photon emitted near the exit face of the cell in the backward direction, while illuminated by the leading edge of the pump pulse, can propagate backwards, nearly all the way to the entrance face of the cell, while all the time the cell remains illuminated by the pump, so that the photon experiences Raman gain all along the cell’s length (see part (a) of Figure 10.15). The experimental set-up shown in part (b) of Figure 10.15 has been used to look for this type of Stokes scatter. Strong stimulated fluorescence has been observed, and the signal is extremely clear, because there is no need to filter out the control pulse: it is propagating in the opposite direction! However the search is still on for Stokes scattering, the signature being that the Stokes signal should tune in frequency as the pump frequency is tuned, rather than remaining at the D2 resonance, as it does currently. Similar arguments would suggest that a Raman quantum memory in which the control and signal pulses are counter-propagating might be efficient, as long as the 10.14 Signal pulse 364 (a) (b) Figure 10.15 Backward Stokes scattering. (a): If the spatial extent of the Raman pump pulse is comparable to the cell length, a backward scattered Stokes photon experiences Raman gain as it propagates backward. (b): A polarizing beamsplitter can be used to deliver the control pulse, and separate the orthogonally polarized stimulated Stokes light that is scattered backwards. This essentially eliminates the background from the Raman pump. We are currently looking for Stokes light using this method. pulses are sufficiently long. The propagation theory of Chapters 4 and 5 is not applicable in this situation, but the numerical model presented in Chapter 6 can be used to show that the memory efficiency indeed remains high in this case. Such an arrangement removes the demanding filtering requirements, and is therefore an appealing possibility. A classical optical memory based on off-resonant Brillouin scattering using counter-propagating pulses in an optical fibre has in fact been implemented [186] , so there is some precedent for this approach. 10.14 Signal pulse Before implementing a single-photon quantum memory, a proof-of-principle demonstration using a weak coherent pulse for the signal field is planned. A single-photon 10.14 Signal pulse 365 source for the signal field is not, therefore, of immediate concern. However it remains challenging to generate the signal field. Currently we have only one laser oscillator. It is therefore necessary to generate the signal pulse from the control. The idea is to sample a small portion of the control field, perhaps using a beamsplitter, and then to apply a frequency modulation to shift the carrier frequency of the sampled pulse by 9.2 GHz, so that this frequency shifted pulse can act as the signal. Raman modulator One way to achieve this frequency modulation is to use the cesium atoms themselves, by inducing strong Stokes scattering of a control pulse. The Stokes sideband can then be sent into a second cesium cell along with the remaining control, where storage can be implemented. However, the Stokes scattering process is somewhat aleatoric: large fluctuations in the intensity of the Stokes light may make it difficult to assess the reliability of the memory. EOM A second possibility is to use an Electro-Optic Modulator (EOM) to apply the frequency shear. This is a device containing an electro-optically active crystal, whose refractive index can be altered by the application of an external voltage. Subjecting the crystal to a sinusoidally varying potential with frequency Ω will imprint the waveform onto the phase of an optical wave passing through the crystal, E(t) −→ E(t) × eiφ sin(Ωt) , (10.23) 10.15 Planned experiment 366 where φ quantifies the amplitude of the phase modulation. Fourier transforming (10.23) reveals the presence of sidebands, separated by the modulation frequency, e E(ω) = ∞ X e + kΩ), sgn(k)|k| J|k| (φ) E(ω (10.24) k=−∞ where the sideband amplitudes are given by Bessel functions. With the choice φ ∼ 0.2, the energy in the first sideband represents around 2% of the total transmitted energy, while all higher order sidebands may be neglected. Use of such a modulator allows one to reproducibly generate a weak signal pulse with the correct frequency shift from the control pulse. A modulator with the ability to operate in the microwave X-band at 9.2 GHz has been ordered from New Focus. 10.15 Planned experiment The current experimental set-up is in a state of flux, we are continually improving our filtering contrast and optical pumping efficiency, in the hope of detecting the strong stimulated Stokes signal that indicates there is sufficient Raman coupling to implement a Raman memory. Plans for the final implementation of the memory are tentative, being contingent on the eventual success of these early stages. However in the spirit of optimism we present in Figure 10.16 below a schematic of the possible layout of an experimental demonstration of a cesium Raman memory. The optical pumping beams are not shown, for clarity, although of course efficient optical pumping is critical. The off-axis geometry for phasematched retrieval described in Chapter 10.15 Planned experiment 367 6 is used, and we assume that the atoms have been prepared by optical pumping in the upper F = 4 state for this purpose. The angle between the control and signal beams is around 2◦ , and we assume that the control is more loosely focussed than the signal (the signal focus can be tightened by expanding the signal beam before it enters the confocal system, but we have not shown the beam expander). Since we seek only to demonstrate the feasibility of the memory, a long memory lifetime is not important, and so the delay between the storage and retrieval control pulses is adjusted by a mechanical delay stage. It is only feasible to move such a stage over a few feet, which corresponds to just a few nanoseconds of variability in the memory storage time, but this is sufficient for our purposes. The quarter wave plate in the control beam following the cell rotates the control polarization through 90◦ (since the control beam traverses it twice, the combined effect being that of a half-wave plate). The rotated control then retrieves the signal field into the orthogonal polarization mode to the polarization mode of the incident signal field. This allows the retrieved signal to be re-directed to a detector using a polarizing beamsplitter. The efficiency of the memory can be quantified by comparing the energy in the retrieved signal field to that of the incident signal. The incident signal pulse is generated from the control using an EOM (see §10.14 above), and an etalon removes the fundamental (the unmodulated light transmitted through the EOM). A Pockels cell is used to reduce the repetition rate of the Ti:Sapphire laser, as described in §10.7. We look forward to overcoming our present difficulties and assembling the above apparatus, or a variation thereupon, in the near future. 10.15 Planned experiment 368 Delay stage Ti:Sapphire Pulse Picker L1 Block E1 Cs cell L2 QWP EOM E2 Retrieved signal Figure 10.16 A possible design for demonstration of a cesium quantum memory. The pulse train from a Ti:Sapphire oscillator passes through a pulse picker to reduce the pulse repetition rate. Consider a single pulse. A beamsplitter redirects a portion of the pulse into an EOM, and the first sideband, shifted by 9.2 GHz is isolated from the output using a Fabry-Perot etalon E1 (see §10.13.3). This is the input signal field. The remainder of the initial pulse is used as the control field. It is directed through the cesium cell at a small angle to the signal pulse using a confocal arrangement (lenses L1 and L2): the signal is (hopefully) stored in the cell. The transmitted control field is sent through a variable delay line, and its polarization is rotated through 90◦ , before being sent back through the cell. The stored signal field is retrieved with the orthogonal polarization to the incident signal, and is sent by a polarizing beamsplitter through an etalon E2 (to remove any residual control) to a detector. Note that we have not shown the optical pumping beams, which are critical. This concludes the thesis. In the next chapter we summarize the results of the present research. Chapter 11 Summary This thesis has been concerned with the problem of storing light. I can recall feeling some puzzlement, lying in bed on a school night, at how completely my room darkened when the light was switched off. Why did you have to keep pouring more and more light into a room? Well, light is an ephemeral beast. But the possibility of a material that remembers its illumination — not with the feeble pallor of glow-inthe-dark paint, but with the unmistakable vigour of a laser pulse — is remarkable. I am not the first to study such media, and the current research was undertaken in the aftermath of the successes of light stopping by EIT. The main contributions of this thesis are theoretical: a fairly general framework for the analysis and optimization of light storage has been developed. The framework provides a unified description of both EIT and Raman storage, and generalizes to tCRIB, lCRIB, AFC and broadened Raman protocols. Further applications of the formalism are expected. The use of the SVD has been crucial to the success of the theoretical programme. Many 370 of the results in the thesis are simple adaptations of well-known facts from linearalgebra to the particular case under study. Another important ingredient of the thesis is numerical simulation. The propagation of optical fields through an atomic ensemble is always described by a set of coupled linear partial differential equations, and these are particularly easy to solve on a modern computer. The ‘take-home’ results are as follows. 1. Any quantum memory is a linear system, with storage and retrieval interactions described by Green’s functions, which are essentially large matrices. 2. The SVD of the Green’s functions provides a complete characterization of the memory. It allows one to immediately identify the optimal input mode. 3. The singular values of the Green’s function are invariant under unitary transformations. This fact can be very useful in analyzing the memory interaction. One can work in the Fourier domain, or indeed in an entirely unfamiliar coordinate system. 4. The efficiency of a memory is limited by its optical depth [133] . 5. Explicit expressions for the Green’s functions describing storage in a Λ-type atomic ensemble are provided. The Rosen-Zener kernel holds whenever the control has a hyperbolic secant temporal profile. The adiabatic kernel holds for all detunings and control pulse shapes that satisfy the adiabatic approximation. The Raman kernel holds for adiabatic interactions that are far-detuned and ‘balanced’. 371 6. In the adiabatic limit, a coordinate transformation links the results for different control profiles. Therefore the SVD only needs to be computed once for a given control pulse energy. Changes to the control profile simple change the coordinate transformation. 7. A Raman memory may be characterized as a multimode beamsplitter interaction between optical and material modes. A single set of modes describes both the transmitted fields and the stored excitations. A single number C characterizes the efficiency of a Raman memory. 8. The Green’s function can always be constructed numerically, so that the optimal input modes and memory efficiencies can be found at any detuning, regardless of whether or not the interaction is adiabatic. 9. Retrieval of the stored excitations can be problematic. Forward retrieval suffers from re-absorption losses. Backward retrieval is not phasematched if the ground and storage states are non-degenerate. Numerical simulations verify that an off-axis geometry allows for efficient backward retrieval with nondegenerate states. Loose focussing of the control field is desirable. 10. The multimode capacity of a quantum memory can be evaluated by considering the SVD of the Green’s function. The multimode scaling of EIT, Raman, tCRIB, lCRIB, AFC, and a broadened Raman protocol is studied. Unbroadened ensembles have poor multimode scaling. Adding an inhomogeneous broadening improves the scaling. The AFC protocol has the best multimode 11.1 Future work 372 scaling of all the protocols studied. 11. If one is not able to shape the signal pulse, optimal storage can still be achieved by instead shaping the control. But to solve the optimization problem, the equations of motion for the memory must be solved numerically. This can be done rather quickly, however. A simple optimization algorithm works well for finding the optimal control profiles. The SVD allows one to verify the optimality of the numerical solutions. This optimality suffers as the interaction becomes less adiabatic. 12. A Raman memory in bulk diamond, based on the excitation of optical phonons, is feasible. It is shown that the equations of motion describing the Raman interaction in diamond have precisely the same form as the Raman equations describing storage in an atomic vapour. The form of the coupling constant C is derived. 13. Attempts to implement Raman storage in cesium vapour have been made, but it is proving difficult even to generate and detect Stokes scattering. I am not a good experimentalist! 11.1 Future work There is a great deal of experimental work to do. It may be that there is a good reason why our attempts to build a Raman memory have been unsuccessful: we should either make the memory work, or find this reason, in the coming year. 11.1 Future work 373 An intriguing theoretical challenge is how to make use of the stored excitations once they are in place. Is it possible to perform computationally interesting operations on the spin wave in an atomic ensemble? Can operations be designed that allow different stored modes in a multimode memory to interact? Work on this front has begun in the literature [17,187,188] , but this is likely to be a rich seam. If you have survived this far, I am very grateful for your attention! I hope that some of the results in this thesis are useful to other workers in the field, even if only as a warning of what not to try. Appendix A Linear algebra Physics is generally concerned with change: the evolution of a system over time, or the response of a system to an external agent. The easiest, and most uninspiring situation to analyze, is when there is no change and no response. Linear algebra is concerned with the much more interesting situation arising when the response depends linearly on some parameter. On the one hand, this is almost always an approximation that only holds for small changes in a parameter, and small responses. So linear algebra rarely provides an exact description. On the other hand, the linear approximation can be successfully applied to nearly every physical system! Linear algebra is therefore useful in almost every branch of physics, as well as in mathematics and science generally. In our case, the linear response of a quantum memory is certainly an approximation, valid when the signal field does not contain too many photons. Here we summarize various concepts that are needed to properly understand the A.1 Vectors 375 singular value decomposition as it pertains to the optimization of quantum storage. There is a significant overlap with the formalism of quantum mechanics, so we will also take this opportunity to review some aspects of that formalism. A clear and comprehensive introduction can be found in Nielsen and Chuang’s quantum information bible [158] . A.1 Vectors A vector is essentially a list of numbers. It also helps to keep in mind the image of a vector as an arrow (see Figure A.1). This analogy cannot always be made with rigour, but it provides a convenient visualization. The numbers comprising a vector are the components of the arrow along the coordinate axes. When writing down these components, implicit reference is therefore always made to some coordinate system. It’s clear that we could rotate the coordinate axes — altering the vector’s components — without changing the arrow in Figure (A.1), and in this sense, a vector transcends its components. Nonetheless, it will be useful to write down the vector components for concreteness. We will use two equivalent sets of notation for a vector labelled ‘v’. Either v, or |vi. The first symbol, in bold face, is in general use. The second — a ket — is an example of ‘Dirac notation’, used only in the context of quantum mechanics. Dirac notation is at times very convenient, and it will help to be able to use these two types of notation interchangeably. The number of components of a vector is called the dimension of the vector. If A.1 Vectors 376 Figure A.1 A vector. On the left is a representation in ‘component form’. On the right the same mathematical object is drawn as an arrow. The direction and length of the arrow are determined by its components α, β. Implicitly, a coordinate system (thinner arrows) is used to define the components. there are n components, the vector is said to be n-dimensional. We will adopt the convention that v is a column vector; v = |vi = v1 v2 . .. . vn (A.1) Vectors can be added together, provided they have the same dimension, v 1 + w1 v +w 2 2 v + w = |vi + |wi = .. . v n + wn . (A.2) A.1 Vectors 377 And a vector can be multiplied by a number, say α, like this αv1 αv 2 αv = α|vi = . . . . αvn (A.3) Using these operations, vectors can be combined together to make new vectors, and it is convenient to think of all these possible vectors as inhabiting a ‘space’, known as an n-dimensional vector space. Our own universe, with three spatial dimensions, can be thought of as a 3-space. A.1.1 Adjoint vectors The components of a vector do not have to be real numbers. In quantum mechanics, and many other applications, they are generally complex numbers. A useful concept in this case is the Hermitian adjoint v † of a vector v. This is simply another vector, this time a row vector, with each component equal to the complex conjugate of the corresponding component of v, v † = hv| = v1∗ v2∗ ... vn∗ . (A.4) The symbol hv| is known, rather unfortunately, as a ‘bra’, for reasons that will become clear. These row vectors (bras) can be added or multiplied by numbers in the same way as column vectors (kets), and so they form their own vector space, A.1 Vectors 378 sometimes known as the adjoint space. Every vector v has a corresponding Hermitian adjoint v † ; every ket has its corresponding bra. And Hermitian conjugation is involutive: The Hermitian adjoint of v † is v again. A.1.2 Inner product Vectors can be multiplied together in ways as various as mathematicians are inventive. The inner product — sometimes scalar product — is defined as the sum of the component-wise products of a bra and a ket with the same dimension, v † w = hv|wi = v1∗ w1 + v2∗ w2 + . . . + vn∗ wn . (A.5) This type of product between two vectors is not another vector; it’s just a number. For some reason it’s rather satisfying to take an inner product, and Paul Dirac’s notation anticipates something of this satisfaction. When a bra hv| encounters a ket |wi they merge to become a ‘braket’ hv|wi, and so fulfill their destiny. It’s quite common to speak of taking the inner product of two kets, |vi and |wi. In this case it is understood that one of the kets has to be replaced by its corresponding bra before using (A.5). Note that hv|wi = hw|vi∗ , so one should be consistent about which of the two kets is replaced. A complex vector space with an inner product defined as in (A.5) is known as a Hilbert space. Quantum mechanics is a theory about vectors in Hilbert space; as such it is extremely simple. It is the reconciliation of this mathematical structure A.1 Vectors 379 with what we know about the real world that makes the theory so difficult to pin down. A.1.3 Norm Having defined the inner product, we can now define the norm of a vector. This is defined as the square root of the inner product of a vector with itself, √ v = ||v|| = || |vi || = v†v = p hv|vi. (A.6) This is a positive, real quantity that grows with the size of the components of v. In fact, substituting in the definition (A.5) and applying Pythagoras’ theorem shows that the norm is simply the length of the arrow representing the vector v. Some further geometrical manipulations reveal that the inner product is related to the angle θ between the arrows representing two vectors, as follows (see Figure A.7), v † w = hv|wi = vw cos θ. (A.7) An immediate consequence of this is that the inner product hv|wi vanishes when θ = π/2, that is, when v is perpendicular to w. Often, the word orthogonal is used instead of perpendicular, especially in the case that the vectors involved are complex, or when they have dimension greater than 3, since then the notion of an angle is less transparent. A.1 Vectors 380 Figure A.2 A.1.4 The inner product of two vectors. Bases We have already mentioned in passing the concept of a coordinate system, with respect to which the components of a vector are defined. We drew the coordinate axes in Figure A.1 as black arrows. The axes themselves are therefore described by a pair of vectors, one pointing along the x-axis; the other along the y-axis. Let’s call them |xi and |yi. If we fix the length (the norm) of these vectors as 1, then we can write any vector v directly in terms of |xi and |yi, vx = vx |xi + vy |yi. |vi = vy (A.8) The set of two vectors {|xi, |yi} is a basis from which we can construct any other 2-dimensional vector. In fact, it’s clear that any two vectors, as long as they point in different directions, can serve as a basis. The nice feature of the set {|xi, |yi} is that these two vectors are orthogonal to each other, hx|yi = 0. This is particularly convenient, because the components of |vi can be found directly by taking inner products, vx = hx|vi, vy = hy|vi. A basis of this kind is almost always preferable to non-orthogonal bases. Such a basis, with mutually orthogonal basis vectors of unit A.2 Matrices 381 norm, is called an orthonormal basis. When we speak of a coordinate system, or coordinate axes, we are implicitly making reference to an orthonormal basis. A.2 Matrices A matrix is essentially an array of numbers, laid out on a rectangular grid, as follows: M11 M12 . . . M1n M M22 . . . M2n 21 M = . .. .. .. . . . . . Mm1 Mm2 . . . Mmn (A.9) The numbers Mij comprising a matrix are known as its elements. The dimension of a matrix is specified by two numbers, the number of rows, and the number of columns in the matrix. In the example (A.9) M has dimension m × n. Just as with vectors, matrices are greater than the sum of their parts: the actual values of the elements of a matrix are not important. To visualize a matrix, one should imagine a process, in which a vector is transformed into another vector (see Figure A.3). For this reason, matrices are sometimes referred to as maps, since they map one vector onto another. The term operator is also used, since a matrix can be viewed as an operation — rotation, or reflection, say — applied to a vector. The way this operation is performed mathematically is via matrix multiplication, written like this, w = M v, or |wi = M |vi. (A.10) A.2 Matrices 382 This multiplication is evaluated by combining the elements of M and the components of v to form w, in the following way. Define a set of column vectors M1j M 2j mj = |mj i = . . . Mmj , (A.11) so that each column of the matrix M is given by one of these vectors, M = m m ... 1 2 m . n (A.12) The vector w is then given by a weighted sum of the mj , with coefficients equal to the components of v, w = v1 m1 + v2 m2 + . . . + vn mn . (A.13) From (A.13) it is clear that the number of columns of M must be the same as the dimension of v for this multiplication to be possible. Thus the number of columns of M sets the dimension of the vectors upon which M can act. Similarly the number of rows of M sets the dimension of the vectors mj , and therefore the dimension of the output vector w. So an m × n matrix is an operator that acts on n dimensional A.2 Matrices 383 vectors to produce m dimensional ones. Figure A.3 A matrix acting on a vector. Here M maps the initial vector v (black) onto the final vector w (red), via a rotation and a ‘dilation’ (length increase). The values of the matrix elements Mij depend on the components of v and w, which in turn depend on the direction of the coordinate axes (thin arrows). Rotating the coordinate axes would change the Mij , but the transformation represented by M would be the same. In this sense, a matrix is more fundamental than its elements. Matrices with the same dimensions can be added together; M + N is just the matrix whose elements are given by the sum of the corresponding elements of M and N . And of course they can be multiplied by numbers. αM is a matrix whose elements are αMij . Incidentally, these properties mean that the space of all matrices is actually also a vector space. But this will not be important for us. Two matrices can be multiplied together to produce a new matrix. In the product M N = Q, each of the column vectors qj of Q are formed from the corresponding column vector nj of N , by combining the column vectors mj of M in a weighted sum like (A.13), with coefficients given by the components of nj . So M acts on each column of N to produce the columns of Q. To multiply a row vector (or a bra) by a matrix, we simply treat the row vector A.2 Matrices 384 as a 1 × n matrix, and apply the above rule. We therefore obtain hv|M = hv|m1 i hv|m2 i . . . hv|mn i . (A.14) Matrices also have Hermitian adjoints. The Hermitian adjoint M † of M is given by swapping the rows and columns of M , and taking the complex conjugate of all its elements, ∗ M11 M∗ 12 † M = . . . M1n∗ ∗ M21 ... ∗ M22 ∗ . . . Mm2 .. . .. . ∗ Mm1 .. . ∗ ∗ M2n . . . Mmn . (A.15) Using this definition, it’s easy to check that Q† = N † M † (note the reversed order of M and N ), and that (M |vi)† = hv|M † . These facts are useful when manipulating expressions involving several matrices. A.2.1 Outer product The outer product of two vectors |vi and |wi is a matrix, written as v1 w1∗ v1 w2∗ ··· v w∗ v w∗ · · · 2 1 2 2 † |vihw| = vw = . .. .. . . . . vn w1∗ vn w2∗ · · · ∗ v 1 wm ∗ v 2 wm .. . ∗ v n wm . (A.16) A.2 Matrices 385 Each column of this matrix is just |vi, multiplied by the corresponding element of hw|, so that its structure can be visualized as that of a row vector hw| with column vectors |vi ‘hanging’ from it. The Dirac notation is very satisfying in this context, since the result of applying the operator |vihw| to a third vector |xi is written like this, |vihw||xi = |vihw|xi = hw|xi|vi. (A.17) That is, the matrix product of |vihw| with |xi is just the same as the inner product of hw| and |xi, multiplied by the vector |vi. If |xi = |wi, the result is w2 |vi. As |xi deviates away from |wi, the inner product hw|xi gets smaller and smaller, until it vanishes, when |xi is orthogonal to |wi. A natural interpretation for the operator (A.16) is therefore as a kind of ‘switch’ that maps an input from |wi to |vi. Operators of this kind are sometimes known as flip operators, or transition operators, in quantum mechanics. Breaking down larger operators into flip operators can often provide valuable insights. A.2.2 Tensor product A further generalization of the outer product is the tensor product. The tensor product is used to combine vector spaces together to produce a new, larger space. Suppose we have an n-dimensional vector space V, and also an m-dimensional space W. The tensor product V ⊗ W of these two spaces would be the space of all vectors with dimension nm. Vectors v and w from the smaller spaces can be combined together via the tensor product to produce a vector v⊗w inhabiting the larger space. A.2 Matrices 386 And similarly matrices M and N acting on the smaller spaces can be combined together to produce an operator M ⊗ N , that acts on vectors in the tensor product space V ⊗ W. The result of applying the combined operator to the combined vector is the same as the result of applying the operators to the vectors separately, and then taking the tensor product: (M ⊗ N )(v ⊗ w) = (M v) ⊗ (N w). (A.18) A common example arising in quantum mechanics is the tensor product of a pair of 2-dimensional vectors, representing the state of a pair of qubits (a pair of electron spins perhaps). Suppose one qubit is in the state labelled |0i, and the other is in the state |1i. The combined state |ψi of both is found by taking the tensor product of these two vectors, |ψi = |0i ⊗ |1i. Sometimes the more compact notation |01i is employed, where the meaning should be clear from the context. But other 4-dimensional vectors, which cannot be represented as tensor products of 2-dimensional vectors, can exist in the 4-dimensional tensor product space. For √ instance, the vector |ψi = (|01i + |10i)/ 2 is a valid state in quantum mechanics (see (1.7) in Section 1.6.4 of Chapter 1). It cannot be written in the form |state 1i ⊗ |state 2i, but it is a 4-dimensional vector, produced by adding together two vectors that can be written in this form. Vectors of this kind, that exist in the tensor product space, but cannot be written as a tensor product of vectors from the component spaces, are known as non-separable. In quantum mechanics, they A.2 Matrices 387 represent states that are entangled. The tensor product of two matrices M (with dimension m × n) and N (with dimension p × q) is found by ‘attaching’ a copy of M to each element of N , as shown below, N11 M ... 11 Np1 M ⊗N = N11 . Mm1 .. Np1 ··· .. . ··· N1q .. . Npq .. . ··· .. .. . ··· N11 · · · . .. M1n . .. Np1 · · · . ··· N1q .. . ··· Npq Mmn N1q .. . Npq .. . . N11 · · · N1q .. .. .. . . . Np1 · · · Npq (A.19) The procedure for vectors is identical; the vectors are just treated as matrices with a single column (or row, in the case of bras). A bit of head scratching will verify that this definition, when combined with standard matrix multiplication (A.13), satisfies the requirement (A.18). An important property of the tensor product is as follows. If {|ii} is an orthonormal basis for one space, and {|ji} is an orthonormal basis for a second space, the set of tensor product vectors {|ii ⊗ |ji} is an orthonormal basis for their tensor product space. A.3 Eigenvalues A.3 388 Eigenvalues Consider a matrix R representing a reflection about the x-axis, as shown in Figure A.4. A vector |1i lying along the x-axis is not changed by the action of this matrix. That is, it is its own reflection. So we have R|1i = |1i. A second vector |2i lying along the y-axis is flipped around by R. Its reflection points in the opposite direction to itself, so R|2i = −|2i. Other vectors are altered in more complicated ways when they are reflected, so that the vector resulting from the application of R is not related to the original vector by a simple numerical factor (1 or −1 in the two cases above). The vectors |1i and |2i are examples of vectors for which the action of R is the same as multiplication by a number. These ‘special’ vectors are known as eigenvectors of R. In general, any matrix M has a set of eigenvectors {|ii}, such that M |ii = λi |ii. (A.20) Here the number λi is the eigenvalue corresponding to the eigenvector |ii. For the example given above, we had λ1 = 1 and λ2 = −1. The eigenvectors and eigenvalues contain all the information required to reconstruct the transformation implemented by M ; they represent the essence of a transformation, and as such they are of paramount importance in linear analysis, and central to quantum mechanics. A.3 Eigenvalues 389 Figure A.4 Eigenvectors and eigenvalues. The matrix R represents reflection in the x-axis (horizontal axis). The eigenvectors of R are those vectors pointing along, or perpendicular to the x-axis, since the application of R to one of these vectors produces the same vector again, multiplied by some number. A.3.1 Commutators In general, matrix multiplication is not commutative. That is, M N 6= N M ; the order in which matrices are multiplied is important. This makes sense when matrices are viewed as representing transformations of vectors (see Figure A.5). Often it is useful to examine the commutator of two matrices, defined by [M, N ] = M N − N M. (A.21) If M and N were just numbers, their commutator would always vanish, but for matrices often it does not. In quantum mechanics, the commutator of different physical quantities may be non-zero, and this non-vanishing of the commutator can be viewed as the source of a great many of the strange features of quantum mechanics. A.3 Eigenvalues 390 3 2 2 1 1 3 Figure A.5 Non-commuting operations. Here M represents a reflection around the y-axis, while N is an anti-clockwise rotation through 90 degrees. The red arrows are numbered in order, with 1 the initial vector, 2 the result after the application of one of the transformations, and 3 the result after both transformations have been applied. On the left, M is applied first, and then N . On the right, N is applied first, followed by M . The results are different because the matrices M and N do not commute. The notation can be counter-intuitive: the product M N represents the application of N first, with M applied afterwards. If the two matrices N , M do commute, then they have common eigenvectors. To see this, suppose that |ui is an eigenvector of N , with eigenvalue λ. If we take the product M N |ui (that is, we apply N to |ui first, and then M ), the result is simply λM |ui. On the other hand, if [M, N ] = 0, we can swap the order of M and N , to get N M |ui. That is, N (M |ui) = λ(M |ui). (A.22) Therefore the vector M |ui is also an eigenvector of N , with the same eigenvalue λ. M |ui must be parallel to |ui, so that M |ui = µ|ui. That is, |ui is also an eigenvector of M , with some new eigenvalue µ. This fact is intimately connected with the epistemology of quantum mechanics. A.4 Types of matrices A.4 391 Types of matrices There are some types of matrix that are particularly important, both for the calculations in this thesis, and for quantum mechanics generally. A.4.1 The identity matrix The identity matrix, often denoted by I, is the matrix equivalent of the number 1. It is the matrix that results in no change when it is multiplied by another matrix — it represents the operation ‘doing nothing’. That is, IM = M I = M . And of course the identity does not change a vector either, I|vi = |vi, hv|I = hv|. The identity matrix is a square matrix (i.e. dimension m×m), with ones along its main diagonal, and zeros everywhere else (the zero elements are left blank below to avoid clutter), 1 I= 1 .. . 1 . (A.23) Sometimes care should be taken to ensure that the correct dimension m of I is used, so that the multiplication is possible. Usually this is quite clear from the context, but the symbol Im can be used when the size of I needs to be specified. A.4 Types of matrices A.4.2 392 Inverse matrix The inverse M −1 of a matrix M is the matrix that ‘undoes’ the action of M . It is the matrix equivalent of a reciprocal. The inverse satisfies the relations M −1 M = M M −1 = I. It is clear that taking the inverse of a matrix is also involutive, since the inverse of an inverse is just the original matrix, (M −1 )−1 = M . If M is rectangular, with m < n, then M describes a map from a larger space into a smaller space, so that some information is inevitably lost, in the sense that there are different vectors in the input space that are mapped to the same vector in the output space. Therefore M cannot have an inverse — it is impossible to ‘undo’ this type of map. It is generally the case that only square matrices, with m = n, have a matrix inverse. It is possible to define a pseudo-inverse, that represents the closest approximation of a true inverse, for any matrix (even rectangular ones), but we will not make use of the pseudo-inverse [189–192] . Calculating the inverse of a matrix can be rather involved, and although an algorithm for inverting 3 × 3 dimensional matrices is taught to students in school, matrix inversion is rarely performed explicitly. Lloyd N. Trefethen is a prominent numerical analyst who teaches a course on computational linear algebra at Oxford University. His reaction to a suggestion that students should consult W. H. Press’s famous book on numerical techniques was The only way to annoy a numerical analyst more than by inverting a matrix, is to use Numerical Recipes. A.4 Types of matrices 393 The formula for the inverse of a 2 × 2 matrix M is simple however, −1 a b M −1 = c d = d −b 1 . ad − bc −c a (A.24) The quantity ad−bc is known as the determinant of the matrix — sometimes denoted by vertical bars, |M | — since it determines whether or not the inverse of M exists: if |M | = 0, the formula for the inverse ‘blows up’. In this case, the matrix does not have in inverse, which implies that there exists some vector |vi such that M |vi = 0. Clearly it is not possible to invert this expression. This provides a convenient way to find the eigenvalues of a matrix. If we want to find |ui such that M |ui = λ|ui, then we must have that (M − λI)|ui = 0, and therefore we require that |M − λI| = 0. A.4.3 Hermitian matrices A Hermitian matrix is equal to its Hermitian adjoint, H = H † . It is the matrix equivalent of a real number, and in fact its eigenvalues are all real numbers. To see this, consider the quantity k = hi|H|ji. On the one hand, using the definition of H, along with (A.20), we have k = (H † |ii)† |ji = (H|ii)† |ji = (λi |ii)† |ji = λ∗i hi|ji. On the other hand, we have k = hi|(H|ji) = λj hi|ji. A.4 Types of matrices 394 Taking the difference of these, we get k − k = (λ∗i − λj )hi|ji = 0. If we set i = j, we must have that λ∗j − λj = 0, since hj|ji > 0 is the square of the norm of |ji. Therefore λj = λ∗j , that is, the eigenvalues of H are real numbers. At the same time, if we set i 6= j, we must have that hi|ji = 0, which means that different eigenvectors of H are all orthogonal to one another. Note that we are free to scale the eigenvectors |ii so that they have length 1. If we do this, the set of eigenvectors {|ii} of a Hermitian matrix is an orthonormal basis. The eigenvectors define a ‘natural’ coordinate system for the space of vectors upon which H acts. And it’s very practical to work with this coordinate system, since the effect of H on each basis vector reduces to multiplication by the corresponding eigenvalue, H|vi = H (v1 |1i + v2 |2i + . . . + vn |ni) = v1 λ1 |1i + v2 λ2 |2i + . . . + vn λn |ni. (A.25) A.4.4 Diagonal matrices A diagonal matrix D is a matrix with zeros everywhere except along its main diagonal, D11 D= . D22 .. . Dmm (A.26) Clearly diagonal matrices must always be square, with n = m. The identity matrix is a diagonal matrix with Djj = 1. Diagonal matrices are very easy to work with. A.4 Types of matrices 395 For example, the square of a diagonal matrix D2 = DD is another diagonal matrix 2 . The inverse D −1 of a diagonal matrix is just another with its elements equal to Djj diagonal matrix with all its elements equal to 1/Djj . Any two diagonal matrices commute with one another, [D1 , D2 ] = 0, and the eigenvalues of a diagonal matrix are just equal to its elements, λj = Djj , with its eigenvectors being the basis vectors of the coordinate system with respect to which the matrix elements are defined. This last property is important. Diagonal matrices are wonderfully simple to manipulate, but surely it is very unlikely that any interesting matrices are diagonal. The point is that all matrices are diagonal matrices (or more correctly, most square matrices), as long as you write them down with reference to the correct coordinate system! This coordinate system is the one defined by the eigenvectors of the matrix, and when written down using this basis, the elements of the matrix are just its eigenvalues. A brief inspection of (A.25) reveals that in fact, when written down with reference to the coordinate system defined by the eigenvectors |ii, H is actually diagonal, with elements H11 = λ1 , H22 = λ2 , etc... For this reason, the process of finding the eigenvalues and eigenvectors of a matrix is sometimes referred to as diagonalization, since this calculation is simply what is required to convert a matrix into a diagonal one. The eigenvalue decomposition can be written as M = W DW −1 , (A.27) where D is a diagonal matrix containing the eigenvalues of M , and where the eigen- A.4 Types of matrices 396 vectors of M comprise the columns of the matrix W . Very efficient algorithms exist for finding this decomposition; the results in this thesis rely heavily on the speed and precision of the LAPACK routines implemented in MATLAB. A.4.5 Unitary matrices A unitary matrix U is a matrix whose inverse is equal to its Hermitian adjoint, U −1 = U † . A unitary matrix represents a rotation in space, so that |wi = U |vi is a vector pointing in a different direction to |vi, but with the same norm — the same length. To see why, consider the norm of |wi, w2 = hw|wi = hv|U † U |vi. But since U † = U −1 , we have that U † U = I, so w2 = hv|I|vi = v 2 . That is, unitary matrices preserve the norm of vectors upon which they act. Figure A.6 A unitary transformation. U represents a rotation from an initial (black) into a new (red) coordinate system. The columns of U are unit vectors comprising an orthonormal basis for the new coordinate system. A rotation can be thought of as a transformation from one orthonormal coordinate system to another, as shown in Figure A.6. Associated with this new coordinate system is an orthonormal basis {|ii}, and these vectors are actually the columns of U , ui = |ui i = |ii. To see this, consider the product K = U † U . From the definition A.4 Types of matrices 397 of the Hermitian adjoint, the rows of U † are the bras hui |, † U = hu1 | .. . hum | . (A.28) Applying the matrix multiplication described in (A.13), we find that each element of the product matrix K is given by an inner product, Kij = hui |uj i. But since U is unitary, K = I, the identity, so that we must have hui |uj i = δij , where δij is the kronecker delta symbol (δij = 1 if i = j, and 0 otherwise). Therefore the column vectors |ui i form an orthonormal basis. If U is applied to a vector |vi pointing along the x-axis, with components v1 = 1, vj6=1 = 0, the result is |u1 i. In the same way, each coordinate axis is mapped by U to a new axis |ui i, and so U represents a rotation into a new orthonormal coordinate system defined by its columns. Incidentally, the above arguments serve to demonstrate that the inner product hu|vi of two vectors |ui and |vi is always independent of the coordinate system used for writing out the components of |ui and |vi. Changing the coordinate system is done by applying a rotation |ui → U |ui, |vi → U |vi, and the inner product is then hu|U † U |vi = hu|vi. Changing coordinates makes no difference. This is to be expected of course, since (A.7) makes no reference to any coordinates. It is worth noting that U † is unitary, if U is. Therefore the columns of U † also form an orthonormal basis, and so the rows of U form an orthonormal basis. Note also that the product of two unitary matrices U , V is also unitary: U V (U V )† = A.4 Types of matrices 398 U V V † U † = U IU † = I. Two rotations composed together can always be thought of as a single rotation. Unitary matrices play a central role in quantum mechanics, and we will encounter them in the optimization of quantum memories. Appendix B Quantum mechanics In this Appendix we give a brief review of the structure of quantum mechanics. This is intended as a pedagogical precursor to Appendix C, on quantum optics. We will make use of the concepts developed in Appendix A. Quantum Mechanics was developed in the early twentieth century, primarily as a theory of atomic physics. In the days before Google, interdisciplinary communication was more difficult, and in fact Werner Heisenberg re-invented matrices in order to formulate his version of quantum theory [193] . The incarnation we present here uses the notation introduced by Paul Dirac [194] , and we follow broadly the excellent account given by Nielsen and Chuang [158] . B.1 Postulates B.1 Postulates B.1.1 State vector 400 In quantum mechanics, the state of a system is described by a ket |ψi. The simplest vector is a 2-dimensional one, and this describes the simplest type of quantum system — a qubit. More complicated systems are described by higher dimensional vectors. B.1.2 Observables Quantities, like energy, momentum or position — any observable that might be measured — are represented by matrices that act on the state vector. These matrices are always Hermitian, and this guarantees that their eigenvalues are real numbers (see Section A.4.3 in Appendix A). In addition, the eigenvectors of Hermitian matrices form an orthonormal basis: they define a coordinate system. B.1.3 Measurements Quantum mechanics provides the following recipe for making predictions about measurements. The observable being measured is assigned to a Hermitian operator H. Making this assignment correctly is left up to the skill and imagination of the physicist. This operator is diagonalized, yielding its eigenvalues {λi } and eigenvectors {|ii}. The eigenvalues are real numbers, and each one represents a possible numerical outcome of the measurement: the number you might see on an oscilloscope screen, for example. Each eigenvalue λi is associated with an eigenvector |ii, and these eigenvectors define a coordinate system. The state vector of the system |ψi is B.1 Postulates 401 written with reference to these coordinates, known as the measurement basis, |ψi = ψ1 |1i + ψ2 |2i + . . . + ψm |mi. (B.1) The probability pi that the measurement yields the result λi is then given by |ψi |2 , the squared magnitude of the ith component of |ψi in the measurement basis. A more compact way to write this is pi = |hψ|ii|2 . (B.2) This is known as the Born rule, after Max Born who proposed it in 1926 [195] . Immediately after the measurement has been completed, the state of the system changes, essentially instantaneously, according to the measurement result, |ψi → |ii. This is known as the collapse postulate. The average value of the measurement result is often useful. This is sometimes called the expectation value of the quantity H, since it is the number one would expect when repeating the measurement many times. The expectation value is given by hHi = P i pi λi , and a bit of thought shows that this is equal to hψ|H|ψi. The fact that hHi = hψ|H|ψi is another convenience of Dirac notation. B.1.4 Dynamics It must always be the case that the probabilities pi sum to unity, P i pi = 1. This just codifies the assertion that we must always get some result from a measurement, even B.1 Postulates 402 if the result is ‘no signal’. Using the Born rule, this means that P 2 i |ψi | = hψ|ψi = 1. That is, the norm of a state vector in quantum mechanics is always exactly equal to 1. The norm can never be altered by any dynamical process, which immediately fixes all dynamics in quantum theory to be unitary. In other words, given some initial state |ψ0 i, and a final state |ψi, we must have |ψi = U |ψ0 i, (B.3) where U is a unitary operator that advances the system from the initial to the final state. Differentiating (B.3) with respect to the time t, we obtain the equation of motion ∂t |ψi = U̇ |ψ0 i = U̇ U † |ψi, (B.4) where the overdot indicates the time derivative of U . Now, the requirement that the norm of |ψi does not change can be expressed by the condition ∂t (hψ|ψi) = 0. Substituting in (B.3) gives hψ0 |U̇ † U + U † U̇ |ψ0 i = 0, (B.5) from which we derive the condition that the operator (U̇ U † ) is skew-Hermitian, meaning that it changes sign under Hermitian conjugation. Any skew-Hermitian B.2 The Heisenberg Picture 403 operator can be represented as the product of the imaginary unit i with a Hermitian operator H, and making this replacement in (B.4) gives us the Schrödinger equation ∂t |ψi = iH|ψi. (B.6) The operator H is known as the Hamiltonian. Schrödinger’s great insight was to identify H as the operator associated with the energy of the system. In (B.6) it is the energy, represented by H, that sets the rate of change of the state vector. Systems with high energy evolve quickly, with fast oscillations, while low energy systems are more sluggish. B.2 The Heisenberg Picture The above discussion was based on the so-called Schrödinger picture, in which the quantum state |ψi evolves in time. It is possible to formulate quantum mechanics differently, and sometimes it is easier to solve a problem by using this different formulation. The results are identical, regardless of how the calculations are done. In Heisenberg’s formulation, the quantum state |ψ0 i of a system at some initial time is fixed. It does not change with time. Instead, the operators acting on the state vector evolve in time. As an example, consider a Hermitian operator A associated with some quantity that we might want to measure. Here we use the symbol A instead of H; we reserve the symbol H for the Hamiltonian from now on. In the Heisenberg picture the operator A depends on the time at which we make the measurement. B.2 The Heisenberg Picture 404 In order for this formulation to work, we must have that the expectation value hAi predicted by either formalism is the same. Denoting the fixed operator in the Schrödinger picture with a subscript S, we must have hψ0 |A|ψ0 i = hψ|AS |ψi, ⇒ hψ0 |A|ψ0 i = hψ0 |U † AS U |ψ0 i, A = U † AS U. ∴ (B.7) That is, the time evolution of an operator in the Heisenberg picture is found by sandwiching the Schrödinger operator between two copies of U , the same operator that generates the time evolution of the state in the Schrödinger picture. Differentiating (B.7) with respect to time, we find ∂t A = U̇ † AS U + U † AS U̇ = U̇ † U A + AU † U̇ . (B.8) Note also that [U † , U̇ ] = 0, since U † U̇ = U † U̇ U U † = U † U U̇ U † = U̇ U † , (B.9) where we used the fact that [U, U̇ ] = 0 (a little thought shows that an operator must always commute with its derivative; see Section A.3.1 in Appendix A). Therefore (B.8) can be re-written in terms of the Hamiltonian, to produce the Heisenberg B.2 The Heisenberg Picture 405 equation ∂t A = i[A, H]. (B.10) This is the fundamental equation of motion in the Heisenberg picture; it plays the same role as the Schrodinger equation does in the Schrodinger picture — generating time evolution. B.2.1 The Heisenberg interaction picture Often we are interested in analysing the behaviour of a system when it is subjected to a weak external field. Of specific relevance in this thesis is the case of an atom illuminated by a laser: the internal electric fields generated by the charges within the atom are much stronger than the electric fields within the laser beam, so the laser acts as a weak external perturbation, on top of the much stronger interactions binding the atom together. In such cases, it is convenient to separate out the strong and weak contributions to the energy of a system. Suppose that we can divide the Hamiltonian into two parts, H = H0 + Hint , where H0 dominates, and Hint represents a comparatively small interaction. The large contribution H0 will make the operator A change very quickly (as can be seen from the form of (B.10), where a large energy produces rapid oscillations in time). This rapid oscillation can obscure any interesting effects arising from the interaction Hamiltonian Hint . To extract e in the following way, these interesting effects, we define a slowly varying operator A e = U0 AU † . A 0 (B.11) B.2 The Heisenberg Picture 406 Here U0 is the time evolution operator associated with the Hamiltonian H0 . That is, U0 satisfies U̇0 U0† = iH0 . Differentiating (B.11) with respect to time, and using the Heisenberg equation (B.10), we find e = U̇0 AU † + U0 (∂t A)U † + U0 AU̇ † , ∂t A 0 0 0 e + U0 (i[A, H0 + Hint ]) U † − iAH e 0, = iH0 A 0 e H0 ] + i[A, e H0 ] + i U0 AHint U † − U0 Hint AU † . = −i[A, 0 0 (B.12) Conveniently, the first two terms cancel. The last term can be re-written in a e = U0 Hint U † , whence we compact form, if we define a modified Hamiltonian H 0 obtain the interaction picture equation of motion e = i[A, e H]. e ∂t A (B.13) Appendix C Quantum optics Quantum optics is the study of the quantum features of light. The theory requires a treatment of ensembles of identical photons, which are easily created and destroyed in their interaction with atoms. Therefore the techniques of quantum field theory must be employed, in order to deal with the creation and destruction of identical particles. In this Appendix we briefly review the quantum mechanical description of the electric field associated with a propagating light beam, before describing the form of the interaction between light and matter. C.1 Modes Classically, light is a transverse electromagnetic wave. Apart from its amplitude, it has three degrees of freedom that must be specified to uniquely determine its properties. These are (i) its polarization, (ii) its frequency and (iii) its propagation direction. C.1 Modes 408 The polarization is the direction along which the electric field oscillates; it is a vector in a plane perpendicular to the propagation direction. It is easy to see that the space of polarizations is simply a 2-dimensional vector space. In fact, due to the possibility of phase delays between different polarization directions, it is actually a complex vector space — a Hilbert space (see Section A.1.2 in Appendix A). Nonetheless it is a 2-dimensional vector space. The same is true for the other degrees of freedom. That is, the space of frequencies is a vector space. It is a space with an uncountably infinite number of dimensions, since the different possible frequencies are infinitely closely spaced, but it is no different in character to the space of polarizations. And similarly for the propagation direction: there are an infinite number of infinitely closely spaced propagation directions, and the set of all of these forms a vector space. Already, in talking of these vector spaces, we have made implicit reference to a basis for each of them. We talk of two perpendicular directions for polarization. Or different directions of propagation. These are labels that we use to keep track of dimensions in a vector space, and they are intuitive and natural. But any basis is as good as any other. For example, instead of talking about different frequencies, we could talk about different arrival times. Or we could think of left and right circular polarizations as the polarization basis. It is useful in quantum optics to be flexible about the basis we use to describe an optical field. A common concept is therefore that of the optical mode. A mode is a member of an orthonormal basis for one of the vector spaces associ- C.1 Modes 409 ated with a light field. So, horizontal polarization is a polarization mode, since it is one of a pair of perpendicular polarizations that form a basis for the space of possible light polarizations. The other, orthogonal mode, is vertical polarization. And a single frequency ω labels a spectral mode. It is orthogonal to another frequency ω 0 , because two plane-waves with these frequencies have a vanishing inner product, Z 0 eiωτ e−iω τ dτ = 0, (C.1) when ω 6= ω 0 . Equivalently, we could label different temporal modes t and t0 . These are orthogonal because two delta-functions with these timings also have a vanishing inner product, Z δ(τ − t)δ(τ − t0 ) dτ = 0, (C.2) when t 6= t0 . An optical mode is a member of a basis for the full space of all possible optical fields. This space of all possible fields is just the tensor product of the vector spaces associated with each degree of freedom. And a basis for the full space is found by taking the tensor product of the bases used for each degree of freedom (see Section A.2.2 in Appendix A). That is, an optical mode is the tensor product of a polarization mode, a spectral mode and a spatial mode. Once a basis of modes is settled upon, it is possible to introduce the concept of a photon. A photon is an excitation of an optical mode. Sometimes it is useful to remember that photons are only defined with respect to a basis of modes. Although C.2 Quantum states of light 410 photons are often contrasted with waves as an embodiment of the particulate nature of light, they do not have to be localized, like tiny bullets. The ‘shape’ of a photon is the shape of the mode of which it is an excitation. C.2 C.2.1 Quantum states of light Fock states Suppose we consider a plane wave optical mode. That is, a mode with a linear polarization (horizontal, say), a single frequency ω, and a single propagation direction k, where k is the wavevector of the mode. An excitation of this mode has a fixed energy, given by the Planck formula E = ~ω, so a single photon in this mode is an eigenstate of the Hamiltonian for the field. Similarly, if we excite two photons in this mode, we have a state with twice the energy, E = 2~ω. This is also an energy eigenstate, but with a different eigenvalue. It follows that these two states must be orthogonal. And by extension, each photon number state is orthogonal to every other photon number state. If we use the notation |ni to denote the state with n photons, we must have hn|mi = δnm . (C.3) Changing the basis of optical modes from plane waves to some other basis cannot change this orthogonality, since the inner product is invariant under unitary transformations. Therefore (C.3) holds generally, for different photon number states of an arbitrary optical mode. C.2 Quantum states of light 411 The orthonormal basis of photon number states |ni, associated with excitations of some given optical mode, is known as the Fock basis for that mode. The photon number states are sometimes known as Fock states, and the space for which they form a basis is accordingly Fock space. The Fock space represents the final degree of freedom associated with an electromagnetic quantum state: the amplitude. That is to say, the more photons in a mode, the more intense the field. Thus the quantum state of an electromagnetic field is fully specified by the tensor product of 4 vector spaces: the polarization, spectral and spatial modes (collectively specifying an optical mode), and finally the Fock space (specifying the photon number: the energy in the field; its brightness). C.2.2 Creation and Annihilation operators A marked difference between optical fields and material systems is the impermanence of photons. Generally the atoms and electrons in a quantum memory are considered to be indestructible. They are not created or destroyed by their interactions. But photons can be absorbed and re-emitted. So we must describe processes that change one Fock state into another — processes that change the number of photons excited into a given mode. This description is accomplished by introducing a creation operator a† , that adds a single photon to an optical mode. The similarity of the symbol ‘†’ for Hermitian conjugation to a ‘+’ sign serves as a useful mnemonic for this operator’s function. The effect of applying a† to an empty optical mode |0i, containing no photons, is to produce the state |1i, with a single photon. C.2 Quantum states of light 412 Further applications of a† add extra photons, with contributions from all possible permutations of arranging these photons (see Figure C.1). These contributions must be included, since photons are bosons, meaning that their state must be unchanged by swapping any pair of photons. The Fock states created by the action of a† are not correctly normalized, so that a numerical factor, accounting for the number of permutations, must be included, (a† )n |0i = √ n!|ni. (C.4) Another way to write this is a† |ni = √ n + 1|n + 1i. (C.5) Taking the norm of (C.5), we have hn|aa† |ni = (n + 1)hn + 1|n + 1i = n + 1, ⇒ aa† |ni = (n + 1)|ni. (C.6) That is, the Hermitian conjugate a = (a† )† is an annihilation operator that removes a photon. And from (C.6) we see that a|ni = √ n|n − 1i. (C.7) C.2 Quantum states of light 413 Note that a ‘kills’ the empty vacuum state, a|0i = 0, which is fortunate, since there cannot be fewer than zero photons in a mode! It is often useful, when manipulating Figure C.1 Symmetrized photons. n applications of the photon creation operator a† to the vacuum state |0i produces a symmetrized n photon state, with contributions from all n! permutations of the n photons. Swapping any two photons leaves the state unchanged, as required by Bose statistics. expression involving these operators, to be able to reverse their ordering. This is done using their commutator which, applying (C.5) and (C.7), is given by hn|[a, a† ]|mi = 0, hn|[a, a† ]|ni = 1, ⇒ [a, a† ] = 1. (C.8) The commutator (C.8) expresses what is known as the canonical commutation relation; commutators of this form are common to creation and annihilation operators for all bosonic fields. Another useful operator is the number operator N = a† a, a Hermitian operator that satisfies the eigenvalue equation N |ni = n|ni, so that N counts the number of photons excited into a particular mode. (C.9) C.3 The electric field C.3 414 The electric field Electric fields are associated with separated charges, while magnetic fields are associated with moving charges. Electrons move rather slowly in most ordinary forms of matter, and accordingly their interaction with light is dominated by its electric component. In this thesis, we treat light fields as if they were purely electric waves, an approximation that is very well satisfied provided that light intensities are not sufficient to produce a relativistic electron plasma. The electric field associated with a beam of light, as might be generated by a laser, can be expressed in terms of the annihilation operators a(ω) associated with plane waves propagating along the beam [107] , Z E(z) = iv g(ω)a(ω)e−iωz/c dω + h.c., (C.10) where z is the longitudinal position along the beam, v is a unit polarization vector in the plane perpendicular to the beam and g(ω) = p ~ω/4π0 Ac is the mode amplitude. Here A is the cross-sectional area of the beam, 0 is the permittivity of free space and c is the speed of light. Note that in principle the electric field is an observable quantity, that we could measure (although at optical frequencies it is not generally possible to directly measure the electric field, at radio frequencies it certainly is feasible). And so E is a Hermitian operator, as expected. The annihilation operators a(ω), labelled by the frequency ω of the mode upon C.4 Matter-Light Interaction 415 which they act, satisfy the commutation relation [a(ω), a† (ω 0 )] = δ(ω − ω 0 ). (C.11) This expresses the fact that operators for different frequency modes do not ‘see’ eachother, so they commute, while when ω = ω 0 , the canonical relation (C.8) is satisfied. The delta function is the appropriate generalization for the case when the modes are labelled by a continuous parameter, such as ω. C.4 Matter-Light Interaction Generally light interacts with matter through electrons. In most quantum memory protocols these are the optically active outer electrons bound to some atoms. We will also consider scattering in a diamond crystal, and here the electrons are more appropriately described as free, or quasi -free particles. The Hamiltonian describing the interactions are slightly different in these two cases; here we briefly review their origin, and the relationship between them. C.4.1 The A.p Interaction The interaction of an electron with the electromagnetic field is found by incorporating the appropriate electromagnetic term, associated with so-called U (1) gauge symmetry, into the Lagrangian density. The effect of this term is to modify the C.4 Matter-Light Interaction 416 momentum p of the electron, p −→ p − eA, (C.12) where e is the electronic charge, and where A is the magnetic vector potential. The potential A is not actually an observable field. The electric and magnetic fields are related to its derivatives, but the absolute value of A is arbitrary to some extent. Different choices for the functional form of A — known as different gauges — produce different Hamiltonians, with differing degrees of calculational convenience; the physical predictions of the theory are unchanged of course. A standard choice of gauge in quantum optics is the Coulomb gauge, which requires that A is divergence free, ∇.A = 0. With this choice, the physical electric and magnetic fields are given, respectively, by E = −∂t A, B = ∇ × A. (C.13) Using (C.13) we can express the potential A in the form (C.10), with the replacement ig(ω) −→ g(ω)/ω. The Hamiltonian for an electron, with mass m, in an electromagnetic field is found by substituting the ‘canonical momentum’, given by (C.12), into the Hamiltonian for a ‘bare’ electron, (p − eA)2 A H= + 0 2m 2 Z E 2 + c2 B 2 dz. (C.14) This is known as the minimal coupling Hamiltonian. The first term is the kinetic C.4 Matter-Light Interaction 417 energy of the electron, with the transformation (C.12) included. The second term, in square brackets, represents the ‘free field’ energy: this is the energy of the electromagnetic fields, in the absence of the electron. The integral extends over all space, or at least, over the entire region occupied by the fields. The contribution from the magnetic field B is very small, but the contribution from the electric field E is more significant. Inserting (C.10) for E shows that the free field energy takes the form R (N + 12 )~ω dω. The term involving the number operator N simply expresses the Planck formula E = ~ω, so that the energy in the field increases with the number of photons excited. The term proportional to 1 2 is known as the zero-point energy: the energy of the vacuum. It is rather unfortunate that this energy is infinite (since it is integrated over all frequencies), but it is possible to work around these technicalities with some mathematical sleight-of-hand, known as renormalization [196] . In any case we will not be concerned with the zero point energy. Multiplying out the first term in (C.14), we obtain a term of the form p2 /2m, which just describes the ‘bare’ kinetic energy of the electron, without the field. There is a term A2 /2m, which describes the field acting back on itself — this type of non-linear back action is generally negligibly small. And there is a term of the form −eA.p/m. This describes the coupling of electronic momenta to the vector potential. In situations where electrons are spread over an extended region, such as in a crystal, this interaction dominates the atom-light coupling. C.4 Matter-Light Interaction C.4.2 418 The E.d Interaction When electrons are well-localized, such as when bound into atoms, a more convenient form of the interaction Hamiltonian can be derived. This is accomplished formally by means of a unitary transformation due to Power, Zienau and Woolley (PZW) [197,198] . In general it is desirable to eliminate explicit reference to the vector potential A in the Hamiltonian, since then the equations are manifestly gauge invariant — it is quite clear that there can be no-dependence on the choice of gauge. The PZW transformation removes A from the Hamiltonian, and introduces interactions between the physical field E and the moments of the atomic charge distribution. To see how this is done, we will need two results. The first is the equal-time commutator of E and A, the amplitudes of the electric field and the vector potential, [A(z), E(z 0 )] = −i~ e δ(z − z 0 ), A0 (C.15) This is easily derived from (C.10) and (C.13) using (C.11). It is well known that in quantum mechanics momentum and position generally satisfy the relation [x, p] = i~, and indeed the form of (C.15) when z = z 0 reflects the fact that in the Coulomb gauge the field E is actually the ‘momentum’ that is conjugate to the ‘coordinate’ A in the electromagnetic Hamiltonian. The second result we need is that eC De−C = D + [C, D], (C.16) C.4 Matter-Light Interaction 419 whenever [C, D] is just a number (i.e. not another operator). Here the exponential of an operator is defined according to the series eC = ∞ X Cn n=1 n! . (C.17) The result (C.16) is straightforward to derive. Consider the product C n D. Using the commutator, we can ‘pull’ the operator D through C, in the following way, C n D = C n−1 (CD) = C n−1 (DC + [C, D]) = (C n−1 D)C + [C, D]C n−1 . (C.18) Repeating this procedure recursively, we obtain C n D = DC n + n[C, D]C n−1 . (C.19) Re-writing the left hand side of (C.16) using the series (C.17), and applying (C.19), we arrive at (C.16). With these preliminaries, we can introduce the PZW transformation. Suppose that the action of the light is to make an optically active electron oscillate; it remains bound to an atom, but it is ‘wiggled’ by the field. This is certainly what we expect would happen classically. The atomic polarization, distinct from the optical polarization, is a useful concept in this situation. It is the ‘dipole moment per unit volume’, where the dipole moment is the product of the electronic charge and displacement. Suppose that the electron, with charge −e, is displaced a distance x C.4 Matter-Light Interaction 420 along the polarization direction v of the incident light field. The dipole moment is d = −exv, and the atomic polarization is P (z) = dδ(z)/A, where the delta function describes a single dipole placed at the position z = 0. To express the interaction energy associated with the atomic polarization, we introduce a unitary transformation of the Hamiltonian, † H → U HU , with iA U = exp ~ Z P (z ).A(z ) dz . 0 0 0 (C.20) This transformation simply changes the coordinate system with respect to which the quantum states |ψi of the atom-light system are defined. Essentially it is nothing more than a cosmetic change, but it has a marked effect on the form of the Hamiltonian. Applying the transformation to the free-field part of (C.14), we have U E 2 U † = (U EU † ).(U EU † ), with Z iA U EU † = vU EU † = v E(z) + P (z 0 )[A(z 0 ), E(z)] dz 0 ~ 1 = E(z) − P (z). 0 (C.21) That is, the PZW transformation adds a component proportional to the electron displacement into the electric field. Using (C.21), the free-field Hamiltonian, neglecting the small contribution from the magnetic field B, becomes A 0 2 Z 2 E dz −→ = 2 Z A 1 0 E − P dz 2 0 Z A 1 2 2 0 E + 2 P dz 0 − d.E(z = 0). 2 0 (C.22) C.4 Matter-Light Interaction 421 The term proportional to E 2 represents the ‘bare’ free-field energy, with no electron present. The P 2 term represents an unimportant ‘self-interaction’ of the electron. But the last term, proportional to E.d, represents the interaction of the physical electric field, at the position of the electron, with the electronic dipole moment. It is known as the electric dipole interaction Hamiltonian, and it serves as the basis for the analysis of all the atomic quantum memory protocols in this thesis. Finally, we note that the electron momentum acquires a component proportional to the vector potential under the PZW transformation, iA p −→ p + ~ = Z A(z 0 )[P (z 0 ), p] dz 0 p + eA(z = 0), (C.23) where this time the commutator of the electronic momentum and position was used, [x, px ] = i~. In the approximation that the wavelength of the light is much longer than the spatial extent of the atom — a limit valid for all interactions at optical frequencies — we can set A ≈ A(z = 0), and then the vector potential A is completely eliminated from the Hamiltonian. Thus when electrons are tightly bound into atoms, the only significant interaction with optical fields occurs through the electric dipole interaction. C.5 Dissipation and Fluctuation C.5 422 Dissipation and Fluctuation In this section we address the issue of loss in quantum systems. Specifically, we seek a theoretical description of the decoherence in a quantum memory: the constituent atoms may emit photons into random directions, or collide with one another, and these processes partially destroy the quantum information stored in the memory. In Chapter 4 we use the Heisenberg picture to describe the propagation of light through a quantum memory, and so we should account for losses using the Heisenberg picture. In the following we use a simple model to show how the equations of motion for a quantum system are modified by the presence of losses. Fortunately it is well known that the results are not significantly altered by refining the model. Our model consists of a single bosonic mode, our system, coupled to a bath of other bosons — a reservoir. We use bosons because their commutation relations are simple, but this model applies rather well to an ensemble quantum memory. As shown in §4.11 in Chapter (4), both the atomic polarization and the spin wave appearing in the equations of motion of an ensemble memory have approximately bosonic commutation relations. In fact, in diamond, the optical phonons that constitute the spin wave really are bosons (see Chapter 9). The reservoir of modes to which the system is coupled could be the electromagnetic field, and indeed this is an excellent description of spontaneous emission losses affecting the atomic polarization. The Hamiltonian contains the free field energy of the bath, and also terms that represent processes in which a photon in the system is destroyed, while a photon in C.5 Dissipation and Fluctuation 423 the reservoir is created, or vice versa. Working in the Heisenberg interaction picture, so that the free-field energy of the system (and zero point energy of the reservoir) is removed, we have Z e = H ωb† (ω)b(ω) dω + κ Z h i a† b(ω) + b† (ω)a dω, (C.24) where b(ω) destroys a photon with frequency ω in the reservoir, and where a destroys a photon in the system. The equal-time commutators of these operators are given by [a, a† ] = 1, and [b(ω), b† (ω 0 )] = δ(ω − ω 0 ). (C.25) Using the Heisenberg equation (B.13), we derive the following equations of motion for the annihilation operators, Z ∂t a = −iκ b(ω) dω, (C.26) ∂t b(ω) = −iωb(ω) − iκa. (C.27) Integrating (C.27) and substituting the result into (C.26), we obtain Z ∂t a(t) = −iκ −iωt e Z b0 (ω) − iκ t e −iω(t−t0 ) 0 a(t ) dt 0 dω, (C.28) 0 where the operators b0 (ω) = b(ω, t = 0) represent the initial state of the reservoir. Performing the frequency integral in the second term produces a delta-function (see C.5 Dissipation and Fluctuation 424 (D.9) in Appendix D), which selects out the time t = t0 , and then (C.28) becomes ∂t a(t) = −γa(t) + F (t), (C.29) where Z F (t) = −iκ b0 (ω)e−iωt dω (C.30) is known as a Langevin noise operator, and where γ = πκ2 is an exponential decay rate. Equations of this form were first used to describe the classical Brownian motion of colloidal particles buffeted by the molecules of a warm fluid. There, the term involving F represents the fluctuating force arising from collisions with the randomly moving molecules, and a similar interpretation is helpful in the present case. The operator F mixes a component of ‘white noise’ into the dynamics of a, which otherwise would simply decay exponentially with a rate γ. That the noise is white, with a flat power spectrum, can be seen from examining its temporal correlation functions. Suppose that initially the reservoir contains some small thermal excitations, so that there are n̄ photons, on average, in a mode with frequency ω. Then we have hb†0 (ω)b0 (ω 0 )i = n̄δ(ω − ω 0 ). (C.31) C.5 Dissipation and Fluctuation 425 The normally ordered correlation function of the noise operator is then given by † 0 Z Z 0 0 n̄δ(ω − ω 0 )ei(ωt−ω t ) dω dω 0 , Z 1 0 = 2πκ2 n̄ × eiω(t−t ) dω 2π hF (t)F (t )i = κ 2 = 2γ n̄δ(t − t0 ), (C.32) (C.33) and the anti-normally ordered correlation function is similarly given by hF (t)F † (t0 )i = 2γ(n̄ + 1)δ(t − t0 ), (C.34) where we used the commutator in (C.25). The noise is therefore delta-correlated, meaning that it is completely random from moment to moment. There is no correlation with earlier times, and the noise changes ‘infinitely quickly’; a signature of unlimited-bandwidth white noise. The infinite bandwidth is a consequence of the fact that the reservoir coupling κ was assumed to be frequency independent. This assumption is known as a Markov approximation, since it means that the dynamics of a does not explicitly depend on its past values. As is clear from the form of (C.29), the Markov approximation gives rise to exponential decay of a: the ubiquity of exponential decay in quantum systems serves to confirm the robustness of this approximation as a model for a wide variety of dissipative processes. The noise F is just sufficient to preserve the expectation value of the equal-time commutator in (C.25), so that a remains a bona fide bosonic operator. This can be C.5 Dissipation and Fluctuation 426 seen by inserting the formal solution to (C.29), −γt a(t) = a(0)e Z t + 0 e−γ(t−t ) F (t0 ) dt0 , (C.35) 0 into the commutator (C.25), Z tZ t 0 00 e−γ(2t−t −t ) h[F (t0 ), F † (t00 )]i dt0 dt00 h[a(t), a† (t)]i = h[a(0), a(0)† ]ie−2γt + 0 0 Z t 0 e2γt dt0 = e−2γt 1 + 2γ 0 = 1. (C.36) This close connection between fluctuations, represented by F , and damping, represented by γ, is a manifestation of the fluctuation-dissipation theorem, discovered first by Einstein. Finally, we note that the solution (C.35) allows us to solve for the time evolution of the number operator for the system, giving the result hN i = ha† (t)a(t)i = e−2γt ha† (0)a(0)i + 1 − e−2γt n̄. (C.37) This shows that in the infinite future, with t −→ ∞, the system relaxes into thermal equilibrium with the reservoir, with N −→ n̄. And this thermal equilibrium condition can often be taken as the initial state of the system, if we want to analyze processes which drive it out of equilibrium. Appendix D Sundry Analytical Techniques In this Appendix we define the Dirac delta function and the unilateral and bilateral Fourier transforms used in this thesis. We discuss some of their properties, and finally we demonstrate some results relating to the Fourier transform of certain Bessel functions. First though, we introduce a useful technique that expedites these calculations. D.1 Contour Integration Contour integration is a powerful method for evaluating integrals that would otherwise be difficult, or impossible, to perform. To see how it works, we consider the integral Z Ix = b f (x) dx a (D.1) D.1 Contour Integration 428 of some smooth function f (x). The anti-derivative F (x) of f is the function that satisfies ∂x F = f , and (D.1) can be simply expressed in terms of F as Ix = F (b) − F (a). That is, the integral of f between two points is given by the change in ‘height’ of its anti-derivative between these same two points (see Figure D.1 (a)). Now suppose that we introduce the possibility of complex arguments for f , so that we consider all the values of f (z), where z = x + iy is an arbitrary complex number. We can now think of f as a two dimensional surface, lying above the (x, y)-plane: as we vary the real and imaginary coordinates x and y, the value of f traces out a characteristic landscape. We could define an integral Z B Iz = f (z) dz (D.2) A from some initial point A = xA + iyA to a final point B = xB + iyB . But now that we are working in a two dimensional plane, the endpoints aren’t enough to specify the integral completely. We need to know the path that we should take to get from A to B. On the other hand, the anti-derivative F (z) also describes some kind of surface in the (x, y)-plane, as shown in part (b) of Figure D.1, and the integral is again given by Iz = F (B) − F (A). So in fact, the integral does not depend on the path taken, it only depends on its endpoints. The proof of this fact is known as the Cauchy integral theorem. Suppose that we integrate along a closed loop, so that the endpoint B is equal to the starting point A. Then F (B) = F (A) and the integral vanishes. This is interesting, since we could build the first integral (D.1) into the D.1 Contour Integration 429 complex integral (D.2), and possibly use this result to draw some conclusions. For instance, suppose we start our integral at A = (a, 0) in the (x, y)-plane, integrate along the real axis (i.e. the line y = 0) to the point (b, 0), and then loop back in a semi-circle to finish where we started at B = A. The straight part of the loop, along the real line, is just equal to Ix . The whole integral must vanish, because we finish where we started. Therefore if we can find an expression for the curved part of the integral — call it Ic — then we must have Ix = −Ic , in order that the two parts of the integral cancel out. In this way, interesting results about real integrals can be derived by considering their behaviour in the complex plane. (a) (b) Figure D.1 Contour integrals. (a) a function f and its antiderivative F . (b) The anti-derivative is extended into the complex plane, and we consider integrating around a loop that includes the original integral along the real line. D.1.1 Cauchy’s Integral Formula So far we have only considered smooth functions. When f contains a discontinuity, the Cauchy integral theorem does not apply. Suppose that we want to integrate the function f (z)/(z − z0 ), which ‘blows up’ when z = z0 . We consider integrating along D.1 Contour Integration 430 a circular loop around the singularity at z0 . We can parameterize the integral using polar coordinates z = z0 + reiθ , with dz = ireiθ dθ. We are free to make the radius r as small as we like, since as above the details of the path taken don’t matter; it only matters that we encircle the singularity. We find I f (z) dz = lim r→0 z − z0 where the symbol H Z 0 2π f (z0 + reiθ ) × ireiθ dθ = 2πi × f (z0 ), reiθ (D.3) indicates integration around a loop. The singularity is like a witness to our passage around the loop, so that we do not quite get back to where we started on completing our roundtrip. The contribution 2πif (z0 ) to the integral is known as the residue associated with the singularity at z = z0 . In general, wherever a function blows up, it leaves a residue that contributes to an integral that encloses it. Differentiating the formula (D.3) with respect to z0 tells us how to handle ‘stronger’ singularities of the form 1/(z − z0 )n , I D.1.2 ∂zn−1 f (z0 ) f (z) 0 dz = 2πi × . n (z − z0 ) (n − 1)! (D.4) Typical example We now give an example of how to use these formulas to evaluate an integral. The method is typical. Consider the integral Z ∞ Ix = −∞ eix dx. x (D.5) D.1 Contour Integration 431 This is equivalent to the contour integral I Iz = lim R→∞ eiz dz, z (D.6) where the integration path is a closed semicircle with radius R, such that the straight portion runs along the real line, as shown in Figure D.2. To see that Iz = Ix , note that the function eiz is damped with a positive imaginary argument, ei(x+iy) = eix e−y . Therefore any contribution to Iz from the curved portion of the path, all of which has y > 0, vanishes in the limit R −→ ∞. Note that if the integrand were conjugated, so that it contained the exponential e−ix , we would have to choose to close the contour in the lower half of the plane, in order that the curved part of the integral should vanish. Now the integral Iz is not zero, because the integrand has a singularity at z = 0. It’s a little inconvenient that this lies exactly on the integration path. We deal with this by moving the singularity inside the integration path by a small amount , in order to apply (D.3); we then take the limit −→ 0 when we are done: I Ix = Iz = lim lim →0 R→∞ eiz dz z − i = lim 2πi × e− (D.7) = 2πi. (D.8) →0 This technique for performing integrals is indispensable in Fourier analysis, as discussed below. And of course it is quite elegant! D.2 The Dirac Delta Function 432 Figure D.2 Upper closure. The integration path is comprised of a straight portion running along the real axis, and a semicircle in the upper half of the complex plane. In the limit of infinite radius R, the straight portion becomes equal to the real integral Ix , and the curved portion vanishes. D.2 The Dirac Delta Function We use the Dirac delta function δ(x) a great deal in this thesis. Where it is not introduced ‘by hand’, prompted by physical arguments, it is most often encountered in the form of an integral over plane waves, δ(x) = 1 2π Z ∞ eikx dk. (D.9) −∞ This integral is not really well defined, but its properties become apparent from the following limit, Z T 1 δ(x) = lim eikx dk T →∞ 2π −T T sinc(xT ) = lim , T →∞ π (D.10) where sinc(θ) = sin(θ)/θ. For each value of T , this describes a narrow peak of width ∼ 1/T and height T , centred at x = 0. As T → ∞, the peak becomes infinitely tall D.2 The Dirac Delta Function 433 and narrow. But its integral remains finite. To see this, consider the integral ∞ −i Ix = sinc(x) dx = 2 −∞ Z ∞ Z −∞ eix dx − x Z ∞ −∞ e−ix dx . x (D.11) The first integral on the right hand side is given by (D.8). Applying the same technique to the second integral, we must use an integration path that is closed in the lower half of the complex plane, as explained in §D.1.2. When we shift the singularity using z → z − i, it moves into the upper half of the plane, so the contour used for the second integral does not enclose the singularity, and therefore the integral vanishes. Therefore we find that Ix = −i {2πi − 0} = π. 2 (D.12) Integrating the delta function then yields Z ∞ 1 δ(x) dx = lim T →∞ π −∞ Z ∞ T sinc(xT ) dx = 1. (D.13) −∞ The Dirac delta function is an ideal ‘spike’, with unit area, and so it has the extremely useful property that Z ∞ f (x)δ(x − x0 ) dx = f (x0 ). −∞ (D.14) D.3 Fourier Transforms D.3 D.3.1 434 Fourier Transforms Bilateral Transform The most familiar Fourier transform is the bilateral transform fe(k) of a function f (z), 1 fe(k) = Fz {f (z)} (k) = √ 2π Z ∞ f (z)eikz dz. (D.15) −∞ The name ‘bilateral’ refers to the lower limit of integration, which is −∞ in this case, so that the whole function f (z), including its values for negative z, is involved in forming the Fourier transform fe. D.3.2 Unitarity √ The factor of 1/ 2π makes the transform unitary, meaning that the norm of the the transform fe is the same as the norm of f , Z ∞ −∞ Z ∞Z ∞Z ∞ 1 0 f (z)f ∗ (z 0 )eik(z−z ) dk dz dz 0 2π Z ∞ −∞ Z ∞ −∞ −∞ 0 = δ(z − z 0 )eik(z−z ) f (z)f ∗ (z 0 ) dz dz 0 −∞ Z−∞ ∞ = |f (z)|2 dz, |fe(k)|2 dk = −∞ where the plane wave expansion (D.9), and the property (D.14) were used. (D.16) D.3 Fourier Transforms D.3.3 435 Inverse The inverse Fourier transform takes essentially the same form, except that the integral kernel is conjugated, f (z) = Fk−1 Z ∞ n o 1 e fe(k)e−ikz dk. f (k) (z) = √ 2π −∞ (D.17) This is easily demonstrated by substituting (D.15) into (D.17) and using (D.9). D.3.4 Shift From the definition (D.15) we can derive a simple formula for the Fourier transform of a shifted function f (z + a), where a is some constant, Fz {f (z + a)} (k) = = Z ∞ 1 √ f (z + a)eikz dz 2π −∞ Z ∞ 1 √ f (x)eik(x−a) dx 2π −∞ = e−ika fe(k), (D.18) where we changed integration variables from z to x = z + a in the second line. This shows that a shift in the coordinate z corresponds to a phase rotation in the Fourier domain. Identical arguments show that for the inverse transform n o Fk−1 fe(k + a) (z) = eiza f (z). (D.19) D.3 Fourier Transforms D.3.5 436 Convolution The bilateral convolution of two functions f and g is defined by the integral 1 f ∗ g(z) = √ 2π Z ∞ 1 f (x)g(z − x) dx = √ 2π −∞ Z ∞ f (z − x)g(x) dx. (D.20) −∞ The Fourier transform of a convolution is given by the product of the Fourier transforms of the convolved functions. That is, Fz {f ∗ g(z)} (k) = = Z ∞Z ∞ 1 f (x)g(z − x)eikz dz dx 2π −∞ −∞ Z ∞Z ∞ 1 f (x)eikx g(y)eiky dxdy 2π −∞ −∞ = fe(k)e g (k), (D.21) where we used a change of variables y = z −x with dy = dz. This shows also that the inverse Fourier transform of a product of two functions is given by the convolution of their individual inverse transforms. D.3.6 Transform of a Derivative The nth derivative ∂zn f (z) of a function can be expressed using the inverse (D.17) as ∂zn f (z) = 1 √ 2π Z ∞ −∞ fe(k)∂zn e−ikz dk n o = Fk−1 (−ik)n fe(k) (z), (D.22) D.4 Unilateral Transform 437 and therefore we must have that Fz {∂zn f (z)} (k) = (−ik)n fe(k). (D.23) That is, the Fourier transform converts differentiation into ordinary multiplication; clearly this is a boon when tackling differential equations. D.4 Unilateral Transform The unilateral Fourier transform is identical to its bilateral counterpart, except that it involves an integral over positive coordinates only, 1 fe(k) = Fz {f (z)} (k) = √ 2π Z ∞ f (z)eikz dz. (D.24) 0 We have used the same notation to denote both types of transform: we will be careful to distinguish between them when the difference is important. This type of transform is appropriate for functions f (z) which vanish, or are undefined, when z < 0. As an example, the function f might describe the temporal response to an interaction at ‘time’ z = 0, in which case causality requires that f vanishes for z < 0. For these types of functions, the unitarity property (D.16) still holds, and the inverse transform is given precisely by (D.17). That is, the inverse is bilateral. D.4 Unilateral Transform D.4.1 438 Shift The shift theorem (D.18) is not directly applicable to the unilateral transform, because of the semi-infinite integral domain. But since the inverse transform is bilateral, the shift theorem (D.19) does apply to the inverse transform. D.4.2 Convolution The unilateral convolution, or causal convolution of two functions f and g is defined similarly to the bilateral convolution (D.20), except that the integral is limited by causal consistency, 1 f ∗ g(z) = √ 2π Z 0 z 1 f (x)g(z − x) dx = √ 2π Z z f (z − x)g(x) dx. (D.25) 0 To calculate the transform of this convolution, we need to take care with the limits of integration, as shown in Figure D.3. We then obtain Fz {f ∗ g(z)} (k) = = = Z ∞Z z 1 f (x)g(z − x)eikx dxdz 2π 0 0 Z ∞Z ∞ 1 f (x)g(z − x)eikz dzdx 2π 0 Z ∞ Zx∞ 1 f (x)eikx g(y)eiky dydx 2π 0 0 = fe(k)e g (k). (D.26) Just as for the bilateral case above, the unilateral Fourier transform of a causal convolution is given by the product of the transforms of the convolved functions. D.4 Unilateral Transform 439 (a) (b) Figure R ∞ R z D.3 Integration limits. (a) Schematic of the integral dx dz. (b) Schematic of the same integral with the order of in0 0 R∞R∞ tegration reversed, 0 x dz dx. Each arrow represents an instance of the inner integral at a fixed value of the outer variable. D.4.3 Transform of a Derivative The real utility of the unilateral transform is in the treatment of derivatives. Consider the unilateral Fourier transform of ∂z f . Integrating parts, 1 √ 2π Z ∞ e 0 ikz Z ∞ i∞ 1 h ikz 1 ∂z f (z) dz = √ e f (z) f (z)ikeikz dz −√ 0 2π 2π 0 1 = −ik fe(k) − √ f (0). (D.27) 2π This is the same result as (D.23) for the bilateral transform, except for the presence of the boundary condition f (0). Note that we assumed f (z −→ ∞) = 0: all the functions we will consider satisfy this property, referred to alternately as boundedness or integrability. Physically, this simply says that the effects of any interaction fall away to nothing as we move away from their source. D.5 Bessel Functions D.4.4 440 Laplace Transform The unilateral Fourier transform occupies a middle ground between the bilateral Fourier transform on the one hand, and the well-known Laplace transform on the other. The Laplace transform Lz {f (z)} (s) is generated from the unilateral Fourier transform by making the replacement ik → s. The Laplace transform is more common since even unbounded functions generally possess a Laplace transform, due to the exponential damping provided by the integral kernel. However this damping destroys the unitarity of the transform. For this reason, the unilateral Fourier transform better suits our purposes in this thesis, since the efficiency of a quantum memory is unchanged by working with this type of transform. D.5 Bessel Functions In this section we define the ordinary and modified Bessel functions that represent the propagators for quantum memories operated adiabatically (see Chapter 5). The nth order ordinary Bessel function of the first kind — denoted by Jn — is defined by the infinite series ∞ X (−1)m z 2m+n Jn (2z) = , m!(m + n)! (D.28) m=0 where n is a non-negative integer. The corresponding modified Bessel funtion In is defined identically, except that the factor of (−1)m is missing from each term of the sum. Conversion between ordinary and modified Bessel functions is reminiscent of D.5 Bessel Functions 441 the relation between trigonometric and hyperbolic ratios, Jn (iz) = in In (z). (D.29) Indeed the J functions have the appearance of decaying cosines, while the I functions all grow exponentially with increasing z. Just as sin(x) and cos(x) take the values 0 and 1, respectively, at the point x = 0, so it is also true that J0 (0) = I0 (0) = 1, while Jn>0 (0) = In>0 (0) = 0. D.5.1 Orthogonality As can be easily shown using the series (D.28), the Bessel functions Jn (z) satisfy the differential equation z∂z (z∂z ) Jn = (n2 − z 2 )Jn . (D.30) Consider the ith zero ai of Jn , which satisfies Jn (ai ) = 0. Defining s = z/ai and X(s) = Jn (ai s), we can re-write (D.30) in the ‘normalized’ form s∂s (s∂s ) X = (n2 − a2i s2 )X, (D.31) with X(1) = 0. We can repeat this procedure for another zero of Jn , say aj , giving s∂s (s∂s ) Y = (n2 − a2j s2 )Y, (D.32) D.5 Bessel Functions 442 where now s = z/aj and Y (s) = Jn (aj s), so that Y (1) = 0. We now multiply (D.32) by X and (D.31) by Y, subtract the two resulting equations and divide through by s, to obtain (a2i − a2j )sXY = ∂s [Xs∂s Y − Y s∂s X] . Integrating this from s = 0 to s = 1, and using the fact that X(1) = Y (1) = 0, the right hand side vanishes. Dividing through by a2i − a2j and converting back into the Jn notation, we find the orthogonality condition 1 Z sJn (ai s)Jn (aj s) ds = 0, when i 6= j. (D.33) 0 Alternatively, writing z = s2 , we have Z 1 √ √ Jn (ai z)Jn (aj z) dz = 0, when i 6= j. (D.34) 0 D.5.2 Memory Propagator In Chapter 5 we are faced with taking the inverse Fourier transform of the function e−ia/k fe(k) = n+1 , k (D.35) where a is some constant and n = 0, 1, 2.... This is done with the help of contour integration, as described at the beginning of this Appendix. Using the series D.5 Bessel Functions 443 expansion of the exponential, we have 1 f (z) = √ 2π " ∞ 1 −∞ k n+1 Z # ∞ X 1 −ia m −ikz e dk. m! k (D.36) m=0 We must therefore perform the integral ∞ e−ikz −∞ k n+m+1 Z Ik = dk. (D.37) We can equate this to the complex integral I Iz = lim lim →0 R→∞ e−ikz dk, (k + i)n+m+1 (D.38) where the integration contour is closed in the lower half of the complex plane, as shown in Figure D.4, and where the regularization shifts the singularity into the interior of the contour. Using Cauchy’s integral formula (D.4), we find ( Iz = lim →0 " ∂ n+m e−ikz −2πi × k (n + m)! = −2πi × (−iz)n+m (n + m)! # ) k=−i . (D.39) Substituting this result into the series (D.36), and comparing with (D.28), we obtain the result z n/2 √ √ f (z) = (−i)n+1 2πΘ(z) Jn 2 az . a (D.40) D.5 Bessel Functions 444 The function Θ(z) is known as the Heaviside step function, defined so that Θ(z) = 0 when z < 0, and Θ(z) = 1 for z > 0. We include it in the above expression to represent its causal nature. That f (z) = 0 for negative z can be seen by considering the integration contour used to evaluate the Fourier transform. When z becomes negative, the integration contour must be closed in the upper half of the complex plane, so that it no longer encloses the singularity, which therefore no longer leaves a residue, and the integral vanishes. Figure D.4 Lower closure. Here the integrand is damped in the lower half of the complex plane, so we close the integration contour in this region. We will also need the inverse transform of (D.35) when n = −1, n o f (z) = Fk−1 e−ia/k (z). (D.41) Again, inserting the series expansion for the exponential, we have Z ∞ 1 X (−ia)m ∞ e−ikz f (z) = √ dk. m 2π m=0 m! −∞ k (D.42) The first term, with m = 0, involves an integral with no singularity. In fact, it is an D.5 Bessel Functions 445 integral over plane waves — a delta function. We separate off this first term, and re-index the remaining terms, 1 f (z) = √ 2π Z ∞ 1 X (−ia)m+1 ∞ e−ikz e−ikz dk + √ dk. 2π m=0 (m + 1)! −∞ k m+1 −∞ Z ∞ (D.43) Then we perform the singular integrals using contour integration, as before, to obtain √ ∞ ia X (−ia)m (−iz)m 2πδ(z) − √ Θ(z) 2π m=0 (m + 1)!m! r √ √ a 2π δ(z) − Θ(z) J1 2 az . = z f (z) = (D.44) We make use of this result in Chapter 5, where it relates transmitted and incident signal fields in a Raman quantum memory. D.5.3 Optimal Eigenvalue Kernel In Chapter (5) we claim that it is possible to derive a certain kernel KA from a storage kernel K by direct integration — the result is used in the work of Gorshkov et al. The Fourier transforms presented above provide one way to confirm it. The two results we need are as follows. First, we need the unilateral Fourier transform of an integral over the product of two storage kernels: Z A = Fz,z 0 0 ∞ √ √ 0 e−ax × J0 (2 bxz)e−uz × J0 (2 cxz 0 )e−vz dx (k, k 0 ). (D.45) D.5 Bessel Functions 446 Using the result (D.40) along with the shift theorem (D.19) we obtain A = = Z ∞ 0 1 e−ibx/(k+iu) eicx/(k +iv) −ax e × × dx 2π 0 k + iu k 0 + iv 1 1 × . 0 2π akk + i(va + c)k + i(ua + b)k 0 − uva − bv − cu (D.46) The second result we need is the unilateral Fourier transform of an anti-normally ordered kernel: o n p 0 B = Fz,z 0 e−α(z+z ) J0 (2 βzz 0 ) (k, k 0 ). (D.47) Transforming over z first, and then z 0 , we get ( ) 0 i e−iβz /(k+iα) −αz 0 B = √ Fz 0 e × (k 0 ) k + iα 2π Z ∞ −[α+iβ/(k+iα)−ik0 ]z 0 i e = dz 0 2π 0 k + iα 1 1 = − × . 2π kk 0 + iα(k + k 0 ) − β − α2 (D.48) (D.49) Comparing these two results, and grinding through some tedious algebra, it is not hard to demonstrate the claimed equality. Appendix E Numerics In this Appendix we review the numerical methods used in this thesis. Our aim is to solve a system of coupled linear partial differential equations (PDEs) in space and time. For pedagogical purposes, we will consider the simple example presented in §3.5 in Chapter 3: A resonant signal pulse with amplitude A propagates through an ensemble of classical atoms, whose response to the optical field is described by the average displacement B of the atomic electrons from their equilibrium positions. The equations of motion are given by ∂τ B = −iαA, and β ∂z A = −i B. α (E.1) (The second equation follows from the first line of (3.34)). Since we are only concerned with the method of numerical solution, we set α = β = 1, and we normalize the τ and z coordinates so that they are dimensionless, with z running from 0 to 1, 448 and τ running from 0 to 10. The boundary conditions are Ain = A(z = 0, τ ) and Bin = B(z, τ = 0): the profile of the incident signal pulse, and the initial atomic excitation, respectively. Generally Ain will take the form of a pulse, reminiscent of a Gaussian, and Bin = 0, if the atoms are prepared in their ground states. To find the solution numerically, we discretize the space and time coordinates on a finite grid, z −→ zj , τ −→ τk , (E.2) and we represent the continuous functions A(z, τ ), B(z, τ ) as matrices A, B whose elements approximate those functions sampled at the grid points, Ajk = A(zj , τk ), Bjk = B(zj , τk ). (E.3) We solve the system (E.1) via the method of lines [199] . The technique is so named because we integrate the spatial derivative ∂z ‘in one go’, whereas we integrate the temporal derivative ∂τ incrementally, stepping forwards in time iteratively. The solution, starting as a series of points in space at τ = 0, evolves gradually forwards in time, and each of the points in space traces out a line, over time, that describes the solution at that position in space. It is a feature of the implementation of the technique that the temporal discretization is much finer than the spatial one, and so the numerics are conducive to this mental picture, which is illustrated schematically in Figure E.1. The spatial derivative is performed using a spectral method, which uses the value of the function at all the spatial points in order to compute E.1 Spectral Collocation 449 Figure E.1 The method of lines. The spatial coordinate is discretized on a coarse grid. The solution at each spatial point traces out a line as the temporal integration proceeds stepwise on a fine grid. the derivative. For smooth functions, this is extremely accurate, even when very few spatial points are used. Before describing the method used for the temporal derivative, we introduce this spectral method more fully. E.1 Spectral Collocation An excellent introduction to this field of numerical analysis is provided in Nick Trefethen’s book Spectral Methods in Matlab [200] — which is hosted on the internet. Other treatments include the books by Gottlieb and Orzag [201] , and by Boyd [202] . The easiest spectral method to understand is that involving a Fourier transform. As shown in §D.3.6 in Appendix D, Fourier transformation of a function f from z to k converts differentiation by z into multiplication by −ik. It is also clear that the operation of taking a Fourier transform, or indeed its inverse, is a linear one. Suppose that we discretize the coordinates z and k, so that the function f is approximated by E.1 Spectral Collocation 450 a vector f , with components fj = f (zj ). In this context the points zj at which f (z) is sampled are known as collocation points. The linearity of the Fourier transform then allows us to represent it as a matrix acting on f . The composition of the operations of taking the Fourier transform, multiplying by −ik, and then taking the inverse Fourier transform, can therefore also be represented as a single matrix, D. An approximation fz to the derivative of f at the collocation points is then given by fz = Df . (E.4) The matrix D is dense, meaning that there are generally very few elements of D that are zero, so all of the elements of f are involved in determining fz . As a result, spectral methods can be very accurate, even if very few collocation points are used. This accuracy comes at the price of requiring a dense matrix multiplication, which is computationally more expensive than a more ‘local’ method — such as the method we employ for the temporal integration (see below). As it happens there is a more efficient way to implement a discrete Fourier transform (DFT), known as the fast Fourier transform (FFT), which was developed by Cooley and Tukey in 1965 [203] . The FFT does not use explicit matrix multiplication, and it is generally employed for Fourier spectral methods. We do not use a Fourier spectral method in this thesis, however. One reason for this is that a DFT works by fitting a finite number of complex exponentials — which are periodic — to the vector f . Since the basis functions are periodic, the vector f must be periodic. Of course any function f (z) defined on a finite domain [0, 1] can be made periodic by ‘gluing’ copies of the E.1 Spectral Collocation 451 function onto either end of the domain, as shown in Figure E.2. But if f (0) 6= f (1), the resulting ‘extended’ function will contain discontinuous jumps, where the copies were glued together. These discontinuities represent a very rapid change in f , so that the spectrum fe(k) contains very high spatial frequencies. Given that f , and the approximated Fourier transform fe, are comprised of only a finite number of collocation points, the accuracy of the DFT is therefore severely compromised by these discontinuities. Fourier spectral methods are not well suited to dealing with functions that are not ‘really’ periodic, for this reason. Since we are modeling the spatial distribution of atomic excitations, which need have no intrinsic periodicity, it behooves us to use a method that is not subject to this limitation. Such a method is provided by polynomial differentiation matrices. (a) (b) Figure E.2 Periodic extension. (a): the function f satisfies f (0) = f (1), so that its periodic extension is smooth. The Fourier spectral derivative of this function is accurate. (b): now f (0) 6= f (1), and the periodic extension of f contains sharp discontinuities, which erode the accuracy of a Fourier spectral derivative. E.1 Spectral Collocation E.1.1 452 Polynomial Differentiation Matrices To generalize spectral differentiation to the case of non-periodic functions, we should use non-periodic basis functions. Instead of fitting the vector f with complex exponentials, as in the Fourier case, we fit f with an algebraic polynomial p(z). We can then easily differentiate this to get p0 (z), and the vector approximating the derivative of f is found by evaluating p0 at the collocation points. The basis functions from which we construct p are specified by requiring that p evaluates to f at the collocation points. We write p in terms of a set of basis functions {pj }, and the elements of f , as p(z) = X fk pk (z). (E.5) k Fixing p(zj ) = fj imposes the condition pk (zj ) = δjk . (E.6) The interpolant p(z) is a polynomial of degree N − 1, if there are N collocation points, and therefore the basis functions pk are also polynomials of degree N − 1. The condition (E.6) means that each of the N collocation points, except for zk , are roots of pk (z), and therefore pk (z) = 1 Y (z − zi ), ak i6=k with ak = Y i6=k (zk − zi ). (E.7) E.1 Spectral Collocation 453 Here ak is just the normalization required so that pk (zk ) = 1. Differentiating the relation (E.5), and setting z = zj gives p0 (zj ) = X p0k (zj )fk . (E.8) k Identifying the elements of fz with p0 (zj ), and using the definition (A.13) in Appendix A for matrix multiplication, we see the elements of the differentiation matrix D are given by Djk = p0k (zj ). (E.9) Taking the logarithm of (E.7) and differentiating, we find p0k (z) = pk (z) X i6=k 1 . z − zk (E.10) The elements of D are then given by the general formulae Djk = P i6=j 1 zj −zi a ak (zjj−zk ) E.1.2 if j = k, (E.11) otherwise. Chebyshev points We are now able to construct a differentiation matrix that works on an arbitrary set of collocation points {zj }. A grid of equally spaced points is the natural choice, but this is in fact disastrously unstable: the interpolant p develops pathological oscillations near the edges of the domain, close to z = 0 and z = 1. This is known E.1 Spectral Collocation 454 as the Runge phenomenon, and it is dealt with by using a set of collocation points that cluster together near the domain edges. To develop an intuition for why this might be, observe that the magnitude of one of the basis functions can be written in the form |pk (z)| = eV (z) , |ak | where V (z) = X ln |z − zi |. (E.12) i6=k Notice that V (z) has the same functional form as the electrostatic potential due to infinite lines of charge intersecting the z-axis at the collocation points zi . Therefore the size of the basis functions is exponentially sensitive to the potential energy associated with the ‘charge distribution’ describing the collocation points. The interpolant p is well-behaved when this energy is constant across the domain [0, 1]. If we allowed the collocation points to move along the z-axis according to the mutual repulsion between them represented by V , they would arrange themselves in a minimum-energy configuration which renders the potential flat, and the interpolant would be stable. In this configuration, the points are clustered together at the domain boundaries, and this explains why choosing a set of collocation points according to this configuration avoids the Runge phenomenon. A stable set of collocation points is provided by the Chebyshev points, which are the projections onto the domain of equally spaced points along a semicircle joining the domain boundaries (see Figure E.3): zj = 1 2 n h io 1 − cos π(j−1) . N −1 (E.13) E.2 Time-stepping 455 By substituting the points (E.13) into the formula (E.11), the appropriate Cheby- (a) (b) collocation points Figure E.3 Chebyshev Points. (a) Chebyshev collocation points are clustered towards the boundary of the domain. They are the downward projections of equally spaced points along a semicircle of radius 12 centred at z = 12 . (b) The method of lines with Chebyshev spectral collocation. shev differentiation matrix can be calculated. We use Matlab for all of our numerical computations, and the differentiation matrices are conveniently generated by a simple script called cheb.m, which is available online, and can be found in Nick Trefethen’s book [200] , along with many detailed examples of its application. Having introduced the method we use for the spatial derivatives, we now discuss the temporal derivatives, before describing how these two techniques are combined to solve the system (E.1). E.2 Time-stepping Suppose that we would like to solve the differential equation f 0 (τ ) = g [f (τ ), τ ] , with f (0) = f0 , (E.14) E.2 Time-stepping 456 where the prime now denotes differentiation with respect to τ . We assume that the boundary condition f0 and the function g are known. Note that g may depend on f . For instance, in our example, B 0 = −iA, but A depends on B. Now, the simplest numerical approach is to make a finite difference approximation to the time derivative. In terms of the discretized functions and coordinates, we have fk+1 − fk δτ = g (fk , τk ) ⇒ fk+1 = fk + δτ g (fk , τk ) , (E.15) where δτ = τk+1 − τk is the time step — assumed to be independent of k. The second line of (E.15) is a recursion relation, relating the future value of f to the present values of f and g. Starting with the boundary condition at τ = 0, we can use this relation to step forward in time, gradually building up the solution for f . This method is known as a first order Euler method, since errors of order δτ accumulate in the numerical solution. In our numerics, we use a second-order Runge-Kutta (RK2) method, which is only slightly more complicated, but accurate up to errors of order δτ 2 . The Runge-Kutta method reduces the errors by using values of the known function g at intermediate points, between τk and τk+1 . The recursion formula therefore requires that we discretize our functions on two grids: a primary grid, with times τk , which we denote with the same notation as used above, and a secondary, intermediate grid, with times τ k = τk + 32 δτ . The second order E.3 Boundary Conditions 457 Runge-Kutta method we employ for time stepping is then written as fk+1 = fk + δτ g (fk , τk ) + 3g f k , τ k , 4 (E.16) where the value of f on the secondary grid is approximated with the first order Euler formula (E.15), so that f k = fk + 23 δτ g (fk , τk ) . E.3 (E.17) Boundary Conditions It is clear how to implement the boundary condition Bin on B using the above time-stepping algorithm. This boundary condition simply tells us the initial values to use, on the first time step. However it is not so clear, from our discussion of spectral methods, how to implement the boundary condition Ain on A, which holds at z = 0 at all times. This is done by incorporating the boundary condition into the dynamical equation for A. First consider the discretized version of the equation ∂z A = −iB, which is DA = −iB, (E.18) where D is a Chebyshev differentiation martrix, which acts on each column of A to produce the corresponding column of −iB. There is no mention of the boundary condition so far. The discretized form of the boundary condition on A can be written E.3 Boundary Conditions 458 as 1 × [A11 A12 A13 · · · A1N ] = [Ain (τ1 ) Ain (τ2 ) Ain (τ3 ) · · · Ain (τN )] . (E.19) That is, 1 multiplied by the first row of A — which corresponds to the values of A at z = z1 = 0 — is equal to the discretized boundary condition. This equation can be ‘built in’ to the dynamical equation (E.18) simply by replacing the first row of D with the first row of the identity matrix I, and by replacing the first row of −iB with the discretized boundary condition. The equation becomes Db A = −iB b , (E.20) where the superscript b — for ‘boundary’ — indicates the modifications h b b b b D11 D12 D13 · · · D1N i = [1 0 0 · · · 0] , h i = i × [Ain (τ1 ) Ain (τ2 ) Ain (τ3 ) · · · Ain (τN )] .(E.21) b b b b B11 B12 B13 · · · B1N and All the other elements of Db and B b are the same as those of D and B. Now that we have fixed the boundary condition for A, we can solve for A in terms of B. Formally, we can invert the modified differentiation matrix Db to get A = −iDb−1 B b . (E.22) E.4 Constructing the Solutions 459 Of course, we do not know all the values of B until we have performed the timestepping, and this requires knowledge of A. In order to get to the solutions for A and B, we must build up B column by column, using the time-stepping, and solving for each column of A in turn using the above procedure. We are now ready to describe how this works. E.4 Constructing the Solutions Let us denote the columns of A by ak , and similarly the columns of B by bk : A= a a ... 1 2 a M , B= b b ... 1 2 b M (E.23) where M is the number of temporal discretization points. The algorithm proceeds as follows. First, we use the boundary condition Bin to set the values of b1 , B11 B 21 . . . BN 1 Bin (z1 ) B (z ) in 2 = .. . Bin (zN ) . (E.24) Then, we solve for a1 using the formula ak = −iDb−1 bbk . (E.25) , E.4 Constructing the Solutions 460 Here bbk is the k th column of B b . That is to say, bbk = bk , except that its first element is replaced by the signal field boundary condition, (bbk )1 = iAin (τk ). Now that we have both b1 and a1 , we implement the first stage of the RK2 iteration, which is to approximate the first time step on the intermediate grid, b1 , using (E.17), bk = bk − i 23 δτ ak . (E.26) Before we can implement the second part of RK2, we must approximate a1 . This is b done by using (E.25) again, this time replacing bbk with bk . That is, b ak = −iDb−1 bk . (E.27) This time, the modified vector contains the signal boundary condition evaluated b at τ = τ k . That is, bk = iAin τk + 23 δτ . Finally, the second part of RK2 is 1 implemented, which provides us with b2 . The formula is bk+1 = bk − i δτ (ak + 3ak ) . 4 (E.28) We have now succeeded in constructing b2 from our knowledge of b1 , and along the way we have found a1 . Iterating this procedure, we can construct b3 and a2 , and then b4 and a3 , and so on. In this way we proceed forward in time until we reach the end of our time domain, at τ = τM = 10. This completes the numerical solution. Here we comment that a dramatic increase in computational speed is achieved E.4 Constructing the Solutions 461 if Gaussian elimination is used instead of matrix inversion in (E.25) and (E.27). Gaussian elimination is an efficient method for solving the matrix equation Ax = y to obtain x, without explicitly calculating the inverse A−1 . In Matlab, this is implemented by the ‘backslash’ operator. For example, the Matlab code for (E.25) might look like A(:,k)=-i*inv(Db)*Bb(:,k), where the ‘inv’ function calls a matrix inversion routine. A much more efficient way to perform the same calculation would be coded as follows, A(:,k)=-i*( Db\Bb(:,k) ). This latter method becomes crucial to the feasibility of our chosen spectral method when modeling complicated dynamics. In order to produce an informative plot, we can use polynomial interpolation in order to replace the coarse Chebyshev grid of spatial collocation points with a finer, equally spaced grid. This utilizes the full accuracy of the spectral method, since the values at the collocation points are implicitly derived from the global interpolant p(z) at each time step. In practice, it is often faster, and quite acceptably accurate, to use a piecewise spline interpolant, which glues together multiple cubic polynomials in a continuous way, to join the collocation points. Routines for implementing this are standard in Matlab. In Figure E.4 we plot example solutions for A and B. Even using only 5 spatial collocation points, the solutions are accurate to ∼ 1%. E.5 Numerical Construction of a Green’s Function 462 Figure E.4 Example solutions. We plot the the squared moduli |A|2 and |B|2 of the solutions for A and B, found using the method of lines, as a function of z and τ . We used N = 5 spatial collocation points, on a Chebyshev grid, and M = 20 equally spaced time steps. The Matlab code runs in 0.02 seconds on a 3 GHz machine. The 2 signal boundary condition is a Gaussian pulse, Ain (τ ) = e−(τ −2.5) , and we assume there are no atomic excitations initially: Bin (z) = 0. The black lines indicate the time evolution of the spatial collocation points. The smooth surfaces are generated by spline interpolation in between these points. The red lines indicate the boundary conditions. The interpolated solutions are accurate to ∼ 1%, even with only 5 collocation points - this rapid convergence is a remarkable feature of spectral collocation. The behaviour is as might be expected: the signal field is absorbed as it propagates through the ensemble, with the transmitted field having undergone significant temporal distortion due to the spectral hole burnt by the atomic absorption line. The atomic excitation grows in time, assuming a roughly exponential shape in space, at the end of the interaction, consistent with Beer’s law. E.5 Numerical Construction of a Green’s Function Now that we are able to solve the system of coupled PDEs describing a quantum memory, for given boundary conditions, we can find the Green’s function — the storage kernel — for the interaction. As described in §5.4 in Chapter 5, this is done E.5 Numerical Construction of a Green’s Function 463 by solving with delta function boundary conditions for the signal field, successively varying the timing of the delta function to construct an approximation to the kernel. As usual for storage, we set Bin (z) = 0. Of course, it is not possible to implement a true delta function numerically. Instead, the delta function Ain (τ ) = δ(τ − τk ) is represented as [A11 A12 · · · A1k · · · A1N ] = [0 0 · · · 1 · · · 0] . (E.29) That is, along the line z = 0, the signal field is identically vanishing at all times, except at τ = τk , where it takes the value 1. Since the equations are linear, the absolute magnitude of Ain is not important, it only matters that the incident signal field is non-zero at just the single time step τ = τk . Due to the causal nature of the interaction, it is not necessary to numerically integrate from τ = 0 up to τ = τk . Since the signal field is zero in this region, and since there is no atomic excitation, the dynamics are trivial in this region. Therefore we only integrate from τ = τk up to τ = τM . Let bkM denote the vector representing the atomic excitation at the end of the interaction at τ = τM , produced by a delta function incident signal at τ = τk . The numerical approximation to the Green’s function is the matrix whose k th column is E.5 Numerical Construction of a Green’s Function 464 bkM , K= b1 M ... bM M b2 M . (E.30) Constructing the Green’s function is a more demanding computation than simply solving an instance of the equations with some particular boundary condition. First, the system of PDEs must be solved multiple times, and second, the sharp nature of the delta function boundary condition generally requires a smaller time step than is required for a smooth boundary condition. Below, in Figure E.5, we plot the numerically constructed Green’s function for our example system, alongside the analytic result, expressed as a Bessel function, which is derived in (3.38) in §3.5 of Chapter 3. With a sufficiently fine grid, the numerical and analytic results are indistinguishable. numerical analytic Figure E.5 A numerically constructed Green’s function. We used N = 30 spatial collocation points, and M = 4000 time steps. The computation takes around 10 minutes on a 3 GHz machine; the agreement with the analytic result is excellent. E.6 Spectral Methods for Two Dimensions E.6 465 Spectral Methods for Two Dimensions In Chapter 6 we describe simulations carried out in three dimensions — two space dimensions, and time. Needless to say, these simulations are much more time consuming than those involving only one spatial dimension. The method we use is essentially identical to that described in §E.4 above, except that now we use spectral collocation to deal with both spatial dimensions: we still use RK2 for the time stepping. In this section, we explain how to extend the spectral method to two dimensions. For concreteness, we will consider solving the simple example system (E.1) using only spectral collocation. The method is rather memory-intensive, and it quickly becomes unwieldy as the problem becomes complex, but it has the appeal of being very direct: we encode all of the dynamics into a single matrix equation, from which we extract the solutions in a single step. To see how this works, consider the following representation of the dynamical equations (E.1), ∂z i A = 0. i ∂τ B (E.31) If we could invert the matrix of derivatives, we might solve for A and B. But this matrix has no unique inverse: boundary conditions are needed. We proceed by discretizing the coordinates, and approximating (E.31) numerically using Chebyshev differentiation matrices to represent the differential operators. We then incorporate the boundary conditions into the resulting matrix equation, much as we did in §E.3 previously. Finally, we extract the solutions for A and B using Matlab’s backslash E.6 Spectral Methods for Two Dimensions 466 operator. In order to represent (E.31) numerically, we must convert A and B into column vectors. First, we discretize both the z and τ coordinates on a Chebyshev grid. We use N spatial collocation points and M temporal collocation points. The functions A and B are then represented as N × M matrices, whose columns describe spatial variation, and whose rows describe evolution in time. Let us denote the columns of A and B by {ak } and {bk }, as before (see (E.23)). We vectorize the matrices A and B by concatenating adjacent columns into a single column vector, of total length NM, a1 a 2 A = vec(A) = . . . aM , b1 b 2 B = vec(B) = . . . bM . (E.32) We now generate the differentiation matrices that will simulate the action of ∂z and ∂τ on A and B. Suppose that Dz is the differentiation matrix designed for the spatial Chebyshev grid defined on [0, 1] (Dz is therefore exactly the same as the matrix D discussed in §E.4). This acts on a column ak of A to produce the required derivative. The matrix Dz , which acts on A as a partial space derivative, is therefore E.6 Spectral Methods for Two Dimensions 467 given by Dz Dz Dz = .. . Dz = I ⊗ Dz , (E.33) where on the right hand side we have used the tensor product notation (see §A.2.2 in Appendix A). This expresses the fact that Dz acts on the spatial part of the vector A (within each column of A), whereas the temporal part (across columns of A) is unaffected. In a similar vein, we assemble the temporal partial derivative ∂τ by first generating a Chebyshev differentiation matrix Dτ for the temporal Chebyshev grid. Since the τ coordinate runs from 0 up to 10, the collocation points are given by τk = 10 × 1 2 n h io 1 − cos π(k−1) . M −1 Because of the factor of 10 appearing here, we have Dτ = (E.34) 1 10 D, where D is an M × M Chebyshev differentiation matrix generated for the domain [0, 1], like the one used for the spatial derivative previously. The operator Dτ representing a partial temporal derivative on B is then given by Dτ = Dτ ⊗ I. (E.35) This operator acts only on the rows B, leaving its columns unaffected. We are now E.6 Spectral Methods for Two Dimensions 468 able to write down our numerical approximation to (E.31), which is Dz i A = 0. i Dτ B (E.36) We can express this compactly as LX = 0, (E.37) in an obvious notation. In order to solve for X — i.e. for A and B — we must insert the boundary conditions for the signal field and atomic excitations. As in §E.3 above, the general approach is to incorporate the equation 1 × ( X at the boundary ) = ( appropriate boundary condition ) . (E.38) For instance, to fix the signal field boundary condition Ain (τ ), we identify those elements of X which describe A at z = 0. This is the first row of the matrix A, corresponding to the elements {1, N + 1, 2N + 1, . . . , (M − 1)N + 1}, (E.39) of X. The left hand side of (E.38) is then implemented in two steps. First, we set the rows of L with row indices given by (E.39) to zero. Second, we set the diagonal elements of L, with both indices given by (E.39), equal to one. The resulting modified E.6 Spectral Methods for Two Dimensions 469 matrix has the property that it maps the signal field at the boundary z = 0 to itself. The right hand side of (E.38) is realized by introducing a vector C into the right hand side of (E.37), with all elements zero, except for those with indices given by (E.39), which are set equal to the discretized boundary condition, C1 = Ain (τ1 ), (E.40) CN +1 = Ain (τ2 ), (E.41) .. . C(M −1)N +1 = Ain (τM ). (E.42) (E.43) The same procedure is used to implement the boundary condition Bin for the atomic excitations. This time the boundary is the first column of B, which is indices 1 to N of B, and therefore indices N M + 1 to N M + N of X. Finally, we are left with a modified equation Lb X = C, (E.44) where the superscript b distinguishes the modified matrix, and where C contains all the relevant boundary conditions. The formal solution is then X = Lb−1 C. (E.45) As mentioned previously, this is calculated much more efficiently if Gaussian elimination, rather than explicit matrix inversion, is used. In Matlab, we invoke the E.6 Spectral Methods for Two Dimensions 470 backslash operator. This completes the numerics. We are now in possession of approximations to both A and B at all spatial and temporal collocation points. Polynomial interpolation, or piecewise spline interpolation, provides us with an accurate representation of the solution for arbitrary z and τ . In Figure E.6 we plot the solutions for A and B found using this method. Figure E.6 Spectral methods in two dimensions. We use N = 5 spatial collocation points, and M = 15 temporal collocation points, to solve the example system (E.1), with the same boundary conditions as shown previously in Figure E.4. The code runs in 0.01 seconds on a 3 GHz machine, and the interpolated solutions shown are identical to those generated using spectral collocation and RK2 time stepping. Appendix F Atomic Vapours In this final Appendix we cover some results pertaining to atomic vapours, and in particular to our ongoing experiments with cesium vapour. We review the behaviour of vapour pressure with temperature, the concept of oscillator strength and its relation to the dipole moment associated with a transition, and the various mechanisms of line broadening. We finish with an analysis of the polarization of Stokes scattering in cesium. F.1 Vapour pressure No solid is infinitely sticky. Although the constituents may be bound tightly, there remains a finite probability that an atom will escape. Every solid is therefore accompanied by a diffuse cloud of free atoms, the pressure of which is known as its vapour pressure. It is this diffuse vapour that is used for the storage medium in our experiments (and many others). The density of the vapour determines its optical F.1 Vapour pressure 472 depth, and with it the achievable storage efficiency (see Chapter 5). To estimate this density, we treat the atoms as a classical statistical ensemble. The probability that an atom occupies a state with energy E is given by the Boltzmann distribution, p(E ) ∼ e−E /kB T . (F.1) The energy E required for an atom to escape depends on the potentials within the solid, and is largely independent of temperature. The ratio between the probabilities for escape at two temperatures T0 , T , is therefore p − E (1/T −1/T0 ) = e kB . p0 (F.2) This is, in fact, a good approximation for the vapour pressure above a solid, if p is now taken to be the pressure. The relation can be derived much more rigourously from the Clausius-Clapeyron equation [204] . Generally the reference state temperature T0 is taken to be the boiling point of the material, since then the vapour pressure p0 must be equal to the ambient pressure at this temperature. The energy E is commonly given as the molar enthalpy of sublimation ∆H. Boltzmann’s constant kB is then switched for the molar gas constant R = 8.31 JK−1 mole−1 . Part (a) of Figure F.1 shows the vapour pressure of thallium — the storage medium used in our preliminary experiments — using the values ∆H = 182 kJmole−1 , T0 = 1746 K and p0 = 5.8 × 105 Pa. Application of the ideal gas law p = nkB T allows us to convert the vapour pressure into the atomic number density, which is also shown in F.1 Vapour pressure 473 the plot. Calculation of the vapour pressure of cesium, which is shown in part (b), is complicated by the fact that cesium melts at 28◦ C. The qualitative behaviour is unchanged, but we use an empirical formula taken from the excellent document by Daniel Steck [184] , which is available online. (a) (b) 0 Pressure (Pa) 15 10 2 10 Cesium 22 10 10 10 −10 10 20 0 10 10 5 10 −20 0 10 5 −30 10 300 18 −2 10 350 400 450 Temperature (K) 10 500 10 10 16 −4 10 300 350 400 450 Temperature (K) 10 500 Number density (m -3 ) 10 Thallium Figure F.1 The vapour pressure of (a) thallium, and (b) cesium. The black curves show the vapour pressure in Pascals (left axis) as a function of temperature, measured in Kelvin. The red curves show the corresponding number density (right axis), calculated using the ideal gas law. As described in §10.9.1 of Chapter 10, a rough estimate of the optical depth of an ensemble is given by d = nλ2 L ∼ n × 10−14 , assuming optical wavelengths and a vapour cell a few centimeters long. From this it is clear that thallium has insufficient density for efficient storage at reasonable temperatures. Raising the temperature further starts to introduce a thermal background into the signal field from blackbody radiation! On the other hand, cesium has an optical depth of order 100, even at room temperature. F.2 Oscillator strengths F.2 474 Oscillator strengths The oscillator strength fjk is a dimensionless measure of the dominance of an atomic transition |ji ↔ |ki, compared to all other possible transitions. The oscillator strengths for all transitions from the state |ji sum to unity, X fjk = 1, ∀j. (F.3) k The relation between the fjk and the dipole matrix elements djk = hj|er|ki may be derived solely from this, and one other condition, which is that fjk ∝ |djk |2 . (F.4) This simply relates the f ’s to the transition probabilities. The following derivation is due to Charles Thiel at Montana State University. To determine the proportionality constant, we insert (F.4) into (F.3). X fjk = 1 = X = X k αjk |hj|er|ki|2 k k e2 αjk × 1 [hj|r|kihk|r|ji + hj|r|kihk|r|ji] . 2 (F.5) The development in the second line seems a little obtuse, but we now make use of the following relation between the matrix elements of the position and momentum F.2 Oscillator strengths 475 operators, pjk = hj|m∂t r|ki = imhj|[r, H]|ki = −imωjk rjk , (F.6) where m is the electron mass, and where ωjk is the frequency splitting between the states |ji, |ki, which are taken to be energy eigenstates of the Hamiltonian H. Substitution of this relation into (F.5) yields 1 = X ie2 αjk [hj|r|kihk|p|ji − hj|p|kihk|r|ji] 2mωjk k = = −ie2 α hj| [r, p] |ji 2m e2 ~α , 2m where in the penultimate line we used the decomposition of the identity I = (F.7) P j |jihj|, and where we somewhat heuristically set αjk = αωjk . In the final line we made use of the canonical commutation relation [r, p] = i~. The resulting expression for the oscillator strength is fjk = 2mωjk |djk |2 . ~e2 (F.8) The inverse relation is useful for calculating the dipole moment from the oscillator strengths listed in data tables. F.3 Line broadening F.3 476 Line broadening There are three processes that increase the absorption linewidth in our cesium vapour: Doppler broadening, pressure broadening and power broadening. An excellent treatment of these effects can be found in The Quantum Theory of Light by Loudon [107] . Here we briefly review the physics of line broadening. F.3.1 Doppler broadening Doppler broadening refers to the variation in the resonance frequencies of atoms moving with different velocities. Consider an atom emitting light of wavelength λ0 while moving at a velocity v (see Figure F.2). Over the course of an optical period T0 , the atom moves so that the wavelength appears compressed, λ = λ0 −vT0 . Using the relations λ0 = 2πc/ω0 , λ = 2πc/ω and T0 = 2π/ω0 , we derive the Doppler shift v δω = −ω0 , c (F.9) where δω = ω − ω0 is the shift in angular frequency caused by the motion of the atom. The spectral intensity is then given by the distribution of atomic velocities in the vapour. An atom with velocity v and mass M has a kinetic energy E = 12 M v 2 . (F.10) F.3 Line broadening 477 Using the Boltzmann distribution (F.1), and substituting for v using (F.9) gives the spectrum 2 I(δω) ∝ e−(δω/γd ) , (F.11) where the Doppler linewidth is given by r γd = ω0 2kB T × . M c (F.12) Figure F.2 The Doppler shift. An atom moving with velocity v ‘catches up’ with the light it emits, so that the wave appears squashed. F.3.2 Pressure broadening Collisions between atoms in a vapour cause disruptions to the wavetrain of light emitted by them. In particular, the phase of the light is randomized by each collision. To understand the effect on the spectrum of the light we should consider the statistics of collisions. An atom with cross sectional area σ travelling at velocity v sweeps out a volume F.3 Line broadening 478 Figure F.3 Collisions in a vapour. An atom with cross section σ travelling with velocity v sweeps out a volume vσdτ in a short time dτ . Any other atom within this volume gets hit! σvdτ in a short time dτ (see Figure F.3). The probability of a collision is just the probability of finding another atom within this volume, which is given by n × σvdτ , where n is the number density of atoms in the vapour. Now consider ps (τ ), the probability that an atom ‘survives’ without colliding for a time τ . We do not know this probability yet, but we can say the following, ps (τ + dτ ) = ps (τ ) × (1 − nσvdτ ) . (F.13) That is, the probability of surviving a further time dτ after τ is given by the probability of surviving for a time τ , and then not colliding during dτ . Taylor expanding (F.13) gives ps (τ ) + ∂τ ps (τ )dτ = ps (τ ) − nσvps (τ )dτ, ⇒ ∂τ ps (τ ) = −nσvps (τ ). (F.14) F.3 Line broadening 479 The survival probability is therefore given by an exponential distribution, ps (τ ) = γp e−γp τ , (F.15) where γp = nσv, and where the preceeding factor of γp ensures that the distribution is correctly normalized. We are now in a position to derive the form of a collision-broadened spectrum. The spectral intensity profile of the emitted light is given by e 2 I(ω) = E(ω) Z 1 ∗ = Fτ √ E (t)E(t + τ ) dt (ω), 2π (F.16) e where E(ω) and E(t) = E0 eiω0 t+φ(t) are the spectral and temporal electric field amplitudes of the light, and where in the second line we have used the convolution theorem (see §D.3.5 in Appendix D). The convolution inside the curly braces is the first order correlation function of E(t). In the absence of a collision, there is no change to E, and this correlation function is constant. After a collision, the phase φ is randomized, and when this is averaged over all atoms, the correlation function vanishes. The average correlation function is thus proportional to the probability that there is no collision in the interval [0, τ ]. Factorizing out the carrier frequency F.3 Line broadening 480 ω0 using the shift theorem (see §D.3.4 in Appendix D), the spectrum is given by I(δω) = = = Z τ |E0 |2 0 0 √ ps (τ )dτ (δω) × Fτ 1 − 2π 0 n o |E0 |2 √ × Fτ e−γp |τ | (δω) 2π γp |E0 |2 , × 2 2π γp + δω 2 (F.17) where in the second line the modulus sign indicates that the correlation function is symmetric in time. (F.17) describes a Lorentzian lineshape with a width γp . Although this spectrum was derived by considering emission, the absorption spectrum is identical, since absorption is simply the time reverse of emission. An estimate of the collision-broadened linewidth is given by r γp ≈ n(πd2atom ) × 2kB T . M (F.18) The number density n can be obtained from the vapour pressure as described in the start of this Appendix. The term in brackets approximates the collision cross-section σ as the area of a circle with radius given by the atomic width datom . The last factor is found by setting the kinetic energy (F.10) equal to the thermal energy kB T and solving for the average atomic velocity. F.3.3 Power broadening In the presence of an intense laser field, an atomic absorption line broadens. One way to understand this effect is to consider that the presence of light at the atomic F.4 Raman polarization 481 resonance frequency stimulates emission from the excited state, and this reduces the lifetime of the state, which introduces a concomitant broadening. Alternatively, consider that the atomic dynamics are governed by Hamiltonian evolution. In the loosest possible sense, the resulting behaviour can be characterized as a mixing of all the frequencies in the Hamiltonian. Thus if there exists an optical frequency ω and a Rabi frequency Ω, there will be driven oscillations at frequencies of ω ± Ω. Such oscillations may be resonant with the atomic transition at ω0 , even when neither frequency is separately. Therefore absorption is possible at large detunings, provided the laser is sufficiently intense. The atomic transition is rendered broader by the laser field. In fact, the power-broadened linewidth is indeed given by γpb = Ω, as a more detailed derivation from the damped optical Bloch equations shows [107] . This effect is essentially the same as the dynamic Stark effect that generates the instantaneous frequency shift in (5.74) in Chapter 5, and the Autler-Townes splitting discussed in §2.3.1 in Chapter 2. F.4 Raman polarization In Chapter 10 we describe the preliminary steps taken toward building a Raman quantum memory in cesium vapour. Here we explain how we arrive at the conclusion that the Stokes light scattered from a linearly polarized pump will be polarized orthogonally to the pump, when far detuned. We consider Raman scattering from the 6S1/2 ↔ 6P3/2 D2 line in cesium, at 852 nm. There are two hyperfine levels in the ground state, with F = 3 and F = 4. F.4 Raman polarization 482 Atoms prepared in one of these levels are transferred by the Raman interaction to the other level, with the transfer mediated by one of the excited states in the 6P3/2 manifold. In the process, a photon from the Raman pump is absorbed, and a Stokes photon is emitted1 . Each photon carries only one quantum of angular momentum, so the intermediate state cannot differ in its angular momentum quantum number by more than 1 from the initial and final states. Of the four hyperfine levels — F = 2, 3, 4, 5 — in the excited state manifold, only the central two with F = 3, 4 are compatible with this requirement, so we consider only these two intermediate levels. Within each hyperfine level, there are 2F + 1 Zeeman sublevels, with magnetic quantum numbers m = −F, −F +1, . . . , F −1, F . We specify one of these states with the notation |F, mi. The magnetic quantum numbers correspond to the component of the atomic angular momentum along some axis, known as the quantization axis. We must choose the direction for this axis before proceeding further. If a magnetic field is applied to the atoms, the only sensible choice is to align the quantization axis with the field. But in the absence of a magnetic field, the choice is arbitrary, and we may choose the quantization direction in whichever way is convenient for our calculations. As mentioned above, a single photon carries one unit of angular momentum, and so it can change the magnetic quantum number by at most ±1, or indeed not at all, if the spin of the photon has no component along the quantization 1 Strictly the term ‘Stokes’ refers to Raman scattered photons with less energy than the pump photons. When they have more energy than the pump, they are commonly termed anti-Stokes photons. In the present case, Stokes scattering only occurs if the atoms are prepared in the lower of the two ground states, with F = 3. The distinction between Stokes and anti-Stokes is not important for us, however. For simplicity, we always refer to the light emitted in Raman scattering as Stokes light, regardless of whether or not its frequency is lower than that of the Raman pump. F.4 Raman polarization 483 axis. Figure F.4 shows how the direction and polarization of a photon with respect to the quantization axis is connected with the changes in m induced by its absorption. quantization axis Figure F.4 Polarization selection rules. Right and left circular polarizations propagating along the quantization axis change m by +1 and −1; they are known as σ+ and σ− polarizations, respectively. It is intuitive that as the electric field follows a corkscrew trajectory, it applies a torque to the atoms about the quantization axis, increasing or decreasing m appropriately. π polarized light does not change m. It is polarized linearly along the quantization axis and propagates at right angles to it, so it cannot induce a turning moment. The orthogonal linear polarization can, however, change m by either +1 or −1. We choose the quantization axis so that the linearly polarized Raman pump is π-polarized. With this choice, absorption of a pump photon cannot change the magnetic quantum number m. Given the initial state |Fi , mi i of an atom, this narrows down the number of possible intermediate states to two: either |Fint = 3, mint = mi i or |Fint = 4, mint = mi i (the only exception is if mi = 0 initially. Then F must change, so there is only one possible intermediate state, with Fint 6= Fi ). The atom now decays from the intermediate excited state into the ‘final’ ground hyperfine level with Ff 6= Fi . There is no restriction on the polarization of the emitted Stokes photon, so there are three possibilities for this latter part of the interaction. The F.4 Raman polarization 484 possible final states are |Ff , mf = mi i and |Ff , mf = mi ±1i. As shown in Figure F.5, each final state |Ff , mf i can be reached from the given initial state |Fi , mi i by two alternative ‘paths’. The first via the excited state with Fm = 3, and the second with Fm = 4. When two different paths connect the same starting and ending points, there is the possibility of interference. The present case is archetypical: π-polarized Stokes emission is forbidden because of destructive interference between the two interaction pathways. To see this, we explicitly evaluate the quantum mechanical amplitude Aππ for π-polarized Stokes emission given a π-polarized pump, given the initial state |Fi , mi i, Aππ (Fi , mi ) = X d(J = 1/2, Fi , mi → J = 3/2, Fm , mm = mi ) × (F.19) Fm =3,4 d(J = 3/2, Fm , mm = mi → J = 1/2, Ff 6= Fi , mf = mi ). Here d denotes a dipole matrix element connecting the states with angular momentum quantum numbers indicated by its arguments. Recall that J = 3/2 describes the intermediate excited states. The selection rules set by fixing both photon polarizations to π keep the magnetic quantum number equal to its initial value throughout the interaction. We have factorized out any dependence of the dipole matrix elements on the radial (or principle) quantum number. That this is always possible is a consequence of the Wigner-Eckhart theorem. There exists a sophisticated and confusing mathematical apparatus for dealing with the resulting ‘reduced’ matrix F.4 Raman polarization 485 elements. In standard notation, we can write [184] d(J, F, m → J 0 , F 0 , m0 ) = (−1)F +J+I+1 p (2F + 1)(2J + 1) J J0 1 F0 F I 0 1F F C(m 0 −m)mm0 , (F.20) j1 j2 J where the curly braces denote a Wigner 6j symbol, and where Cm is a Clebsch1 m2 M Gordan coefficient. Here I = 7/2 is the cesium nuclear spin. The 1’s appearing in various places represent the angular momentum of the photon involved in the transition. Angular momentum conservation requires that the lower indices of the Clebsch-Gordan coefficient obey the sum rule m1 + m2 = M . The symbols in (F.20) can be evaluated by standard routines in Mathematica. Matlab routines have also been written, and are available freely on the web. Performing the sum in (F.19) using (F.20), we find that Aππ is identically zero for all choices of the initial state (i.e. either Fi = 3 or 4, and any value of mi ). No doubt there is a deep grouptheoretical reason for this, but I do not know what it is! Nonetheless the calculation shows that Stokes scattered photons emerge with the orthogonal linear polarization to that of the pump. A very fortunate outcome for the experimenter who wishes to filter the weak Raman signal out from the bright pump. In evaluating the scattering amplitude, we have given equal weight to both pathways. This implicitly assumes that the optical fields are sufficiently far detuned from the excited state manifold that the small energy splitting between the intermediate hyperfine levels — around 200 MHz — makes a negligible difference to the Raman coupling. If instead we tune into resonance with one of the levels, we can neglect F.4 Raman polarization 486 -3 -2 -1 0 1 2 3 -4 -3 -2 -1 0 1 2 3 4 -3 -2 -1 0 1 2 3 -4 -3 -2 -1 0 1 2 3 4 Figure F.5 Alternative scattering pathways. Each initial state is coupled by the Raman interaction to a final state via two possible interaction pathways; the first involves an intermediate excited state with Fm = 3 (black-blue), the second with Fm = 4 (black-red). The figure illustrates this with the example of an atom initially prepared in the state Fi = 4, mi = 1. Absorption of a π-polarized Raman pump photon couples this state to the intermediate states |Fm = 3, mm = 1i and |Fm = 4, mm = 1i. Finally emission of a Stokes photon reduces the magnetic quantum number by one: both paths end with the final state |Ff = 3, mf = 0i. scattering involving the other level, and there is no longer any interference. Isolating just a single term from the sum in (F.19), we calculate a Stokes π-polarization of 40% on resonance with the Fm = 3 state, averaged over all possible initial states. This averaging assumes an unpolarized ensemble, with atoms populating all Zeeman substates in the initial hyperfine level equally. If we tune into resonance with the Fm = 4 state, on average only 26% of the Stokes light is π-polarized. These calculations show that even on resonance, there remains a significant proportion of the Stokes scattering that is polarized orthogonally to the pump, so that polarization filtering is always feasible. Note that these conclusions hold equally well for the polarizations of the the signal and control fields in a cesium quantum memory, which has at its heart the very F.4 Raman polarization 487 same Raman interaction. Therefore a vertically polarized signal field can be stored by a horizontally polarized control pulse, when both are tuned far from resonance. If we choose a circularly polarized Raman pump — this time aligning the quantization axis with the pump propagation direction for convenience — we find that, far from resonance, all of the Stokes light is emitted into the same circular polarization as the pump. That is, right-circular scatters into right-circular, and left-circular into left-circular. This is rather unfortunate, because it precludes the use of what might otherwise have been a rather ingenious ‘trick’2 : Suppose that the atomic ensemble is prepared in the the state |Fi = 4, mi = 4i. Then a σ+ -polarized photon can only couple to a state with mm = 5, which lies within the Fm = 5 hyperfine level in the excited state. This level cannot participate in the Raman interaction (see the discussion at the start of this section), and so a σ+ -polarized photon cannot act as a Raman pump. If the ensemble is used as a quantum memory, then a σ+ polarized control pulse cannot cause spontaneous Stokes scattering. This eliminates the unwanted ‘noise’ process mentioned in §4.6 of Chapter 4 (see part (b) of Figure 4.3). Unfortunately, the signal photon must be σ− -polarized, and then the memory coupling vanishes because quantum interference between the two possible scattering pathways destroys any coupling between orthogonal circular polarizations. This explains why we make no attempt to spin-polarize our cesium ensemble: it is more trouble than it is worth! 2 This was suggested by Jean Dalibard when he came to Oxford to give a colloquium. Bibliography [1] CE Shannon. A mathematical theory of communication. ACM SIGMOBILE Mobile Computing and Communications Review, 5(1):3–55, 2001. [2] G.E. Moore. Cramming more components on to integrated circuits. Electronics Journal, 38(8):114–117, 1965. [3] R.P. Feynman. Feynman Lectures on Computation. Da Capo Press, 1996. [4] D. Deutsch. Quantum theory, the church-turing principle and the universal quantum computer. Proceedings of the Royal Society of London. Series A, Mathematical and Physical Sciences (1934-1990), 400(1818):97–117, 1985. [5] Benjamin Schumacher. Quantum coding. Phys. Rev. A, 51(4):2738–2747, Apr 1995. [6] A. M. Steane. Error correcting codes in quantum theory. Phys. Rev. Lett., 77(5):793–797, Jul 1996. [7] A. R. Calderbank and Peter W. Shor. Good quantum error-correcting codes exist. Phys. Rev. A, 54(2):1098–1105, Aug 1996. [8] W.K. Wootters and W.H. Zurek. A single quantum cannot be cloned. Nature, 299(5886):802–803, 1982. [9] M.A. Nielsen and I.L. Chuang. Quantum Computation and Quantum Information. Cambridge University Press, 2000. [10] A. Steane. The ion trap quantum information processor. Applied Physics B: Lasers and Optics, 64(6):623–643, 1997. [11] M. V. Gurudev Dutt, L. Childress, L. Jiang, E. Togan, J. Maze, F. Jelezko, A. S. Zibrov, P. R. Hemmer, and M. D. Lukin. Quantum register based on individual electronic and nuclear spin qubits in diamond. Science, 316(5829):1312– 1316, 2007. [12] L. Childress, MV Gurudev Dutt, JM Taylor, AS Zibrov, F. Jelezko, J. Wrachtrup, PR Hemmer, and MD Lukin. Coherent Dynamics of Coupled Electron and Nuclear Spin Qubits in Diamond, 2006. BIBLIOGRAPHY 489 [13] E. Knill, R. Laflamme, and G. J. Milburn. A scheme for efficient quantum computation with linear optics. Nature, 409(6816):46–52, 2001. [14] N. Yoran and B. Reznik. Deterministic linear optics quantum computation with single photon qubits. Phys. Rev. Lett., 91(3):037903, Jul 2003. [15] Michael A. Nielsen. Optical quantum computation using cluster states. Physical Review Letters, 93(4):040503, 2004. [16] J. M. Raimond, M. Brune, and S. Haroche. Manipulating quantum entanglement with atoms and photons in a cavity. Rev. Mod. Phys., 73(3):565–582, Aug 2001. [17] Karl Tordrup, Antonio Negretti, and Klaus Molmer. Holographic quantum computing. Physical Review Letters, 101(4):040501, 2008. [18] C.A. Muschik, I. de Vega, D. Porras, and J.I. Cirac. Quantum processing photonic states in optical lattices. Arxiv preprint quant-ph/0611093, 2006. [19] S. D. Barrett, P. P. Rohde, and T. M. Stace. Scalable quantum computing with atomic ensembles, 2008. [20] N. Gisin, G. Ribordy, W. Tittel, and H. Zbinden. Quantum cryptography. Reviews of Modern Physics, 74(1):145–195, 2002. [21] RL Rivest, A. Shamir, and L. Adleman. A method for obtaining digital signatures and public-key cryptosystems. Communications of the ACM, 21(2):120– 126, 1978. [22] M.A. Nielsen and I.L. Chuang. Quantum Computation and Quantum Information. Cambridge University Press, 2000. [23] P.W. Shor. Algorithms for quantum computation: Discrete logarithms and factoring. Proceedings of the 35th Annual Symposium on Foundations of Computer Science, 124, 1994. [24] C.Y. Lu, D.E. Browne, T. Yang, and J.W. Pan. Demonstration of Shor’s quantum factoring algorithm using photonic qubits. Arxiv preprint arXiv:0705.1684, 2007. [25] BP Lanyon, TJ Weinhold, NK Langford, M. Barbieri, DFV James, A. Gilchrist, and AG White. Experimental Demonstration of a Compiled Version of Shor’s Algorithm with Quantum Entanglement. Physical Review Letters, 99(25):250505, 2007. [26] L.M.K. Vandersypen, M. Steffen, G. Breyta, C.S. Yannoni, R. Cleve, and I.L. Chuang. Experimental Realization of an Order-Finding Algorithm with an NMR Quantum Computer. Physical Review Letters, 85(25):5452–5455, 2000. BIBLIOGRAPHY 490 [27] C.H. Bennett, G. Brassard, et al. Quantum cryptography: Public key distribution and coin tossing. Proceedings of IEEE International Conference on Computers, Systems, and Signal Processing, 175, 1984. [28] W. Heisenberg. The Physical Principles of the Quantum Theory. Courier Dover Publications, 1949. [29] JS Bell. On the Einstein-Podolsky-Rosen paradox. Physics (US) Discontinued with Vol. 4, no. 1, 1968, 1, 1964. [30] Artur K. Ekert. Quantum cryptography based on bell’s theorem. Phys. Rev. Lett., 67(6):661–663, Aug 1991. [31] Jian-Wei Pan, Dik Bouwmeester, Harald Weinfurter, and Anton Zeilinger. Experimental entanglement swapping: Entangling photons that never interacted. Phys. Rev. Lett., 80(18):3891–3894, May 1998. [32] D. Deutsch, A. Ekert, R. Jozsa, C. Macchiavello, S. Popescu, and A. Sanpera. Quantum Privacy Amplification and the Security of Quantum Cryptography over Noisy Channels. Physical Review Letters, 77(13):2818–2821, 1996. [33] W. Dür, H.-J. Briegel, J. I. Cirac, and P. Zoller. Quantum repeaters based on entanglement purification. Phys. Rev. A, 59(1):169–181, Jan 1999. [34] Charles H. Bennett, Herbert J. Bernstein, Sandu Popescu, and Benjamin Schumacher. Concentrating partial entanglement by local operations. Phys. Rev. A, 53(4):2046–2052, Apr 1996. [35] Jian-Wei Pan, Christoph Simon, Caslav Brukner, and Anton Zeilinger. Entanglement purification for quantum communication. Nature, 410(6832):1067– 1070, 2001. [36] Christoph Simon and Jian-Wei Pan. Polarization entanglement purification using spatial entanglement. Phys. Rev. Lett., 89(25):257901, Dec 2002. [37] L.M. Duan, MD Lukin, JI Cirac, and P. Zoller. Long-distance quantum communication with atomic ensembles and linear optics. Nature, 414:413–418, 2001. [38] Nicolas Sangouard, Christoph Simon, Bo Zhao, Yu-Ao Chen, Hugues de Riedmatten, Jian-Wei Pan, and Nicolas Gisin. Robust and efficient quantum repeaters with atomic ensembles and linear optics. Physical Review A (Atomic, Molecular, and Optical Physics), 77(6):062301, 2008. [39] L. Jiang, J. M. Taylor, and M. D. Lukin. Fast and robust approach to longdistance quantum communication with atomic ensembles. Physical Review A (Atomic, Molecular, and Optical Physics), 76(1):012301, 2007. BIBLIOGRAPHY 491 [40] Zeng-Bing Chen, Bo Zhao, Yu-Ao Chen, Jorg Schmiedmayer, and Jian-Wei Pan. Fault-tolerant quantum repeater with atomic ensembles and linear optics. Physical Review A (Atomic, Molecular, and Optical Physics), 76(2):022329, 2007. [41] Nicolas Sangouard, Christoph Simon, Jiri Minar, Hugo Zbinden, Hugues de Riedmatten, and Nicolas Gisin. Long-distance entanglement distribution with single-photon sources. Physical Review A (Atomic, Molecular, and Optical Physics), 76(5):050301, 2007. [42] Bo Zhao, Zeng-Bing Chen, Yu-Ao Chen, Jorg Schmiedmayer, and Jian-Wei Pan. Robust creation of entanglement between remote memory qubits. Physical Review Letters, 98(24):240502, 2007. [43] L.M. Duan, JI Cirac, and P. Zoller. Three-dimensional theory for interaction between atomic ensembles and free-space light. Physical Review A, 66(2):23818, 2002. [44] Peter J. Mosley, Jeff S. Lundeen, Brian J. Smith, Piotr Wasylczyk, Alfred B. U’Ren, Christine Silberhorn, and Ian A. Walmsley. Heralded generation of ultrafast single photons in pure quantum states. Physical Review Letters, 100(13):133601, 2008. [45] Axel Kuhn, Markus Hennrich, and Gerhard Rempe. Deterministic singlephoton source for distributed quantum networking. Phys. Rev. Lett., 89(6):067901, Jul 2002. [46] Christian Kurtsiefer, Sonja Mayer, Patrick Zarda, and Harald Weinfurter. Stable solid-state source of single photons. Phys. Rev. Lett., 85(2):290–293, Jul 2000. [47] A. Beveratos, S. Kühn, R. Brouri, T. Gacoin, J. P. Poizat, and P. Grangier. Room temperature stable single-photon source. The European Physical Journal D - Atomic, Molecular, Optical and Plasma Physics, 18(2):191–196, 2002. [48] I. I. Rabi, J. R. Zacharias, S. Millman, and P. Kusch. A new method of measuring nuclear magnetic moment. Phys. Rev., 53(4):318, Feb 1938. [49] H. Mabuchi and A. C. Doherty. Cavity quantum electrodynamics: Coherence in context. Science, 298(5597):1372–1377, 2002. [50] P. Rabl, D. DeMille, J. M. Doyle, M. D. Lukin, R. J. Schoelkopf, and P. Zoller. Hybrid quantum processors: Molecular ensembles as quantum memory for solid state circuits. Physical Review Letters, 97(3):033003, 2006. [51] J. I. Cirac, P. Zoller, H. J. Kimble, and H. Mabuchi. Quantum state transfer and entanglement distribution among distant nodes in a quantum network. Phys. Rev. Lett., 78(16):3221–3224, Apr 1997. BIBLIOGRAPHY 492 [52] M. Hennrich, T. Legero, A. Kuhn, and G. Rempe. Vacuum-stimulated raman scattering based on adiabatic passage in a high-finesse optical cavity. Phys. Rev. Lett., 85(23):4872–4875, Dec 2000. [53] EM Purcell. Spontaneous emission probabilities at radio frequencies. Phys. Rev, 69(681):166, 1946. [54] AD Boozer, A. Boca, R. Miller, TE Northup, and HJ Kimble. Reversible State Transfer between Light and a Single Trapped Atom. Physical Review Letters, 98(19):193601, 2007. [55] J. Bochmann, M. Muecke, G. Langfahl-Klabes, C. Erbel, B. Weber, H. P. Specht, D. L. Moehring, and G. Rempe. Fast Excitation and Photon Emission of a Single-Atom-Cavity System. Phys. Rev. Lett., 67(13):1727–1730, Sep 2008. [56] J. P. Reithmaier, G. Sek, A. Loffler, C. Hofmann, S. Kuhn, S. Reitzenstein, L. V. Keldysh, V. D. Kulakovskii, T. L. Reinecke, and A. Forchel. Strong coupling in a single quantum dot-semiconductor microcavity system. Nature, 432(7014):197–200, 2004. [57] A. Kiraz, P. Michler, C. Becher, B. Gayral, A. Imamoğlu, L. Zhang, E. Hu, WV Schoenfeld, and PM Petroff. Cavity-quantum electrodynamics using a single InAs quantum dot in a microdisk structure. Applied Physics Letters, 78:3932, 2001. [58] A. ImamogŻlu, DD Awschalom, G. Burkard, DP DiVincenzo, D. Loss, M. Sherwin, and A. Small. Quantum Information Processing Using Quantum Dot Spins and Cavity QED. Physical Review Letters, 83(20):4204–4207, 1999. [59] R.J. Young, S.J. Dewhurst, R.M. Stevenson, P. Atkinson, A.J. Bennett, M.B. Ward, K. Cooper, D.A. Ritchie, and A.J. Shields. Single electron-spin memory with a semiconductor quantum dot. Arxiv preprint arXiv:0706.2143, 2007. [60] A. André, D. Demille, JM Doyle, MD Lukin, SE Maxwell, P. Rabl, RJ Schoelkopf, and P. Zoller. A coherent all-electrical interface between polar molecules and mesoscopic superconducting resonators. Nature Physics, 2(9):636–642, 2006. [61] M.K. Tey, S.A. Aljunid, F. Huber, B. Chng, Z. Chen, G. Maslennikov, and C. Kurtsiefer. Interfacing light and single atoms with a lens. Arxiv preprint arXiv:0804.4861, 2008. [62] M.K. Tey, Z. Chen, S.A. Aljunid, B. Chng, F. Huber, G. Maslennikov, and C. Kurtsiefer. Strong interaction between light and a single trapped atom without a cavity. Arxiv preprint arXiv:0802.3005, 2008. BIBLIOGRAPHY 493 [63] M. Sondermann, R. Maiwald, H. Konermann, N. Lindlein, U. Peschel, and G. Leuchs. Design of a mode converter for efficient light-atom coupling in free space. Applied Physics B: Lasers and Optics, 89(4):489–492, 2007. [64] K.-J. Boller, A. Imamolu, and S. E. Harris. Observation of electromagnetically induced transparency. Phys. Rev. Lett., 66(20):2593–2596, May 1991. [65] Michael Fleischhauer, Atac Imamoglu, and Jonathan P. Marangos. Electromagnetically induced transparency: Optics in coherent media. Reviews of Modern Physics, 77(2):633, 2005. [66] J. H. Eberly, M. L. Pons, and H. R. Haq. Dressed-field pulses in an absorbing medium. Phys. Rev. Lett., 72(1):56–59, Jan 1994. [67] S. H. Autler and C. H. Townes. Stark effect in rapidly varying fields. Phys. Rev., 100(2):703–722, Oct 1955. [68] Lene Vestergaard Hau, S. E. Harris, Zachary Dutton, and Cyrus H. Behroozi. Light speed reduction to 17 metres per second in an ultracold atomic gas. Nature, 397(6720):594–598, 1999. [69] M. Fleischhauer and M. D. Lukin. Dark-state polaritons in electromagnetically induced transparency. Phys. Rev. Lett., 84(22):5094–5097, May 2000. [70] Chien Liu, Zachary Dutton, Cyrus H. Behroozi, and Lene Vestergaard Hau. Observation of coherent optical information storage in an atomic medium using halted light pulses. Nature, 409(6819):490–493, 2001. [71] K. S. Choi, H. Deng, J. Laurat, and H. J. Kimble. Mapping photonic entanglement into and out of a quantum memory. Nature, 452:67–71, March 2008. [72] Julien Laurat, Chin wen Chou, Hui Deng, Kyung Soo Choi, Daniel Felinto, Hugues de Riedmatten, and H J Kimble. Towards experimental entanglement connection with atomic ensembles in the single excitation regime. New Journal of Physics, 9(6):207, 2007. [73] T. Chaneliere, D. N. Matsukevich, S. D. Jenkins, S.-Y. Lan, R. Zhao, T. A. B. Kennedy, and A. Kuzmich. Quantum interference of electromagnetic fields from remote quantum memories. Physical Review Letters, 98(11):113602, 2007. [74] Irina Novikova, Nathaniel B. Phillips, and Alexey V. Gorshkov. Optimal light storage with full pulse shape control, 2008. [75] CV Raman and KS Krishnan. A new type of secondary radiation. Nature, 121(3048):501, 1928. [76] A. E. Kozhekin, K. Mølmer, and E. Polzik. Quantum memory for light. Phys. Rev. A, 62:033809, 2000. BIBLIOGRAPHY 494 [77] J. Nunn, I. A. Walmsley, M. G. Raymer, K. Surmacz, F. C. Waldermann, Z. Wang, and D. Jaksch. Mapping broadband single-photon wave packets into an atomic memory. Physical Review A (Atomic, Molecular, and Optical Physics), 75(1):011401, 2007. [78] Mattias Nilsson and Stefan Krll. Solid state quantum memory using complete absorption and re-emission of photons by tailored and externally controlled inhomogeneous absorption profiles. Optics Communications, 247(4-6):393– 403, 2005. [79] S. A. Moiseev and S. Kröll. Complete reconstruction of the quantum state of a single-photon wave packet absorbed by a doppler-broadened transition. Phys. Rev. Lett., 87(17):173601, Oct 2001. [80] SA Moiseev, VF Tarasov, and BS Ham. Quantum memory photon echo-like techniques in solids. Journal of Optics B: Quantum and Semiclassical Optics, 5(4):S497–S502, 2003. [81] S. A. Moiseev and B. S. Ham. Photon-echo quantum memory with efficient multipulse readings. Phys. Rev. A, 70(6):063809, Dec 2004. [82] B. Kraus, W. Tittel, N. Gisin, M. Nilsson, S. Kroll, and J. I. Cirac. Quantum memory for nonstationary light fields based on controlled reversible inhomogeneous broadening. Physical Review A (Atomic, Molecular, and Optical Physics), 73(2):020302, 2006. [83] AL Alexander, JJ Longdell, MJ Sellars, and NB Manson. Photon Echoes Produced by Switching Electric Fields. Physical Review Letters, 96(4):43602, 2006. [84] G. Hétet, JJ Longdell, AL Alexander, PK Lam, and MJ Sellars. Electro-Optic Quantum Memory for Light Using Two-Level Atoms. Physical Review Letters, 100(2):23601, 2008. [85] M. U. Staudt, S. R. Hastings-Simon, M. Nilsson, M. Afzelius, V. Scarani, R. Ricken, H. Suche, W. Sohler, W. Tittel, and N. Gisin. Fidelity of an optical memory based on stimulated photon echoes. Physical Review Letters, 98(11):113601, 2007. [86] E. Desurvire, C.R. Giles, and J.L. Zyskind. Erbium-doped fiber amplifier, June 25 1991. US Patent 5,027,079. [87] Nicklas Ohlsson, R. Krishna Mohan, and Stefan Krll. Quantum computer hardware based on rare-earth-ion-doped inorganic crystals. Optics Communications, 201(1-3):71–77, 2002. [88] E. Fraval, M. J. Sellars, and J. J. Longdell. Dynamic decoherence control of a solid-state nuclear-quadrupole qubit. Physical Review Letters, 95(3):030506, 2005. BIBLIOGRAPHY 495 [89] B. R. Judd. Optical absorption intensities of rare-earth ions. Phys. Rev., 127(3):750–761, Aug 1962. [90] G. Hétet, JJ Longdell, MJ Sellars, PK Lam, and BC Buchler. Bandwidth and Dynamics of the Gradient Echo Memory. Arxiv preprint arXiv:0801.3860, 2008. [91] Mikael Afzelius, Christoph Simon, Hugues de Riedmatten, and Nicolas Gisin. Multi-Mode Quantum Memory based on Atomic Frequency Combs, 2008. [92] Luı́s E. E. de Araujo, Ian A. Walmsley, and C. R. Stroud. Analytic solution for strong-field quantum control of atomic wave packets. Phys. Rev. Lett., 81(5):955–958, Aug 1998. [93] L. E. E. de Araujo and I. A. Walmsley. Quantum control of molecular wavepackets: An approximate analytic solution for the strong-response regime. Journal of Physical Chemistry A, 103(49):10409–10416, 1999. [94] Luı́s E. E. de Araujo and Ian A. Walmsley. Quantum control of rydberg wave packets in the strong-response regime. Phys. Rev. A, 63(2):023401, Jan 2001. [95] L.E.E. de Araujo and I.A. Walmsley. Analytic solution for quantum control of atomic and molecular wavepackets. JOURNAL OF OPTICS B QUANTUM AND SEMICLASSICAL OPTICS, 5(1):27–42, 2003. [96] J. Oreg, G. Hazak, and JH Eberly. Multilevel inversion schemes in and beyond the adiabatic limit. Physical Review A, 32(5):2776–2783, 1985. [97] S. Guérin, S. Thomas, and H. R. Jauslin. Optimization of population transfer by adiabatic passage. Phys. Rev. A, 65(2):023409, Jan 2002. [98] B. W. Shore, K. Bergmann, A. Kuhn, S. Schiemann, J. Oreg, and J. H. Eberly. Laser-induced population transfer in multistate systems: A comparative study. Phys. Rev. A, 45(7):5297–5300, Apr 1992. [99] B. Lauritzen, S. R. Hastings-Simon, H. de Riedmatten, M. Afzelius, and N. Gisin. State Preparation by Optical Pumping in Erbium Doped Solids using Stimulated Emission and Spin Mixing. Messenger of Mathematics, 19:1–5, 2008. [100] Hugues de Riedmatten, Mikael Afzelius, Matthias U. Staudt, Christoph Simon, and Nicolas Gisin. A solid-state light-matter interface at the single-photon level. Nature, 456(7223):773–777, 12 2008. [101] S.L. Braunstein and P. van Loock. Quantum information with continuous variables. Reviews of Modern Physics, 77(2):513–577, 2005. [102] Samuel L. Braunstein and Peter van Loock. Quantum information with continuous variables. Reviews of Modern Physics, 77(2):513, 2005. BIBLIOGRAPHY 496 [103] T. C. Ralph. Continuous variable quantum cryptography. Phys. Rev. A, 61(1):010303, Dec 1999. [104] C. Silberhorn, TC Ralph, N. Lütkenhaus, and G. Leuchs. Continuous Variable Quantum Cryptography: Beating the 3 dB Loss Limit. Physical Review Letters, 89(16):167901, 2002. [105] Stefano Pirandola, Stefano Mancini, Seth Lloyd, and Samuel L. Braunstein. Continuous-variable quantum cryptography using two-way quantum communication. Nat Phys, advanced online publication:–, 2008. [106] Lijian Zhang, Christine Silberhorn, and Ian A. Walmsley. Secure quantum key distribution using continuous variables of single photons. Physical Review Letters, 100(11):110504, 2008. [107] Rodney Loudon. The Quantum Theory of Light. Oxford University Press, 2004. [108] E. Wigner. On the quantum correction for thermodynamic equilibrium. Phys. Rev., 40(5):749–759, Jun 1932. [109] T. Hirano, K. Kotani, T. Ishibashi, S. Okude, and T. Kuwamoto. 3 dB squeezing by single-pass parametric amplification in a periodically poled KTiOPO4 crystal. Optics Letters, 30(13):1722–1724, 2005. [110] RJ Senior, GN Milford, J. Janousek, AE Dunlop, K. Wagner, H.A. Bachor, TC Ralph, EH Huntington, and CC Harb. Observation of a comb of optical squeezing over many gigahertz of bandwidth. Optics Express, 15(9):5310–5317, 2007. [111] Henning Vahlbruch, Moritz Mehmet, Simon Chelkowski, Boris Hage, Alexander Franzen, Nico Lastzka, Stefan Gossler, Karsten Danzmann, and Roman Schnabel. Observation of squeezed light with 10-db quantum-noise reduction. Physical Review Letters, 100(3):033602, 2008. [112] Y. Eto, T. Tajima, Y. Zhang, and T. Hirano. Observation of quadrature squeezing in a?ˆ(2) nonlinear waveguide using a temporally shaped local oscillator pulse. Optics Express, 16(14):10650–16657, 2008. [113] Alex Abramovici, William E. Althouse, Ronald W. P. Drever, Yekta Gursel, Seiji Kawamura, Frederick J. Raab, David Shoemaker, Lisa Sievers, Robert E. Spero, Kip S. Thorne, Rochus E. Vogt, Rainer Weiss, Stanley E. Whitcomb, and Michael E. Zucker. Ligo: The laser interferometer gravitational-wave observatory. Science, 256(5055):325–333, 1992. [114] Carlton M. Caves. Quantum-mechanical noise in an interferometer. Phys. Rev. D, 23(8):1693–1708, Apr 1981. BIBLIOGRAPHY 497 [115] H. Yuen and J. Shapiro. Optical communication with two-photon coherent states–part i: Quantum-state propagation and quantum-noise. Information Theory, IEEE Transactions on, 24(6):657–668, Nov 1978. [116] T. Fernholz, H. Krauter, K. Jensen, J. F. Sherson, A. S. Sorensen, and E. S. Polzik. Spin squeezing of atomic ensembles via nuclear-electronic spin entanglement. Physical Review Letters, 101(7):073601, 2008. [117] Vivi Petersen, Lars Bojer Madsen, and Klaus Molmer. Magnetometry with entangled atomic samples. Physical Review A (Atomic, Molecular, and Optical Physics), 71(1):012312, 2005. [118] D. Leibfried, M. D. Barrett, T. Schaetz, J. Britton, J. Chiaverini, W. M. Itano, J. D. Jost, C. Langer, and D. J. Wineland. Toward heisenberg-limited spectroscopy with multiparticle entangled states. Science, 304(5676):1476– 1478, 2004. [119] C. F. Roos, M. Chwalla, K. Kim, M. Riebe, and R. Blatt. /‘designer atoms/’ for quantum metrology. Nature, 443(7109):316–319, 2006. [120] P. J. Windpassinger, D. Oblak, P. G. Petrov, M. Kubasik, M. Saffman, C. L. Garrido Alzar, J. Appel, J. H. Muller, N. Kjaergaard, and E. S. Polzik. Nondestructive probing of rabi oscillations on the cesium clock transition near the standard quantum limit. Physical Review Letters, 100(10):103601, 2008. [121] Masahiro Kitagawa and Masahito Ueda. Nonlinear-interferometric generation of number-phase-correlated fermion states. Phys. Rev. Lett., 67(14):1852–1854, Sep 1991. [122] Brian Julsgaard, Jacob Sherson, J. Ignacio Cirac, Jaromir Fiurasek, and Eugene S. Polzik. Experimental demonstration of quantum memory for light. Nature, 432(7016):482–486, 2004. [123] K. Hammerer, A.S. Sorensen, and E.S. Polzik. Quantum interface between light and atomic ensembles. Not yet published, 2008. [124] Jun Liu, Songcan Chen, and Xiaoyang Tan. Fractional order singular value decomposition representation for face recognition. Pattern Recogn., 41(1):378– 395, 2008. [125] T. N. Palmer, R. Buizza, F. Molteni, Y.-Q. Chen, and S. Corti. Singular vectors and the predictability of weather and climate. Philosophical Transactions: Physical Sciences and Engineering, 348(1688):459–475, 1994. [126] Guifré Vidal. Entanglement of pure states for a single copy. Phys. Rev. Lett., 83(5):1046–1049, Aug 1999. [127] E. Beltrami. Sulle funzioni bilineari. Giornale di Matematiche ad Uso degli Studenti Delle Universita, 11:98–106, 1873. BIBLIOGRAPHY [128] C. Jordan. Sur la réduction des formes bilinéaires. l’Academie des Sciences, Paris, 78:614–617, 1874. 498 Comptes Rendus de [129] JJ Sylvester. A new proof that a general quadric may be reduced to its canonical form (that is, a linear function of squares) by means of a real orthogonal substitution. Messenger of Mathematics, 19:1–5, 1889. [130] E. Schmidt. Zur Theorie Der Linearen Und Nichtlinearen Integralgleichungen. I Teil: Entwicklung WillkuLicher Funktion Nach Systemen Vorgeschriebener. Mathenatische Annalen, 63:433–476, 1907. [131] H. Bateman. A formula for the solving function of a certain integral equation of the second kind. Transactions of the Cambridge Philosophical Society, 20:179– 187, 1908. [132] C. Eckart and G. Young. The approximation of one matrix by another one of lower rank. Psychometrika, 1:211–218, 1936. [133] Alexey V. Gorshkov, Axel Andre, Mikhail D. Lukin, and Anders S. Sorensen. Photon storage in lambda-type optically dense atomic media. ii. free-space model. Physical Review A, 76:033805, 2007. [134] J.M. Supplee. Optical Bloch equations: Informal motivation without the Schrödinger equation. American Journal of Physics, 68:180, 2000. [135] T.H. Boyer. Electrostatic potential energy leading to a gravitational mass change for a system of two point charges. American Journal of Physics, 47(2):129–31, 1979. [136] B.R. Holstein and A.R. Swift. Elementary derivation of the radiation field from an accelerated charge. American Journal of Physics, 49:346, 1981. [137] R.P. Feynman, R.B. Leighton, M. Sands, and EM Hafner. The Feynman Lectures on Physics; Vol. I. American Journal of Physics, 33:750, 1965. [138] A. Kalachev. Quantum storage on subradiant states in an extended atomic ensemble. Physical Review A, 76(4):43812, 2007. [139] M. G. Raymer and J. Mostowski. Stimulated Raman scattering: Unified treatment of spontaneous initiation and spatial propagation. Phys. Rev. A, 24(4):1980–1993, 1981. [140] Alexey V. Gorshkov, Axel Andre, Mikhail D. Lukin, and Anders S. Sorensen. Photon storage in lambda-type optically dense atomic media. ii. free-space model. Physical Review A (Atomic, Molecular, and Optical Physics), 76(3):033805, 2007. [141] R.W. Boyd. Nonlinear Optics. Academic Press, 2003. BIBLIOGRAPHY 499 [142] N. Rosen and C. Zener. Double Stern-Gerlach Experiment and Related Collision Phenomena. Physical Review, 40(4):502–507, 1932. [143] P. Pechukas and J.C. Light. On the Exponential Form of Time-Displacement Operators in Quantum Mechanics. The Journal of Chemical Physics, 44:3897, 2004. [144] RT Robiscoe. Extension of the Rosen-Zener solution to the two-level problem. Physical Review A, 17(1):247–260, 1978. [145] S. Klarsfeld and JA Oteo. Magnus approximation in the adiabatic picture. Physical Review A, 45(5):3329–3332, 1992. [146] WR Salzman. An alternative to the magnus expansion in time-dependent perturbation theory. The Journal of Chemical Physics, 82:822, 1985. [147] Samuel L. Braunstein. Squeezing as an irreducible resource. Physical Review A (Atomic, Molecular, and Optical Physics), 71(5):055801, 2005. [148] W. Wasilewski and MG Raymer. Pairwise entanglement and readout of atomic-ensemble and optical wave-packet modes in traveling-wave Raman interactions. Physical Review A, 73(6):63816, 2006. [149] W. Wasilewski, AI Lvovsky, K. Banaszek, and C. Radzewicz. Pulsed squeezed light: Simultaneous squeezing of multiple modes. Physical Review A, 73(6):63819, 2006. [150] Irina Novikova, Alexey V. Gorshkov, David F. Phillips, Anders S. Sorensen, Mikhail D. Lukin, and Ronald L. Walsworth. Optimal control of light pulse storage and retrieval. Physical Review Letters, 98(24):243602, 2007. [151] Nathaniel B. Phillips, Alexey V. Gorshkov, and Irina Novikova. Optimal light storage in atomic vapor, 2008. [152] K. Surmacz, J. Nunn, K. Reim, K. C. Lee, V. O. Lorenz, B. Sussman, I. A. Walmsley, and D. Jaksch. Efficient spatially resolved multimode quantum memory. Physical Review A (Atomic, Molecular, and Optical Physics), 78(3):033806, 2008. [153] C. Simon, H. de Riedmatten, M. Afzelius, N. Sangouard, H. Zbinden, and N. Gisin. Quantum repeaters with photon pair sources and multimode memories. Phys. Rev. Lett., 98:190503, 2007. [154] O. A. Collins, S. D. Jenkins, A. Kuzmich, and T. A. B. Kennedy. Multiplexed memory-insensitive quantum repeaters. Physical Review Letters, 98(6):060502, 2007. BIBLIOGRAPHY 500 [155] J. Nunn, K. Reim, K. C. Lee, V. O. Lorenz, B. J. Sussman, I. A. Walmsley, and D. Jaksch. Multimode memories in atomic ensembles. Physical Review Letters, 101(26):260502, 2008. [156] J. H. Eberly. Schmidt Analysis of Pure-State Entanglement. ArXiv Quantum Physics e-prints, August 2005. [157] R. Grobe, K. Rzazewski, and JH Eberly. Measure of electron-electron correlation in atomic physics. Journal of Physics B: Atomic, Molecular, and Optical Physics, 27(16):L503–L508, 1994. [158] M. A. Nielsen and I. L. Chuang. Quantum Computation and Quantum Information. Cambridge University Press, 2004. [159] G.N. Watson. A Treatise on the Theory of Bessel Functions. Cambridge University Press, 1995. [160] Nicolas Sangouard, Christoph Simon, Mikael Afzelius, and Nicolas Gisin. Analysis of a quantum memory for photons based on controlled reversible inhomogeneous broadening. Physical Review A, 75:032327, 2007. [161] JJ Longdell, G. Hétet, MJ Sellars, and PK Lam. Analytic treatment of CRIB Quantum Memories for Light using Two-level Atoms. Arxiv preprint arXiv:0807.0067, 2008. [162] G. Hetet, J. J. Longdell, M. J. Sellars, P. K. Lam, and B. C. Buchler. Multimodal properties and dynamics of gradient echo quantum memory. Physical Review Letters, 101(20):203601, 2008. [163] Alexey V. Gorshkov, Axel Andre, Mikhail D. Lukin, and Anders S. Sorensen. Photon storage in lambda-type optically dense atomic media. iii. effects of inhomogeneous broadening. Physical Review A, 76:033806, 2007. [164] G. Hétet, M. Hosseini, B. M. Sparkes, D. Oblak, P. K. Lam, and B. C. Buchler. Photon echoes generated by reversing magnetic field gradients in a rubidium vapor. Opt. Lett., 33(20):2323–2325, 2008. [165] S.A. Moiseev and W. Tittel. Raman-Echo Quantum Memory. 2008. [166] Alexey V. Gorshkov, Tommaso Calarco, Mikhail D. Lukin, and Anders S. Sorensen. Photon storage in lambda-type optically dense atomic media. iv. optimal control using gradient ascent. Physical Review A (Atomic, Molecular, and Optical Physics), 77(4):043806, 2008. [167] C. Kittel and P. McEuen. Introduction to solid state physics. Wiley New York, 1986. [168] PG Klemens. Anharmonic Decay of Optical Phonons. 148(2):845–848, 1966. Physical Review, BIBLIOGRAPHY 501 [169] A. Debernardi, S. Baroni, and E. Molinari. Anharmonic Phonon Lifetimes in Semiconductors from Density-Functional Perturbation Theory. Physical Review Letters, 75(9):1819–1822, 1995. [170] M.S. Liu, L.A. Bursill, S. Prawer, and R. Beserman. Temperature dependence of the first-order Raman phonon line of diamond. Physical Review B, 61(5):3391–3395, 2000. [171] FC Waldermann, B.J. Sussman, J. Nunn, VO Lorenz, KC Lee, K. Surmacz, KH Lee, D. Jaksch, IA Walmsley, P. Spizziri, et al. Measuring phonon dephasing with ultrafast pulses using Raman spectral interference. Physical Review B, 78(15), 2008. [172] A. Cantarero, C. Trallero-Giner, and M. Cardona. Excitons in one-phonon resonant raman scattering: Deformation-potential interaction. Phys. Rev. B, 39(12):8388–8397, Apr 1989. [173] W. Pötz and P. Vogl. Theory of optical-phonon deformation potentials in tetrahedral semiconductors. Physical Review B, 24(4):2025–2037, 1981. [174] GD Whitfield. Theory of Electron-Phonon Interactions. Physical Review Letters, 2(5):204–205, 1959. [175] H. Frohlich. Interaction of electrons with lattice vibrations. Proceedings of the Royal Society of London. Series A, Mathematical and Physical Sciences (1934-1990), 215(1122):291–298, 12 1952. [176] Kun Huang. On the interaction between the radiation field and ionic crystals. Proceedings of the Royal Society of London. Series A, Mathematical and Physical Sciences (1934-1990), 208(1094):352–365, 09 1951. [177] K. Rzazewski and RW Boyd. Equivalence of interaction Hamiltonians in the electric dipole approximation. Journal of Modern Optics, 51(8):1137–1147, 2004. [178] M. O. Scully and M. S. Zubairy. Quantum Optics. Cambridge University Press, 1997. [179] R. P. Feynman. Forces in molecules. Phys. Rev., 56(4):340–343, Aug 1939. [180] W. Hayes, R. Loudon, and J.F. Scott. Scattering of Light by Crystals. American Journal of Physics, 47:571, 1979. [181] R. Loudon. Theory of the First-Order Raman Effect in Crystals. Proceedings of the Royal Society of London. Series A, Mathematical and Physical Sciences (1934-1990), 275(1361):218–232, 1963. [182] R. Loudon. Theory of Stimulated Raman Scattering from Lattice Vibrations. Proceedings of the Physical Society, 82:393–400, 1963. BIBLIOGRAPHY 502 [183] A.D. Greentree, P. Olivero, M. Draganski, E. Trajkov, J.R. Rabeau, P. Reichart, B.C. Gibson, S. Rubanov, S.T. Huntington, D.N. Jamieson, et al. Critical components for diamond-based quantum coherent devices. JOURNAL OF PHYSICS CONDENSED MATTER, 18(21):825, 2006. [184] Daniel A. Steck. Cesium D line data. 2008. [185] J.L. Picque. Hyperfine optical pumping of cesium with a CW GaAs laser. Quantum Electronics, IEEE Journal of, 10(12):892–897, 1974. [186] Zhaoming Zhu, Daniel J. Gauthier, and Robert W. Boyd. Stored Light in an Optical Fiber via Stimulated Brillouin Scattering. Science, 318(5857):1748– 1750, 2007. [187] Ashok K. Mohapatra, Mark G. Bason, Bjorn Butscher, Kevin J. Weatherill, and Charles S. Adams. A giant electro-optic effect using polarizable dark states. Nat Phys, 4(11):890–894, 11 2008. [188] D. E. Chang, V. Gritsev, G. Morigi, V. Vuletic, M. D. Lukin, and E. A. Demler. Crystallization of strongly interacting photons in a nonlinear optical fibre. Nat Phys, 4(11):884–889, 11 2008. [189] G. Golub and W. Kahan. Calculating the singular values and pseudo-inverse of a matrix. Milestones in Matrix Computation: The Selected Works of Gene H. Golub with Commentaries, 2007. [190] GW Stewart. On Scaled Projections and Pseudo-inverses. University of Maryland, 1988. [191] A.E. Albert. Regression and the Moore-Penrose pseudoinverse. Academic Press, New York, 1972. [192] TNE Greville. Some Applications of the Pseudoinverse of a Matrix. SIAM Review, 2:15, 1960. [193] M. Born and P. Jordan. On quantum mechanics. Z. Phys, 34(11-12):858–888, 1925. [194] P.A.M. Dirac. The Principles of Quantum Mechanics. Oxford University Press, 1958. [195] M. Born. Quantenmechanik der Stoßvorgänge. Hadrons and Nuclei, 38(11):803–827, 1926. Zeitschrift für Physik A [196] D.J. Amit and V. Martin-Mayor. Field theory, the renormalization group, and critical phenomena. World Scientific Singapore, 1984. [197] RG Woolley. Gauge invariant wave mechanics and the Power-Zienau-Woolley transformation. Journ. Phys. A, 13:2795, 1980. BIBLIOGRAPHY 503 [198] M. Babiker and R. Loudon. Derivation of the power-zienau-woolley hamiltonian in quantum electrodynamics by gauge transformation. Proceedings of the Royal Society of London. Series A, Mathematical and Physical Sciences, 385(1789):439–460, 1983. [199] S.C. Reddy and L.N. Trefethen. Stability of the method of lines. Numerische Mathematik, 62(1):235–267, 1992. [200] L.N. Trefethen. Spectral Methods in MATLAB. Society for Industrial Mathematics, Philadelphia, 2000. [201] D. Gottlieb and S.A. Orszag. Numerical Analysis of Spectral Methods: Theory and Applications. Society for Industrial Mathematics, 1977. [202] J.P. Boyd. Chebyshev and Fourier Spectral Methods. Courier Dover Publications, 2001. [203] J.W. Cooley and J.W. Tukey. An algorithm for the machine computation of complex Fourier series. Mathematics of Computation, 19(90):297–301, 1965. [204] RG Gareev, FA Arslanov, and GG Telyashev. THEORETICAL PRINCIPLES. Chemistry and Technology of Fuels and Oils, 37(4), 2001.