The Critical Points of Coherent Information on the Manifold of Positive Definite Matrices
by Alireza Tehrani
A Thesis presented to The University of Guelph
In partial fulfilment of requirements for the degree of Master of Science in Mathematics
Guelph, Ontario, Canada c Alireza Tehrani, January, 2020 ABSTRACT
THE CRITICAL POINTS OF COHERENT INFORMATION ON THE MANIFOLD OF POSITIVE DEFINITE MATRICES
Alireza Tehrani Advisors: University of Guelph, 2020 Dr. Bei Zeng Dr. Rajesh Pereira
The coherent information of quantum channels plays a important role in quantum infor- mation theory as it can be used to calculate the quantum capacity of a channel. However, it is a non-linear, non-differentiable optimization problem. This thesis discusses that by restricting to the space of positive definite density matrices and restricting the class of quan- tum channels to be strictly positive, the coherent information becomes differentiable. This allows the computation of the Riemannian gradient and Hessian of the coherent information. It will be shown that the maximally mixed state is a critical point for the n-shot coherent information of the Pauli, dephrasure and Pauli-erasure channels. In addition, the classifica- tion of the maximally mixed state as a local maxima/minima and saddle-point will be solved for the one shot coherent information. The hope of this work is to provide a new avenue to explore the quantum capacity problem. Dedication
To those who have inspired or encouraged me: Brother, Father, Randy, Farnaz and Paul.
iii Acknowledgements
I want to express my gratitude and thanks to Dr. Zeng for the guidance, advice and direction that was provided and for giving me the opportunity to work in quantum information theory and to be part of IQC and the Peng Cheng Laborotory. It has been a inspiring experience to witness the level of enthusiasm with respect to science that a person could have. In addition to Dr.Pereira, whom made sure I was on the right path, answering any questions and providing new suggestions. I’m thankful for the time spent doing research and words can never be enough to express the kindness that was shown. Couldn’t ask for better office mates, David, Eric, Katrina, Momina and Ningping thanks for the time spent together.
iv Contents
Abstract ii
Dedication iii
Acknowledgements iv
List of Figures viii
1 Introduction 1 1.1 Introduction to Problem ...... 1 1.1.1 Contributions ...... 2 1.2 Overview of Chapters ...... 3
2 Coherent Information 5 2.1 von Neumann Entropy ...... 5 2.1.1 Classical Entropy ...... 6 2.1.2 Density Matrices ...... 7 2.1.3 Quantum/von Neumann Entropy ...... 8 2.2 Coherent Information ...... 11 2.2.1 Quantum Channels ...... 11 2.2.2 Motivation ...... 15 2.2.3 Coherent Information of a channel ...... 17 2.3 Properties ...... 19 2.3.1 Data Processing and Error-Correction ...... 19 2.3.2 Quantum Capacity ...... 20 2.3.3 Lipschitz ...... 21 2.3.4 Conjecture: Optima on Positive Definite Density Matrices ...... 22 2.4 Known Results ...... 22 2.4.1 Amplitude-Damping Channel ...... 23 2.4.2 Pauli Channel ...... 23 2.4.3 Dephrasure Channel ...... 24 2.4.4 Degradable and Antidegradable Quantum Channels ...... 25
v 3 Introduction to Manifold Theory and The Gradient/Hessian 27 3.1 Preliminaries to Manifold Theory ...... 28 3.1.1 Manifolds ...... 28 3.1.2 Smooth Maps ...... 33 3.1.3 Tangent Space ...... 36 3.1.4 Differential ...... 37 3.1.5 Tangent Bundle ...... 43 3.1.6 Submanifolds ...... 44 3.1.7 Riemannian Manifolds ...... 46 3.2 Riemannian Gradient and Hessian ...... 48 3.2.1 Euclidean Gradient ...... 49 3.2.2 Riemannian Gradient ...... 50 3.2.3 Hessian ...... 53
4 Gradient/Hessian of Coherent Information on Positive Definite Matrices 59 4.1 Strictly Positive Quantum Channels ...... 60 4.1.1 Definition ...... 60 4.1.2 Dense ...... 63 4.1.3 Examples ...... 64 4.2 Differentiability of Coherent Information ...... 64 4.2.1 Differentiability of Entropy Exchange and Coherent Information . . . 65 4.3 Gradient and Hessian ...... 66 4.3.1 Gradient ...... 66 4.3.2 Hessian ...... 71
5 Critical Points and Local Maximas/Minimas 76 5.1 n-shot Coherent Information ...... 77 5.1.1 Critical Points of Product States ...... 77 5.2 Unital Channels ...... 80 5.3 Pauli Channel ...... 83 5.3.1 Dephasing Channel ...... 87 5.3.2 Depolarizing Channel ...... 88 5.4 Dephrasure Channel ...... 95 5.5 Pauli Erasure Channel ...... 103
6 Conclusion And Further Work 107 6.1 Further Work ...... 109
Bibliography 112
vi A Matrix Calculus 115 A.1 Matrix Functions ...... 115 A.2 Frechet and Gateux Derivatives ...... 117 A.3 Matrix Exponential and Logarithm ...... 118 A.4 Power Series ...... 120
vii List of Figures
2.1 The initial model and assumptions of the coherent information...... 16
5.1 Coherent Information of Depolarizing Channel Evaluated at the Maximally Mixed State...... 91 5.2 Maximum of Coherent Information of Depolarizing Channel...... 92 5.3 von Neumann entropy of the optimal solution of maximum of coherent infor- mation ...... 93 5.4 The eigenvalue of the Hessian at the Maximally Mixed state of coherent in- formation...... 94
viii Chapter 1
Introduction
1.1 Introduction to Problem
One of the most challenging problems of Quantum Information is the evaluation of quantum capacity Q(N ) of a quantum channel N , the maximal amount of quantum information that can be sent through N . It was shown in [22], [11], and [28], known as the LSD theorem, that it can be calculated as I (N ⊗n) Q(N ) = lim c , n→∞ n c where Ic(N ) = maxρ∈D S(N (ρ)) − S(N (ρ)) is the coherent information of the quantum ⊗n channel N and D is the space of all density matrices. The term Ic(N ) is called the n-shot coherent information. It is the quantum analogue of mutual information found in classical information theory. Unlike the classical case, quantum capacity does not have a single letter expression [27]. This is due to the existence of superadditivity
⊗n Ic(N ) > nIc(N ), for some channel use n. Note that in general, for all n the following inequality holds ⊗n Ic(N ) ≥ nIc(N ). It was shown in 2005 by Devetak and Shor [10], that if the quantum channel is degradable then the quantum capacity has a single letter formula as Q(N ) = Ic(N ). This is due to ⊗n the fact that for all n, the coherent information is subadditive Ic(N ) = nIc(N ). If the quantum channel is antidegradable, then the quantum capacity is known to be zero. The notion of approximate degradable quantum channels were introduced in 2017 and bounds
1 on the quantum capacities were found [32] which were further studied in [18]. The quantum capacity is generally an extremely difficult problem. It is shown in 2015, that an unbounded number of channel uses of optimizing the coherent information may be required to find non-zero quantum capacity [7]. Due to this result, the problem shifted to understanding examples of when superadditivity occurs. This is further studied in the following papers [21] (2017) and [19] (2018), respectively. The overall goal of this thesis is to investigate the critical points (ie when its gradient is zero) of the n-shot coherent information in a coordinate invariant way. This is done by first investigating when the coherent information becomes a smooth map on the space of positive definite density matrices D++. The reasoning behind this restriction is that the global optima is conjectured to generically be positive definite. Additionally, the set of positive definite density matrices is dense inside the space of all density matrices and it has a smooth manifold structure. This allows the calculation of the gradient and Hessian of the coherent information. It will have the advantage of being coordinate invariant and will be shown that critical points can be completely studied on the space of positive definite matrices M++ (rather than D++).
1.1.1 Contributions
The contributions of this thesis is as follows.
• The restriction of the coherent information to the manifold of postive definite matrices and the conditions on the quantum channel N and the complementary channel N c to insure smoothness of coherent information is explored. This is heavily tied to the notion of a strictly positive map [4]. This is discussed in chapter four.
• The Riemannian Gradient of the coherent information has been solved within the dense class of strictly positive quantum channels. It will be shown that a positive definite
density matrix ρ is a critical point of the coherent information Ic of a strictly positive quantum channel N iff −N † log(N (ρ)) + (N c)† log(N c(ρ)) ∈ span{I},
where log is the matrix logarithm, N c is the complementary channel, N † is the adjoint of a channel and I is the identity matrix. This is the statement of Theorem 4.3.2.
2 • It will be shown that the coherent information Ic(N , ρ) can be written using the gra- dient term above, ie
Ic(N , ρ) = hgrad(Ic(N , ρ))|ρi,
† where hA|Bi = T r(A B) is the Frobenius inner product and the gradient grad(Ic(N , ρ)) is −N †(log(N (ρ))) + N c†(log(N c(ρ))) on the manifold of positive definite matrices. It will be shown that the coherent information is a positively homogeneous function of degree zero.
• In addition, the Riemannian Hessian Hess(Ic(N , ρ)) of the coherent information Ic at a positive definite density matrix ρ is solved. It is a linear map from trace zero Hermitian matrices H0 to itself such that † c † c Hess(Ic(N , ρ))[V ] = P − N d logN (ρ) N (V ) + (N ) d logN c(ρ) N (V ) ,
where V is a trace zero Hermitian matrix, P is the orthogonal projection from Her- 0 mitian matrices H to trace zero, Hermitian matrices H and d logX : H → H is the differential of the matrix logarithm. This is a statement of Theorem 4.3.5.
• It will be shown that critical points ρ of k-shot coherent information become critical points of nk-shot coherent information via the product state ρ⊗n. This is the topic in section 5.1.
• The maximally mixed state is shown to always be a critical point for n-shot coherent information for the Pauli and Pauli+erasure channel. Indicating the role it has as a bifurcation point. This is explored in chapter five.
• In addition, the eigenvalues of the Hessian at the maximally mixed state are solved indicating when the maximally mixed state is a local maxima/minima or saddle point for the single shot coherent information. The result for dephasing and dephrasure channels match those in literature and the results for the depolarizing channel match computational results.
1.2 Overview of Chapters
Chapter two introduces the coherent information of a quantum channel. It will begin with a discussion of von Neumann entropy, following the coherent information and its known
3 properties. It will finish with known results on the coherent information with respect to amplitude-damping channel, and dephrasure channel. Chapter three introduces first the basics of smooth manifold theory needed to show that the positive definite density matrices D++ is a manifold and that its tangent space, the space of all directional derivaties, is isomorphic to the space of trace zero Hermitian matri- ces. Finally, the minimal requirements of Riemannian manifold theory is shown to formally define the gradient and Hessian with emphasis placed on submanifold that are isometrically embedded into a Euclidean space. All examples illustrate applications pertaining to the von Neumann entropy as a smooth map on the manifold D++. Chapter four first introduces the conditions needed on the quantum channel N and the complementary channel N c such that the coherent information becomes a smooth map on the manifold D++. It will be shown that the class of strictly positive quantum channel, ie it is positive definite invariant, will insure that the channel entropy S(N ) is smooth. Additionally, it will be shown that strictly positive quantum channels are dense within the class of quantum channels, known before in [35]. The complementary channel N c is going to be related to the Gram matrix, which then will be shown is strictly positive when the Kraus
operators {Ai} of N are linearly independent. Finally the chapter ends with the calculation of the gradient and Hessian of the coherent information alongside discussing their properties pertaining to critical points and local maximas/minimas. It will show that the gradient is invariant under positive scalar multiplication of density matrices and that the coherent information can be written in terms of the gradient, linking it to the Euler’s Homogenous function Theorem. Chapter five applies the results of the previous chapter to different classes of quantum channels. It will begin with a discussion of n-shot coherent information and how critical points can be formed via product states of critical points of ’lower’ shot coherent information. The main result is that the maximally mixed state is always a critical point for the n-shot coherent information of the Pauli, dephrasure and Pauli-erasure channels. The eigenvalues of the Hessian of these channels will be solved at the maximally mixed state for the one- shot coherent information. The results obtained from the dephrasure channel will match the results obtained from the paper [19], validating the formulas for the gradient and Hessian. For the depolarizing channel with channel parameter p, it will be shown that when p ∈ [0, 0.07055) the maximally mixed state is a local maxima, p ∈ {0.07055} will be a saddle- 1 point and p ∈ (0.07055, 3 ] will be a local minima for the one shot coherent information.
4 Chapter 2
Coherent Information
In classical information theory, Claude Shannon’s famous paper [27] introduces the notion of information and entropy, the expectation value of the information content of a probability distribution. In turn, he defines the mutual information of two probability distributions in order to quantify the notion of channel capacity, a measure of how much information passes through a classical communication channel. The quantum analogue of mutual information is known as the coherent information of a channel. Similar to mutual information, this can be used to define the quantum analogue of channel capacity, known as the quantum capacity. This measures the maximal amount of quantum information that passes through a quantum communication channel. This chapter is about the definition and properties of coherent information. Section 2.1 introduces quantum analogue of entropy alongside its properties. After doing so, the coherent information of a channel is motivated and defined in section 2.2. The coherent information has applications to familiar concepts from classical information theory, such as data-processing inequality, error-correction, and the quantum capacity. The coherent information will be shown to be a Lipschitz function which satisfies Levy’s lemma in section 2.3. The last section 2.4 introduces examples of quantum channels where results on either quantum capacity or the coherent information is known and serve as examples used in the later chapters of this thesis.
2.1 von Neumann Entropy
This section introduces the notion of classical entropy as the measure of disorder of a prob- ability distribution. The properties of classical entropy and the joint classical entropy will
5 be presented. The section will conclude with the quantum analogue of entropy known as the von Neumann entropy. Its properties and joint entropy will also be introduced with its connection to quantum entanglement. For the purposes of this thesis, only discrete random variables and finite-dimensional Hilbert spaces are presented.
2.1.1 Classical Entropy
Let X : E → R be a discrete random variable over a finite set of outcomes E. Suppose pX (X(e)) is a probability distribution over the random variable X, that gives the probability of a event e in E occurring. For simplicity, we will defined the probability of event e occurring as pX (e) instead of pX (X(e)).
The information content iX (e) of a event e in random variable X measures the amount of information in bits needed to represent e. Claude Shannon showed in [27] that the infor- mation content is defined by iX (e) = − log2(pX (e)), where we identify the logarithm base two of zero to be zero. The classical entropy H(X) of a discrete random variable X is defined to be the ex- pectation value EX [iX ] of the information content over X. This is precisely H(X) = P e∈E pX (e) log2(pX (e)). The classical entropy H(X) can be thought of as a measure of the disorder of X or the average number of bits needed to represent the random variable X. The following is a series of properties that the classical entropy has. Proofs of these are found in [34].
Theorem 2.1.1. Consider the classical entropy H(X) over a discrete random variable X with probability distribution pX . The following property holds for the classical entropy:
1. The classical entropy is non-negative over all discrete random variables X.
2. The classical entropy is concave over any two discrete random variables X1 and X2, ie
αH(X1) + (1 − α)H(X2) ≤ H(αX1 + (1 − α)X2).
3. The classical entropy is zero only when the probability distribution pX is one for a event e and zero everywhere else.
4. The maximum of classical entropy occurs when all events e have the same probability 1 ie pX (e) = |E| , where |E| is the cardinality of the set of all outcomes E.
6 5. The classical entropy is continuous function as a function over the space of all proba- bility distributions of events E.
Given two discrete random variables, X and Y , the joint probability distribution is
denoted as pX,Y . The classical entropy H(X,Y ) of two random variables is defined to be the expectation value of the information content of the joint probability distribution. This is referred as the joint entropy of two random variables X and Y .
The conditional probability pX|Y =y(x) is the probability of event x over the r5andom variable X if the event y for the random variable Y occurred. The conditional entropy H(X|Y = y) over the realization of Y = y is the entropy over the conditional probability
distribution pX|Y =y. The conditional entropy H(X|Y ) is the following (Definition 10.2.1 in [34]): X X H(X|Y ) = − pX,Y (x, y) log(pX|Y =y(x)). x y
2.1.2 Density Matrices
Density operators are the analogue of probability distributions on quantum states. |ψihψ| is the density operator with probability one that |ψi is obtained. Every density operator ρ can P be written as a convex-sum of pure-states |ψihψ| ∈ L(H). In other words, ρ = λi|ψiihψi| P such that λi = 1. The space of all density matrices D is the convex hull of all quantum states in H ie
X X D = conv{|ψihψ| : |||ψi|| = 1} = ci|ψihψ| : ci = 1 and |||ψi|| = 1 .
We have the following characterization of density operators which can be found in any standard quantum information book [34].
Theorem 2.1.2. A operator ρ ∈ L(H) is a density operator iff it is a trace-one, positive semidefinite Hermitian operator.
A density operator ρ that is rank-one implies it is a pure-state ie ρ = |ψihψ|. A density operator that is rank k can be written as a convex sum of k-pure states with no fewer pure- states can be removed. A maximal-rank density operator ρ is one whose rank is equivalent to dimension of H.
7 2.1.3 Quantum/von Neumann Entropy
Let H be a k-dimensional Hilbert space whose unit vectors represent the quantum states that can model a single particle. Define a discrete random variable X : H → R over the Hilbert space such that only a finite set in H has non-zero measure.
This defines an noise ensemble of quantum state, ie a set {(pX (i), |ψii)} where pX (i) is
the probability of obtaining the quantum state |ψii. The density matrix ρ of this ensemble
can be defined to be the expectation value over all quantum states |ψii representing as a
bounded linear operator in L(H) as |ψiihψi| (via Riesz representation Theorem), ie
X ρ = EX [|ψiihψi|] = pX (i)|ψiihψi|.
The von Neumann entropy S(ρ) of a density matrix ρ is defined to be the trace of the P classical entropy of the probability distribution pX , ie S(ρ) = i pX (i) log2(pX (i)). Similar to classical entropy, this measures how ”mixed/disorder” the ensemble is. This is equivalent to the definition proposed above. Suppose one has a density matrix ρ inside the space of linear bounded operators L(H) on a Hilbert space H, ie a trace-one, positive semidefinite Hermitian matrix. The unitary diagonalization of hermitian matrices is needed:
Theorem 2.1.3 (Unitary Diagonalization of Hermitian Matrices). Let X be a Hermitian matrix. Then X admits a spectral/unitary decomposition as X = UΣU †, where U is a unitary matrix and Σ is a diagonal matrix with real entries.
From this theorem, one has a diagonalization of the form ρ = UΣU †, where U are
unitary operators on H that represents the eigenvectors |ψii and Σ is a diagonal matrix with
eigenvalues λi. The trace of a density matrix is the sum of eigenvalues and being positive semidefinite matrix requires that all eigenvalues are greater than or equal to zero. Hence,
the eigenvalues λi of ρ is a number between zero and one, whose sum over all eigenvalues
is one. This exactly represents the probability of obtaining the eigenvector state |ψii. The
density matrix ρ represents the ensemble {(λi, |ψii)} of quantum states with probability λi
of having |ψii. The von Neumann entropy of ρ is defined to be the classical entropy over density matrix based on its unitary diagonalization. The previous result motivates the following definition of quantum entropy.
Definition 2.1.1 (von Neumann entropy). Given a density matrix ρ of L(H) for some † Hilbert space H. Consider the diagonalization ρ = UΣU of ρ, with eigenvalues λi.
8 The von Neumann entropy S : D → R over the space of all density matrices D is defined to be the classical entropy of its diagonalization,
S(ρ) = H({λi}). (2.1)
Alternatively, the von Neumann entropy can be defined using the trace-operator and the logarithm of the matrix as follows,
S(ρ) = T r(ρ log(ρ)). (2.2)
where the logarithm of a matrix log(ρ) is defined on the unitary diagonalization of ρ as a matrix function. In other words, the matrix logarithm is defined as follows, log(ρ) = log(UΣU †) = U log(Σ)U †, where UΣU † is the unitary diagonalization of ρ and log(Σ) is the matrix where logarithm base two is applied to each of its eigenvalue. It is assumed that the logarithm of zero, ie log(0) = 0.
The alternative definition 2.2 is equivalent to 2.1 by the following.
S(ρ) = T r(ρ log(ρ)) (Definition) = T r((UΣU †) log(UΣU †)) (Diagonalization) = T r(UΣU †U log(Σ)U †) (Definition of log) = T r(UΣ log(Σ)U †) (Definition of Unitary) = T r(U †UΣ log(Σ)) (Cyclic Property of Trace) = T r(Σ log(Σ)) (Definition of Unitary) X = λi log(λi) (Trace is sum of eigenvalues)
Similar to classical entropy, the von Neumann entropy shares many of the same properties. Proofs of which can be found in [34].
Theorem 2.1.4. The von Neumann entropy S : D → R over the space of density matrices D has the following properties for all density matrix ρ.
1. The von Neumann entropy is non-negative over all density matrices, S(ρ) ≥ 0.
2. von Neumann entropy is concave, ie let σ be another density matrix and β ∈ [0, 1] :
S(βρ + (1 − β)σ) ≤ βS(ρ) + (1 − β)S(σ)
9 3. The maximum of von Neumann occurs at the maximally mixed state I/d, where d is the the dimension of H.
4. Consider a metric on D, based on the trace norm, as ||ρ − σ|| = T r p(ρ − σ)†(ρ − σ) .
The von Neumann entropy is continuous with respect to the topology on D induced from that metric.
The last point actually holds for all norms on the space of bounded linear operators L(H) of Hilbert space H. This is because the Hilbert space considered here are finite-dimensional and hence so is L(H). It is then known that all norms are equivalent on finite-dimensional Hilbert spaces (see Theorem 5.10.6 in [25]).
Joint Quantum Entropy
A quantum state representing a n-particle is described as a unit vector of the tensor product Nn ⊗n of single-particle Hilbert spaces, ie H ⊗ · · · ⊗ H := i=1 H := H . Similarly, the n- particle density operator is described as a trace one, positive semidefinite and Hermitian Nn operator in L( i=1 H). Choosing a basis representation on H induces a basis representation Nn on L( i=1 H) to obtain a n-particle density matrix ρ1···n. Consider the following to define a description of k-particles that arose from a n-particle description. The partial trace T rj over the jth particle is defined to be a function from Nn the n-particle description L( i=1 H) to a function from the (n − 1)-particle description Nn K L( i=1,i6=j H). The partial trace has the following action where we define {|li}l=1 to be a orthonormal basis set of the jth particle, k-dimensional Hilbert space H,
K X ⊗(j−1) ⊗n−j ⊗(j−1) ⊗n−j T rj(ρ1···n) = I ⊗ hl| ⊗ I ρ1···n I ⊗ |li ⊗ I . l=1
The joint quantum entropy of n-particles is defined to be quantum entropy of ρ1···n as defined earlier. Furthermore, the conditional quantum entropy over k-particles from n- particles is defined to be the von Neumann entropy of T rk···nρ1···n where (n − k) particles are traced-out. The following are properties of the joint quantum entropy. Proofs can be found in [34].
10 Theorem 2.1.5 (Properties of Joint Quantum Entropy.). Let HA and HB be two finite- dimensional Hilbert space.
1. Let ρ ∈ L(HA) and σ ∈ L(HB) be two density matrices. Then the von Neumann entropy S is additive on ρ ⊗ σ, ie
S(ρ ⊗ σ) = S(ρ) + S(σ).
2. Let |φAQi be a unit vector in HA ⊗ HB. Consider its partial trace on system A and Q, respectively:
ρA = T rQ(|φihφ|AQ)
ρQ = T rA(|φihφ|AQ)
Then S(ρA) = S(ρQ) while S(|φihφ|AQ) = 0.
2.2 Coherent Information
This section will first introduce and motivate the coherent information of a channel with respect to a input quantum state. The final subsection will then introduce the coherent information of a channel and show that it is always a positive quantity. Before doing so, the notion of a quantum channel, isometric extensions and their adjoints needs to be formally addressed.
2.2.1 Quantum Channels
Denote H to be a finite-dimensional Hilbert space. Denote L(HA, HB) to be the space of bounded linear operators between HA and HB. This subsection is largely based on [34]. Quantum channels are a type of model for the evolution of density matrices.
Definition 2.2.1 (Axiomatic Definition of Quantum Channels.). A linear map N : L(HA) →
L(HB) is said to be a quantum channel if it satisfies the following:
1. It is linear ie N (λρ + σ) = λN (ρ) + N (σ).
2. It is trace-perserving ie T r(N (ρ)) = T r(ρ).
11 3. It is completely-positive, ie the map (IR ⊗ N ) maps positive semidefinite operators to positive semidefinite operators for all reference systems R.
The next theorem is a more concrete representation of a quantum channel.
Definition 2.2.2 (Kraus Representation of Quantum Channel). A map N : L(HA) →
L(HB) is a quantum channel iff there exists operators {Ai ∈ L(HA, HB)} such that
X † N (ρ) = AiρAi
P † where ρ ∈ L(HA) and Ai Ai = I, where I is the identity map on HA. Each operator Ai is called a Kraus operator. The minimal number of Kraus operators is called the Choi rank.
The analogue of purification of a quantum state is the isometric extension.
Theorem 2.2.1 (Isometric Extension). Let N : L(HA) → L(HB) be a quantum channel.
Let HE be the environment Hilbert space that has no dimension higher than the Choi-rank. † The isometric extension of N is a isometry operator U : HA → HB ⊗ HE (ie U U = I) such that † N (ρ) = T rE UρU holds for all ρ ∈ L(HA).
If {Ai} are the Kraus operators for N , a isometric extension can be obtained as U = P Ai ⊗ |iiE, where {|iiE} forms a orthonormal basis on HE.
Alternatively, given a isometric extension U, then the Kraus operators Ai can be defined as
Ai := (IB ⊗|iihi|E)U, where IB is the identity operator on L(HB) and {|iiE} is a orthonormal
basis on L(HE).
The isometric extension can be used to define the complementary channel, which is a mapping of the system to the environment.
Definition 2.2.3 (Complementary Channel). Let N : L(HA) → L(HB) be a quantum
channel. Let HE be the environmental system whose dimension is no larger than the Choi- rank of N . c The complementary channel is a map N : L(HA) → L(HE) such that for any isometric
extension U : HA → HB ⊗ HE the following holds:
c † N (ρ) = T rB UρU
12 for all ρ ∈ L(HA).
The complementary channel has a simple form when the Kraus operators are {Ai}. P Choose a isometric extension as U = i Ai ⊗ |iiE, then
c † N (ρ) = T rB UρU X † = T rB (Ai ⊗ |iiE)ρ(Aj ⊗ hj|E) ij X † = T rB (Ai ⊗ |iiE)(ρ ⊗ IE)(Aj ⊗ hj|E) ij X † = T rB (AiρAj) ⊗ |iihj|E ij X X † = (hk| ⊗ IE) (AiρAj) ⊗ |iihj|E (|ki ⊗ IE) k ij X † = T r(AiρAj)|iihj|E. ij
c Hence in basis representation |iiE, the complementary channel is the matrix N with entries c † c N ij = T r(AiρAj). This matrix N is called the entropy exchange matrix (for this thesis it will sometimes refer entropy exchange as the entropy of the entropy exchange matrix S(N c)).
Definition 2.2.4 (Adjoint of a quantum channel). Let N : L(HA) → L(HB) be a quantum † channel. The adjoint of the channel N is the unique linear operator from L(HB) to L(HA) such that hN †(σ)|ρi = hσ|N (ρ)i
for all ρ ∈ L(HA) and σ ∈ L(HB).
The adjoint of a quantum channel N with Kraus operators {Ai} can trivially be seen † P † to be N (ρ) = Ai ρAi. The adjoint of the complementary channel is a little harder to characterize.
Theorem 2.2.2 (Adjoint of complementary channel). Let N : L(HA) → L(HB) be a quan- c tum channel. Let N : L(HA) → L(HE) be the complementary channel. Then the adjoint c† c N : L(HE) → L(HA) of N is
c† X † N (σ) = hi|σ|jiAi Aj. ij
13 Proof. The proof will be done straight from definition.
c † X † hσ|N (ρ)i = T r(σ T r(AiρAj)|iihj|E) ij X † † † = T r(AiρAj)T r(σ |iihj|E) (Tr is linear and T r(AiρAj) is a number.) ij X † † = T r(AiρAj)T r(hj|σ |ii) ( Trace cyclic) ij X † † † = T r(AiρAj)hj|σ |ii ( hj|σ |ii is a number) ij X † † † = T r(hj|σ |iiAiρAj)( hj|σ |ii is a number) ij X † † = T r(Ajhj|σ |iiAiρ) ( Trace cyclic) ij X † † = T r( Ajhj|σ |iiAiρ) ( Trace linear) ij X † † † = T r( hj|σ |iiAjAiρ)( hj|σ |ii is a number) ij = hN c†(σ)|ρi
The next proof shows what the adjoint of the channel is in terms of the isometric exten- sion.
Theorem 2.2.3 (Adjoint Channel with Isometric Extension). Let N : L(HA) → L(HB) be a quantum channel. Let U : HA → HB ⊗ HE be the isometric extension of N . Then the † adjoint channel N : L(HB) → L(HA) is
† † N (σ) = U σB ⊗ IE U,
where IE : HE → HE is the identity channel on HE. c † Similarly, the adjoint of the complementary channel (N ) : L(HE) → L(HA) is
c † † (N ) (σ) = U (IB ⊗ σ)U,
where σ ∈ L(HE) and IB is the identity channel on HB.
14 Proof. Let σ ∈ L(HB) and ρ ∈ L(HA). Let {|ii} for a orthonormal basis for environ- † ment system HE. From theorem 2.2.1, the channel N (ρ) is T rE(UρU ). Starting with the definition of the adjoint channel by definition.
hN c(σ)|ρi = hσ|N (ρ)i † = hσ|T rE(UρU )i X † = hσ| (IB ⊗ hi|E)UρU (IB ⊗ |iiE)i ( Definition of partial trace) i X † = hU (IB ⊗ |iiE)σ(IB ⊗ hi|E)U|ρi (Trace is cyclic and linear) i † X = hU (σ ⊗ |iihi|E)U|ρi (σ is in L(HB)) i † = hU (σ ⊗ IE)U|ρi ( Definition of Identity.)
† † Hence, the adjoint channel N (σ) is U (σ ⊗ IE)U. The exact proof can be shown by re- placing the partial trace over system E with system B to show the adjoint channel of the complementary channel.
2.2.2 Motivation
Consider the following scenario, first motivated in [26]. There are three systems, two of which are the internal system A and Q which are accessible to the experimenter and the third is the external system E modeling the environment which is inaccessible to the experimenter. Each one of these systems is modeled as a Hilbert space HA, HQ and HE, respectively. Lastly, consider a noisy quantum channel NA : L(HA) → L(HA) that acts on system A, but not system Q. From Theorem 2.2.1, every quantum channel can be associated to a isometry operator UAE : HA ⊗ HE → HA acting on a larger system AE that includes the system A together with the environment system E [31].
Suppose initially, the experimenter constructs a pure state |ψiAQ ∈ HA ⊗HQ of system A and Q. Suppose the environment starts off as a pure state |0iE, which respect to some basis element |0i in HE. The total initial description of all systems A, Q and E is |ψiAQ ⊗ |0iE ∈
HA ⊗ HQ ⊗ HE. This is reflected in the following figure 2.2.2.
The coherent information of N with respect to |ψiAQ is a measure on how well the entanglement between system A and system Q from |ψiAQ is preserved if system A goes through the noisy channel NA. Since von Neumann entropy is one way of measuring en-
15 |ψAQi (IQ ⊗ UAE)(|ψAQi ⊗ |0Ei
UAE |0Ei
Figure 2.1: The initial model and assumptions of the coherent information. tanglement. Then it can be formulated by looking at the von Neumann entropy of system
A from |ψAQihψAQ| after going through noisy channel NA (by taking the partial trace with respect to system Q). Then comparing it to the von Neumann entropy of the AQ system
|ψAQihψAQ| after going the channel (IQ ⊗ NA). Denoting the initial state of system A as ρA = T rQ |ψAQihψAQ| . It is the following formula,