Binary Subspace Chirps
Tefjol Pllaha*, Olav Tirkkonen*, and Robert Calderbank†
*Department of Communications and Networking, Aalto University, Finland †Department of Electrical and Computer Engineering, Duke University, NC, USA Emails: {tefjol.pllaha, olav.tirkkonen}@aalto.fi, [email protected]
Abstract We describe in details the interplay between binary symplectic geometry and quantum computation, with the ultimate goal of constructing highly structured codebooks. The Binary Chirps (BCs) are Complex Grassmannian Lines in N = 2m dimensions used in deterministic compressed sensing and random/unsourced multiple access in wireless net- works. Their entries are fourth roots of unity and can be described in terms of second order Reed-Muller codes. The Binary Subspace Chirps (BSSCs) are a unique collection of BCs of ranks ranging from r = 0 to r = m, embedded in N dimensions according to an on-off pattern determined by a rank r binary subspace. This yields a codebook that is asymptotically 2.38 times larger than the codebook of BCs, has the same minimum chordal distance as the codebook of BCs, and the alphabet is minimally extended from {±1, ±i} to {±1, ±i, 0}. Equivalently, we show that BSSCs are stabilizer states, and we characterize them as columns of a well-controlled collection of Clifford matrices. By con- struction, the BSSCs inherit all the properties of BCs, which in turn makes them good candidates for a variety of applications. For applications in wireless communication, we use the rich algebraic structure of BSSCs to construct a low complexity decoding algorithm that is reliable against Gaussian noise. In simulations, BSSCs exhibit an error probability comparable or slightly lower than BCs, both for single-user and multi-user transmissions.
1 Introduction
Codebooks of complex projective (Grassmann) lines, or tight frames, have found application in multiple problems of interest for communications and information processing, such as code division multiple access sequence design [1], precoding for multi-antenna transmissions [2] and network coding [3]. Contemporary interest in such codes arise, e.g., from deterministic com- pressed sensing [4–8], virtual full-duplex communication [9], mmWave communication [10], and random access [11]. One of the challenges/promises of 5G wireless communication is to enable massive machine- arXiv:2102.12384v1 [cs.IT] 24 Feb 2021 type communications (mMTC) in the Internet of Things (IoT), in which a massive number of low-cost devices sporadically and randomly access the network [12]. In this scenario, users are assigned a unique signature sequence which they transmit whenever active [13]. A twin use-case is unsourced multiple access, where a large number of messages is transmitted infrequently. Polyanskiy [12] proposed a framework in which communication occurs in blocks of N channel uses, and the task of a receiver is to identify correctly L active users (messages) out of 2B, with one regime of interest being N = 30, 000,L = 250, and B = 100. Ever since its introduction, there have been several follow-up works [11,14–17], extensions to a massive MIMO scenario [18] where the base station has a very large number of antennas, and a discussion on the fundamental limits on what is possible [19]. Given the massive number of to-be-supported (to-be-encoded) users (messages), the design criteria are fundamentally different and one simply cannot rely on classical multiple-access
1 Figure 1. Interplay of binary world and complex world. Prior art is depicted in yellow. The contributions of this paper are depicted in green. See also [21, 24].
channel solutions. For instance, interference is unavoidable since it is impossible to have orthogonal signatures/codewords. Additionally, given that there is a small number of active user, the interference is limited. Thus, the challenge becomes to design highly structured codebooks of large cardinality along with a reliable and low-complexity decoding algorithm. Codebooks of Binary Chirps (BCs) [5, 20] provide such highly structured Grassmannian line codebook in N = 2m dimensions with additional desirable properties. All entries come from a small alphabet, being a fourth root of unity, and can be described in terms of sec- ond order Reed-Muller (RM) codes. RM codes have the fascinating property that a Walsh- Hadamard measurement cuts in half the solution space. This yields a single-user decoding complexity of O(N log2 N), coming from the Walsh-Hadamard transform and number of re- quired measurements. Additionally, the number of codewords is reasonably large, growing as √ 3+log N √ 2m(m+3)/2 = N 2 , while the minimum chordal distance is 1/ 2. We expand the BC codebook to the codebook of Binary Subspace Chirps (BSSCs) by collectively considering all BCs in S = 2r dimensions, r = 0, . . . , m, in N = 2m dimensions. That is, given a BC in S = 2r dimensions, we embed it in N = 2m dimensions via a unique on- off pattern determined by a rank r binary subspace. Thus, a BSSC is characterized by a sparsity r, a BC part parametrized by a binary symmetric matrix Sr ∈ Sym(r; 2) and a binary vector m b ∈ F2 , and a unique on-off pattern parametrized by a rank r binary subspace H ∈ G(m, r; 2); see (79) for the formal definition. The codebook of BSSCs inherits all the desirable properties of BCs, and in addition, it has asymptotically√ about 2.384 more codewords. Thus, an active device with a rank r signature will transmit α/ 2r, α ∈ {±1, ±i} during time slots determined by the rank r subspace H, and it will be silent otherwise. This resembles the model of [9], in which active devices can also be used (to listen) as receivers during the off-slots. Given the structure of BSSCs, a unified rank, on-off pattern, and BCs part (in this order) estimation technique is needed. In [21], a reliable on-off pattern detection was proposed, which made use of a Weyl-type transform [22] on m qubit diagonal Pauli matrices; see (103). The algorithm can be described with the common language of symplectic geometry and quantum computation. The key insight here is to view BSSCs as common eigenvectors of maximal sets of commuting Pauli matrices, commonly referred as stabilizer groups. Indeed, we show that BSSCs are nothing else but stabilizer states [23], and their sparsity is determined by the diagonal portion of the corresponding stabilizer group; see Corollaries 2 and 3. We also show that each BSSC is a column of a unique Clifford matrix (86), which itself is the common eigenspace of a unique stabilizer group (101); see also Theorem 1. The interplay between the binary world and the complex world is depicted in Figure 1.
2 Making use of these structural results, the on-off pattern detection of [21] can be general- ized to recover the BC part of the BSSC, this time by using the Weyl-type transform on the off-diagonal part of the corresponding stabilizer group. This yields a single-user BSSC recon- struction as described in Algorithm 2. In [24], we added Orthogonal Matching Pursuit (OMP) to obtain a multi-user BSSCs reconstruction (see Algorithm 3) with reliable performance when there is a small number of active users. As the number of active users increases, so does the interference, which has a quite destructive effect on the on-off pattern. However, state-of- the-art solutions for BCs [8, 17, 25] such as slotting and patching, can be used to reduce the interference. Preliminary simulations show that BSSCs exhibit a lower error probability than BCs. This is because BSSCs have fewer closest neighbors on average than BCs. In addition, BSSCs are uniformly distributed over the sphere, which makes them optimal when dealing with Gaussian noise. Throughout, the decoding complexity is kept at bay from the underlying symplectic geom- etry. The sparsity, the BC part, and the on-off pattern of a BSSC can be described in terms of the Bruhat decomposition (22) of a symplectic matrix. Indeed, the unique Clifford matrix (86) of which a BSSC is a column, is parametrized by a coset representative (24) as described in Lemma 1. In turn, such coset representative determines a unique stabilizer group (101). We use this interplay to reconstruct a BSSC by reconstructing the stabilizer group that stabilizes 2 the given BSSC. This alone reduces the complexity from O(N ) to O(N log2 N). The paper is organized as follows. In Section 2 we review the basics of binary symplectic geometry and quantum computation. In order to obtain a unique parametrization of BSSCs, we use Schubert cells and the Bruhat decomposition of the symplectic group. In Section 3 we lift the Bruhat decomposition of the symplectic group to obtain a decomposition of the Clifford group. Additionally, we parametrize those Clifford matrices whose columns are BSSCs. In Section 4 we give the formal definition of BSSCs, along with their algebraic and geometric properties. In Sections 5 and 6 we present reliable low complexity decoding algorithms, and discuss simulation results. We end the paper with some conclusions and directions future research.
1.1 Conventions
All vectors, binary or complex, will be columns. F2 denotes the binary field, GL(m; 2) denotes the group of binary m×m invertible matrices, and Sym(m; 2) denotes the group of binary m×m symmetric matrices. We will denote matrices (resp., vectors) with upper case (resp., lower case) bold letters. AT will denote the transpose and A−T will denote the inverse transposed. cs (A) and rs (A) will denote the column space and the row space of A respectively. Since all our vectors are columns, we will typically deal with column spaces, except when we work with notions from quantum computation, where row spaces are customary. Im will denote the m × m matrix (complex or binary). G(m, r; 2) =∼ GL(m; 2)/GL(r; 2) denotes the binary m Grassmannian, that is, the set of all r-dimensional subspaces of F2 . U(N) denotes the set of unitary N × N complex matrices and † will denote the conjugate transpose of a matrix.
2 Preliminaries
In this section we will introduce all preliminary notions needed for navigating the connection between the 2m dimensional binary world and the 2m dimensional complex world, as depicted in Figure 1. The primary bridge used here is the well-known homomorphism (62) and the Bruhat decomposition of the symplectic group. We focus on cosets of the symplectic group modulo the semidirect product GL(m; 2) o Sym(m; 2). These cosets are characterized by a rank r = 0, . . . , m and a binary subspace H ∈ G(m, r; 2), which we will think of as the column space of an m × r binary matrix in column reduced echelon form. We will use Schubert cells as a formal and systematic approach. This also provides a framework for describing well-known facts from binary symplectic geometry (e.g., Remark 4). Finally, Subsection 2.3 discusses common notions from quantum computation.
3 2.1 Schubert Cells Here we discuss the Schubert decomposition of the Grassmannian G(m, r; 2) with the respect to the standard flag {0} = V0 ⊂ V1 ⊂ · · · ⊂ Vm, (1) m where Vi = span{e1,..., ei} and {e1,..., em} is the standard basis of F2 . Fix a set of r indices I = {i1, . . . , ir} ⊂ {1, . . . , m}, which, without loss of generality, we assume to be in increasing order. The Schubert cell CI is the set of all m × r matrices that have 1 in leading positions (ij, j), 0 on the left, right, and above each leading position, and every other entry is free. This is simply the set of all binary matrices in column reduced echelon form with leading positions I. By counting the number of free entries in each column one concludes that
r X dim CI = (m − ij) − (r − j). (2) j=1 Fix H ∈ G(m, r; 2), and think of it as the column space of a m × r matrix H. After column operations, it will belong to some cell CI , and to emphasize this fact, we will denote it as HI . Schubert cells have a well-known duality theory which we outline next. Let HgI T be such that (HI ) HgI = 0. Of course cs (HgI ) ∈ G(m, m − r; 2). Let Ie := {1, . . . , m}\I and put C := {H | H ∈ C }. There is a bijection between {C } and {C } , fI gI I I fI |I|=r Je |J |=r realized by reverting the rows and columns of HgI and identifying Je = {i1, . . . , im−r} with J := {m − i + 1, . . . , m − i + 1}. With this identification, we will denote H the unique b 1 m−r Ie element of cell C that is equivalent with H , obtained by reverting the rows and columns of Ie gI HgI : H = P H P , (3) Ie ad,mgI ad,m−r where Pad is the antidiagonal matrix in respective dimensions. Each cell has a distinguished element: II ∈ CI will denote the identity matrix Im restricted to I, that is, the unique element in CI that has all the free entries 0. Note that II has as jth column the ijth column of Im, and thus its non-zero rows form Ir. In particular if |I| = m then I = I . We also have I ∈ C . With this notation one easily verifies that I m Ie Ie (I )TH = I , (I )TI = 0, (H )T I = I . (4) I I r I Ie gI Ie m−r
In addition, HI can be completed to an invertible matrix P := H I ∈ GL(m; 2). (5) I I Ie
Note that when II is completed to an invertible matrix it gives rise to a permutation matrix. T Next, (4) along with the default equality (HI ) HgI = 0 implies that
−T h i PI = II HgI . (6)
Let us describe this framework with an example. Example 1. Let m = 3 and r = 2. Then 1 0 u 1 C = 0 1 , C^ = v , C =∼ C = v , {1,2} {1,2} {d3} {1} u v 1 u 1 0 u 0 C = u 0 , C^ = 1 , C =∼ C = 1 , {1,3} {1,3} {d2} {2} 0 1 0 u 0 0 1 0 C = 1 0 , C^ = 0 , C =∼ C = 0 . {2,3} {2,3} {d1} {3} 0 1 0 1
4 Let us focus on I = {1, 3}. Then CI is constructed directly by definition, that is, in column reduced echelon form with leading positions 1 and 3. Whereas, CfI is constructed so that T (HI ) HgI = 0. Then we revert the rows and columns (only rows in this case) to obtain the last object where we identify1 {2} = {^1, 3} with {d2} ≡ {2}. In this case, as we see from above, there is only one free bit. This yields two subspaces/- matrices HI , which when completed to an invertible matrix as in (5) yield
1 0 0 1 0 0 Pu=0 = 0 0 1 , Pu=1 = 1 0 1 . (7) 0 1 0 0 1 0
Then one directly computes
1 0 0 1 0 1 −T −T Pu=0 = 0 0 1 , Pu=1 = 0 0 1 . (8) 0 1 0 0 1 0
Compare (8) with (6); the first two columns are obviously II , whereas the last column is precisely C ≡ C with rows reverted. Note here that when all the free bits are zero then {d2} {2} the resulting P is simply a permutation matrix, and in this case P−T = P.
2.2 Bruhat Decomposition of the Symplectic Group 2m We first briefly describe the symplectic structure of F2 via the symplectic bilinear form
T T h a, b | c, d is := b c + a d. (9)
One is naturally interested in automorphisms that preserve such symplectic structure. It follows T directly by the definition that a 2m × 2m matrix F preserves h • | • is iff FΩF = Ω where
0 I Ω = m m . (10) Im 0m
We will denote the group of all such symplectic matrices F with Sp(2m; 2). Equivalently,
AB F = ∈ Sp(2m; 2) (11) CD
T T T T iff AB , CD ∈ Sym(m; 2) and AD + BC = Im. It is well-known that
m 2 Y |Sp(2m; 2)| = 2m (4i − 1). (12) i=1
Consider the row space H := rs [AB] of the m × 2m upper half of a symplectic matrix T T F ∈ Sp(2m; 2). Because AB is symmetric one has [AB]Ω[AB] = 0 and thus h x | y is = 0 ⊥ for all x, y ∈ H. We will denote ( • ) s the dual with respect to the symplectic inner prod- uct (9). It follows that H ⊆ H⊥, that is, H is self-orthogonal or totally isotropic. Moreover, H is maximal totally isotropic because dim H = m and thus H = H⊥. The set of all self- dual/maximal totally isotropic subspaces is commonly referred as the Lagrangian Grassman- nian L(2m, m; 2) ⊂ G(2m, m; 2). It is well-known that
m Y |L(2m, m)| = (2i + 1). (13) i=1
1In this specific case there is no need for identification, but this is only a coincidence. For different choices of I one needs a true identification.
5 For reasons that will become clear latter on we are interested in decomposing symplectic matrices into more elementary symplectic matrices, and we will do this via the Bruhat decom- position of Sp(2m; 2). While the decomposition holds in a general group-theoretic setting [26], here we give a rather elementary approach; see also [27]. We start the decomposition by writing
m [ Sp(2m; 2) = Cr, (14) r=0 where AB Cr = F = ∈ Sp(2m; 2) rank C = r . (15) CD In Sp(2m; 2) there are two distinguished subgroups: P 0 SD := FD(P) = P ∈ GL(m; 2) , (16) 0 P−T IS SU := FU (S) = S ∈ Sym(m; 2) . (17) 0 I
Let P be the semidirect product of SD and SU , that is,
P = {FD(P)FU (S) | P ∈ GL(m; 2), S ∈ Sym(m; 2)}. (18)
Note that the order of the multiplication doesn’t matter since
T FD(P)FU (S) = FU (PSP )FD(P), (19)
T and PSP is again symmetric. It is straightforward to verify that P = C0, and that in general
Cr = {F1FΩ(r)F2 | F1, F2 ∈ P}, (20) where Im|−r Im|r FΩ(r) = , (21) Im|r Im|−r with Im|r being the block matrix with Ir in upper left corner and 0 else and Im|−r = Im −Im|r. Note here that Ω = FΩ(m) and ΩFΩ(r)Ω = FΩ(m−r). Then it follows by (20) (and by (19)) that every F ∈ Sp(2m; 2) can be written as
F = FD(P1)FU (S1)FΩ(r)FU (S2)FD(P2). (22)
The above constitutes the Bruhat decomposition of a symplectic matrix; see also [24, 28]. Remark 1. It was shown in [29] that a symplectic matrix F can be decomposed as
T F = FD(P1)FU (S1)ΩFΩ(r)FU (S2)FD(P2). (23)
2 T If we, instead, decompose ΩF as in (23) and insert Ω = I2m between FD(P1) and FU (S1), we see that (23) is reduced to (22). This reduction from a seven-component decomposition to a five-component decomposition is beneficial in quantum circuits design [28, 30]. In what follows we will focus on the right action of P on Sp(2m; 2), that is, the right cosets in the quotient group Sp(2m; 2)/P. It is an immediate consequence of (22) and (19) that a coset representative will look like
FD(P)FU (S)FΩ(r), (24) for some rank r, invertible P, and symmetric S. However, two different invertibles P may yield representatives of the same coset. We make this precise below.
6 Lemma 1. A right coset in Sp(2m; 2)/P is uniquely characterized by a rank r, an r × r m symmetric matrix Sr ∈ Sym(r), and a r-dimensional subspace H in F2 . Proof. Write a coset representative F as in (24). This immediately determines r. Next, write S in a block form Sr X S = T , (25) X Sm−r where Sr, Sm−r are symmetric. Denote Sfr, Sbm−r ∈ Sym(m; 2) the matrices that have Sr and Sm−r in upper left and lower right corner respectively and 0 otherwise. Put also Ir 0 Xe = T . (26) X Im−r With this notation we have
FU (S)FΩ(r) = FU (Sfr)FΩ(r)FU (Sbm−r)FD(Xe ). (27)
In other words FU (S)FΩ(r) and FU (Sfr)FΩ(r) belong to the same coset. Now consider an invertible Pr 0 Pe = . (28) 0 Pm−r It is also straightforward to verify that
FU (Sfr)FΩ(r)FD(Pe) = FU (Sfr)FD(Pb)FΩ(r) −1 −T = FD(Pb)FU (Pb SfrPb )FΩ(r), where P−T 0 Pb = r , (29) 0 Pm−r and the second equality follows by (19). Thus FD(P1)FU (S1)FΩ(r), where
−1 −T P1 := PPb, S1 := Pb SfrPb (30) represents the same coset. Note that the transformation (30) doesn’t change the column space m C (that is, the lower left corner of F), which is an r-dimensional subspace in F2 .
Next, using Schubert cells will choose a canonical coset representative. We will use the same notation as in the above lemma. Let r and Sfr be as above. To choose P, think of the r-dimensional subspace H from the above lemma as the column space of a matrix H, which belongs to some Schubert cell CI . We will use the coset representative
FO(PI , Sr) := FD(PI )FU (Sfr)FΩ(r), (31) where PI is as in (5). Let F ∈ Sp(2m; 2) be in block from as in (11), and assume it is written as
−T F = FD(P )FU (Sfr)FΩ(r)FD(M)FU (S). (32)
T Multiplying both sides of (32) on the left with FD(P ) and on the right with FU (S), and then comparing respective blocks we obtain
T P A = (Sfr + Im|−r)M, (33) T T −T P AS = P B + Im|rM , (34) −1 P C = Im|rM, (35) −1 −1 −T P CS = P D + Im|−rM , (36)
7 Algorithm 1 Bruhat Decomposition of Symplectic Matrix Input: A symplectic matrix F. 1. Block decompose F to A, B, C, D as in (11). 2. r = rank (C). 3. Find P as in (5) from cs (C). −1 4. Mup is the first r rows of P C. T 5. Mlo is the last m − r rows of P A. M 6. M = up . Mlo T −1 7. Sfr = P AM + Im|−r. 8. Sr is the upper left r × r block of Sfr. −1 −T 9. Nup is the first r rows of P D + Im|−rM . T −T 10. Nlo is the last m − r rows of P B − Im|rM . N 11. S = M−1 up . Nlo Output: r, P, Sr, M, S
which we can solve for M, Sfr and S, while assuming that we know F (and implicitly P which can be determined by the column space of the lower-left block of F). First we find M. For this, recall that Sfr has nonzero entries only on the upper left r × r block. Thus, it follows by (33) that the last m − r rows of M coincide with the last m − r rows of PTA. Similarly, it follows from (35) that the first r rows of M coincide with the first r rows of P−1C. With M in hand we have T −1 Sfr = P AM + Im|−r. (37) By using (35) in (36) we see that the first r rows of MS coincide with first r rows of P−1CS. Similarly, by using (33) in (34), we see that the last m − r rows of MS coincide with the last m−r rows of PTAS. Multiplication with M−1 yields S. We collect everything in Algorithm 1, which gives not only the Bruhat decomposition but also a canonical coset representative. We end this section with a few remarks. Remark 2. One can follow an analogous path by considering left action of P on Sp(2m; 2). This follows most directly by the observation that if F = FD(P)FU (S)FΩ(r) is a right coset −1 −1 representative then F = FΩ(r)FU (S)FD(P ) is a left coset representative. Remark 3. Note that for the extremal case r = m, a coset representative as in (31) is completely determined by a symmetric matrix S ∈ Sym(m; 2), since in this case, as one would recall, PI = II = Im. Remark 4. Directly from the definition we have
m 2 Y |P| = |GL(m; 2)| · |Sym(m; 2)| = 2m (2i − 1), (38) i=1 which combined with (12) yields
m Y |Sp(2m; 2)/P| = (2i + 1) = |L(2m, m)|. (39) i=1
The above is of course not a coincidence. Indeed, Sp(2m; 2) acts transitively from the right on L(2m, m). Next, consider rs 0m Im ∈ L(2m, m). If a symplectic matrix F as in (11) fixes this space, then C = 0 and A is invertible. Additionally, because F is symplectic to start with, we obtain D = A−T and ABT =: S is symmetric. Thus B = SAT, and F ∈ P. That is,
8 P is the stabilizer (in a group action terminology) of rs 0m Im ∈ L(2m, m). The mapping Sp(2m; 2)/P −→ L(2m, m), given by h i T −1 FO(PI , Sr) 7−→ rs Im|rPI (Im|rSfr + Im|−r)PI (40) is well-defined (because, as mentioned, the upper half of a symplectic matrix is maximal isotropic). It is also injective, and thus bijective due to cardinality reasons. Of course one can have many bijections but we choose this one due to Theorem 1.
2.3 The Heisenberg-Weyl Group m 2 Fix N = 2 , and let {e0, e1} be the standard basis of C , which is commonly referred as the m m computational basis. For v = (v1, . . . , vm) ∈ F2 set ev := ev1 ⊗ · · · ⊗ evm . Then {ev | v ∈ F2 } is the standard basis of CN =∼ (C2)⊗m. The Pauli matrices are 0 1 1 0 I , σ = , σ = , σ = iσ σ . (41) 2 x 1 0 z 0 −1 y x z
m For a, b ∈ F2 put a1 b1 am bm D(a, b) := σx σz ⊗ · · · ⊗ σx σz . (42) Directly by definition we have
bTv D(a, 0)ev = ev+a and D(0, b)ev = (−1) ev, (43) and thus, the former is a permutation matrix whereas the latter is a diagonal matrix. Then
T D(a, b)D(c, d) = (−1)b cD(a + c, b + d). (44)
Thanks to (44) we have
T T D(a, b)D(c, d) = (−1)b c+a dD(c, d)D(a, b). (45)
In turn, D(a, b) and D(c, d) commute iff
T T h a, b | c, d is := b c + a d = 0, (46) that is, iff (a, b) and (c, d) are orthogonal with respect to the symplectic inner product (9). Also thanks to (44), the set
k m HWN := {i D(a, b) | a, b ∈ F2 , k = 0, 1, 2, 3} (47) is a subgroup of U(N) and is called the Heisenberg-Weyl group. We will also call its elements Pauli matrices as well. Directly from the definition, we have a surjective homomorphism of groups 2m k ΨN : HWN −→ F2 , i D(a, b) 7−→ (a, b). (48) ∼ ∗ Its kernel is ker ΨN = {±IN , ±iIN } = Z4. We will denote HWN := HWN / ker ΨN the ∗ projective Heisenberg-Weyl group, and the induced isomorphism ΨN . 2m Note that h • | • is defines a nondegenerate bilinear form in F2 that translates commutativ- 2m ity in HWN to orthogonality in F2 . A commutative subgroup S ⊂ HWN is called a stabilizer ⊥s group if −IN ∈/ S. Thus, for a stabilizer S, thanks to (46) we have ΨN (S) ⊆ ΨN (S) [31,32]. r In addition, because ΨN restricted to a stabilizer is an isomorphism, we have that |S| = 2 iff dim ΨN (S) = r. We will think of ΨN (S) as the row space of a full rank matrix [AB] where both A and B are r × m binary matrices. We will write
T T r E(A, B) := {E(x A, x B) | x ∈ F2}, (49)
9 T where E(a, b) := ia bD(a, b). Combining this with (43) and (44) we obtain
aTb X bTv T E(a, b) = i (−1) ev+aev . (50) m v∈F2 2m Next, if rs [AB] is self-orthogonal in F2 then E(A, B) is a stabilizer. Moreover, ΨN (E(A, B)) = rs [AB], which yields a one-to-one correspondence between stabilizers in 2m HWN and self-orthogonal subspaces in F2 . It also follows that a maximal stabilizer must have 2m elements. Thus there is a one-to-one correspondence between maximal stabilizers and Lagrangian Grassmannians L(2m, m) ⊂ G(2m, m). Of particular interest are maximal stabilizers
m XN := E(Im, 0m) = {E(a, 0) | a ∈ F2 }, (51) m ZN := E(0m, Im) = {E(0, b) | b ∈ F2 }, (52) which we naturally identify with XN := ΨN (XN ) = rs [Im 0m] and ZN := ΨN (ZN ) = rs [0m Im]. What follows holds in general for any stabilizer, but for our purposes, we need only focus on the maximal ones. Let S = E(A, B) ⊂ HWN be a maximal stabilizer and let {E1,..., Em} be an independent generating set of S (that is, span{ΨN (E1),..., ΨN (Em)} = ΨN (S)). Consider the complex vector space [33]
N V (S) := {v ∈ C | Eiv = v, i = 1, . . . , m}. (53) It is well-known (see, e.g., [34]) that dim V (S) = 2m/|S| = 1. A unit norm vector that generates it is called stabilizer state, and with a slight abuse of notation is also denoted by V (S). Because we are disregarding scalars, it is beneficial to think of a stabilizer state as Grassmannian line, that is, V (S) ∈ G(CN , 1). Next, m Y IN + Ei 1 X Π := = E (54) S 2 N i=1 E∈S is a projection onto V (S). m d1 dm Given a stabilizer as above, for any d ∈ F2 , {(−1) E1,..., (−1) Em} also describes a stabilizer Sd. Similarly to (54) put m di Y IN + (−1) Ei Π := Sd 2 i=1 (55) 1 X T = (−1)d xE(xTA, xTB). N m x∈F2 m It is readily verified that {ΠSd | d ∈ F2 } are pair-wise orthogonal, and a stabilizer group determines a resolution of the identity X IN = ΠSd . (56) m d∈F2 Thus every such projection determines a one-dimensional subspace which with another abuse of notation (see also Remark 5 below) we call a stabilizer state. Remark 5. A stabilizer state as in (53) is the unit norm vector that is fixed by the stabilizer. m T T Now for every E ∈ S = E(A, B) there exists a unique x ∈ F2 such that E = E(x A, x B). m T T dTx For d ∈ F2 , consider the map χd : E(x A, x B) 7−→ (−1) . Then ΠSd projects onto N V (Sd) := {v ∈ C | Ev = χd(E)v for all E ∈ Sd}, (57) that is, the state that under the action of E is scaled by χd(E). Then of course V (S0) = V (S) m where 0 ∈ F2 . In addition, the map χd is a linear character of S, which has led to non-binary considerations [35].
10 Remark 6. Let {E1,..., Em} be an independent generating set of a maximal stabilizer S and consider Sd. By [34, Prop. 10.4] it follows that for each i = 1, . . . , m, there exists Gi ∈ HWN † † d1 dm such that Gi EiGi = −Ei and Gi EjGi = Ej for i 6= j. Now put Gd := G1 ··· Gm . Then
† GdΠS Gd = ΠSd . (58)
m N It follows that {V (Sd) | d ∈ F2 } is an orthonormal basis of C . In [36] the authors used a similar insight to construct maximal sets of mutually unbiased bases.
3 Clifford Group
The Clifford group in N dimensions [37] is defined to be the normalizer of HWN in U(N) modulo U(1): † CliffN = {G ∈ U(N) | GHWN G = HWN }/U(1). (59) ∼ The reason one quotients out U(1) = {αIN | |α| = 1}, is to obtain a finite group. In this case ∗ HWN is a normal subgroup of CliffN . 2m 2m Let {e1,..., e2m} be the standard basis of F2 , and consider G ∈ CliffN . Let ci ∈ F2 be such that † GE(ei)G = ±E(ci). (60)
Then the matrix FG whose ith row is ci is a symplectic matrix such that
† T GE(c)G = ±E(c FG) (61)
2m for all c ∈ F2 . Based on (61) we obtain a group homomorphism
Φ : CliffN −→ Sp(2m; 2), G 7−→ FG, (62)
∗ with kernel ker Φ = HWN [30]. This map is also surjective; see Section 3.1 where specific ∗ 2m preimages are given. From (12) and (48) (|HWN | = 2 ) follows that
m m2+2m Y i |CliffN | = 2 (4 − 1). (63) i=1
† −1 Remark 7. Since Φ is a homomorphism we have that Φ(G ) = FG and as a consequence † T −1 G E(c)G = ±E(c FG ). We will make use of this simple observation later on to determine when a column of G is an eigenvector of E(c). This interplay with symplectic geometry provides an exponential complexity reduction in various applications. Here, we will focus on efficiently computing common eigenspaces of maximal stabilizers.
The phase and Hadamard matrices
1 0 1 1 1 GP = and H2 = √ (64) 0 i 2 1 −1
3 are easily seen to be in the Clifford group Cliff2. Some authors also include (GP H2) = exp(πi/4)I2 [32], which in our case would disappear as a scalar quotient. Thus (63) differs by a factor of 1/8 of what is commonly considered as the cardinality of the Clifford group; see A003956 at oeis.org. For our purposes the phases are irrelevant.
3.1 Decomposition of the Clifford Group In this section we will make use of the Bruhat decomposition of Sp(2m; 2) to obtain a decom- position of CliffN . To do so we will use the surjectivity of Φ from (62) and determine preimages
11 of coset representatives from (31). The preimages of symplectic matrices FD(P), FU (S), and FΩ(r) under Φ are the unitary permutation matrix
GD(P) := ev 7−→ ePTv, (65) vTSv mod 4 GU (S) := diag i , (66) m v∈F2 ⊗r GΩ(r) := (H2) ⊗ I2m−r , (67) respectively. We refer the reader to [30, Appendix I] for details. Remark 8. Note that directly by the definition of the Hadamard matrix we have
1 vTw HN := GΩ(m) = √ [(−1) ]v,w∈ m . (68) 2m F2 Whereas, for any r = 1, . . . , m, one straightforwardly computes
vTw G (r)·Z(m, r) = [(−1) · f(v, w, r)] m , (69) Ω v,w∈F2 ⊗m−r where Z(m, r) := I2r ⊗ σz is the diagonal Pauli that acts as σz on the last m − r qubits, and m Y f(v, w, r) = (1 + vi + wi). (70) i=r+1 Note that the value of f will be 1 precisely when v and w coincide in their last m−r coordinates and 0 otherwise. It follows that f is identically 1 when r = m and f is the Kronecker function δv,w when r = 0. We will use f to determine the sparsity or rank of a Clifford matrix/stabilizer state. Of course r = m corresponds to fully occupied objects with only nonzero entries; see also Remarks 12 and 13 for the extreme cases of r = 0, 1. Example 2 (Example 1 continued). Let us reconsider the invertible matrices from (7). Recall that there we had m = 3, r = 2. Here we will construct the Cliffords corresponding to the 2 canonical coset representative (40), with Sr = 02×2. For the case u = 0 one computes T GD√(Pu=0) as in (66), and multiplies it (from the right) by GΩ(2) as in (67) (we will omit 1/ 22) and and then by Z(3, 2) to obtain + 0 + 0 + 0 + 0 + 0 − 0 + 0 − 0 0 − 0 − 0 − 0 − 0 − 0 + 0 − 0 + Gu=0 = . (71) + 0 + 0 − 0 − 0 + 0 − 0 − 0 + 0 0 − 0 − 0 + 0 + 0 − 0 + 0 + 0 −
As mentioned, (65) by definition yields a permutation matrix. Thus Gu=0 is nothing else but GΩ(2) = H2 ⊗ H2 ⊗ I2, with its rows permuted accordingly, and a possible sign introduced to its columns by the diagonal matrix Z(3, 2) = I4 ⊗ σz. Similarly, for the case u = 1, one obtains + 0 + 0 + 0 + 0 + 0 − 0 + 0 − 0 0 − 0 − 0 − 0 − 0 − 0 + 0 − 0 + Gu=1 = . (72) 0 − 0 − 0 + 0 + 0 − 0 + 0 + 0 − + 0 + 0 − 0 − 0 + 0 − 0 − 0 + 0 We will discuss how the {+, −, 0} patterns are correlated later on.
2See Section 4.1 for why we consider the transpose instead of P itself.
12 Let us now return to the Clifford group. The Bruhat decomposition (22) of Sp(2m; 2) already gives a decomposition of CliffN . However, in order to have a concise approach one has to be a bit careful. In this section we will write G = Φ−1(F), where the equality is taken ∗ −1 modulo the center HWN nZ8. In other words, we will disregard the central part of Φ (F) and consider only the Clifford part. The cyclic group Z8 of order 8 comes into play to accommodate 8th roots of unity coming out of products GU (S)GΩ(r). This setup, yet again, confirms the importance of ∗ CliffN := {exp(iπk/4)G | k ∈ Z8, G ∈ CliffN }. (73)
Let G = {GD(P)GU (S) | P ∈ GL(m; 2), S ∈ Sym(m; 2)} be the preimage of P from (18). For obvious reasons, it is referred as the Hadamard-free group; see also [38]. As for the case of the symplectic group, this group acts from the right on matrices of the form
GD(P1)GU (S1)GΩ(r)GU (S2)GD(P2) (74) and thus, a coset representative would look like
GF := GD(P1)GU (S1)GΩ(r), (75)
One is interested on coset representatives. To understand this, it is enough to understand the preimage of generators of GL(m; 2) and Sym(m; 2). Let us start with the former, which can be generated by two elements [39]. Namely, it can be generated P := Im + E12 where E12 is the elementary (binary) matrix with 1 in position (1, 2) and 0 elsewhere and the cyclic permutation matrix Πcycl acting as the permutation (12 ··· m). It is of interest to consider a larger set of generators. Let Πi,j be a transposition matrix. Then of course P along with all m the Πi,j generate GL(m; 2). While Πi,j swaps dimensions i and j in F2 , it is easily seen that −1 2 ⊗m Φ (FD(Πi,j)) swaps the tensor dimensions i and j in (C ) . Moreover
1 0 0 0 −1 0 1 0 0 Φ (FD(P)) = ⊗ IN−4. (76) 0 0 0 1 0 0 1 0
The above 4 × 4 matrix is known in quantum computation as the controlled-NOT (CNOT) quantum gate. In itself, the CNOT gate is of form GD(P) where
1 1 P = P−1 = . (77) 0 1
T m For Sym(m; 2) we consider matrices Sv := v v where v ∈ F2 is a vector with at most two non-zero entries. Then −1 1 Φ (FU (Sv)) = √ (IN + iE(0, v)). (78) 2 Note that when v has exactly one non-zero entry in position j, the jth tensor dimension will contain the phase matrix GP = exp(−iπ/4)(I2 + iσz) form (64). On the other hand, when v has exactly two non-zero entries (78) gives rise to GCZ (GP ⊗ GP ) in tensor dimensions i and j, where GCZ = diag(1, 1, 1, −1). The latter is known in quantum computation as the controlled-Z (CZ) quantum gate, and it is of form GU (I2). In conclusion, the Bruhat decomposition of Sp(2m; 2) directly yields some fundamental quantum gates as described above. Similar ideas were used in [28] where the depth of stabilizer circuits was considered. Classically, there exist several decompositions of the symplectic group, which in principle would yield a decomposition of the Clifford group. Particularly important in quantum computation is the decomposition into symplectic transvections [40]; see [41, 42] for a detailed description and [43] for using these ideas to sample the Clifford group.
13 4 Binary Subspace Chirps
Binary subspace chirps (BSSCs) were introduced in [21] as a generalization of binary chirps (BCs) [5]. In this section we describe the geometric and algebraic features of BSSCs, and use their structure to develop a reconstruction algorithm. For each 1 ≤ r ≤ m, subspace N H ∈ G(m, r; 2), and symmetric Sr ∈ Sym(r) we will define a unit norm vector in C as follows. Let HI ∈ CI be such that cs (HI ) = H, as described in Section 2.1. Then HI is m completed to an invertible P := PI as in (5). For all b, a ∈ F2 define
1 T −T −1 T −1 wH,Sr (a) = √ ia P SP a+2b P af(b, P−1a, r), (79) b 2r where S ∈ Sym(m; 2) is the matrix with Sr on the upper left corner and 0 elsewhere, f is as in (70), and the arithmetic in the exponent is done modulo 4. To avoid heavy notation however we will omit the upper scripts. Then we define a binary subspace chirp to be
N wb := wb(a) m ∈ C . (80) a∈F2
Note that when r = m we have P = Im and f is the identically 1 function. Thus, we obtain the binary chirps [5]. Directly from the definition (and the definition of f) it follows that wb(a) 6= 0 precisely when b and P−1a coincide in their last m − r coordinates. Making use of the structure of P as in (5) we may conclude that wb(a) 6= 0 iff
T HgI a = bm−r, (81)
m−r where bm−r ∈ F2 consists of the last m − r coordinates of b. Note that since rank HI = r r r it follows that (81) has 2 solutions, and this in turn implies that wb has 2 non-zero entries. In particular wb is a unit norm vector. Concretely, making use (4) we see that the solution space of (81) is given by {x := I b + H x | x ∈ r}. (82) e Ie m−r I F2
We say that the rank r determines the sparsity of wb and the subspace H (equivalently HI ) determines the on-off pattern of wb.
T T T Remark 9. Fix a subspace chirp wb, and write b = [br bm−r]. Then wb(a) 6= 0 iff a is as r in (82) for some x ∈ F2. Making use of (6) and (4) we obtain x P−1a = , (83) bm−r
T −T −1 T and as a consequence a P SP a = x Srx where Sr is the (symmetric) upper-left r × r block of S. Thus the nonzero entries of wb are of the form
wt(b ) (−1) m−r T T x Sr x+2br x wb(x) = √ i (84) 2r
r −1 for x ∈ F2. Note that there is a slight abuse of notation where we have identified x with P a (thanks to (83) and the fact that b is fixed). Above, the function wt( • ) is just the Hamming weight which counts the number of non-zero entries in a binary vector. We conclude that the on-pattern of a rank r binary subspace chirp is just a binary chirp in 2r dimensions; compare (84) with [5, Eq. (5)]. It follows that all lower-rank chirps are embedded in 2m dimensions, which along with all the chirps in 2m dimensions yield all the binary subspace chirps. As discussed, the embeddings are determined by subspaces.
14 4.1 Algebraic Structure of BSSCs In what follows we fix a rank r, invertible P, and symmetric S. Recall that S contains a r × r T symmetric in its upper left corner and 0 otherwise. Next, let F := FΩ(r)FU (S)FD(P ) and T let GF = GD(P )GU (S)GΩ(r). With this notation we have Φ(GF) = F. Recall also that m N −1 {ea | a ∈ F2 } is the standard basis of C . With a substitution u := P a we have 1 X wb = √ wb(a)ea r 2 m a∈F2
1 X uTSu bTu = √ i (−1) f(b, u, r)ePu r 2 m u∈F2
T 1 X uTSu bTu = GD(P ) · √ i (−1) f(b, u, r)eu r 2 m u∈F2 T = GD(P )GU (S)GΩ(r)Z(m, r)eb (85)
= GF · Z(m, r)eb, (86) where (85) follows by (69). Note that in (86), the diagonal Pauli Z(m, r) only ever introduces an additional sign on columns of GF. Thus, the binary subspace chirp wb is nothing else but the bth column of GF, up to a sign. However, as mentioned, for our practical purposes a sign (or even a complex unit) is irrelevant. Example 3 (Examples 1 and 2 continued). Let us consider the case u = 0, and for simplicity, 3 let us set the symmetric S to be the zero matrix , so that GU (S) is the identity matrix. The on-off pattern of the resulting BSSCs is governed by the r = 2 dimensional subspace T T T T H = {000 , 100 , 001 , 101 } = cs (H{1,3}). The above argument tells us that these BSSCs are precisely the columns of Gu=0 from (71). One verifies this directly using the definition (79). Moreover, the structure of the on-off patterns is completely determined by (82). Indeed, we see in (71) two on-off patterns: one determined by H (if I b ∈ H) and one determined by its Ie m−r coset4 (if I b ∈/ H). In our specific case, we have I = {1, 3}, and thus I = I = 010T. Ie m−r Ie {2} Thus I b ∈ H iff b = 0 iff b ∈ {000T, 010T, 100T, 110T}, which corresponds to columns Ie m−r m−r {1, 3, 5, 7}. Additionally, within each of these columns, the on-off pattern is again governed by H. Indeed, the non-zero entries in these columns are in positions/rows indexed by H, that is, {1, 2, 5, 6} – precisely as described by (81). Since cosets form a partition it follows immediately that columns indexed by different cosets are orthogonal. Orthogonality of columns within each coset is a bit more delicate to see directly. We will further discuss the general structure of on-off patterns in Section 4.2. Equation (40) gives a one-to-one correspondence between canonical coset representatives and maximal stabilizers. Above we mentioned that BSSCs are columns of Clifford matrices parametrized by such coset representatives. The last piece of the puzzle is found by simulta- neously diagonalizing the commuting matrices of a maximal stabilizer. We make this precise in the following. m Theorem 1. Let F and GF be as above. The set {wb | b ∈ F2 } consisting of the columns T −1 of GF is the common eigenspace of the maximal stabilizer E(Im|rP , (Im|rS + Im|−r)P ) from (40).
Proof. Consider the matrix G := GF parametrized by the symplectic matrix F, and recall that wb is the bth column of GF. It follows from Remark 7 that the columns of G are the eigenspace of E(x, y) iff G†E(x, y)G = ±E([x, y]TF−1) (87) −1 is diagonal. Recall also that E(x, y) is diagonal iff x = 0, and observe that FΩ(r) = FΩ(r). T Thus, GΩ(r) will be the common eigenspace of the maximal stabilizer S iff ±E([x y] FΩ(r))
3 GU (S) does not affect the on-off pattern at all. 4There are exactly 2 = 23/22 cosets since H has dimensions 2.
15 is diagonal for all E(x, y) ∈ S. Then it is easy to see that such maximal stabilizer is E(Im|r, Im|−r). Next, if w is an eigenvector of E(c) then
Gw = ±GE(c)w = ±GE(c)G†Gw = ±E(cTΦ(G))Gw implies that Gw is an eigenvector of E(cTΦ(G)). The proof is concluded by computing T [Im|r Im|−r]FU (S)FD(P ).
Remark 10. Note that for r = m one has E(Im|r, Im|−r) = E(Im, 0m) and GΩ(r) = HN . Thus the above theorem covers the well-known fact that HN is the common eigenspace of N the maximal stabilizer XN = E(Im, 0m). It is also well-known that the standard basis of C (that is, IN ) is the common eigenspace of the maximal stabilizer ZN = E(0m, Im) of diagonal Paulis. This is of course consistent with the aforesaid fact since [0m Im]Ω = [Im 0m] and −1 HN = Φ (Ω). In this extremal case we also have PI = Im and Sfr = S ∈ Sym(m; 2). So the above theorem also covers [44, Lem. 11] which (in the language of this paper) says that the common eigenspace of E(Im, S) is GU (S)HN . Remark 11. Theorem 1 is a closed form realization of a more general fact. Let S be a maximal 2m stabilizer and let S = rs [AB] ⊂ F2 be its corresponding isotropic subspace. Consider also the diagonal Paulis ZN and its corresponding subspace ZN = rs [0m Im]. Then, by [30, † † Alg. 1] there exists G ∈ CliffN such that GSG = ZN . In other words, G simultaneously diagonalizes S, and moreover, the respective diagonal is a Pauli. In the symplectic domain, it follows by [30, Thm. 25] that there are precisely 2m(m+1)/2 symplectic solutions to the equation [AB]F = [0m Im]. m Qm r Corollary 1. The set of all binary subspace chirps VBSSC consists of 2 r=1(2 + 1) unit m(m+3)/2 norm vectors, while the set of all binary chirps VBC consists of 2 unit norm vectors. The ratio of codebook cardinalities is |VBSSC|/|VBC| ≈ 2.384. Proof. By Lemma 1 we need only consider coset representatives (31), which are in bijection (40) with the set of maximal stabilizers L(2m, m). Now the first statement follows by (13). Next, recall that the binary chirps correspond to those binary subspace chirps for which r = m, and thus, P = Im. In other words they are parametrized by a symmetric matrix m S ∈ Sym(m; 2) and b ∈ F2 .
The size of the codebook VBSSC can be also straightforwardly deduced from the charac- terization of Lemma 1. Indeed, each rank zero BSSC has precisely 20 = 1 non-zero entry. m m Thus there are 2 rank zero BSSCs, parametrized only by a column index b ∈ F2 . On the other hand, rank r BSSCs are characterized by S ∈ Sym(r; 2) and H ∈ G(m, r; 2). Of course, |Sym(r; 2)| = 2r(r+1)/2. Whereas the size of the Grassmannian is given by the 2-binomial coefficient, that is r−1 m Y 1 − 2m−i |G(m, r; 2)| = = . (88) r 1 − 2i+1 2 i=0 Thus, we have
m m X m Y |V | = 2m · 2r(r+1)/2 = 2m · (2r + 1), (89) BSSC r r=0 2 r=1 where the last equality is simply the 2-binomial theorem [45]. Corollary 2. Each binary subspace chirp is a stabilizer state. The converse is also true.
Proof. Stabilizer states can be defined equivalently as the orbit of e0 under the action of CliffN ; see [46] for instance. Then the first statement follows by (86). The converse is true due to cardinalities.
16 (a) r = 0. (b) r = 1.
(c) r = 2 (BCs).
Figure 2. BSSCs in N = 4 dimensions. White = 0, Blue = 1, Cyan = -1, Red = i, Magenta = −i.
Corollary 3. Let S be a maximal stabilizer. Then the stabilizer state V (S) is a rank r BSSC m−r iff |S ∩ ZN | = 2 . Proof. By Corollary 2 we know that V (S) is a BSSC of rank r, which in turn is stabilized by the maximal stabilizer of Theorem 1. Such stabilizer has precisely 2m−r diagonal Paulis; see also (101).
We mentioned that the extremal case r = m gives the codebook VBC. Before discussing general on-off patterns, we consider the lower-end extremal cases r = 0, 1.
Remark 12. Let r = 0. In this case we again have P = Im and S = 0m. In addition f(v, w, 0) = δv,w. Thus, from (79) we see that wb(a) 6= 0 iff a = b, in which case we have wt(b) wb(a) = (−1) . This can also be seen from (86). Indeed, since GΩ(0) = IN we have GF = IN . Note also that Z(m, 0) = σz ⊗ · · · ⊗ σz = E(0, 1) is the common eigenspace of the maximal stabilizer E(Im, Im), as established by Theorem 1.
T m Remark 13. Let r = 1. In this case, either S = 0m or S = e1e1 , where e1 ∈ F2 is the first standard basis vector.√ It follows that, up to a Pauli matrix, GU (S) is either IN or the transvection (IN + iZ1)/ 2 where Z1 = E(0, e1) √has σz on the first qubit and identity elsewhere; see also (78). Similarly GΩ(1) = (X1 + Z1)/ 2 is another transvection. Thus, rank one BSSCs are columns of transvections, permuted by some Clifford permutation GD(P). See [41, 42] for more on transvections transvections.