arXiv:hep-th/9208005v1 3 Aug 1992. DAMTP/92-20. QUANTUM RANDOM WALKS AND TIME-REVERSAL. S. Majid1. Department of Applied Mathematics.
DAMTP/92-20
QUANTUM RANDOM WALKS AND TIME-REVERSAL
arXiv:hep-th/9208005v1 3 Aug 1992
S. Majid1 Department of Applied Mathematics & Theoretical Physics University of Cambridge Cambridge CB3 9EW, U.K. March 1992
ABSTRACT Classical random walks and Markov processes are easily described by Hopf algebras. It is also known that groups and Hopf algebras (quantum groups) lead to classical and quantum diffusions. We study here the more primitive notion of a quantum random walk associated to a general Hopf algebra and show that it has a simple physical interpretation in quantum mechanics. This is by means of a representation theorem motivated from the theory of Kac algebras: If H is any Hopf algebra, it may be realised in Lin(H) in such a way that ∆h = W (h ⊗ 1)W −1 for an operator W . This W is interpreted as the time evolution operator for the system at time t coupled quantum-mechanically to the system at time t + δ. Finally, for every Hopf algebra there is a dual one, leading us to a duality operation for quantum random walks and quantum diffusions and a notion of the coentropy of an observable. The dual system has its time reversed with respect to the original system, leading us to a CTP-type theorem.
1
Introduction
A Hopf algebra is an algebra H over a field k, equipped with algebra homomorphisms ∆ : H → H ⊗ H (the comultiplication) and ǫ : H → k (the counit) such that (∆ ⊗ id)∆ = (id ⊗ ∆)∆,
(ǫ ⊗ id)∆ = id = (id ⊗ ǫ)∆.
(1)
In addition, there is a linear map S : H → H (the antipode) obeying ·(S ⊗ id)∆ = 1ǫ = ·(id ⊗ S). The axioms (1) are obtained by writing out the axioms of an algebra and reversing the arrows. Thus the algebra multiplication · : H ⊗ H → H is supplemented by a ‘reversed-multiplication’ ∆. This means that the algebra axioms are supplemented in such a way as to restore some 1
SERC Fellow and Drapers Fellow of Pembroke College, Cambridge
1
kind of time-reversal or input-output symmetry. Hopf algebras are perhaps among the simplest systems where this kind of input-output symmetry can be formulated. The goal of this paper is to use Hopf algebras to explore this general notion of input-output symmetry in a precise setting, namely that of random walks and Markov processes. Thus our algebras H will always be thought of as algebras of observables (i.e. as if obtained from boolean algebras or from probability spaces, or quantum algebras of observables in the noncommutative case). In this context, asking that the algebra is a Hopf algebra really does restore an input-output symmetry to the system. An element h ∈ H is viewed as a random variable and ∆h ∈ H ⊗ H expresses a linear combination of composite random variables whose joint evaluation would lead to an evaluation of h. Thus, while algebras are traditionally used in formulations of (intuitionistic) deductive logic, with the multiplication expressing necessity of a conclusion, the presence of a Hopf algebra puts us into a framework of modal logic, with the comultiplication expressing (in some sense) the notion of ‘possibility’ to supplement the notion of ‘necessity’. We will examine these issues in a probabilistic setting rather than a logical one, based on classical and quantum random processes. The idea of thinking about Hopf algebras in this context is not a new one. At least in the context of stochastic calculus, it is well known that Hopf algebras provide a tool for formulating random walks and generalizing them. Indeed, this was one of the motivations behind the development of Hopf algebras some years ago, and in more recent times has lead to quantum diffusions and other processes based on Hopf algebras[1][19][4]. From a mathematical point of view there is no particular obstruction to going ahead and formulating the same constructions when the Hopf algebra is an arbitrary (possibly non-commutative) one. In particular, random walks make sense even when the Hopf algebra is non-commutative. On the other hand, it is not at all clear a priori if such generalized random walks are anything more than mathematical deformations, i.e. if they are physical quantum processes. It is this question that we address in the present paper. Our main result in this direction is in Section 3 and is an algebraic operator realization 2
theorem for arbitrary Hopf algebras. Some of the ideas behind the realization theorem come from the theory of Kac algebras, see[9][3][7], but a feature of our new treatment is that neither the ∗-structure nor a Hilbert space need be built in from the start. Thus, our construction should be of interest to Hopf algebraists and in the theory of discrete random processes, as well as in the quantum context. In a quantum mechanical context it means that every Hopf algebra H leads in a reasonably convincing way to a quantum random walk based on a quantum evolution operator W (which we build from H). This is such that a step of the random walk on H, considered actively as an operation on H, consists of the following. First, we embed h ∈ H as h ⊗ 1 ∈ H ⊗ H. Here the first copy of H is the algebra of observables at time t and the second copy is the algebra of observables at time t + δ (i.e. one step further in time). Then we evolve h ⊗ 1 by the quantum evolution operator W of the joint system H ⊗ H. Finally, we take an expectation value in the first copy of H to leave us in the second copy at t + δ. This is the quantum step of a quantum random walk. Our theorem asserts that a random walk on any Hopf algebra can always be put into this form. Section 4 then proceeds to study the impact of Hopf algebra duality in this context. Every Hopf algebra has a dual one, and hence every such quantum random walk has a dual one. We find that it is truly time-reversed with respect to the original. The ideas here are based on the notion of observable-state symmetry (quantum ‘Mach principle’) developed in the quantumgravitational context in [13][12][17]. We begin in Section 2 with a brief introduction to how Hopf algebra methods can be used to do classical random walks. Nothing in Section 2.1 should be new to experts, however we have not been able to find a suitable treatment elsewhere. We show how to use the Hopf algebra of the real line to easily derive the diffusion equation for Brownian motion from the limit of a random walk given by stepping to the left or right with probability p, (1 − p). Section 2.2 shows how straightforward q-deformation techniques lead immediately to a q-deformed random walk on R and a limiting q-Brownian process. The latter result can be compared with the quantum stochastic process related to the Azema martingale and based on a similar (but different) Hopf 3
algebra in [19]. Our more primitive treatment as the limit of a discrete walk seems to be new. Let us note that the question of when a (quantum) algebra of observables is a Hopf algebra has already been studied by the author in a very different context through a series of papers [13][14][16][12]. We studied particles moving on curved spaces and found that not every background metric admits this possibility of a Hopf algebra structure for the quantization of a particle in it. For example, for a differentiable 1+1 dimensional quantum dynamical system we looked for Hopf algebras of self-dual type and found that the only possibility was for the background to be that of a black-hole type metric[12]. Mathematically, the existence of a non-cocommutative Hopf algebra structure on the (non-commutative) quantum algebra of observables, corresponds to a non-Abelian group-structure on ‘phase-space’. Here we regard the quantum algebra of observables in the fashion of non-commutative geometry as if it is the ring of functions on some space. The phase space does not exist in an ordinary sense for in this case the functions on it would be commutative. Physically, the reason is that the position and co-ordinate functions in phase space can no longer be determined simultaneously because of Heisenberg’s uncertainty principle. Thus, the Hopf algebra structure corresponds to a non-Abelian group structure on phase space but in a language general enough to allow the space to be a quantum space rather than a classical one. The non-Abelianess means that the phase-space as a (quantum) space is curved, and hence led to models exhibiting a unification of quantum effects with gravitational ones. These models in [13][12] also exhibited a remarkable duality phenomenon, based on Hopf algebra duality and interchanging microscopic (quantum) physics with macroscopic (gravitational physics)[17]. As a byproduct, we were led to a large class of non-commutative and non-cocommutative Hopf algebras which were quite different from the more popular quasitriangular Hopf algebras (quantum groups) of Drinfeld and Jimbo etc. related to Yang-Baxter equations and such topics. Thus, it is the same mathematical structure, namely an algebra of observables that is a Hopf algebra, which has these two interpretations, (i) in terms of gravitational physics as in the authors previous work and (ii) in terms of probability theory as here. Either application alone would lead us to study Hopf algebras and the fact that the same structure serves both indicates 4
the possibility of a remarkable unification of the two topics. Roughly speaking, generalized (quantum) Riemannian geometry can be reformulated in terms of the language of random walks, and vice-versa. This is one of the long-term motivations behind the present work, and will be further developed elsewhere. Here we concentrate only on the probabilistic interpretation. Both (i) and (ii) are quite different again from the standard interpretation of quasitriangular Hopf algebras in the theory of inverse scattering - where the quantum group is not at all a quantum algebra of observables but rather a quantum symmetry. This role of quantum symmetry represents a third potential unification made possible by Hopf algebras, but not something that we shall develop here. On the other hand, there is a tangential connection between related quantum spin chains and non-commutative Markov processes which can be mentioned. Acknowledgments I thank K.R. Parthasarathy and several others for valuable discussions during the December 1991 meeting in Oberwolfach.
2
Random Walks and Markov Processes via Hopf Algebras
In this preliminary section we explain in elementary terms how Hopf algebras can be used to describe random walks and Markov processes. Even in simple and well-known cases they encode the required computation in a very direct way. The reason for this is quite fundamental as explained in the introduction.
2.1
Brownian Motion
We begin by recalling some standard ideas from classical probability. To be concrete we focus on R (the real line) as our probability space. A probability density function for us is simply a pointwise positive function ρ ∈ L1 (R) such that
R
dxρ(x) = 1. We also allow atomic measures
(δ-functions). We can use such a ρ to define a random walk as follows: at x0 = 0, xi+1 = xi + X where X is chosen randomly with probability distribution ρ. The routine question to ask is: what is the probability distribution after n steps, i.e. of xn ? Clearly it is given by a probability
5
density ρn (say) defined by n
ρ (x) =
Z
y1 +···yn =x
dy1 · · · dyn ρ(y1 ) · · · ρ(yn ).
(2)
Note that in asking after the system at various steps, we are always asking the same mathematical question, namely the distribution of a random variable on the real line. Thus we take the point of view that the question, which we denote X, is not changing with each step. Rather, it is the probability distribution of X which is changing with each step. This X is the abstract position observable of the random walk, i.e. ‘where the particle is at’ (statistically speaking). Its expectation value and moments etc are changing with each time step. The distribution (2) immediately gives < 1 >ρn =< 1 >nρ = 1,
< X 2 >ρn = n(< X 2 >ρ − < X >2ρ )+n2 < X >2ρ
< X >ρn = n < X >ρ ,
(3) where < > denotes expectation value in the marked state. From this we see that the mean position is increasing with each step, as would be expected. Also, if we define Y (n) = n1 X then < Y (n) >ρn =< X >ρ is independent of n while its variance < Y (n)2 >ρn − < Y (n) >2ρn =
1 n (
ρ − < X >2ρ ) decreases. Similarly the higher variances decrease, so that the rescaled variables Y (n) become more and more sharply peaked. This is the basis of the central limit theorem. This elementary computation can be done by means of Hopf algebras as follows. Firstly, we regard the abstract question X (roughly speaking) as an element of L∞ (R) (the Hopf-von Neumann algebra of functions on the real line). It is defined as the coordinate function X(x) = x. Now, a probability distribution assigns to any power X n , or more precisely any bounded function f (X) an expectation value < f (X) >ρ =
R
dxρ(x)f (x). This gives a positive linear functional φ
on L∞ (R), corresponding to ρ via φ(f ) =< f (X) >ρ . Positive means φ(f ∗ f ) > 0 for all f 6= 0 and corresponds to ρ real and positive. The normalization corresponds to φ(1) = 1. Thus we can work directly with the Hopf algebra L∞ (R) in place of random variables, and φ a (normal, unital) state on L∞ (R) in place of ρ. This is a familiar point of view in the context of classical 6
mechanics (and quantum mechanics) and L∞ (R) here is the algebra of observables of the system. The Hopf algebra structure on L∞ (R) is (∆f )(x, y) = f (x + y), where L∞ (R) ⊗ L∞ (R) = L∞ (R × R). It expresses the group law on R in algebraic terms. The co-ordinate function, for example, has the additive comultiplication (∆X)(x, y) = X(x+y) = x+y = (X ⊗ 1+1 ⊗ X)(x, y) or as elements of the Hopf algebra, ∆X = X ⊗ 1 + 1 ⊗ X.
(4)
The group law played a key role in (2) and likewise, ∆ plays the corresponding role in defining a new state φn . Indeed, the comultiplication on any Hopf algebra defines a multiplication in its dual, and it is this multiplication (the convolution product) that we need. Explicitly, (φψ)(f ) = (φ ⊗ ψ)(∆f ),
e.g. φn (f ) = (φ⊗ n )(∆n−1 f ).
(5)
where ∆n−1 denotes ∆ applied n − 1 times. We will come to the formal definition of a Hopf ∗-algebra in Section 2.3, but in our example (and in general) the properties of the ∗-structure are such as to ensure that if φ is positive then φn is also positive etc, i.e. also a state. Associativity of this convolution algebra is ensured by the first of (1), while ǫ in (1) is the identity state in the algebra. In these terms, the computation of (3) looks like < 1 >ρn < X >ρn ρn
= φn (1) = (φ⊗ n )∆n−1 1 = (φ ⊗ · · · ⊗ φ)(1 ⊗ · · · ⊗ 1) = φ(1)n = 1 = (φ⊗ n )∆n−1 X = (φ ⊗ · · · ⊗ φ)(X ⊗ · · · ⊗ 1 + · · · + 1 ⊗ · · · ⊗ X) = nφ(X) = n < X >ρ ⊗n
= (φ
)∆
n−1
X
m
= (φ ⊗ · · · ⊗ φ)(
X
i1 +···+in =m
=
X
!
!
m X i1 ⊗ · · · ⊗ X in ) i1 · · · in
X m m φ(X i1 ) · · · φ(X in ) = i1 · · · in i1 · · · in
!
< X i1 >ρ · · · < X in >ρ
Thus we reproduce the results above in this new language. The computation in this form uses nothing other than the two equations (4) and (5). Not only does it bypass unpleasant convolution integrals as in (2), but it is conceptually rather cleaner also. The conceptual picture is that expectation value of X after n steps is simply the expectation value of ∆n−1 X in the tensor 7
product system L∞ (R)⊗ n , φ⊗ n . The tensor product system is the system for n independent random variables, each with the same distribution φ, and the iterated comultiplication embeds the algebra of observables for our one particle into this system. If we denote these co-ordinate functions that generate our n-fold tensor product by Xi = 1 ⊗ · · · ⊗ X ⊗ · · · ⊗ 1
(6)
(embedded in the i’th position), then the random variable X of our one particle embeds as X1 + X2 + · · · + Xn . Regarding ∆n−1 as understood, we can simply write X = X1 + · · · + Xn . Here the Xi are n independent random variables. To demonstrate the power of this formalism further, let us compute a specific example. In terms of ρ we take an atomic measure with ρ = pδa + (1 − p)δ−a , i.e. peaked as δ-functions at a, −a. In terms of φ we have φ(f ) = (pφa + (1 − p)φ−a )(f ) = pf (a) + (1 − p)f (−a)
(7)
where φa (f ) = f (a) is the linear map on L∞ (R) given by evaluation at a. We can also introduce the linear map D(f ) = f ′ (0). It is not a state but we can still view it as a (densely defined) map on L∞ (R) and make the convolution product etc as above, i.e. we formally view it in the convolution algebra. Then the form of ∆ in (5) gives D n (f ) = (D ⊗ · · · ⊗ D)(∆n−1 f ) =
∂ ∂ |0 · · · |0 f (x1 + · · · + xn ) = f (n) (0). ∂x1 ∂xn
(8)
Hence in this convolution algebra we have the expansion φa = ǫ + aD +
a2 2 D + ··· 2!
(9)
(Taylors theorem). We are now ready to compute the system after n steps (and its limit) as described by the states a2 ct tα 1 2 φn = (ǫ + 2a(p − )D + D 2 + · · ·)n = (ǫ + D + D 2 + · · ·)n → φ∞ = etαD +ctD 2 2! n n where t, c, α are defined by t = nδ,
a2 2
=
tα n
= αδ and 2a(p − 21 ) =
and n → ∞ (or δ → 0) with t, c, α fixed. 8
tc n
(10)
= δc and we send a → 0
This limiting state describes our random walk after an infinite number of steps, with the step viewed as a step in time of size δ, which tends to zero, i.e. it is the continuous limit of the random walk. We can still evaluate our observables in this limit, φ∞ (f ) = (etαD
2 +ctD
d 2
)(f ) = (etα dx
d +ct dx
f )(0)
(11)
and we can also compute the distribution ρ∞ corresponding to φ∞ . Formally, if φ is any state we can formally get back the corresponding density by ρ(x) = φ(δx ) (the δ-function at x, which of course has to be approximated to lie in L∞ (R)). We have d 2
ρ∞ (y) = φ∞ (δy ) = (etα dx
d +ct dx
1
δy )|x=0 = (4παt)− 2 e−
(y−ct)2 4αt
(12)
where the right hand side is the unique solution G(y, t) of the diffusion equation
∂ ∂t G(y, t)
=
2
∂ ∂ ∞ is that the latter α ∂y 2 G(y, t) − c ∂y G(y, t). Another way to see that this must coincide with ρ
is characterized by
R
Z
ρ∞ (x)f (x) = etαD (
∂ ∞ ρ )(x)f (x)dx ∂t
2 +ctD
(f ) for all f : Differentiating this
= ((αD2 + cD)etαD
2 +ctD
we have
)(f )
d2 d + c )f (x)dx 2 dx dx Z d2 d ∞ = ((α 2 − c )ρ (x))f (x)dx dx dx =
Z
∂ ∂t
ρ∞ (x)(α
(13)
Where (D ⊗ id)∆f (X) = f ′ (X). Our derivation of (12) can be compared with more standard derivations based on Stirling’s formula etc. In our case, we used only a limit of the form (1 +
Z n n)
→ eZ in deriving (10).
These elementary computations demonstrate the usefulness and directness of this approach based on Hopf algebras. A remarkable fact can now be observed: in these computations we do not need to assume that the Hopf algebra is commutative! Thus, if H is any Hopf algebra, the convolution product (5) for elements φ, ψ of H ∗ make perfectly good sense. We can also take continuum limits by expanding φ and using the same limiting procedure as in (10) without assuming commutativity. These remarks fit well into the overall context of ‘non-commutative’ or ‘quantum’ probability, where we replace L∞ (Ω) for a classical probability space Ω by a noncommutative von Neumann algebra: we can do random walks by working with not-necessarily 9
commutative Hopf-von Neumann algebras. Such generalized random walks can be called ‘noncommutative’ or ‘quantum’ random walks. This situation can be viewed as some of the underlying motivation behind, for example, the notion of a quantum stochastic process. Examples of the latter are indeed known in connection with Hopf algebras[19], though how they might be obtained as the limit of random walks seems to be less well studied, but see for example [11]. Moreover, whether such generalized noncommutative random walks can be viewed as actual quantum mechanical processes is also an open question. These are two of the issues to be addressed in Sections 3,4.
2.2
q-Brownian Motion
We now give a non-commutative example. The above random walk leads to Brownian motion (given by the diffusion equation): the non-commutative example will constitute a q-deformation of it, and can be called q-Brownian motion. A related example has been studied as a quantum stochastic process in [19][18] under the heading of the Azema process. Our more primitive treatment as the limit of a random walk seems to be new and in any case is based on a different Hopf algebra. Note that a ∗-structure was implicit in the above (the states are taken positive) and this continues to be so in our setting now even though the algebra is non-commutative. As in quantum mechanics, the observables of most interest are the self-adjoint ones and states should be positive in the sense φ(h∗ h) ≥ 0 for all h in the algebra. We assume they are normalised as φ(1) = 1. For our Hopf algebra H we take generators X, g, g−1 and relations, comultiplication, counit, antipode and ∗-structure gX = qXg,
gg−1 = 1 = g−1 g,
∆g = g ⊗ g,
∆X = X ⊗ g−1 + g ⊗ X,
ǫg = 1, ǫX = 0. (14)
SX = −q −1 X,
Sg = g−1 ,
X ∗ = X,
g∗ = g,
q ∗ = q −1 .
(15)
This is a standard Hopf algebra [21] but note that q is a parameter of modulus 1 for the ∗structure that we need. The elements X m gi for m ∈ Z+ , i ∈ Z are a basis. The comultiplication 10
should be compared with (4). If φ is any state as above, the distribution φn after n steps is computed from (5) using the comultiplication ∆. It is given by embedding X in H ⊗ n via ∆n−1 and applying φ to each H. From (14) we have ∆n−1 g = g ⊗ · · · ⊗ g,
∆n−1 X = X ⊗ g−1 ⊗ · · · ⊗ g−1 +g ⊗ X ⊗ g−1 ⊗ · · · ⊗ g−1 +· · ·+g ⊗ · · · g ⊗ X. (16)
If we write the right hand expression as ∆n−1 X = X1 +X2 · · ·+Xn where X1 = X ⊗ g−1 · · · ⊗ g−1 , X2 = g ⊗ X ⊗ g−1 · · · ⊗ g−1 etc, we see that the random variable describing the position after n steps is the sum of n random variables Xi embedded in H ⊗ n . They are not, however, independent in a usual sense. Instead, Xi Xj = q 2 Xj Xi ,
i>j
(17)
from the relations in (14). This shows up when we look at higher moments, where we need to compute ∆n−1 X m . This is a standard computation in connection with the above Hopf algebra. We have ∆n−1 X m =
X
i1 +···+in
[m]q2 ! X1i1 · · · Xnin , ! ! · · · [i ] [i ] 2 2 n q =m 1 q
[i]q =
1 − qn 1−q
(18)
using the relations (17). Proposition 2.1 The expectation values after n steps are < g >φn =< g >nφ , < X 2 >φn =< X 2 >φ
< X >φn =< X >φ
< g >nφ − < g−1 >nφ < g >φ − < g−1 >φ
[n]φ − [n]φ < g2 >nφ − < g−2 >nφ + [2]q < Xg >φ < g−1 X >φ 2 −2 < g >φ − < g >φ < g2 >φ − < g−2 >φ
These moments reduce to the usual results (3) for a random walk on R if we set q → 1 and ±2 > ∼< g >±2 . g → 1 in a strong sense with < g±1 >φ ∼< g >±1 φ φ φ and < g
Proof
The expectation of g and X follows at once from (18). For the expectation of X 2 we
have to write out ∆X 2 =
Pn
P 2 in−j . Summing over j, we obtain for the second term in < X 2 > n the result g2 >i−1 φ φ φ < g
[2]q2 < Xg >φ < g−1 X >φ times the expression n−1 X
i−1 φ
1− < g−2 >φn−i 1− < g−2 >φ n−1 n−1 −2 2 < g−2 >φ (< g >φ − < g >φ ) − >φ )(1− < g2 >φ ) (1− < g−2 >φ ) (< g2 >φ − < g−2 >φ )
1− < g2 >φn−1 (1− < g−2
= (< g2 >φ − < g−2 >φ )−1
1− < g2 >nφ 1− < g−2 >nφ − 1− < g2 >φ 1− < g−2 >φ
!
after suitable reorganization of the partial fractions. Note that if we write < g±2 >φ = 1 ± 2δ + O(δ) and take δ → 0 then this expression tends to
n(n−1) 2
as in Section 2.1 ⊔ ⊓
We can also compute the continuous limit of a random walk on this Hopf algebra, as follows. For simplicity we assume that q is not a root of unity. For our state φ we take φ(gi f (X)gj ) = pf (q
i−j 2
a) + (1 − p)f (−q
i−j 2
a)
(19)
for a ∈ R and 0 ≤ p ≤ 1. This reduces to the choice in Section 2.1 when i = j. Note that φ((f (X)gi )∗ f (X)gi ) = φ(gi f¯(X)f (X)gi ) = p|f (a)|2 + (1 − p)|f (−a)|2 so that φ is positive. Here, and below, we can concentrate on f given by polynomials or suitable power-series in the X. As in Section 2.1, this algebraic approach is largely for convenience: underlying it is a Hopfvon Neumann algebra similar to an extension of L∞ (R) above (this is needed to make sense of some of the exponentials below). In order to approximate φ we make use of a well known operator of ‘q-differentiation’ ∂q cf[2] to define an operator Dq : H → H by 1
D q gi f (X)gj = gi− 2
1 1 i−j f (qX) − f (q −1 X) j− 1 i−j g 2 q 2 = gi− 2 (∂q f )gj− 2 q 2 −1 (q − q )X
12
(20)
where we have extended the action to g in such a way that Dq is a ∗-preserving and (formally) 1
a completely positive operator. Here the g− 2 is for notational convenience: it should be understood on one side or the other as a factor g−1 (in view of the relations (14)). In particular, we have D q gi X m gj = [[m]]q gi−1 X m−1 gj q
i−j 2
q
m−1 2
,
[[m]]q =
q m − q −m . q − q −1
(21)
Motivated by this we take a linear functional Dq : H → C given by evaluation of Dq at X = 0, g = 1 and consider its iterated convolution product according to (5). We have Dq (gi X m gj ) = δm,1 q
Dqn (gi X m gj )
i−j 2
(22)
= (Dq ⊗ · · · Dq ) (gi ⊗ · · · ⊗ gi )(∆n−1 X m )(gj ⊗ · · · ⊗ gj ) = δn,m [m]q2 !q
i−(n−1+j) 2
q
i−1−(n−2+j) 2
···q
i−(n−1)−j 2
= δn,m [[m]]q !q
i−j m 2
(23)
using (16) to obtain (22) and then (18),(22) to obtain (23). We also need a (well known) q-exponential eq defined with coefficients 1/[[i]]q ! in place of the usual 1/i!. From (23) we conclude in the convolution algebra that q )(gi f (X)gj ) = f (q (eaD q
i−j 2
a).
(24)
Hence we can approximate φ just as before, by a2 ct tα 1 Dq2 + · · ·)n = (ǫ + Dq + Dq2 + · · ·)n , φn = (ǫ + 2a(p − )Dq + 2 [[2]]q n n where t, c, α are defined by t = nδ,
a2 [[2]]q
tα n
=
and 2a(p − 21 ) =
tc n
2
φ∞ = etαDq +ctDq (25)
and we send a → 0 and n → ∞
(or δ → 0) with t, c, α fixed as before. We can then evaluate φ∞ on elements of our algebra to compute expectation values. We can also try to introduce a corresponding probability density ρ∞ defined via ∞
φ (f (X)) =
Z
∞
ρ∞ (x)f (x)dx,
∀f
(26)
−∞
with respect to ordinary integration (say). This means that we are visualizing our deformed random walk with respect to the usual picture of R, which is embedded in H as a subalgebra. 13
This is not the only possibility (one could, for example, use here a q-integration). Note also that φ∞ (gi f (X)gi ) = φ∞ (f (X))
(27)
from (23), so (26) determines φ∞ completely. Proceeding with (26) and differentiating with respect to t we have Z
(
∂ ∞ ρ )(x)f (x)dx ∂t
2
= ((αDq2 + cDq )etαDq +ctDq )(f (X)) = φ∞ ((αD 2q + cD q )f (X)) = =
Z
Z
(ρ∞ (x)α∂q2 + cρ∞ (x)∂q )f (x)dx (α∂q2 ρ∞ (x) − c∂q ρ∞ (x))f (x)dx.
(28)
where we used (Dq ⊗ id)∆ = Dq .
(29)
This is easily proven from (16) and (22). We also used (27) and the identity Z
f (qx) − f (q −1 x) dx (q − q −1 )x Z Z Z h(q −1 x′ ) h(qx′ ) h(qx) − h(q −1 x) ′ ′ ′ ′ = f (x)dx (30) f (x )dx − f (x )dx = − −1 ′ −1 ′ (q − q )x (q − q )x (q − q −1 )x h(x)
for all analytic functions h, f such that the contours can be rotated for the changes of variable (so that ∂q† = −∂q with respect to this L2 inner product on suitable test functions). Thus, ρ∞ is characterized as the solution of ∂ ∞ ρ (x, t) = α∂q2 ρ∞ (x, t) − c∂q ρ∞ (x, t). ∂t
(31)
This is a q-deformed diffusion equation. Note that unlike the undeformed case, however, our φ∞ does not involve ordinary differentiation but rather the non-local operator of q-differentiation, and hence there is no reason to think that a smooth solution ρ∞ to (31) should exist along familiar Gaussian lines. It can perhaps be interpreted stochastically. Our elementary derivation of q-Brownian motion can be compared with the treatment of the Azema martingale[18] and its quantum stochastic process in [19]. 14
2.3
Transition Operators
The last example should convince the reader that it is interesting to consider random walks on non-commutative Hopf algebras, even if only as deformations of standard ones. Note also that in the above random walks, we are not required to use the same state φ for the distribution of each step of the walk. We could just as easily have a collection of states {φ1 , φ2 , · · · , φn } to be used successively in each step of the random walk. The distribution after n steps is given by the convolution product φ1 φ2 · · · φn , i.e. the tensor product φ1 ⊗ φ2 · · · ⊗ φn applied to the image via ∆n−1 in H ⊗ n . Such a random walk with distinct steps is said to be non-stationary. Closely related to random walks are Markov processes, and these too can be built on Hopf algebras. In fact, for our purposes they are the same structure from an equivalent ‘active’ point of view. We briefly explain this now. Thus, given a linear functional φ on H, we define Tφ : H → H by Tφ = (φ ⊗ id)∆
(32)
This is equivalent to φ, and also recovers the convolution multiplication (5) as composition of operators ǫ ◦ Tφ = φ,
φψ = ǫ ◦ Tψ Tφ
(33)
on using the counity and coassociativity axioms (1) of ∆, ǫ. Thus, our all-important convolution product is just composition of a corresponding operator Tφ : H → H. It can be called the Markov transition operator corresponding to φ. In terms of the Markov transition operator, the system evolves actively by Tφ : H → H and the final expectation values after n steps are obtained by applying ǫ to the evolved observables. Thus, if we have a random walk with varying states φ1 , · · · , φn at each step, the corresponding expectations after n steps are < f >φ1 φ2 ···φn = ǫ ◦ Tφn · · · Tφ1 (f ).
(34)
An algebra H equipped with a step-evolution operator, or operators T : H → H is a (possibly non-commutative) Markov process. We see that this is just an ‘active’ way of thinking about 15
random walks, in which the observables rather than the states evolve. We will say more about this in the next section. In the above example, the role of the counit was in the evaluation at X = 0, g = 1 (see (14)). Thus, D q = TDq is the operator (q-differentiation) corresponding to the functional Dq = ǫ ◦ Dq that was used, see (29). Likewise, the transition operator Tφ∞ corresponding to the limiting state is 2
Tφ∞ = etαD q +ctD q , i.e. the q-deformed time evolution. This formalism of random walks and transition operators works for general Hopf algebras over a field k. Finally, we formalize the situation regarding the ∗-structure needed in a quantum mechanical setting. Recall that a ∗-algebra means an algebra over k = C equipped with an antilinear involutive anti-algebra homomorphism ∗. Then a Hopf ∗-algebra is a ∗-algebra H such that ∆h∗ = (∆h)∗ ⊗ ∗ ,
ǫ(h∗ ) = ǫ(h),
(S ◦ ∗)2 = id.
(35)
Such ∗-structures on Hopf algebras have been emphasised by [22], among others. In this case (for φ positive), the transition operator Tφ is a completely positive map [1]. Our example in Section 2.2 fits into this setting and Dq is completely positive, at least formally. Also, if H is a Hopf algebra then H ∗ (defined in a suitable way) is also a Hopf algebra and the comultiplication in one is determined by the multiplication in the other. This is just the origin of the convolution product (5), but is also works the other way with < ∆φ, h ⊗ g >=< φ, hg > and < Sφ, h >=< φ, Sh > with the comultiplication and antipode of H ∗ coming from H (for this reason we write the evaluation symmetrically as < , > and say that H, H ∗ are dually paired Hopf algebras[15, Sec. 1]). In the Hopf ∗-algebra case we have in addition < φ∗ , h >= < φ, (Sh)∗ >.
(36)
It follows at once from this and (35) that Tφ (h)∗ = T(Sφ)∗ (h∗ ) 16
(37)
so that Tφ is ∗-preserving iff φ is anti-self-adjoint in the sense φ∗ = Sφ
(38)
in H ∗ . For our example of Section 2.2, the state Dq is anti-self-adjoint in this way so that Dq was necessarily ∗-preserving, as easily seen by direct computation. In summary, positivity of φ corresponds to complete positivity of the corresponding transition operator, while anti-selfadjointness corresponds to the transition operator being ∗-preserving. Let us note finally that Markov processes (like random walks and stochastic processes) can be defined more generally than those based on Hopf algebras, for example in the context of spin chains in [4]. Also, their continuum limit is related to a process of dilation of operators[10].
3
Operator Realization of Hopf Algebras
We have seen in the last section how any Hopf algebra H equipped with a linear functional φ can be interpreted formally as leading to a generalized random walk or Markov process. In the random walk interpretation the distribution of an observable h ∈ H after n steps is given by embedding h into H ⊗ n as ∆n−1 h and applying the expectation value φ to each factor. The factors are the steps of the random walk and ∆n−1 h is understood literally as the linear combination of all the ways to arrive at h via the elements in H ⊗ n viewed as successive steps. In this section we explore this interpretation further by proving a general result about the coproduct of an arbitrary Hopf algebra. Although we are primarily interested in the Hopf ∗-algebra situation needed for quantum mechanics, our first result is more general and works over an arbitrary field or commutative ring k. Perhaps this more general case will be needed in discrete applications as well as in attempts to unify the present considerations with Planck-scale physics as mentioned in the Introduction. If we have to drop classical geometry at the Planck scale, then it is likely that we will have to drop classical functional analysis too (based on Rn ) at some point, and do everything algebraically. While the result is known in some form at a Hopf ∗-algebra level[3], our purely algebraic level seems to be more novel. 17
Theorem 3.1 Let H be a Hopf algebra and view H ⊂ Lin(H) by h 7→ h⊲ = Lh (the left regular representation Lh (g) = hg). In the algebra Lin(H ⊗ H) ⊇ Lin(H) ⊗ Lin(H) there is an invertible element W such that ∆h = W (h ⊗ 1)W −1 ,
Sh = (ǫ ⊗ id) ◦ W −1 (h ⊗( )),
W12 W13 W23 = W23 W12 .
Here W12 = W ⊗ 1, W23 = 1 ⊗ W in Lin(H ⊗ 3 ) ⊇ Lin(H)⊗ 3 (similarly for W13 ). Let H ∗ ⊂ Lin(H) by φ 7→ φ⊲ = Rφ∗ (the left coregular representation Rφ∗ = (id ⊗ φ)∆). Then viewed in this algebra we also have, where defined, ∆φ = W −1 (1 ⊗ φ)W,
Sφ = (id ⊗ φ) ◦ W −1 (( ) ⊗ 1)).
In the finite-dimensional case the subalgebras H ⊂ Lin(H) and H ∗ ⊂ Lin(H) together generate all of Lin(H). Proof
Note that the set of linear maps Lin(H ⊗ H) contains Lin(H) ⊗ Lin(H) in the standard
way but may be larger in the infinite-dimensional case. Explicitly, W, W −1 are defined as such linear maps H ⊗ H → H ⊗ H by W (g ⊗ h) = Here ∆g =
P
X
g(1) ⊗ g(2) h,
W −1 (g ⊗ h) =
X
g(1) ⊗(Sg(2) )h.
g(1) ⊗ g(2) is a standard notation[20]. For brevity, we will omit the
first verify the identities
(39) P
signs. We
(W (h ⊗ 1)W −1 )(g ⊗ g′ ) = W (h ⊗ 1)⊲(g(1) ⊗(Sg(2) )g) = W (hg(1) ⊗(Sg(2) )g′ ) = h(1) g(1)(1) ⊗ h(2) g(1)(2) (Sg(2) )g′ = h(1) g ⊗ h(2) g′ = (∆h)⊲(g ⊗ g′ ), (W −1 (1 ⊗ φ)W )(g ⊗ g′ ) = W −1 (1 ⊗ φ)⊲(g(1) ⊗ g(2) g′ ) = W −1 (g(1) ⊗ g(2)(1) g′ (1) ) < φ, g(2)(2) g′ (2) > = g(1)(1) ⊗(Sg(1)(2) )g(2)(1) g′ (1) ) < φ, g(2)(2) g′ (2) >= g(1) ⊗ g′ (1) < φ, g(2) g′ (2) >= φ(1) ⊲g ⊗ φ(2) ⊲g′ . Similarly for the antipodes. As for the equations satisfied by W itself, we evaluate on H ⊗ H ⊗ H as W12 W13 W23 (g ⊗ h ⊗ f ) = W12 W13 (g ⊗ h(1) ⊗ h(2) f ) = W12 (g(1) ⊗ h(1) ⊗ g(2) h(2) f ) 18
= g(1)(1) ⊗ g(1)(2) h(1) ⊗ g(2) h(2) f = g(1) ⊗ g(2)(1) h(1) ⊗ g(2)(2) h(2) f = W23 (g(1) ⊗ g(2) h ⊗ f ) = W23 W12 (g ⊗ h ⊗ f ). For the last part we show that every operator H → H arises by actions ⊲ of H and H ∗ on H, at least when H is finite-dimensional. In this case every linear operator can be viewed as an element of H ⊗ H ∗ acting on H in the usual way by evaluation, namely (h ⊗ φ)(g) = h < φ, g >. We have to represent this by elements of H, H ∗ acting via ⊲. Indeed, as operators in Lin(H) we find (h ⊗ φ) = h(S −1 ea (1) )⊲(< φ, ea (2) > f a ⊲( )) where ea is a basis of H and f a a dual basis. Thus h(S −1 ea(1) )⊲(< φ, ea(2) > ⊗ f a ⊲g) = h(S −1 ea(1) )g(1) < φ, ea(2) >< f a , g(2) >= h(S −1 g(2)(1) )g(1) < φ, g(2)(2) >= h < φ, g > as required. We used coassociativity and the properties of S −1 as skew-antipode. ⊔ ⊓ If W is any invertible operator obeying the pentagon identity W12 W13 W23 = W23 W12 then it is easy to see that ∆(h) = W (h ⊗ 1)W −1 for any operator h will always be coassociative and an algebra homomorphism, cf[9]. If we can arrange also for a counit (typically by restricting our operators to some subalgebra) then this gives a Hopf algebra. A Hopf algebra of this type is very concrete, being realised as operators on some vector space, and hence very suitable for physical applications. On the other hand, it might be thought that only very special Hopf algebras could be obtained concretely in this way, but our theorem says that every Hopf algebra can be realized concretely by acting on itself. Even more, it says that both the Hopf algebra and its dual can be realized concretely at the same time as subalgebras of operators on the same space. For a quantum-mechanical setting, as well as to make contact with the existing theory of Hopf-von Neumann and Kac algebras, we consider now the situation when H is a Hopf ∗-algebra. To do this, it is helpful to cast the last part of the preceding theorem as a statement about Weyl algebras. If H, H ∗ are dually paired Hopf algebras, we define w(H) (the Weyl algebra of H) to be the semidirect product H>⊳H ∗ where H ∗ acts on H by the left action ⊲ above. The multiplication is defined in the linear space H ⊗ H ∗ by (h ⊗ φ)(g ⊗ ψ) =
X
hg(1) ⊗ φ(1) ψ < φ(2) , g(2) > . 19
(40)
Both H, H ∗ are subalgebras. Incidentally, this is symmetric in a certain sense between H, H ∗ ; one has w(H ∗ )∼ =w(H)op at least if H has a bijective antipode (which we assume). Now, if H is a Hopf ∗-algebra then so is H ∗ and w(H) becomes a ∗-algebra by (h ⊗ φ)∗ = (1 ⊗ φ∗ )(h∗ ⊗ 1).
(41)
This is the only possibility when we bear in mind that (h ⊗ φ) = (h ⊗ 1)(1 ⊗ φ). It is easy to see that it indeed gives a ∗-algebra and that H, H ∗ are ∗-subalgebras. Corollary 3.2 Let H be a finite-dimensional Hopf algebra. Then Lin(H)∼ =w(H) as algebras. The left hand side has the usual composition of linear maps. Hence if H is a Hopf ∗-algebra, Lin(H) becomes a ∗-algebra. The first part follows at once from the last part of the theorem, where h ⊗ φ in w(H)
Proof
built on H ⊗ H ∗ corresponds to h⊲(φ⊲( )) in Lin(H). Conversely, from the proof there, the element corresponding to h ⊗ φ ∈ Lin(H) (acting in the usual way) is h(S −1 ea(1) ) ⊗ < φ, ea(2) > f a ∈ w(H) built on H ⊗ H ∗ . Hence Lin(H) is a ∗-algebra inherited from the ∗-algebra structure of w(H). Since H, H ∗ are subalgebras, this is such that (h⊲)∗ = h∗ ⊲ and (φ⊲)∗ = φ∗ ⊲ as operators in Lin(H). ⊔ ⊓
Proposition 3.3 In the setting of Theorem 3.1, suppose that H is a finite-dimensional Hopf ∗-algebra and let Lin(H) have the ∗-structure generated by H, H ∗ (i.e. from w(H)). With respect to this ∗-structure, the fundamental operator W is unitary, W ∗ ⊗ ∗ = W −1 . The right-invariant integral
R
on H defines a sesquilinear form (g, h) =
R
g∗ h = (h, g) which is
compatible with the ∗-structure on Lin(H) induced by H, H ∗ in the sense Z
∗ ′
(h⊲g) g =
Z
∗
∗
′
g (h ⊲g ),
Z
∗ ′
(φ⊲g) g =
20
Z
g∗ (φ∗ ⊲g′ ),
g, g′ , h ∈ H, φ ∈ H ∗ .
For the first part we begin by writing W in terms of such elements from H, H ∗ . Indeed,
Proof
W (g ⊗ g′ ) = g(1) ⊗ g(2) g′ = g(1) < f a, g(2) > ⊗ ea g′ = f a ⊲g ⊗ ea ⊲g′ , where ea is a basis of H and f a is a dual basis. So W = f a ⊲ ⊗ ea ⊲ and hence , by definition, W ∗ ⊗ ∗ = f a∗ ⊲ ⊗ ea ∗ ⊲ = f ′a ⊲ ⊗(Se′a )⊲ where e′a = S −1 ea ∗ , f ′a = f a∗ are a new mutually dual basis. Dropping the primes, the action of the result is f a ⊲g ⊗(Sea )⊲g′ = g(1) < f a, g(2) > ⊗(Sea )g′ = g(1) ⊗(Sg(2) )g′ = W −1 (g ⊗ g′ ). For the second part we show that this ∗-structure on Lin(H) really is an adjoint operation R
with respect to ( , ). The right integral is characterized by (id ⊗ )∆ = 1 scale. This means at once that (S )∗ = R
R
R
and is unique up to
since (S )∗ is also a right integral and has the same R
normalization, and this implies that ( , ) is hermitian as stated. For the operators h⊲ we have (h⊲g, g′ ) = (h⊲g)∗ g′ = (hg)∗ g′ = R
R
R
g∗ h∗ g′ =
R
Z
(g(1) )∗ < φ, g(2) >h =
Z
harder is for the operators φ⊲, (φ⊲g, h) = = = = = =
Z
Z
Z
Z
Z
g
Z ∗
(φ⊲g)∗ h =
(1)
h=
Z
g∗ (h∗ ⊲g′ ) = (g, (h⊲)∗ g′ ) is automatic. Rather
g∗ (1) < φ, (g∗ (2) )∗ >h
g∗ (1) h(1) < S −1 φ∗ (3) , g∗ (2) >< (S −1 φ∗ (2) )φ∗ (1) , h(2) >
g∗ (1) h(1) < (S −1 φ∗ (2) )(1) , g∗ (2) >< (S −1 φ∗ (2) )(2) , h(2) >< φ∗ (1) , h(3) > g∗ (1) h(1) < S −1 φ∗ (2) , g∗ (2) h(2) >< φ∗ (1) , h(3) > (g∗ h(1) )(1) < S −1 φ∗ (2) , (g∗ h(1) )(2) >< φ∗ (1) , h(2) > g∗ h(1) < φ∗ , h(2) >=
Z
g∗ (φ∗ ⊲h) = (g, (φ⊲)∗ h).
Here the second equality is from the definition of φ⊲g = g(1) < φ, g(2) >. The third equality is from the definition of a Hopf ∗-algebra with regard to ∆ ◦ ∗. The fourth equality is from the relation between the ∗ structures in H, H ∗ and (S ◦ ∗)2 = id. The fifth equality inserts some factors that collapse to ǫ(h(2) ) via the fact that S −1 is a skew-antipode for H. The sixth equality writes the multiplication in H ∗ in terms of the comultiplication in H, and uses coassociativity and that S −1 is an anticoalgebra map. The seventh equality writes the comultiplication in H ∗ in terms of H to combine some factors, while the eighth equality uses that ∆ is an algebra homomorphism. By these manipulations we are ready, in the ninth equality, to use that 21
R
is a
right integral, leading to the required result. ⊔ ⊓ In a functional-analytic context we can take for H a von-Neumann algebra, and in place of Lin(H) we take B(Hφ ), the bounded operators on the Hilbert space Hφ determined by a state or weight φ. H is embedded in this by the GNS construction. With φ =
R
(the integral
on H), we have that the ∗-algebra structure on Lin(H) becomes the usual adjoint operation † on B(Hφ ), so that W † = W −1 . The structure in Theorem 3.1 and Proposition 3.3 is then characteristic of a Kac algebra[9]. This was historically limited to S 2 = id, but this is not needed when formulated along the lines above, cf [3]. Although we will work algebraically, we keep in mind this Hopf-von Neumann or Kac algebra setting when we consider quantum mechanical examples. One complication in the Hopf-von Neumann or Kac algebra setting is that the counit is typically unbounded. As explained in the introduction, the significance of these algebraic results for our present purposes is that they give a direct ‘quantum mechanical’ interpretation of the random walk associated to a Hopf algebra H and a linear functional φ. In terms of the Markov transition operator we see that
Tφ (h) = (φ ⊗ id) W (h ⊗ 1)W −1 .
(42)
Here φ should be thought of as giving the expectation value over the first copy of H in H ⊗ H ⊂ Lin(H) ⊗ Lin(H). Thus our random walk, considered actively as a Markov process consist of the following. First, we embed h ∈ H as h ⊗ 1 ∈ H ⊗ H. Here the first copy of H is the algebra of observables at time t and the second copy is the algebra of observables at time t + δ (i.e. one step further in time). Then we evolve h ⊗ 1 by the quantum evolution operator W of the joint system H ⊗ H. Finally, we take a conditional expectation value in the first copy of H to leave us in the second copy at t + δ. This represents ‘forgetting’ the details of where the system might have been at the now unobserved time t. Thus Tφ is the quantum step of a quantum random walk in a reasonably physical way.
22
4
Duality, Coentropy and Time-Reversal
In this section we explore the implications of Hopf algebra duality for the interpretation of quantum random walks on Hopf algebras. A parallel observable-state symmetry in the context of quantum-gravity was explored in [12]. The point of [12] was that when an algebra of observables is a Hopf ∗-algebra, then the dual H ∗ is also a (Hopf) ∗-algebra. Thus we can, in principle, regard H ∗ instead as the algebra of observables of some dual quantum system. The original observables are now regarded as elements of H ∗∗ i.e. linear combinations of states from the dual point of view. Roughly speaking (and not worrying about positivity) the same expectation value φ(h) of observable h in state φ is regarded from the dual point of view as h(φ), the expectation of φ in state h. We gave several examples in [12][16] based on quantum particles on homogeneous spaces. The classical data was a pair of groups (G1 , G2 ) (the momentum and position groups respectively) acting on each other in a compatible way. The dual Hopf (von Neumann-Kac) algebra was of the same type with the roles of G1 , G2 interchanged. In the dual picture the quantum particle moves along orbits of G2 in G1 instead of orbits of G1 in G2 . Thus, interesting models certainly exist in which the dual system has just as good a physical interpretation as the original system and for which the implications of this observable-state symmetry can be explored. Our considerations will be at a formal algebraic level (there exist algebraic examples too[14]) but the most natural setting that we have in mind is this Hopf-von Neumann or Kac algebra one. To explore this duality in the present context we fix H a Hopf ∗-algebra and φ a state on it. We have seen that φ generates a quantum random walk with transition operator Tφ . Note that there is clearly an arrow of time built into this interpretation as we successively take each step of the walk. It can be expressed by the fact that the completely positive operator Tφ necessarily increases the entropy of φn , as follows. Firstly, the relative entropy S(φ, ψ) ≤ 0 is defined between two states φ, ψ in a standard way. For the algebra of bounded operators on a Hilbert space, states are of the form φ =
23
P
i si
, ψ =
j rj
P
< ψj | |ψj > (convex linear combinations of pure states) and S(φ, ψ) =
X
(−si log si + si log rj )| < φi |ψj > |2 .
(43)
i,j
This definition extends to abstract von Neumann and C ∗ algebras as well as into certain algebraic situations. For recent work see [8][6] and elsewhere. For our purposes below we need only the standard abstract properties of the entropy without worrying too much about its detailed definition. Its heuristic interpretation, as discussed in [5][8] is that eS (φ,ψ) is the probability per unit measurement of the system appearing to be in state φ when it is in state ψ. In very general terms it measures the ratio of the impurity of φ to that of ψ. For the entropy of a single state an obvious choice for us is to take the entropy relative to ǫ or the entropy of ǫ relative to the state. For our random walk we have 0 ≥ S(ǫ, φn ) = S(ǫ⊗ n ◦ ∆n−1 , φ⊗ n ◦ ∆n−1 ) ≥ S(ǫ⊗ n , φ⊗ n )
(44)
since ∆n−1 is a completely positive map (being a ∗-algebra homomorphism) and using (1). There is a similar identity S(φn , ǫ) ≥ S(φ⊗ n , ǫ⊗ n ) although this is less interesting because S(φ, ǫ) etc tend to be −∞ or 0 (this is the case classically). These inequalities give some information about the relative entropy between ǫ and φn . We can also get incremental information in the form 0 ≥ S(φn , φn+1 ) = S(φn−1 ◦ Tφ , φn ◦ Tφ ) ≥ S(φn−1 , φn ) ≥ · · · ≥ S(ǫ, φ)
(45)
since Tφ is completely positive. This says that as n grows, the states φn of our random walk change more and more slowly in the sense that the probability for the system to still appear in state φn when it is in the next φn+1 , increases. It also says that the impurity of φn , which is generally less than that of φn+1 , tends towards the latter. In fact, the reason for this is that as the walk evolves, the states are becoming more and more dissipated or impure, but this however, is in some sense bounded by the right integral assume is positive). To see this we note that Z
◦Tφ = (φ ⊗
Z
)∆ = φ(1) 24
Z
R
=
from Proposition 3.3 (which we
Z
(46)
R
by the right invariance of . Hence 0 ≥ S(φn+1 ,
Z
) = S(φn ◦ Tφ ,
Z
) = S(φn ◦ Tφ ,
Here the entropy of φn relative to
R
Z
◦Tφ ) ≥ S(φn ,
Z
) ≥ · · · ≥ S(ǫ,
R
(47)
. It also becomes more and
more impure as its degree of impurity approaches the degree of impurity of R
)
indeed increases with each step and in this sense the
state becomes more and more similar (before any rescaling) to
integral
Z
R
as claimed. This
represents a kind of ‘maximal entropy’ or ‘maximally impure’ state, and our random
walk evolves towards it. In the classical case of functions on a finite group, the integral is precisely the maximally impure state consisting of the average of all the pure states (which are given by evaluation at the various group elements), while ǫ is a pure state (given by evaluation R
at the group identity element). In this case it is easy to compute that S(ǫ, ) = − log |Ω| where R
|Ω| is the order of the group Ω. There is a similar formula for S(ǫ, ) for general Hopf algebras and we see that it provides a lower bound for S(φn , ) . The origin of this increase of entropy relative to
R
R
is that whereas the evolution of the joint
system by W in (42) is invertible, the conditional expectation given by evaluation in the state φ represents forgetting information. This is the reason that repeated application of Tφ is entropy increasing in this way. Moreover, we demonstrated in Section 2.2 precisely how this discrete arrow of time leads in a limit to a more familiar continuous arrow of time. Now for a fixed H we can also consider a ∈ H as an element of H ∗∗ , and hence this induces a random walk on H ∗ (a Hopf algebra dually paired with H). We need to assume that this a is positive as a linear functional on H ∗ , Definition 4.1 An element a in a Hopf ∗-algebra H is copositive if < a, φ∗ φ >≥ 0 for all φ ∈ H ∗ . It is normalised if ǫ(a) = 1. For example, if a is grouplike and anti-self-adjoint in the sense Sa = a∗ , then it is copositive. Convex linear combinations a=
X
si ai ,
Sai = a∗i ,
i
25
∆ai = ai ⊗ ai
of such elements ai are also copositive. We say that a copositive observable is pure if it is not the convex linear combination of copositives. Now, if a is a normalized copositive element then the random walk induced on H ∗ has transition operator Ta : H ∗ → H ∗ . This random walk when viewed actively as an irreversible evolution among the observables of the dual system (i.e. among elements of H ∗ ) can be referred to H where it appears passively as evolution of the states of the dual system. This is given by the adjoint operators Ta∗ : H → H. We compute < φ, Ta∗ (h) >=< Ta (φ), h >=
X
< φ(1) , a >< φ(2) , h >=< φ, ah >,
i.e.
Ta∗ = La : H → H. (48)
Thus, the transition operator of the dual random walk corresponds to left multiplication by a when viewed in terms of the original system. Note that information is still being lost with each step even in this picture. An immediate corollary of the remarks in Section 2.3 (at least in the finite-dimensional case) is Corollary 4.2 An element a ∈ H is copositive iff the operator of left-multiplication La : H → H is the adjoint of a completely positive map. This interpretation of a copositive element a ∈ H as leading to a quantum random walk in the dual, leads us to define the notion of coentropy of an element of H. Definition 4.3 Let a, b ∈ H be copositive elements. The relative coentropy S ∗ (a, b) is the relative entropy of a, b as linear functionals on H ∗ . The familiar properties of entropy now appear in dual form for the coentropy. For example, a ◦ Tb = ba as linear functionals on H ∗ (see (48)) means that S ∗ (ba, bc) ≥ S ∗ (a, c)
(49)
∗ for three copositive elements. Likewise, the heuristic interpretation of S ∗ (a, b) is that eS (a,b)
should be thought of as the probability per unit trial of mistaking the observable or random variable b for a when examined on random states of the system. In very general terms it is the 26
ratio of the impurity of a to that of b. An example of a copositive element is provided by the left integral element in our Hopf ∗-algebra. This is an element Λ ∈ H such that hΛ = ǫ(h)Λ for all h, and at least in the finite-dimensional case this exists and is unique up to scale. It is anti-self-adjoint according to SΛ = Λ∗ , and in nice cases is copositive in the above sense. It corresponds to the right integral R
above for .
R
on H ∗ and is maximally impure in the way that we have seen
We now consider H equipped with a copositive element a. It induces a random walk in H ∗ , but when referred to H this process is the left multiplication operator. Thus, from the point of view where H is regarded as a quantum algebra of observables, the second process appears not as a random walk on H but as what we will call a creation process. Recall that in any GNS representation of H, the element 1 in H leads to a vacuum vector |0 > and La turns this into another vector a|0 >, created by a. Repeating this for a sequence a1 , · · · an gives the vector an an−1 · · · a1 |0 > . To be concrete here, one can keep in mind the GNS representation given by the integral
R
as
in Proposition 3.3. In summary, a geometrical process involving the group structure (comultiplication) of H ∗ corresponds from the point of view of H to a quantum process involving multiplication in the quantum algebra of observables. This was the point of view of [12] also. This quantum ‘creation process’ also has a simple classical meaning, at least if we do not worry about the copositivity or normalization requirements. In the classical case the algebra of observables is the algebra of functions on a space Ω. Among them, the characteristic functions of subsets of Ω correspond to the observables or random variables that specify membership of the subsets. A sequence a1 , a2 , · · · of such characteristic functions is then a creation process that successively specifies more and more conjunctions as we successively multiply more and more characteristic functions. The first step of the creation process is 1 (the identity function), the next step a1 .1 specifies membership of one subset, a2 .a1 .1 further specifies an additional membership, etc. This models then the way that classical concepts or observables are created
27
as a series of specifications. We have seen that this elementary process corresponds to a random walk in the dual. Finally, our comments above about the entropy of a random walk now become comments about coentropy in a creation process. Using (49), the analogue of (47) is 0 ≥ S ∗ (an+1 , Λ) ≥ S ∗ (an , Λ) ≥ · · · ≥ S ∗ (1, Λ).
(50)
This suggests that for a typical copositive element a the sequence 1, a, a2 , · · · typically tends to Λ (and perhaps converges to something sensible after rescaling according to a dual central limit theorem). In general terms it becomes more and more impure as it approaches Λ in this coentropic sense. We can easily see this phenomenon in the classical case of the algebra of functions on a (say discrete) group Ω, as follows. In this case the copositives are convex linear combinations of normalised traces (characters) of representations of Ω, with the pure copositives being the traces of the irreducible representations. The normalization is ǫ(a) = a(e) = 1 where e is the group identity. The integral Λ is given by the Kronecker δ-function at e. For Abelian groups we have a =
P
i si ai
where the ai are the one-dimensional representations of the group,
i.e. the ‘plane waves’. If we proceed (with care) to the continuous case Ω = R then the pure copositives are just the pure frequency waves aω = eıω( ) . Our assertion that typical copositive elements give creation processes tending to the integral is then something familiar. For example, we can take the convex linear combination a=
1 ıω( ) 1 −ıω( ) e + e = cos ω( ) 2 2
(this just corresponds to Brownian motion for the dual random walk as in (7)). It is well-known that squaring such cosine waves introduces higher harmonics. Moreover, raising to higher and higher powers an makes the wave more and more impure as further harmonics are introduced until we have something resembling the integral Λ. This is the maximally impure wave containing all frequencies, i.e. the δ-function. This is a simple physical interpretation of (50) in the classical case.
28
If we do not worry about the copositivity condition, we can also give another picture of this general phenomenon in terms of membership of subsets as mentioned above. For this we consider a non-stationary creation process given by a family of subsets containing e (so that their characteristic functions a1 , a2 , · · · are correctly normalised). These indeed typically tend to {e} (corresponding to Λ) as the only point definitely in their joint intersection. This is a second ‘boolean’ interpretation of (50). We are now ready to consider a system (H, φ, a) consisting of a Hopf ∗-algebra, a state φ and a normalized copositive element a in H. The first element induces a random walk in H. The second induces a creation process in H. In terms of these elementary considerations, we can express our observable-state duality in this context as a ‘CTP’-type proposition. Proposition 4.4 Let (H, φ, a) be a Hopf ∗-algebra equipped with state φ (giving a random walk) and a normalised copositive element a (giving a creation process). Then (H ∗ , a, φ) has the same interpretation but with the algebra of observables given by H ∗ rather than H and the roles of a, φ interchanged. More generally, let (H, φ1 , φ2 , · · · , φn , a1 , a2 , · · · , am ) be a non-stationary random walk with successive distributions φ1 , φ2 , · · · for each step, and a non-stationary creation process successively creating the vectors a1 |0 >, a2 a1 |0 >, · · · . Then (H ∗ , am , am−1 , · · · , a1 , φn , φn−1 , · · · , φ1 ) has the same interpretation with the roles of H, H ∗ , a, φ interchanged and the order reversed, i.e. successively applying steps am , am−1 , · · · and creating vectors φn |ǫ >, φn−1 φn |ǫ >, · · ·. Here the vacuum vector |ǫ > in H ∗ corresponds to the counit ǫ. Proof
The first part summarises the discussion above. We saw in (48) that the random
walk on H ∗ , when referred to H is the left regular representation. Similarly with the roles of H, H ∗ reversed. We are of course expressing nothing other than the self-duality of the axioms of a Hopf ∗-algebra. For the second part we have to note carefully the order in (34). The operator T as a map T : H ∗ → Lin(H) is an anti-representation. Its adjoint therefore leads to a map L : H ∗ → Lin(H ∗ ) (the creation process on H ∗ ) which is a representation rather than 29
an antirepresentation (since taking adjoints reverses the order). Note that this reversal is not a feature of our conventions but a genuine aspect of the interpretation of Tφ as explained in Section 2.3 and of the (quantum) creation process as discussed above. ⊔ ⊓ Thus, the combined random-walk-creation process in H can be viewed equally well as a combined process in H ∗ , but with the time direction built into the processes on H ∗ reversed relative to the processes on H. Thus the symmetry between observables and states when the quantum algebra of observables is a Hopf algebra[12] also involves a time reversal. Another way to see this reversal is to return to Theorem 3.1. There we see also that the comultiplications of both H and H ∗ can be understood as conjugation by a joint evolution operator. For the former (which leads to the random walk in H) this is by embedding h as h ⊗ 1 and conjugation by W ( )W −1 . For the latter (which leads to a random walk in H ∗ ) it is by embedding φ as 1 ⊗ φ (hence with a left-right or parity reversal) and conjugation by W −1 ( )W (hence with the inverse joint evolution operator). We note that this time-reversal phenomenon associated with Hopf algebra duality is reminiscient of CTP-invariance in particle physics. It suggests a relationship between Hopf algebra duality and particle-antiparticle duality which will be developed further elsewhere.
References [1] L. Accardi, M. Sch¨ urmann, and M. von Waldenfels. Quantum independent increment processes on superalgebras. Math. Zeit., 198:451–477, 1988. [2] G.E. Andrews. q-series: Their development and application in analysis, number theory, combinatorics, physics and computer algebra. Technical Report 66, AMS, 1986. [3] S. Baaj and G. Skandalis. Unitaires multiplicatifs et dualite pour les produits croisse de C ∗ -algebres. Maths preprint, Paris, 1990. [4] P. Biane. Some properties of quantum Bernoulli random walks. In Quantum Probability and Related Topics VI (Proc. Trento, 1989), pages 193–203. World Sci. 30
[5] O. Bratteli and D. W. Robinson. Operator Algebras and Quantum Statistical Mechanics, volume II. Springer-Verlag, 1979. [6] A. Connes, H. Narnhofer, and W. Thirring. Dynamical entropy of C ∗ algebras and von Neumann algebras. Comm. Math. Phys., 112:691–719, 1987. [7] A. Van Daele and S. Van Keer. The Yang-Baxter and pentagon equation. Preprint, 1991. [8] M. Donald. On the relative entropy. Comm. Math. Phys., 105:13–34, 1986. [9] M. Enock and J-M. Schwartz. Une dualite dans les algebres de von Neumann. Bull. Soc. Math. Fr. Mem., 44:144, 1975. [10] B. K¨ ummerer. Survey on the theory of non-commutative stationary Markov processes. Preprint. [11] J.M. Lindsay and K.R. Parthasarathy. The passage from random walks to diffusion in quantum probability ii. Indian J. Stat. A, 50:151–170, 1988. [12] S. Majid. Hopf algebras for physics at the Planck scale. J. Classical and Quantum Gravity, 5:1587–1606, 1988. [13] S. Majid. Non-commutative-geometric Groups by a Bicrossproduct Construction. PhD thesis, Harvard mathematical physics, 1988. [14] S. Majid. Physics for algebraists: Non-commutative and non-cocommutative Hopf algebras by a bicrossproduct construction. J. Algebra, 130:17–64, 1990. [15] S. Majid. Quasitriangular Hopf algebras and Yang-Baxter equations. Int. J. Modern Physics A, 5(1):1–91, 1990. [16] S. Majid. Hopf-von Neumann algebra bicrossproducts, Kac algebra bicrossproducts, and the classical Yang-Baxter equations. J. Funct. Analysis, 95:291–319, 1991.
31
[17] S. Majid. Principle of representation-theoretic self-duality. Phys. Essays, 4(3):395–405, 1991. [18] K.R. Parthasarathy. Azema martingales and quantum stochastic calculus. In R.R. Bahadur, editor, Proc. R.C. Bose Memorial Symposium. To appear. [19] M. Sch¨ urmann. White noise on involutive bialgebras. In Quantum Probability and Related Topics VI (Proc. Trento, 1989), pages 401–419. World Sci. [20] M.E. Sweedler. Hopf Algebras. Benjamin, 1969. [21] E. Taft. Proc. Nat. Acad. Sci. (USA), 1971. [22] S.L. Woronowicz. Compact matrix pseudo groups. Comm. Math. Phys., 111:613–665, 1987.
32