Accessibility statement: these notes will eventually be WCAG2.1AA compliant, once I finish writing them. I tested it out using NVDA, a screen reader. It works best if you uncheck “graphic” and “clickable” in the settings. I am also testing out dyslexia–friendly fonts right now. You can also download the TeX source file.

1 About this webpage

These lecture notes were created with Ximera, an interactive textbook platform hosted by Ohio State University. The Ximera Project is funded 2024-2026 (with no other external funding) by a $2,125,000 Open Textbooks Pilot Program grant from the federal Department of Education.

I am still finding typos and mistakes, please so take this webpage as a draft for now. It is updated with all three lectures. The next update(s) will include appropriate references. To load the most updated version, click the orange “update” button at the top of the page. If it is not there, then you are reading the most up-to-date version. The button looks like this:

PIC

Recordings of the lectures are available on Oleg Zaboronski’s webpage.

Funding was provided by The London Mathematical Society and Lancaster University, University of Bristol, University of Edinburgh, Warwick University, Queen Mary University of London, and Ohio State University.

2 Introduction

Broadly speaking, integrable probability is a branch of probability theory which studies models which have exact solutions. For this reason, these models are often called exactly solvable. From these exact solutions, one can derive precise universal asymptotics, such as the famed Tracy–Widom distribution. The origin of these solutions are usually due to algebraic symmetries underlying the model. In these set of lecture notes, we will introduce the relevant algebraic background with a probabilistic researcher as the target audience.

For historical context, the terminology originates in integrable systems from Hamiltonian mechanics. In Hamiltonian mechanics, the phase space is represented as a smooth manifold with even dimension \(2n,\) with coordinates denoted \(q_1,\ldots ,q_n,p_1,\ldots ,p_n\) for the momentum and position. An integrable system is a system with \(n\) conserved quantities which are mutually Poisson commuting, and the Liouville–Arnold theorem states that the equations of motion can be solved in quadratures. For an explicit example, the harmonic oscillator in one dimension (imagine an object attached to a frictionless spring) is integrable, and the conserved quantity is the total energy. In contrast, the three–body problem is not integrable, and its solutions are notoriously non–exact.

During the 1980s, many Soviet mathematicians and physicsts introduced quantum mechanics into integrable systems. In quantum mechanics, quantities such as position, momentum and energy became operators which generally do not commute; additionally, the values of these quantities only take discrete, “quantized” values. In this context, the concept of “conserved quantities” becomes “commuting operators,” and values of the quantities are eigenvalues of eigenstates. At the time, their approaches were called the “quantum inverse scattering method.” Since then, the method has been generalized to abstract algebraic objects, such as Hopf algebras.

These notes will introduce one such algebraic object, known as quantum groups, with a particular focus on the Yang–Baxter equation. The exposition will use probability and mathematical physics as a motivation. The goal is for a reader with a probability background to be able to read contemporary (as of 2024) research papers in integrable probability.

3 ASEP, XXZ, and \(U_q(sl_2)\)

To begin to motivate the notes, we first introduce the asymmetric simple exclusion process (ASEP) and its relationship to the Heisenberg XXZ model and the quantum group \(U_q(sl_2).\)

3.1 Definition of ASEP

In ASEP, particles randomly jump on a lattice, which we assume to be one–dimensional. At most one particle may occupy a site, and jumps to occupied sites are blocked (hence the term “exclusion”). Jumps are nearest neighbour (hence the term “simple”). If the jumps are continuous–time exponential clocks with left rates \(\alpha \) and right rates \(\beta ,\) then let \(q=\sqrt {\beta /\alpha }\) denote the asymmetry parameter. If \(q\neq 1,\) then the model is asymmetric (sometimes partially asymmetric), while for \(q=1\) the model is symmetric. For \(q=0\) or \(q=\infty \) the model is called totally asymmetric. The symmetric exclusion process (without the word simple) can be defined more generally on an arbritrary graph. In princple, so can the asymmetric exclusion process, although this is not as well studied.

PIC

The generator of the simple exclusion process can be explicitly written. In the most elementary case where there are two lattice sites, then there are four possible configurations. Associate to each particle the vector \([1\ 0]\) and to each hole (i.e. a non–particle) the vector \([0 \ 1].\) Tensoring the vectors together, we can associate to each configuration a canonical basis element of the four–dimensional vector space \(\mathbb {C}\otimes \mathbb {C}.\) The field here is chosen to be the complex numbers because it is algebraically closed, although of course probabilities are real numbers.

PIC

With that set up, the generator is then a \(4\times 4\) matrix \[ \alpha \left ( \begin{array}{cccc} 0 & 0 & 0 & 0 \\ 0 & -1 & 1 & 0 \\ 0 & q^2 & -q^2 & 0 \\ 0 & 0 & 0 & 0 \end{array} \right ). \] The constant \(\alpha \) can be viewed as a time rescaling, so it can be removed without loss of generality.

In a more general setting, where there are \(N\) lattice sites, the generator can be defined from the above \(4\times 4\) matrix, which we now denote \(\mathcal {L}.\) The generator is now a \(2^N \otimes 2^N\) and acts on \((\mathbb {C}^2)^{\otimes N}.\) Define \(\mathcal {L}_{i,i+1}\) by \[ \mathcal {L}_{i,i+1} = (\mathrm {Id}_2)^{i-1} \otimes \mathcal {L} \otimes (\mathrm {Id}_2)^{N-1}. \] With this notation, the generator is \[ \sum _{i=1}^{N-1} \mathcal {L}_{i,i+1}. \] Similar notation using subscripts will be used throughout these notes.

3.2 ASEP and Yang–Baxter Equation

The corresponding notion of integrability in the quantum case comes from Yang–Baxter equation. It was shown by Baxter that when a matrix solves the Yang–Baxter equation, there are corresponding “commuting transfer matrices,” which are analogous to conserved quantities in the classical setting.

Using the same notation as in the previous section, we say that a matrix \(R\) solves the braided Yang–Baxter Equation if: \[ R_{12}R_{23}R_{12} = R_{23}R_{12}R_{23}. \] Somewhat confusingly, this equation is also sometimes called the braid equation or just the Yang–Baxter equation. An equivalent formulation is \[ \check {R}_{12}\check {R}_{13}\check {R}_{23} = \check {R}_{23}\check {R}_{13}\check {R}_{12} \] where \(\check {R} = P \circ R\) and \(P(x \otimes y) = y \otimes x\) permutes the tensor power (exercise left to the reader).

A visual representation of YBE is given by this image:

PIC

In some settings, the YBE occurs in the scattering of particles with quantum mechanics considerations, such as the QISM (quantum inverse scattering method.)

Given an arbitrary matrix \(R,\) one can check that it satisfies the YBE through direct computation. However, it maybe more helpful to consider some simple examples first.

If \(P: V \otimes V \rightarrow V \otimes V\) denotes the permutation operator which sends \(u \otimes v\) to \(v \otimes u,\) then the braided Yang–Baxter equation becomes an identity of transpositions: \[ (1\ 2) \circ (2 \ 3 ) \circ (1 \ 2 ) = (2 \ 3) \circ (1 \ 2) \circ (2 \ 3). \] This identity holds, since both sides equal the permutation \((1\ 3).\) An even more simple solution is the identity matrix.

A probabilist may remark that a permutation matrix and an identity matrix are both examples of a stochastic matrix, albeit somewhat trivial examples. It may then be natural to try to find more general stochastic matrices which solve the Yang–Baxter equation. The “simplest” generalization occurs when \(V\) is two–dimensional, with basis \(e_1,e_2.\) Using exclusion processes as a prototype, we can define the operator \(R_{\alpha \beta }\) by \[ R_{\alpha \beta }(e_1 \otimes e_1) = e_1 \otimes e_1, \quad R_{\alpha \beta }(e_2\otimes e_2) = e_2 \otimes e_2 \] and \begin{align*} R_{\alpha \beta }(e_1 \otimes e_2) &= (1-\alpha )e_1 \otimes e_2 + \alpha e_2 \otimes e_1,\\ R_{\alpha \beta }(e_2 \otimes e_1) &= \beta e_1\otimes e_2 + (1-\beta )e_2 \otimes e_1. \end{align*}

Inserting this matrix into the Yang–Baxter equation and evaluating at \(e_1 \otimes e_1 \otimes e_2,\) one finds that \(\alpha (1-\beta )=0\) is a necessary requirement for a solution to YBE. Evaluating \( R_{\alpha ,1}\) and \(R_{0,\beta }\) at \(e_1\otimes e_2 \otimes e_1\) and \(e_2 \otimes e_1 \otimes e_1\) yields no aditional equations. (Exercise left to the reader)

One could also evaluate YBE at the three vectors \(e_2 \otimes e_1 \otimes e_1,e_1\otimes e_2 \otimes e_1\) and \(e_1 \otimes e_1 \otimes e_2,\) but instead we use another symmetry of ASEP. This is called particle–hole involution. In words, in an ASEP with drift to the right, the holes evolve as an ASEP with drift to the left. Symbolically, we define an involution \(T\) of \(V\) which switches \(e_1\) and \(e_2.\) Then \[ T^{\otimes 2} R_{\alpha ,\beta }T^{\otimes 2} = R_{\beta ,\alpha }, \quad \quad T^{\otimes 2} R_{\alpha ,\beta } = R_{1-\alpha ,1-\beta }, \quad \quad R_{\alpha ,\beta }T^{\otimes 2} = R_{1-\beta ,1-\alpha }. \] A mildly helpful observation here is that \(T\) is stochastic, so its composition with other stochastic matrices is also stochastic. Using these identities, one immediately verifies YBE for \(R_{\alpha ,0}\) and \(R_{1,\beta }.\) Furthermore, these two solutions are related up to the particle–hole involution.

One notices that \(P\circ R_{\alpha ,0}-\mathrm {id}_4\) is the generator of ASEP at two lattice sites. It is then natural to ask whether or not addition by constants affects integrability. The answer is “no” (otherwise integrable probability would not exist). As some intuition, let us suppose that the generator has an explicit set of eigenvectors and eigenvalues. In general, an arbitrary Markov process will have complicated eigenvectors and eigenvalues, but for integrable models we expect “nice” eigenvectors and eigenvalues, from which probabilistic information can be extracted. Then addition by constants leaves the eigenvectors unaffects and only shifts the eigenvalues, leaving them just as “nice.” Thus, it is fair to say that ASEP is integrable, at least on two lattice sites.

At this juncture, it is reasonable to object that nothing has been said about the integrability of ASEP on arbitrarily many lattice sites. We will put this question on hold, as the next section’s discussion of the XXZ model will lay the foundation for answering this question.

3.3 XXZ model

Having established that ASEP is an integrable model at two lattice sites, it is natural to ask about other integrable models that look similar to ASEP, so that we can avoid re–inventing the wheel. Thus we now turn our attention to a related model, called the quantum Heisenberg model. These were introduced by Werner Heisenberg to incorporate quantum mechanics into magnetism. Each lattice site has a microscopic magnetic dipole, which can either be up or down. If there are \(N\) lattice sites, then the Hamiltonian acts on the \(2^N\)–dimensional vector space \((\mathbb {C}^2)^{\otimes N}.\) So, at the very least, the state space of the XXZ model is the same as the ASEP state space.

To define the model, first recall that Pauli spin matrices: \[ \sigma ^1 = \left (\begin{array}{cc} 0 & 1 \\ 1 & 0 \end{array}\right ), \quad \quad \sigma ^2 = \left ( \begin{array}{cc} 0 & -i \\ i & 0 \end{array}\right ), \quad \quad \sigma ^3 = \left ( \begin{array}{cc} 1 & 0 \\ 0 & -1 \end{array}\right ). \] The notation \(\sigma ^x,\sigma ^y,\sigma ^z\) are often used instead. Note that these are traceless and are Hermitian, the latter property being important in physics (since Hermitian matrices have real eigenvalues, and these eigenvalues correspond to physical quantities). Letting \[ \sigma ^a_j = (\mathrm {Id})^{\otimes j-1} \otimes \sigma ^a \otimes (\mathrm {Id})^{\otimes N-j} \] so that \(\sigma ^a_j\) acts on the \(j\)–th lattice site while fixing the others, the Hamiltonian is then \[ \frac {-1}{2}\sum _{j=1}^N \left ( J_x \sigma ^1_j \sigma ^1_{j+1} + J_y \sigma ^2_j \sigma ^2_{j+1} + J_z \sigma ^3_j\sigma ^3_{j+1} - h\sigma ^3_j\right ). \] where \(J_x,J_y,J_z\) are some coupling constants and \(h\) is the external field. We impose periodic boundary conditions, so that \(\sigma ^a_{N+1}=\sigma ^a_1.\) If \(J_x,J_y,J_z\) are distinct then one obtains the XYZ model. If \(J_x=J_y \neq J_z\) then this is the XXZ model; if \(J_x=J_y=J_z\) then it is the XXX model.

Each summand in the Hamiltonian can be written as a four by four matrix: \[ \left (\begin{array}{cccc} J_z + 2h & 0 & 0 & J_x-J_y \\ 0 & -J_z & J_x + J_y & 0 \\ 0 & J_x + J_y & -J_z & 0 \\ J_x - J_y & 0 & 0 & J_z-2h \end{array}\right ) \] If \(J_x=J_y,\) as assumed in the XXZ model, then the top–right and bottom–left entries are zero. In this case, the non–zero entries are at least the same as in the ASEP generator. Recalling the previous section’s heuristic that adding by constants does not change integrability, we can set \(J_z=1,h=0\) to obtain the matrix \[ \mathrm {Id}_4 + \left ( \begin{array}{cccc} 0 & 0 & 0 & 0 \\ 0 & -2 & 2J_x & 0 \\ 0 & 2J_x & -2 & 0\\ 0 & 0 & 0 & 0 \end{array}\right ). \]

At this point, it is worth noticing that the YBE is a statement about operators on vector spaces, which should not depend on the choice of bases on the vector spaces. In other words, conjugation (which corresponds to a change of basis) should not affect integrability. As it turns out, conjugation by \[ \left ( \begin{array}{cccc} 1 & 0 & 0 & 0 \\ 0 & 1 & \gamma -1 & 0 \\ 0 & 0 & \gamma & 0 \\ 0 & 0 & 0 & 1 \end{array} \right ) \] results in the ASEP generator on two sites. (Exercise left to the reader).

3.4 \(sl_2\) symmetry

Next, we introduce the \(sl_2\) symmetry of ASEP. This was first observed in the seminal paper of [1].

So far, we have determined that ASEP is integrable and is conjugate to the XXZ Hamiltonian (assuming that we are ignoring the question of arbitrary lattice sizes). As it turns out, the Pauli matrices form a basis of the real Lie algebra \(su_2\) corresponding to the real compact Lie group \(SU(2).\) Here, a real Lie group means that the underlying field is the real numbers, not that all the entries are real (consider the unit circle in the complex plane as a real Lie group, for example).

The complex variant of \(SU(2)\) is the (non–compact) Lie group \(SL(2,\mathbb {C})\) of traceless \(2 \times 2\) matrices with complex entries. For a variety of reasons, some of which answer historical, most mathematicians are introduce to the Lie algebra \(sl_2\) rather than \(su_2,\) so we will use that notation.

By definition, \(sl_2\) is the Lie algebra with basis denoted \(e,f,h\) and brackets \[ [e,f]=h, \quad [h,e]=2e, \quad [h,f]=-2f. \] Note that the explicit matrices \(E_{12},E_{21},E_{11}-E_{22}\) satisfy these relations, where \(E_{ij}\) denotes the matrix with a \(1\) at the \(i,j\)–entry and a 0 everywhere else.

Returning to the XYZ model, one can calculate that the XXX model has \(SU(2)\) symmetry (exercise to the reader) in the sense that \[ \sum _{j=1}^N \sigma ^a_j \] commutes with the Hamiltonian of the XXX model for \(a=x,y,z.\) However, for the XXZ model, the commutation only works for \(a=z.\) Therefore, whatever algebra is underlying the XXZ model (and hence that of ASEP), we know that it can not be as simple as \(SU(2).\) Futhermore, since XXZ and ASEP have a free parameter, the algebra should be a one–parameter deformation of \(SU(2).\) This leads into the next topic of quantum groups.

4 Quantum Group \(U_q(sl_2)\)

The Drinfel’d–Jimbo quantum groups are quantizations of the Lie groups. Rather confusingly, the quantum groups are not themselves groups.

Before diving into the rigorous definition, one notes where the quantization occurs. Using probabilistic intuition, we expect that the SSEP should have an underlying \(sl_2\) symmetry, whereas the ASEP should have the quantum group symmetry. However, SSEP and ASEP on a single lattice site are identical, so the quantization must require multiple lattice sites. We’ve established that in this case, the vector space is \((\mathbb {C}^{\otimes 2})^{\otimes N}.\) This raises the question: how does \(sl_2\) act on tensor powers?

To answer this question, we take an aside about Lie algebras, which are tangent spaces to Lie groups at the origin. What this implies is that for a Lie group \(G\) with a Lie algebra \(g,\) we have that \[ X \in g \text { implies } \exp (tX) \in G \text { for all } t\in \mathbb {R} \] and \[ X = \frac {d}{dt} \exp (tX) \Big |_{t=0}. \] Setting \(g=g_X(t)=\exp (tX),\) the natural action of a group on tensor products is \[ g(v \otimes w)= gv \otimes gw. \] Taking derivatives and heuristically using Leibniz’s rule for derivative of products, we obtain \[ X(v \otimes w) = Xv \otimes w + v \otimes Xw. \] This is expressed through the co–product \(\Delta : g \rightarrow g \otimes g\) defined by \[ \Delta (X) = 1 \otimes X + X \otimes 1. \]

Since it is the co–product that allows for the algebra to act on multiple lattice sites (i.e. tensor powers), the quantization occurs in the co–product, not the algebra itself! As it turns out, the right deformation is \[ \Delta (e) = q^h \otimes e + e \otimes 1, \quad \Delta (f) = 1 \otimes f + f \otimes q^{-h}, \quad \Delta (h)=h\otimes 1 + 1\otimes h, \] where we interpret \(q^h\) as a formal power series in \(h.\) Note that we do not deform the co–product of \(h,\) which is consistent with \(\sigma ^z\) commuting with the XXZ model.

Depending on the preferences of the author, the generator \(h\) may be replaced with \(k:=q^h\) with the co–product becoming \[ \Delta (k) = k \otimes k. \] (Exercise to the reader why that is the same co–product). In words, \(h\) is primitive and \(k\) is group–like.

4.1 Co-algebras and bi–algebras

While I presented the co–product as the object that allows for multiple lattice sites, that is of course not how it is presented to algebraists. A more common description is that a co–algebra is dual to an algebra, with the co–product dual to the product. This subsection will provide a rigorous definition of a co–algebras.

One motivation is that the space of functions on an algebra is a co–algebra. The Lie algebra is actually itself a space of differential operators on a Lie group (it is the direction derivative), so is a space of functions on the algebra of smooth functions on the Lie group. Thus, you would expect a Lie algebra to have a co–algebra structure.

Recall that an algebra \(A\) is a vector space with a multiplication operator \[ m: A \otimes A \rightarrow A \] which assumed to be bilinear. It is also associative in the sense that \[ m(x \otimes m(y \otimes z)) = m(m(x \otimes y) \otimes z), \] which essentially means that \( x(yz) = (xy)z.\) This is equivalently stated as \[ m \circ (m \otimes \mathrm {id}) = m \circ (\mathrm {id} \otimes m). \] Additionally, there is an identity element \(1 \in A\) satisfying \[ m(1 \otimes x) = m(x \otimes 1) = x. \] This can be written as an inclusion from the underlying field \(F\) to the algebra \(A.\)

Before defining co–algebras, it will be useful to draw certain commutative diagrams. The associativity can be written as [Picture]

If one reverses the arrow of this diagram, which is also fortunately easy on Tikz, one obtains

[Picture]

The condition of co–associativity means that this diagram commutes.

In a similar vein, the unit element can be understood as an inclusion from the base field to the algebra such that the following diagram commutes:

[Picture]

Reversing the direction of the arrows, one obtains

[Picture]

A co–algebra is a vector space with a co–product and co–unit satisfying the above properties. For \(U_q(\mathfrak {sl}_2)\) the co–unit is \[ \epsilon (e)=\epsilon (f)=0, \quad \epsilon (q^h)=1. \]

A bi–algebra is an algebra and co–algebra such that the co–product and co–unit are also algebra homomorphisms.

A Hopf algebra is a bi–algebra with an anti–homomorphism \(S\) such that this diagram commutes:

[Picture]

Finally, I get to \(R\)–matrices. A Hopf algebra \(H\) is co–commutative if the coproduct \(\Delta \) equals its reversal \(\Delta =:=P \circ \Delta \) where \(P: H \otimes H \rightarrow H \otimes H\) maps \(x\otimes y\) to \(y \otimes x.\) A Hopf algebra is almost co–commutative if there is an invertible element \(R \in H \otimes H\) such that \(R \Delta (x) = \Delta '(x) R\) for all \(x \in H.\)

An almost co–commtuative Hopf algebra is quasi–triangular if the additional relations hold for the element \(R\): \[ (\Delta \otimes \mathrm {id})(R) = R_{13}R_{23}, \quad (\mathrm {id} \otimes \Delta )(R) = R_{13}R_{12}. \] In a quasi–triangular Hopf algebra, the \(R\)–matrix satisfies the Yang–Baxter equation.

5 Explicit solutions to YBE

5.1 Affine Lie algebras

Having shown that solutions to Yang–Baxter Equation can come from Hopf algebras, we now introduce a spectral–dependent Yang–Baxter Equation that comes from affine Lie algebras.

Given a finite–dimensional Lie algebra \(\mathfrak {g}\), define its affinization by \[ \hat {\mathfrak {g}} = \mathfrak {g} \otimes _{\mathbb {C}} \oplus \mathbb {C}[t,t^{-1}] \oplus \mathbb {C}c \] where \(c\) is a central element. In words, this is a one–dimensional central extension of the loop algebra over \(\mathfrak {g}.\) The Lie bracket is then defined by \[ [a\otimes t^n, b\otimes t^m] = [a,b] \otimes t^{n+m} + (a,b)n\delta _{m,-n}c \] where \((\cdot ,\cdot )\) is the Cartan–Killing form on \(\mathfrak {g}.\) This Lie algebra is now infinite–dimensional, which makes its representation theory tricker.

There is an analogous quantum group \(\mathcal {U}_q(\mathfrak {g}),\) which has an \(R\)–matrix satisfying Yang–Baxter equation. For modules where \(c\) acts as \(0\) (level \(0\) representations), we can define an evaluation module by takng a module of \(\mathfrak {g}\) and then letting \(t\) act as multiplication by \(z\). If the element \(R\) now acts on evaluation modules, then it depends on a parameter which we call the spectral parameter. The Yang–Baxter equation now reads \[ \check {R}_{12}(z) \check {R}_{13}(zw) \check {R}_{23}(w) = \check {R}_{23}(w) \check {R}_{13}(zw) \check {R}_{12}(z). \] If we attempt to find \(4\times 4\) solutions to the spectral–dependent YBE analogous to ASEP, we get a two–parameter family of solutions, which end up being the weights of the stochastic six vertex model.

5.2 Fusion

In general, we do not really want to brute force solutions to the Yang–Baxter equation in arbitrary representations. Here, I present a way to construct new solutions from old solutions.

In linear algebra, for any vector space \(V\), a symmetric tensor is an element of \(V^{\otimes N}\) preserved by any \(P_{i,i+1}.\) For quantum groups, however, we have a \(q\)–deformation in the tensor powers so that \(P\) and \(\Delta \) no longer commute. However, there is still a symmetric tensor power \[ \mathrm {Sym}^N V \subseteq V^{\otimes N}. \]

In lieu of an extensive algebraic discussion, we present it in probabilistic language. Recall Pitman–Rogers intertwining: suppose there are two Markov kernels \(\Lambda \) and \(\Phi \) such that \(\Lambda \Phi =\mathrm {Id}\). Let \(Q=\Lambda P\Phi \) and assume \(\Lambda P = Q\Lambda .\) Then \(\Phi \) defends a Markov projection from \(P\) to \(Q\):

PIC

More specifically, if \(X(t)\) is a Markov process on state space \(S\) and \(\Phi :S\rightarrow \hat {S}\) is deterministic, with a Markov kernel \(\Lambda : \hat {S}\rightarrow S,\) then \(\Phi (X(t))\) is a Markov process on \(\hat {S}.\)

This procedure is sometimes called fusion.

This intertwining can be constructed algebraically, but instead we draw a metaphor with stationary measures. We want to show that \[ \Lambda P = \Lambda P \Phi \Lambda . \] If for all \(v\), we have that \(v\Lambda \) is stationary under \(P\), i.e. \(v\Lambda P = v\Lambda \), then the above intertwining becomes \[ v\Lambda = v\Lambda \Phi \Lambda \] which then follows from \(\Lambda \Phi =\mathrm {Id}.\)

Calculating the fused vertex weights then requires orthogonal polynomials (please reference the Appendix).

Question: do these weights satisfy YBE? In fact, they might not! At least for some Lie algebras. But they seem to for type \(A.\)

5.3 Construction with Central Elements

From here, let \(k=q^h.\)

We define some subalgebras by letting \(U^{> 0}\) be generated by \(e,\) and \(U^{<0}\) generated by \(f\) and \(U^0\) generated by \(k\). Analogously, \(U^{\geq 0}\) is generated by \(e,k\) and \(U^{\leq 0}\) is generated by \(f,k\). There are central elements, sometimes called Casimir elements, that have the form \[ C \in U^0 + U^{< 0}U^{> 0}. \]

One reason for introducing these subalgebras is because of the highest weight representations. We call \(V\) a highest weight representation if there exists a vector \(v,\) called the highest weight vector, such that \[ U^{>0}v=0 \] and \(V\) is generated by \(U^{\leq 0}v.\) Then the action of the Casimir on the highest weight vector is given by \[ Cv = av \] for some complex number \(a.\) Also let \(\{v_{\mu }:\mu \in \mathcal {B}\}\) denote a basis of \(V,\) where each \(v_{\mu }\) can be expressed as \(F^{\mu }v\) for some \(F^{\mu } \in U^{\leq 0}.\)

We define the Hamiltonian \(H\) by taking the co–product of \(C\) on tensor powers of representations: \[ H = \Delta (C) \vert _{V \otimes V}. \] We want to find a diagonal matrix \(G\) such that \(G^{-1}HG-c\mathrm {Id}\) is the generator of a continuous–time Markov process. As a necessary condition, we need that \(G^{-1}HG\) has rows which sum to \(c\); non–negativity is more difficult to ensure. One way to achieve this is with an eigenvector of eigenvalue \(c\). Namely, assume that \(g\) is a vector such that \(Hg=cg.\) If \(g\) has the form \[ g=\sum _{\boldsymbol {\mu }=(\mu _1,\mu _2)\in \mathcal {B}\times \mathcal {B}} g(\boldsymbol {\mu }) v_{\mu _1}\otimes v_{\mu _2} \] then the diagonal matrix \(G(\boldsymbol {\mu },\boldsymbol {\mu })=g(\boldsymbol {\mu })\) will satisfy the necessary requirement, as long as \(g(\boldsymbol {\mu })\) is always nonzero.

As it turns out, \(g=v\otimes v\) (the tensor power of the highest weight vector) will satsify \[ Hg=a^2g \] due to the form of the co–product and the central element. However, this choice will have zero entries. So instead we take \(g\) of the form \[ \sum _{\boldsymbol {\mu }=(\mu _1,\mu _2)} c(\mu _1,\mu _2)\Delta (F^{\mu _1}F^{\mu _2}) v\otimes v \] for some appropriate choice of coefficients \(c(\mu _1,\mu _2)\) (reference the appendix for more detailed calculations).

6 Appendix

This section will cover topics that were not included in the in–person lectures, but are still worth mentioning.

I may make this a separate webpage later.

6.1 RTT Construction

6.2 Orthogonal polynomial vertex weights

6.3 Hecke algebras

References

[1]   G. Schütz, Duality relations for asymmetric exclusion processes, J Stat Phys 86, 1265–1287 (1997).

2024-11-20 17:16:41