This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

The Pentagram Integrals on Inscribed Polygons

Richard Evan Schwartz    and Serge Tabachnikov  Supported by N.S.F. Research Grant DMS-0072607

1 Introduction

The pentagram map is a geometric iteration defined on polygons. This map is defined in practically any field, but it is most easily described for polygons in the real projective plane. Geometrically, the pentagram map carries the polygon PP to the polygon QQ, as shown in Figure 1.

[Uncaptioned image]

Figure 1: The pentagram map

This pentagram map is always defined for convex polygons, and generically defined for all polygons. The pentagram map commutes with projective transformations and induces a generically defined map TT on the space 𝒬n{\cal Q\/}_{n} of cyclically labelled 111Technically, one needs to consider the square of the map in order to get a canonically defined map on labelled nn-gons. However, if one is willing to break symmety, preferring (say) left over right, then the map itself is defined on labelled nn-gons. nn-gons modulo projective transformations. TT is the identity map on 𝒬5{\cal Q\/}_{5} and has period 22 on 𝒬6{\cal Q\/}_{6}. The map TT is not periodic on 𝒬n{\cal Q\/}_{n} for n7n\geq 7.

The pentagram map also defines a map on a slightly larger space 𝒫n{\cal P\/}_{n}. The space 𝒫n{\cal P\/}_{n} is the space of twisted nn-gons modulo projective transformations. A twisted nn-gon is a map ϕ:𝑹𝑹𝑷2\phi:\mbox{\boldmath{$R$}}\to\mbox{\boldmath{$R$}}\mbox{\boldmath{$P$}}^{2}, such that

ϕ(k+n)=Mϕ(k)k.\phi(k+n)=M\circ\phi(k)\hskip 30.0pt\forall k.

Here MM is a real projective transformation, called the monodromy of the twisted polygon. The space 𝒫n{\cal P\/}_{n} contains the space 𝒬n{\cal Q\/}_{n} as a codimension 88 submanifold.

This pentagram map is studied for pentagons in [M]. We studied the general case in [S1], [S2], and [S3]. Finally, in [OST1] (see [OST2] for a short version) we showed that the pentagram map is a completely integrable system on 𝒫n{\cal P\/}_{n}. We think that the pentagram map is a completely integrable system on 𝒬n{\cal Q\/}_{n} as well, but there are certain technical details we need to resolve before establishing this fact.

In [S3] we constructed polynomial functions

O1,,O[n/2],On,E1,,E[n/2],En:𝒫n𝑹O_{1},...,O_{[n/2]},O_{n},E_{1},...,E_{[n/2]},E_{n}:{\cal P\/}_{n}\to\mbox{\boldmath{$R$}}

which are invariant under the pentagram map. In [OST1] these functions are recognized a complete list of integrals for the completely integrable system. There is an invariant Poisson structure, and the above functions Poisson-commute with respect to this Poisson structure. We call these functions the monodromy invariants, and we will construct them in the next section.

The purpose of this paper is to prove the following result.

Theorem 1.1

Ok(P)=Ek(P)O_{k}(P)=E_{k}(P) for any twisted nn-gon PP that is inscribed in a conic section and any k=1,,[n/2],nk=1,...,[n/2],n.

The result, in particular, applies to closed inscribed polygons. Theorem 1.1 boils down to a countably infinite family of polynomial identities. The polynomials involved are somewhat reminiscent of the symmetric polynomials, but they have somewhat less symmetry then these. We noticed this result by doing computer experiments. One novel feature of the theorem is that we discovered not just the result but also the proof by way of computer experimentation. We wrote a Java applet to aid us with the combinatorics of the proof. This applet is available on the first author’s website. 222http://www.math.brown.edu/\simres

While our proof is mainly combinatorial, we think that perhaps there should be a better proof based on geometry. Accordingly, we will describe our polynomials in three ways – geometrically, combinatorially, and in terms of determinants. We will only use the combinatorial description in our proof, but we hope that the other descriptions might ring a bell for some readers.

The pentagram map seems to interact nicely with polygons that are inscribed in conic sections. We mention our paper [ST], in which we describe some finite configuration theorems, à la Pappus, that we discovered in connection with the pentagram map and inscribed polygons.

Here is an overview of the paper. In §2 we define the monodromy invariants. In §3 we reduce Theorem 1.1 to a combinatorial problem. In §4 we solve this combinatorial problem.

The second author would like to thank Brown University for its hospitality during his sabbatical; he was also supported by NSF grant DMS-0072607.

2 The Monodromy Invariants

The papers [S3] and [OST1] give a good account of the monodromy invariants. The notation in the two papers is slightly different, and we will follow [OST1]. We will give three descriptions of these invariants, one geometrical, one combinatorial, and one based on determinants. The reader only interested in the proof of Theorem 1.1 need only pay attention to the combinatorial definition. As we say in the introduction, we mention the other definitions just in the hopes that it will ring a bell for some reader.

Because we are not using the geometric or determinental definitions in our proof, we will not include the arguments that show the equivalence of the various definitions. The paper [S3] has a proof that the geometric and combinatorial definitions coincide.

2.1 The Geometric Definition

First of all, we have the cross ratio

[t1,t2,t3,t4]=(t1t2)(t3t4)(t1t3)(t2t4).[t_{1},t_{2},t_{3},t_{4}]=\frac{(t_{1}-t_{2})(t_{3}-t_{4})}{(t_{1}-t_{3})(t_{2}-t_{4})}.

Suppose that ϕ\phi is a twisted nn-gon, with monodromy MM. We let vi=ϕ(i)v_{i}=\phi(i). The label ii in Figure 2 denotes viv_{i}, and similarly for the other labels.

[Uncaptioned image]

Figure 2: vertex labels

We associate to each vertex viv_{i} two numbers:

xi=[vi2,vi1,((vi2,vi1)(vi,vi+1),((vi2,vi1)(vi+1,vi+2)],x_{i}=[v_{i-2},v_{i-1},((v_{i-2},v_{i-1})\cap(v_{i},v_{i+1}),((v_{i-2},v_{i-1})\cap(v_{i+1},v_{i+2})],
yi=[vi+2,vi+1,((vi+2,vi+1)(vi,vi1),((vi+2,vi+1)(vi1,vi2)].y_{i}=[v_{i+2},v_{i+1},((v_{i+2},v_{i+1})\cap(v_{i},v_{i-1}),((v_{i+2},v_{i+1})\cap(v_{i-1},v_{i-2})].

Here (a,b)(a,b) denotes the line determined by points aa and bb. For instance, xix_{i} is the cross ratio of the 44 white points in Figure 2. We call the invariants x1,y1,x2,y2,x_{1},y_{1},x_{2},y_{2},... the corner invariants. These invariants form a periodic sequence of length 2n2n. That is xk+n=xkx_{k+n}=x_{k} and yk+n=yky_{k+n}=y_{k} for all kk.

We define

On=i=1nxi;En=i=1nyi.O_{n}=\prod_{i=1}^{n}x_{i};\hskip 30.0ptE_{n}=\prod_{i=1}^{n}y_{i}.

The other monodromy invariants are best defined in an indirect way. Recall that MM is the monodromy of our twisted polygon ϕ\phi. We lift MM to an element of GL3(𝑹)GL_{3}(\mbox{\boldmath{$R$}}) which we also denote by MM. We define

Ω1=trace3(M)det(M);Ω2=trace3(M1)det(M1).\Omega_{1}=\frac{{\rm trace\/}^{3}(M)}{{\rm det\/}(M)};\hskip 30.0pt\Omega_{2}=\frac{{\rm trace\/}^{3}(M^{-1})}{{\rm det\/}(M^{-1})}.

These quantities are independent of the lift of MM and only depend on the conjugacy class of MM. Finally, these quantities are invariant under the pentagram map.

We define

Ω~1=On2EnΩ1;Ω~2=OnEn2Ω2.\widetilde{\Omega}_{1}=O_{n}^{2}E_{n}\Omega_{1};\hskip 30.0pt\widetilde{\Omega}_{2}=O_{n}E_{n}^{2}\Omega_{2}.

It turns out that these quantities are polynomials in the corner invariants. The remaining monodromy invariants are suitably weighted homogeneous parts of these polynomials.

We have a basic rescaling operation

Rt(x1,y1,x2,y2,)=(tx1,t1y1,tx2,t1y2,).R_{t}(x_{1},y_{1},x_{2},y_{2},...)=(tx_{1},t^{-1}y_{1},tx_{2},t^{-1}y_{2},...).

We say that a polynomial in the corner invariants has weight kk if

Rt(P)=tkP.R_{t}^{*}(P)=t^{k}P.

Here RtR_{t}^{*} denotes the obvious action of RtR_{t} on polynomials. In [S3] we show that

Ω~1=k=1[n/2]Ok;Ω~2=k=1[n/2]Ek.\widetilde{\Omega}_{1}=\sum_{k=1}^{[n/2]}O_{k};\hskip 30.0pt\widetilde{\Omega}_{2}=\sum_{k=1}^{[n/2]}E_{k}.

Here OkO_{k} and EkE_{k} are the weight kk polynomials in each sum, and [n/2][n/2] denotes the floor of n/2n/2.

2.2 The Combinatorial Definition

Now we describe the combinatorial formulas for these invariants. Again, the paper [S3] has a proof that the description below coincides with the one in the previous section. In everything we say, the indices are taken cyclically, mod nn.

We introduce the monomials

Xi:=xiyixi+1.X_{i}:=x_{i}\,y_{i}\,x_{i+1}. (1)

The monodromy polynomial OkO_{k} is built out of the monomials xix_{i} for i=1,,ni=1,...,n and XjX_{j} for j=1,..,nj=1,..,n. We call two monomials consecutive if they involve consecutive or coinciding variables xix_{i}. More explicitly, we have the following criteria:

  1. 1.

    XiX_{i} and XjX_{j} are consecutive if j{i2,i1,i,i+1,i+2};j\in\left\{i-2,\,i-1,\,i,\,i+1,\,i+2\right\};

  2. 2.

    XiX_{i} and xjx_{j} are consecutive if j{i1,i,i+1,i+2};j\in\left\{i-1,\,i,\,i+1,\,i+2\right\};

  3. 3.

    xix_{i} and xjx_{j} are consecutive if j{i1,i,i+1}j\in\{i-1,i,i+1\}.

Let OO be a monomial obtained by the product of the monomials XiX_{i} and xjx_{j}, that is,

O=Xi1Xisxj1xjt.O=X_{i_{1}}\cdots{}X_{i_{s}}\,x_{j_{1}}\cdots{}x_{j_{t}}.

Such a monomial is called admissible if no two of the monomials are consecutive. For every admissible monomial, define the weight |O|=s+t|O|=s+t and the sign sign(O)=(1)s\mathrm{sign}(O)=(-1)^{s}. One then has

Ok=|O|=ksign(O)O;k{1,2,,[n2]}.O_{k}=\sum_{|O|=k}\mathrm{sign}(O)\,O;\hskip 30.0ptk\in\left\{1,2,\ldots,\left[\frac{n}{2}\right]\right\}.

For example, if n5n\geq 5 we obtain the following two polynomials:

O1=i=1n(xixiyixi+1),O2=i=1n(xixi+2xiyixi+1xi+3),O_{1}=\sum_{i=1}^{n}\left(x_{i}-x_{i}\,y_{i}\,x_{i+1}\right),\qquad O_{2}=\sum_{i=1}^{n}\left(x_{i}\,x_{i+2}-x_{i}\,y_{i}\,x_{i+1}\,x_{i+3}\right),

The same formula works for EkE_{k}, if we make all the same definitions with xx and yy interchanged. More precisely, one builds the polynomials EkE_{k} from the monomials yiy_{i} and Yj:=yj1xjyjY_{j}:=y_{j-1}x_{j}y_{j} with the same restriction that no consecutive monomials are allowed and the same definitions of the weight and sign.

We note that the dihedral symmetry

σ(xi)=yi,σ(yi)=xi\sigma(x_{i})=y_{-i},\ \sigma(y_{i})=x_{-i} (2)

interchanges the polynomials OkO_{k} and EkE_{k}.

2.3 The Determinantal Definition

Now we describe determinantal formulas for the monodromy invariants; these formulas did not appear in our previous papers on the subject.

For positive integers k>lk>l, we define the four-diagonal determinant

Flk=|1xkXk100011xk1Xk200011xk2Xk20011xl+10011|,F_{l}^{k}=\left|\begin{array}[]{ccccccc}1&x_{k}&X_{k-1}&0&0&\dots&0\\ 1&1&x_{k-1}&X_{k-2}&0&\dots&0\\ 0&1&1&x_{k-2}&X_{k-2}&\dots&0\\ \dots&\dots&\dots&\dots&\dots&\dots&\dots\\ 0&\dots&\dots&\dots&1&1&x_{l+1}\\ 0&\dots&\dots&\dots&0&1&1\end{array}\right|,

where xix_{i} and yiy_{i} are the corner invariants and XiX_{i} is as in (1). By convention,

Fk+2k=0,Fk+1k=1,Fkk=1.F_{k+2}^{k}=0,\ F_{k+1}^{k}=1,\ F_{k}^{k}=1.

Then one has the following formula for the monodromy invariants OkO_{k}:

i=0[n/2]Oi=F1n+F0n1F1n1+xnynx1F2n1.\sum_{i=0}^{[n/2]}O_{i}=F_{1}^{n}+F_{0}^{n-1}-F_{1}^{n-1}+x_{n}y_{n}x_{1}F_{2}^{n-1}. (3)

Similarly, for EkE_{k}, define

Gpq=|1yp+1Yp+200011yp+2Yp+300011yp+3Yp+40011yq0011|.G_{p}^{q}=\left|\begin{array}[]{ccccccc}1&y_{p+1}&Y_{p+2}&0&0&\dots&0\\ 1&1&y_{p+2}&Y_{p+3}&0&\dots&0\\ 0&1&1&y_{p+3}&Y_{p+4}&\dots&0\\ \dots&\dots&\dots&\dots&\dots&\dots&\dots\\ 0&\dots&\dots&\dots&1&1&y_{q}\\ 0&\dots&\dots&\dots&0&1&1\end{array}\right|.

Then one has:

i=0[n/2]Ei=G0n1+G1nG1n1+ynx1y1G1n2.\sum_{i=0}^{[n/2]}E_{i}=G_{0}^{n-1}+G_{1}^{n}-G_{1}^{n-1}+y_{n}x_{1}y_{1}G_{1}^{n-2}. (4)

Formulas (3) and (4) simplify if one considers an open version of the monodromy invariants: instead of having a periodic “boundary condition” xi+n=xi,yi+n=yix_{i+n}=x_{i},y_{i+n}=y_{i}, assume that xi=0x_{i}=0 for i0i\leq 0 and in+1i\geq n+1, and yi=0y_{i}=0 for i1i\leq-1 and ini\geq n. With this “vanishing at infinity” boundary conditions, the monodromy invariants are given by a single determinant:

i=0[n/2]Oi=F0n,i=0[n/2]Ei=G1n1.\sum_{i=0}^{[n/2]}O_{i}=F_{0}^{n},\quad\sum_{i=0}^{[n/2]}E_{i}=G_{-1}^{n-1}.

We shall encounter an open version of monodromy invariants in Section 4.3.

3 Reduction to the Puzzle

A non-degenerate conic in 𝑹𝑷2\mbox{\boldmath{$R$}}\mbox{\boldmath{$P$}}^{2} can be identified with 𝑹𝑷1\mbox{\boldmath{$R$}}\mbox{\boldmath{$P$}}^{1} by way of the stereographic projection from a point of the conic. This identification is unique up to a projective transformation of the real projective line. That is, a different choice of the center of projection amounts to a projective transformation of 𝑹𝑷1\mbox{\boldmath{$R$}}\mbox{\boldmath{$P$}}^{1}.

Recall from the introduction that a twisted nn-gon is a map ϕ:𝑹𝑹𝑷2\phi:\mbox{\boldmath{$R$}}\to\mbox{\boldmath{$R$}}\mbox{\boldmath{$P$}}^{2}, such that

ϕ(k+n)=Mϕ(k)k.\phi(k+n)=M\circ\phi(k)\hskip 30.0pt\forall k.

Here MM is a real projective transformation, called the monodromy of the twisted polygon. Given a twisted nn-gon, we let vj=ϕ(j)v_{j}=\phi(j). These are the vertices of the twisted nn-gon.

If (,v1,v0,v1,)(\dots,v_{-1},v_{0},v_{1},\dots) is an inscribed twisted polygon, we can consider the vertices viv_{i} as points of the real projective line and talk about their cross-ratios which are uniquely defined. Referring to the cross-ratio on the projective line, we set:

pi=1[vi2,vi1,vi,vi+1].p_{i}=1-[v_{i-2},v_{i-1},v_{i},v_{i+1}].

In the next lemma we express the corner invariants of an inscribed polygon in terms of the quantities pip_{i}. Once we specify the cross ratios {pi}\{p_{i}\}, we produce an inscribed twisted polygon having corner invariants {(xi,yi)}\{(x_{i},y_{i})\}. Thus we have a map (pi)(xi,yi)(p_{i})\mapsto(x_{i},y_{i}). We denote this map by FF.

Lemma 3.1

One has:

xi=[vi2,vi1,vi,vi+2],yi=[vi2,vi,vi+1,vi+2]x_{i}=[v_{i-2},v_{i-1},v_{i},v_{i+2}],\quad y_{i}=[v_{i-2},v_{i},v_{i+1},v_{i+2}]

and the map FF is given by the formula

xi=1pipi+1,yi=1pi+1pi.x_{i}=\frac{1-p_{i}}{p_{i+1}},\quad y_{i}=\frac{1-p_{i+1}}{p_{i}}. (5)

Proof: Consider Figure 3. Using the projection from point vi+1v_{i+1}, we obtain:

xi=[vi2,vi1,A,B]=[(vi+1vi2),(vi+1vi1),(vi+1A),(vi+1B)]=[vi2,vi1,vi,vi+2].x_{i}=[v_{i-2},v_{i-1},A,B]=[(v_{i+1}v_{i-2}),(v_{i+1}v_{i-1}),(v_{i+1}A),(v_{i+1}B)]=[v_{i-2},v_{i-1},v_{i},v_{i+2}].

A similar projection from vi1v_{i-1} yields the formula for yiy_{i}. The expression for xix_{i} in terms of pip_{i} and pi+1p_{i+1} follows now from the identity

[vi2,vi1,vi,vi+2]=[vi2,vi1,vi,vi+1]1[vi1,vi,vi+1,vi+2],[v_{i-2},v_{i-1},v_{i},v_{i+2}]=\frac{[v_{i-2},v_{i-1},v_{i},v_{i+1}]}{1-[v_{i-1},v_{i},v_{i+1},v_{i+2}]},

and likewise for yiy_{i}. \spadesuit

[Uncaptioned image]

Figure 3: Proof of Lemma 3.1

The dihedral group also acts on cyclic sequences (pi)(p_{i}) by cyclic permutations and the orientation-reversing involution σ(pi)=p1i\sigma^{\prime}(p_{i})=p_{1-i}. It follows from (5) that Fσ=σFF\circ\sigma^{\prime}=\sigma\circ F where σ\sigma is the involution (2). Hence FF is a dihedrally equivariant map.

After xix_{i} and yiy_{i} are replaced by pip_{i} via (5), the polynomials OkO_{k} and EkE_{k} become Laurent polynomials in the variables pip_{i}. The identity En=OnE_{n}=O_{n} obviously holds since both sides equal (1pi)/pi\prod(1-p_{i})/p_{i}. We need to prove that Ek=OkE_{k}=O_{k} for k=1,,[n/2]k=1,...,[n/2]. The strategy is to show that every monomial in the variables pip_{i} appears in EkE_{k} and OkO_{k} with the same coefficient. Since the map FF is dihedrally equivariant and the orientation-reversing involution on the variables (xi,yi)(x_{i},y_{i}) interchanges EkE_{k} and OkO_{k}, it suffices to show that if two monomials in the variables pip_{i} are related by orientation-reversing symmetry (for example, the involution σ\sigma^{\prime}) then they appear in OkO_{k} with the same coefficients.

Let us compute the monomials in the variables pp’s that appear in the polynomial OkO_{k}. Using (5), one finds:

xi=1pi+1pipi+1x_{i}=\frac{1}{p_{i+1}}-\frac{p_{i}}{p_{i+1}} (6)

and

Xi=1pipi+1pi+2+1pi+1pi+2+21pipi+221pi+2pi+1pipi+2+pi+1pi+2.-X_{i}=-\frac{1}{p_{i}p_{i+1}p_{i+2}}+\frac{1}{p_{i+1}p_{i+2}}+2\frac{1}{p_{i}p_{i+2}}-2\frac{1}{p_{i+2}}-\frac{p_{i+1}}{p_{i}p_{i+2}}+\frac{p_{i+1}}{p_{i+2}}. (7)

We see that the variables pp’s that appear in xix_{i} involve two indices, ii and i+1i+1, and the those in XiX_{i} involve three indices, i,i+1i,i+1 and i+2i+2. Pictorially, these terms can be represented as follows: for xix_{i}, see Figure 4, and for XiX_{i}, see Figure 5. In these figures, the presence of each term pp in the numerator and denominator is represented by a shaded square, and its absence by a white square.

[Uncaptioned image]

Figure 4: pictorial representation of (6)

[Uncaptioned image]

Figure 5: pictorial representation of (7)

According to Section 2.2, the polynomial OkO_{k} is the sum of all admissible products of kk terms, and each term is either xix_{i} or (Xj)(-X_{j}). The admissibility condition is that the monomials xix_{i} or XjX_{j} involved are sufficiently distant; what it means precisely is that the respective tiles in Figures 4 and 5, corresponding to these terms, do not overlap. This is a crucial observation.

To summarize, each monomial in OkO_{k}, after the substitutions (6) and (7), is represented by a collection of kk tiles depicted in Figures 4 and 5, taken with the product of their respective coefficients. The tiles, that occupy two or three consecutive positions, are placed around a circle having NN positions available (if we are concerned with twisted NN-gons). There may be empty positions left between the tiles.

As a final preparation for the next section, we introduce the following notation: letter AA denotes a shaded square in the lower row of a tile (pip_{i} in denominator), letter BB a shaded square in the upper row (pip_{i} in numerator), and letter XX an empty column in a tile. Thus, the two tiles in Figure 4 correspond to the words XAXA and BABA, and the six tiles in Figure 5 to AAA,XAA,AXA,XXA,ABAAAA,XAA,AXA,XXA,ABA and XBAXBA, respectively. We also use letter XX to mean an empty slot between the tiles.

In the next section we reformulate Theorem 1.1 as a statement about a kind of puzzle involving words in letters {A,X,B}\{A,X,B\}. After we formulate the combinatorial statement, we prove it. The combinatorial result we prove implies Theorem 1.1.

4 The Puzzle

4.1 The Main Result

Now we are going to extract the main combinatorial information from the discussion at the end of the last section.

We fix some integer N>0N>0 and consider the set of length NN lists in the letters {A,X,B}\{A,X,B\}. We consider two lists the same if they are cyclic permutations of each other. We say that a cyclic sentence is an equivalence class of such strings. To illustrate our notation by way of example, (AXABA)(AXABA) denotes the equivalence class of AXABAAXABA. Here N=5N=5. We let 𝒮\cal S denote the set of such sentences, with the dependence on NN understood.

We single out certain strings of letters, and to each of these special words we assign a coefficient and a weight. Here is the list.

wordcoefficientweightX10XA11XAA11XBA11AXA21AAA11BA11ABA11XXA21\matrix{{\rm word\/}&{\rm coefficient\/}&{\rm weight\/}\cr X&1&0\cr XA&1&1\cr XAA&1&1\cr XBA&1&1\cr AXA&2&1\cr AAA&-1&1\cr BA&-1&1\cr ABA&-1&1\cr XXA&-2&1\cr}

We say that a parsing of a cyclic sentence is a description of the cyclic sentence as a concatenation of words. For example

(ABA/XA);(XAA/BA)(ABA/XA);\hskip 30.0pt(XAA/BA)

are the only two parsings of (ABAXA)(ABAXA). We define the coefficient of a parsing to be the product of the coefficients of the word. We define the weight of the parsing to be the sum of the weights of the words. Both parsings above have coefficient 1-1 and weight 22.

For each cyclic sentence SS, we define c(S,w)c(S,w) to be the total sum of the coefficients of the weight ww parsings of SS. For instance, when S=(ABAXA)S=(ABAXA), we have c(S,2)=2c(S,2)=-2 and otherwise c(S,w)=0c(S,w)=0. For a more streamlined equation, we define

|S|=w=0c(S,w)tw.|S|=\sum_{w=0}^{\infty}c(S,w)t^{w}.

Here |S||S| is a polynomial in tt that encodes all the coefficients at once. For example

|(ABAXA)|=2t2.|(ABAXA)|=-2t^{2}.

Let S¯\overline{S} denote the cyclic sentence obtained by reversing SS. In view of Section 3, Theorem 1.1 is equivalent to the following result.

Theorem 4.1 (Cyclic)

We have |S|=|S¯||S|=|\overline{S}| for all cyclic sentences SS.

In light of the work in the previous section, Theorem 4.1 implies Theorem 1.1. The rest of this chapter is devoted to proving Theorem 4.1.

4.2 The Tight Puzzle

Before we tackle Theorem 4.1, we slightly modify our puzzle for the sake of convenience.

Lemma 4.2

Suppose that (W)(W) contains the string XBXB. Then |W|=0|W|=0.

Proof: No word ends in BB, and so the string XBXB must continue as XBAXBA. The occurence of /X/BA/.../X/BA/... in any parsing contributes weight 11 and coefficient 1-1 whereas the occurence of /XBA/.../XBA/... contributes weight 11 and coefficient 11. If we have a parsing that involves /XBA/.../XBA/... we can create a new parsing by replacing the last /XBA/.../XBA/... with /X/BA/.../X/BA/.... These two parsings have the same weight and opposite coefficient, and thereby cancel each other in the total sum. \spadesuit

By Lemma 4.2, we can simply throw out any strings that contain XBXB, and we may drop the word XBAXBA from our list of words. There is a similar cancellation involving XXAXXA. Within a parsing, the occurence of /X/XA/.../X/XA/... has weight 11 and coefficient 11 whereas the occurence of /XXA/.../XXA/... has weight 11 and coefficient 2-2. If we have a parsing that involves /XXA/.../XXA/... we can create a new parsing by replacing the last /XXA/.../XXA/... with /X/XA/.../X/XA/.... The new parsing cancels out “half” of the original. Thus, we may consider an alternate puzzle where the parsing /X/XA//X/XA/ is forbidden and the word list is

wordcoefficientweightX10XA11XAA11AXA21AAA11BA11ABA11XXA11\matrix{{\rm word\/}&{\rm coefficient\/}&{\rm weight\/}\cr X&1&0\cr XA&1&1\cr XAA&1&1\cr AXA&2&1\cr AAA&-1&1\cr BA&-1&1\cr ABA&-1&1\cr XXA&-1&1\cr}

All we have done is dropped XBAXBA from the list and changed the weight of XXAXXA from 2-2 to 1-1. We call this last puzzle the tight puzzle. Establishing Theorem 4.1 for the tight puzzle is the same as establishing these results for the original one.

4.3 The Open Version

As an intermediate step to proving Theorem 4.1, we state a variant of the result. We consider bi-infinite strings in the letters {A,B,X}\{A,B,X\}, where there are only finitely many AAs and BBs. We say that two such strings are equivalent if one of them is a shift of the other one. We say that an open sentence is an equivalence class of such strings. We use finite strings to denote sentences, with the understanding that the left and right of the finite string is to be padded with an infinite number of XXs. Thus, ABAXAABAXA refers to the bi-infinite sentence XXABAXAXX...XXABAXAXX.... We define parsings just as in the cyclic case. For instance, here are all the parsings of this sentence

  • /ABA/XA/(1)/ABA/XA/(-1)

  • /XXA/BA/XA/(1)/XXA/BA/XA/(1)

(Recall that we have forbidden /X/XA/X/XA.) The first of these have weight 22 and the last one has weight 33. We have put the coefficients next to the parsings in each case. Our notation is such that the left and right sides of each expression are padded with /X/X/.../X/X/.... Based on the list above, we have

|ABAXA|=t2+t3.|ABAXA|=-t^{2}+t^{3}.

Here is the variant of Theorem 4.1 for open sentences.

Theorem 4.3 (Open)

We have |S|=|S¯||S|=|\overline{S}| for all open sentences SS.

Theorem 4.1 implies Theorem 4.3 in a fairly direct way. For instance, suppose we are interested in proving Theorem 4.3 for an open sentence SS. We say that the span of SS is the combinatorial distance between the first and last non-XX letter of SS. For instance, the span of ABAXAABAXA is 44. Supposing that SS has span ss, we simply create a cyclic sentence of length (say) s+10s+10 by padding the nontrivial part of SS with XXs and then taking the cyclic equivalence class. Call this cyclic sentence SS^{\prime}. We clearly have

|S|=|S|=|S¯|=|S¯|.|S|=|S^{\prime}|=|\overline{S}^{\prime}|=|\overline{S}|.

The middle equality is Theorem 4.1. The end inequalities are obvious.

Now we turn to the proof of Theorem 4.3. In the next result, WW stands for a finite string in the letters A,B,XA,B,X.

Lemma 4.4 (Right Identities)

The following identities hold.

  1. 1.

    |WAAA|+t|W|=0|WAAA|+t|W|=0.

  2. 2.

    |WXAA|t|W|=0|WXAA|-t|W|=0.

  3. 3.

    |WXXA|+t|W|=0|WXXA|+t|W|=0.

  4. 4.

    |WAXA|t|WA|2t|W|=0|WAXA|-t|WA|-2t|W|=0.

  5. 5.

    |WABA|+t|WA|+t|W|=0|WABA|+t|WA|+t|W|=0.

Proof: Consider Identity 1. Any parsing of WAAAWAAA must have the form W/AAAW/AAA. But AAAAAA has weight 11 and coefficient 1-1. Hence c(WAAA,w)=c(W,w1)c(WAAA,w)=-c(W,w-1). Also c(WAAA,0)=0c(WAAA,0)=0. Identity 1 follows immediately from this. Identity 2 and Identity 3 have the same proof.

Consider Identity 4. There are two kinds of parsings of WAXAWAXA. One kind has the form W/AXAW/AXA and the other kind has the form /WA/XA/WA/XA. Note that AXAAXA has weight 11 and coefficient 22 and XAXA has weight 11 and coefficient 11. From this, we see that c(WAXA,w)=2c(W,w1)+c(WA,w1)c(WAXA,w)=2c(W,w-1)+c(WA,w-1) for all ww. Identity 4 follows immediately. Identity 5 has the same proof. \spadesuit

Discussion: If Theorem 4.3 really holds, then the “reverses” of all the identities above should always hold. Let’s consider an example in detail. The reverse of Identity 2 above is

|AAXW|t|W|=0,|AAXW|-t|W|=0,

for all strings WW. However, taking W=ABAW=ABA, the weight 33 parsings of AAXWAAXW are

  • /XXA/AXA/BA/(2)/XXA/AXA/BA/(2).

  • /XAA/XA/BA/(1)/XAA/XA/BA/(-1).

As usual, our convention is to leave off the words /X/X/.../X/X/... on both sides. Adding up the coefficients, we see that c(AAXW,3)=1c(AAXW,3)=1. At the same time, the only weight 22 parsing of WW is

  • /XXA/BA/(1)/XXA/BA/(1)

Hence c(W,2)=1c(W,2)=1. This accords with our supposed equality, but the 33 parsings in the one case don’t obviously cancel out the 22 parsings in the other. In Lemma 4.4, the various parsings matched up and cancelled each other in an obvious way. However, this does not happen for the reverse identities. Nonetheless, we will prove the reverse identities of Lemma 4.4.

Lemma 4.5 (Left Identities)

The following identities hold.

  1. 1.

    |AAAW|+t|W|=0|AAAW|+t|W|=0.

  2. 2.

    |AAXW|t|W|=0|AAXW|-t|W|=0.

  3. 3.

    |AXXW|+t|W|=0|AXXW|+t|W|=0.

  4. 4.

    |AXAW|t|AW|2t|W|=0|AXAW|-t|AW|-2t|W|=0.

  5. 5.

    |ABAW|+t|AW|+t|W|=0|ABAW|+t|AW|+t|W|=0.

Proof: We will prove Identity 5. The other identities have the same proof. First of all, we check computationally that Identity 5 holds (say) for all strings WW having length at most 33.

Suppose now that WW is a shortest word for which we don’t know the result of this lemma. We know that WW has length at least 33, so we can write W=VRW=VR, where RR has length 33 and ends in AA. Consider the case when R=AXAR=AXA.

By induction, we have

|ABAV|+t|AV|+t|V|=0.|ABAV|+t|AV|+t|V|=0. (8)
|ABAVA|+t|AVA|+t|VA|=0.|ABAVA|+t|AVA|+t|VA|=0. (9)

Using Identity 4 of Lemma 4.4 (three times) we have

t|VAXA|=t2|VA|+2t2|V|.t|VAXA|=t^{2}|VA|+2t^{2}|V|. (10)
t|AVAXA|=t2|AVA|+2t2|AV|.t|AVAXA|=t^{2}|AVA|+2t^{2}|AV|. (11)
|ABAVAXA|=t|ABAVA|+2t|ABAV|.|ABAVAXA|=t|ABAVA|+2t|ABAV|. (12)

When we add together the right hand sides of Equations 10, 11, 12, we get 0, thanks to Equations 8 and 9. Hence, when we add the left hand sides of Equations 10, 11, 12, we also get 0. But this last sum is exactly the identity we wanted to prove.

A similar argument works when RR is any of the other 33-letter strings that appear in Lemma 4.4. The only case we haven’t considered is the case when R=XBAR=XBA, but these strings are forbidden. \spadesuit

Now that we have Lemma 4.4 and Lemma 4.5, our proof of Theorem 4.3 goes by induction. First of all, we check Theorem 4.3 for all strings having span at most 33. Suppose then that WW is the shortest open sentence for which we do not know Theorem 4.3. We can write W=VRW=VR where RR is some string of length 33 that ends in AA.

Let’s consider the case when R=XAAR=XAA. Then we have

|W|=|VXAA|=t|V|=t|V¯|=|AAXV¯|=|W¯|.|W|=|VXAA|=t|V|=t|\overline{V}|=|AAX\overline{V}|=|\overline{W}|.

The second equality is Identity 2 of Lemma 4.4. The third equality is the induction assumption. The fourth equality is Identity 2 of Lemma 4.5. A similar argument works when RR is any of the 33 letter strings in Lemma 4.4. The final case, R=XBAR=XBA, is forbidden.

This completes the proof of Theorem 4.3.

4.4 The Cyclic Case

We need to mention another convention before we launch into the cyclic case. Besides cyclic and open sentences, there is one more case we can consider. We introduce the notation [W][W] to denote an open word whose parsings cannot be created by padding XXs onto the left and right of WW. We will illustrate what we have in mind by way of example. Setting W=ABAXAW=ABAXA, the parsings of the open string WW are

/AXA/BA/;/XXA/X/ABA./AXA/BA/;\hskip 30.0pt/XXA/X/ABA.

However, the second parsing involves two XXs that have been padded onto the left of WW. Only the first parsing of WW is also a parsing of the locked string [W][W]. We let |[W]||[W]| be the polynomial that encodes the weights and coefficients of all the parsings of [W][W].

Lemma 4.6

Theorem 4.1 holds for any cyclic word WW with no XX in it.

Proof: To avoid some messy notation, we will consider an example. The example is sufficiently complex that it should illustrate the general proof. Suppose that

W=(BA2BA5BA1BA7).W=(BA^{2}BA^{5}BA^{1}BA^{7}). (13)

Here, for instance A2=AAA^{2}=AA. Any parsing of WW must have the breaks

W=(BA/A1BA/A4BA/A0BA/A6).W=(BA/A^{1}BA/A^{4}BA/A^{0}BA/A^{6}).

The point is that we must have a break after each BABA. From this, we see that

|W|=|[A6BA]|×|[ABA]|×|[A4BA]|×|[A0BA]|.|W|=|[A^{6}BA]|\times|[ABA]|\times|[A^{4}BA]|\times|[A^{0}BA]|.

To get the list of exponents on the right hand side of this product, we simply decrement each exponent in Equation 13 by one. But we would get the same list of exponents (perhaps in a different order) when considering the reverse word W¯\overline{W}. \spadesuit

Below we prove the following result.

Lemma 4.7

The relation

|(WX)|+2|(WB)||W|=0|(WX)|+2|(WB)|-|W|=0

holds for all open words WW.

Lemma 4.7 allows us to finish the proof of Theorem 4.1. Our proof goes by induction on the number of XsXs in the word WW. Let WW be a word having the smallest number of XXs, for which we do not know Theorem 4.1.

After cyclically permuting the letters in WW, we can write W=VXW=VX. By Lemma 4.7, we have

|(W)|=|(VX)|=2|(VB)|+|V|.|(W)|=|(VX)|=-2|(VB)|+|V|.

Applying Lemma 4.7 to W¯\overline{W}, we have

|(W¯)=|(V¯X)|=2|(V¯B)|+|V¯|=2|VB¯|+|V¯|.|(\overline{W})=|(\overline{V}X)|=-2|(\overline{V}B)|+|\overline{V}|=-2|\overline{VB}|+|\overline{V}|.

Setting Y=VBY=VB, we have

|(W)|=2|(Y)|+|V|=2|(Y¯)|+|V¯|=|(W¯)|.|(W)|=-2|(Y)|+|V|=-2|(\overline{Y})|+|\overline{V}|=|(\overline{W})|.

The middle equality comes from Theorem 4.3 (to handle VV) and the induction assumption (to handle YY.)

4.5 Some Auxilliary Relations

It remains only to prove Lemma 4.7. We will establish some auxilliary relations in this section, and then use them in the next section to prove Lemma 4.7.

Lemma 4.8

For any string WW,

|XXW|=|W|.|XXW|=|W|.

Proof: This is a tautology. \spadesuit

Lemma 4.9

For any string WW,

|W|=|[XW]||[AXW]|.|W|=|[XW]|-|[AXW]|.

Proof: The proof makes use of the right identities from Lemma 4.4 and is similar to the proof of Theorem 4.3. One first checks the statement for all strings of span three, and then argues inductively on the span. The induction step is proved using the right identities from Lemma 4.4 that hold verbatim for locked strings as well.

To illustrate the idea, we assume that WW ends with AXAAXA, that is, W=VAXAW=VAXA. Then, by Identity 4 of Lemma 4.4,

|W|=|VAXA|=t|VA|+2t|V|,|W|=|VAXA|=t|VA|+2t|V|,
|[XW|=|[XVAXA]|=t|[XVA]|+2t|[XV]|,|[XW|=|[XVAXA]|=t|[XVA]|+2t|[XV]|,

and

|[AXW]|=|[AXVAXA]|=t|AXVA|+2t|[AXV]|.|[AXW]|=|[AXVAXA]|=t|AXVA|+2t|[AXV]|.

By the induction assumption,

|VA|=|[XVA]||[AXVA]|,|V|=|[XV]||[AXV]|,|VA|=|[XVA]|-|[AXVA]|,\ \ |V|=|[XV]|-|[AXV]|,

and the result follows for WW. \spadesuit

Lemma 4.10

For any string WW,

|(AXAW)|=2t|[W]|+|[XAWA]|.|(AXAW)|=2t|[W]|+|[XAWA]|.

Proof: Either a parsing of the cyclic word (AXAW)(AXAW) contains the string AXAAXA, or there is a break after the first AA in (AXAW)(AXAW). The former case corresponds to the first term, 2t|[W]|2t|[W]|, and the latter case to the second, |[XAWA]||[XAWA]|. \spadesuit

Lemma 4.11

For any string WW,

|(ABAW)|=|[WABA]|.|(ABAW)|=|[WABA]|.

Proof: There must be a break after second AA in (ABAW)(ABAW), and this provides a one-to-one correspondence between the parsings of (ABAW)(ABAW) and [WABA][WABA]. \spadesuit

4.6 Proof of Lemma 4.7

Lemma 4.12

Lemma 4.7 holds if WW does not start and end with AA.

Proof: This is a case-by-case analysis. Suppose that W=VXW=VX for some word VV. Then (WX)=(VXX)(WX)=(VXX). Since we must have a break between VV and XXXX, it follows that

|(VXX)|=|XXV|=|V|=|W|.|(VXX)|=|XXV|=|V|=|W|.

(the second equality holds by Lemma 4.8). On the other hand,

|(WB)|=|(VXB)|=0|(WB)|=|(VXB)|=0

since the combination X/BAX/BA is prohibited. The claim of Lemma 4.7 follows. Similarly, if W=XVW=XV for some word VV then (WX)=(VXX)(WX)=(VXX), and the same argument applies. If W=VBW=VB then then each term in the equality of Lemma 4.7 vanishes. The same holds if W=BVW=BV. Finally, if WW ends in BB, all terms in Lemma 4.7 are trivial. \spadesuit

The only remaining case is when W=AVAW=AVA for some word VV. What we need to prove is

|(AXAV)|+2|(ABAV)||AVA|=0.|(AXAV)|+2|(ABAV)|-|AVA|=0. (14)

By Lemma 4.10,

|(AXAV)|=2t|[V]|+|[XAVA]|.|(AXAV)|=2t|[V]|+|[XAVA]|.

By Lemma 4.11,

|(ABAV)|=|[VABA]|.|(ABAV)|=|[VABA]|.

By Lemma 4.9,

|AVA|=|[XAVA]||[AXAVA]|.|AVA|=|[XAVA]|-|[AXAVA]|.

Therefore, the left hand side of (14) equals

2t|[V]|+|[XAVA]|+2|[VABA]||[XAVA]|+|[AXAVA]|.2t|[V]|+|[XAVA]|+2|[VABA]|-|[XAVA]|+|[AXAVA]|. (15)

By Identity 5 of Lemma 4.4 for locked words,

|[VABA]|=t|[VA]|t|[V]|.|[VABA]|=-t|[VA]|-t|[V]|.

Finally,

|[AXAVA]|=2t|[VA]||[AXAVA]|=2t|[VA]|

since a parsing of [AXAVA][AXAVA] must start with AXAAXA. It follows that (15) equals

2t|[V]|+|[XAVA]|2t|[VA]|2t|[V]||[XAVA]|+2t|[VA]|=0,2t|[V]|+|[XAVA]|-2t|[VA]|-2t|[V]|-|[XAVA]|+2t|[VA]|=0,

as needed.

This completes the proof of Lemma 4.7.

5 References

[M] Th. Motskin, The Pentagon in the projective plane, and a comment on Napier’s rule, Bull. Amer. Math. Soc. 52 (1945), 985-989.

[OST1] V. Ovsienko, R. E. Schwartz, S. Tabachnikov, The Pentagram Map: A Discrete Integrable System, Comm. Math. Phys. 2010 (to appear), math arXiv:0810.5605.

[OST2] V. Ovsienko, R. E. Schwartz, S. Tabachnikov, Quasiperiodic motion for the pentagram map, Electr. Res. Announ. Math. 16 (2009), 1–8.

[S1] R. Schwartz, The Pentagram Map J. Experiment. Math. 1 (1992), 71–81.

[S2] R. Schwartz, Recurrence of the Pentagram Map, J. Experiment. Math. 110 (2001), 519–528.

[S3] R. Schwartz, Discrete Monodromy, Pentagrams, and the Method of Condensation, J. Fixed Point Theory Appl. 3 (2008), 379–409.

[ST] R. Schwartz, S. Tabachnikov, Elementary Surprises in Projective Geometry, Math. Intelligencer, 2010 (to appear), math arXiv:0910.1952.