The Pentagram Integrals on Inscribed Polygons
1 Introduction
The pentagram map is a geometric iteration defined on polygons. This map is defined in practically any field, but it is most easily described for polygons in the real projective plane. Geometrically, the pentagram map carries the polygon to the polygon , as shown in Figure 1.
![[Uncaptioned image]](https://cdn.awesomepapers.org/papers/a0d51e10-27f0-4310-bcdc-ca2694054fa4/x1.png)
Figure 1: The pentagram map
This pentagram map is always defined for convex polygons, and generically defined for all polygons. The pentagram map commutes with projective transformations and induces a generically defined map on the space of cyclically labelled 111Technically, one needs to consider the square of the map in order to get a canonically defined map on labelled -gons. However, if one is willing to break symmety, preferring (say) left over right, then the map itself is defined on labelled -gons. -gons modulo projective transformations. is the identity map on and has period on . The map is not periodic on for .
The pentagram map also defines a map on a slightly larger space . The space is the space of twisted -gons modulo projective transformations. A twisted -gon is a map , such that
Here is a real projective transformation, called the monodromy of the twisted polygon. The space contains the space as a codimension submanifold.
This pentagram map is studied for pentagons in [M]. We studied the general case in [S1], [S2], and [S3]. Finally, in [OST1] (see [OST2] for a short version) we showed that the pentagram map is a completely integrable system on . We think that the pentagram map is a completely integrable system on as well, but there are certain technical details we need to resolve before establishing this fact.
In [S3] we constructed polynomial functions
which are invariant under the pentagram map. In [OST1] these functions are recognized a complete list of integrals for the completely integrable system. There is an invariant Poisson structure, and the above functions Poisson-commute with respect to this Poisson structure. We call these functions the monodromy invariants, and we will construct them in the next section.
The purpose of this paper is to prove the following result.
Theorem 1.1
for any twisted -gon that is inscribed in a conic section and any .
The result, in particular, applies to closed inscribed polygons. Theorem 1.1 boils down to a countably infinite family of polynomial identities. The polynomials involved are somewhat reminiscent of the symmetric polynomials, but they have somewhat less symmetry then these. We noticed this result by doing computer experiments. One novel feature of the theorem is that we discovered not just the result but also the proof by way of computer experimentation. We wrote a Java applet to aid us with the combinatorics of the proof. This applet is available on the first author’s website. 222http://www.math.brown.edu/res
While our proof is mainly combinatorial, we think that perhaps there should be a better proof based on geometry. Accordingly, we will describe our polynomials in three ways – geometrically, combinatorially, and in terms of determinants. We will only use the combinatorial description in our proof, but we hope that the other descriptions might ring a bell for some readers.
The pentagram map seems to interact nicely with polygons that are inscribed in conic sections. We mention our paper [ST], in which we describe some finite configuration theorems, à la Pappus, that we discovered in connection with the pentagram map and inscribed polygons.
Here is an overview of the paper. In §2 we define the monodromy invariants. In §3 we reduce Theorem 1.1 to a combinatorial problem. In §4 we solve this combinatorial problem.
The second author would like to thank Brown University for its hospitality during his sabbatical; he was also supported by NSF grant DMS-0072607.
2 The Monodromy Invariants
The papers [S3] and [OST1] give a good account of the monodromy invariants. The notation in the two papers is slightly different, and we will follow [OST1]. We will give three descriptions of these invariants, one geometrical, one combinatorial, and one based on determinants. The reader only interested in the proof of Theorem 1.1 need only pay attention to the combinatorial definition. As we say in the introduction, we mention the other definitions just in the hopes that it will ring a bell for some reader.
Because we are not using the geometric or determinental definitions in our proof, we will not include the arguments that show the equivalence of the various definitions. The paper [S3] has a proof that the geometric and combinatorial definitions coincide.
2.1 The Geometric Definition
First of all, we have the cross ratio
Suppose that is a twisted -gon, with monodromy . We let . The label in Figure 2 denotes , and similarly for the other labels.
![[Uncaptioned image]](https://cdn.awesomepapers.org/papers/a0d51e10-27f0-4310-bcdc-ca2694054fa4/x2.png)
Figure 2: vertex labels
We associate to each vertex two numbers:
Here denotes the line determined by points and . For instance, is the cross ratio of the white points in Figure 2. We call the invariants the corner invariants. These invariants form a periodic sequence of length . That is and for all .
We define
The other monodromy invariants are best defined in an indirect way. Recall that is the monodromy of our twisted polygon . We lift to an element of which we also denote by . We define
These quantities are independent of the lift of and only depend on the conjugacy class of . Finally, these quantities are invariant under the pentagram map.
We define
It turns out that these quantities are polynomials in the corner invariants. The remaining monodromy invariants are suitably weighted homogeneous parts of these polynomials.
We have a basic rescaling operation
We say that a polynomial in the corner invariants has weight if
Here denotes the obvious action of on polynomials. In [S3] we show that
Here and are the weight polynomials in each sum, and denotes the floor of .
2.2 The Combinatorial Definition
Now we describe the combinatorial formulas for these invariants. Again, the paper [S3] has a proof that the description below coincides with the one in the previous section. In everything we say, the indices are taken cyclically, mod .
We introduce the monomials
(1) |
The monodromy polynomial is built out of the monomials for and for . We call two monomials consecutive if they involve consecutive or coinciding variables . More explicitly, we have the following criteria:
-
1.
and are consecutive if
-
2.
and are consecutive if
-
3.
and are consecutive if .
Let be a monomial obtained by the product of the monomials and , that is,
Such a monomial is called admissible if no two of the monomials are consecutive. For every admissible monomial, define the weight and the sign . One then has
For example, if we obtain the following two polynomials:
The same formula works for , if we make all the same definitions with and interchanged. More precisely, one builds the polynomials from the monomials and with the same restriction that no consecutive monomials are allowed and the same definitions of the weight and sign.
We note that the dihedral symmetry
(2) |
interchanges the polynomials and .
2.3 The Determinantal Definition
Now we describe determinantal formulas for the monodromy invariants; these formulas did not appear in our previous papers on the subject.
For positive integers , we define the four-diagonal determinant
where and are the corner invariants and is as in (1). By convention,
Then one has the following formula for the monodromy invariants :
(3) |
Similarly, for , define
Then one has:
(4) |
Formulas (3) and (4) simplify if one considers an open version of the monodromy invariants: instead of having a periodic “boundary condition” , assume that for and , and for and . With this “vanishing at infinity” boundary conditions, the monodromy invariants are given by a single determinant:
We shall encounter an open version of monodromy invariants in Section 4.3.
3 Reduction to the Puzzle
A non-degenerate conic in can be identified with by way of the stereographic projection from a point of the conic. This identification is unique up to a projective transformation of the real projective line. That is, a different choice of the center of projection amounts to a projective transformation of .
Recall from the introduction that a twisted -gon is a map , such that
Here is a real projective transformation, called the monodromy of the twisted polygon. Given a twisted -gon, we let . These are the vertices of the twisted -gon.
If is an inscribed twisted polygon, we can consider the vertices as points of the real projective line and talk about their cross-ratios which are uniquely defined. Referring to the cross-ratio on the projective line, we set:
In the next lemma we express the corner invariants of an inscribed polygon in terms of the quantities . Once we specify the cross ratios , we produce an inscribed twisted polygon having corner invariants . Thus we have a map . We denote this map by .
Lemma 3.1
One has:
and the map is given by the formula
(5) |
Proof: Consider Figure 3. Using the projection from point , we obtain:
A similar projection from yields the formula for . The expression for in terms of and follows now from the identity
and likewise for .
The dihedral group also acts on cyclic sequences by cyclic permutations and the orientation-reversing involution . It follows from (5) that where is the involution (2). Hence is a dihedrally equivariant map.
After and are replaced by via (5), the polynomials and become Laurent polynomials in the variables . The identity obviously holds since both sides equal . We need to prove that for . The strategy is to show that every monomial in the variables appears in and with the same coefficient. Since the map is dihedrally equivariant and the orientation-reversing involution on the variables interchanges and , it suffices to show that if two monomials in the variables are related by orientation-reversing symmetry (for example, the involution ) then they appear in with the same coefficients.
Let us compute the monomials in the variables ’s that appear in the polynomial . Using (5), one finds:
(6) |
and
(7) |
We see that the variables ’s that appear in involve two indices, and , and the those in involve three indices, and . Pictorially, these terms can be represented as follows: for , see Figure 4, and for , see Figure 5. In these figures, the presence of each term in the numerator and denominator is represented by a shaded square, and its absence by a white square.
According to Section 2.2, the polynomial is the sum of all admissible products of terms, and each term is either or . The admissibility condition is that the monomials or involved are sufficiently distant; what it means precisely is that the respective tiles in Figures 4 and 5, corresponding to these terms, do not overlap. This is a crucial observation.
To summarize, each monomial in , after the substitutions (6) and (7), is represented by a collection of tiles depicted in Figures 4 and 5, taken with the product of their respective coefficients. The tiles, that occupy two or three consecutive positions, are placed around a circle having positions available (if we are concerned with twisted -gons). There may be empty positions left between the tiles.
As a final preparation for the next section, we introduce the following notation: letter denotes a shaded square in the lower row of a tile ( in denominator), letter a shaded square in the upper row ( in numerator), and letter an empty column in a tile. Thus, the two tiles in Figure 4 correspond to the words and , and the six tiles in Figure 5 to and , respectively. We also use letter to mean an empty slot between the tiles.
4 The Puzzle
4.1 The Main Result
Now we are going to extract the main combinatorial information from the discussion at the end of the last section.
We fix some integer and consider the set of length lists in the letters . We consider two lists the same if they are cyclic permutations of each other. We say that a cyclic sentence is an equivalence class of such strings. To illustrate our notation by way of example, denotes the equivalence class of . Here . We let denote the set of such sentences, with the dependence on understood.
We single out certain strings of letters, and to each of these special words we assign a coefficient and a weight. Here is the list.
We say that a parsing of a cyclic sentence is a description of the cyclic sentence as a concatenation of words. For example
are the only two parsings of . We define the coefficient of a parsing to be the product of the coefficients of the word. We define the weight of the parsing to be the sum of the weights of the words. Both parsings above have coefficient and weight .
For each cyclic sentence , we define to be the total sum of the coefficients of the weight parsings of . For instance, when , we have and otherwise . For a more streamlined equation, we define
Here is a polynomial in that encodes all the coefficients at once. For example
Let denote the cyclic sentence obtained by reversing . In view of Section 3, Theorem 1.1 is equivalent to the following result.
Theorem 4.1 (Cyclic)
We have for all cyclic sentences .
4.2 The Tight Puzzle
Before we tackle Theorem 4.1, we slightly modify our puzzle for the sake of convenience.
Lemma 4.2
Suppose that contains the string . Then .
Proof: No word ends in , and so the string must
continue as .
The occurence of in any parsing contributes
weight and coefficient whereas the occurence of
contributes weight and coefficient .
If we have a parsing that involves
we can create a new parsing by replacing
the last with . These two
parsings have the same weight and opposite coefficient,
and thereby cancel each other in the total sum.
By Lemma 4.2, we can simply throw out any strings that contain , and we may drop the word from our list of words. There is a similar cancellation involving . Within a parsing, the occurence of has weight and coefficient whereas the occurence of has weight and coefficient . If we have a parsing that involves we can create a new parsing by replacing the last with . The new parsing cancels out “half” of the original. Thus, we may consider an alternate puzzle where the parsing is forbidden and the word list is
All we have done is dropped from the list and changed the weight of from to . We call this last puzzle the tight puzzle. Establishing Theorem 4.1 for the tight puzzle is the same as establishing these results for the original one.
4.3 The Open Version
As an intermediate step to proving Theorem 4.1, we state a variant of the result. We consider bi-infinite strings in the letters , where there are only finitely many s and s. We say that two such strings are equivalent if one of them is a shift of the other one. We say that an open sentence is an equivalence class of such strings. We use finite strings to denote sentences, with the understanding that the left and right of the finite string is to be padded with an infinite number of s. Thus, refers to the bi-infinite sentence . We define parsings just as in the cyclic case. For instance, here are all the parsings of this sentence
-
•
-
•
(Recall that we have forbidden .) The first of these have weight and the last one has weight . We have put the coefficients next to the parsings in each case. Our notation is such that the left and right sides of each expression are padded with . Based on the list above, we have
Here is the variant of Theorem 4.1 for open sentences.
Theorem 4.3 (Open)
We have for all open sentences .
Theorem 4.1 implies Theorem 4.3 in a fairly direct way. For instance, suppose we are interested in proving Theorem 4.3 for an open sentence . We say that the span of is the combinatorial distance between the first and last non- letter of . For instance, the span of is . Supposing that has span , we simply create a cyclic sentence of length (say) by padding the nontrivial part of with s and then taking the cyclic equivalence class. Call this cyclic sentence . We clearly have
The middle equality is Theorem 4.1. The end inequalities are obvious.
Now we turn to the proof of Theorem 4.3. In the next result, stands for a finite string in the letters .
Lemma 4.4 (Right Identities)
The following identities hold.
-
1.
.
-
2.
.
-
3.
.
-
4.
.
-
5.
.
Proof: Consider Identity 1. Any parsing of must have the form . But has weight and coefficient . Hence . Also . Identity 1 follows immediately from this. Identity 2 and Identity 3 have the same proof.
Consider Identity 4.
There are two kinds of parsings of . One kind
has the form and the other kind has the form
. Note that has weight and
coefficient and has weight and
coefficient . From this, we see that
for
all . Identity 4 follows immediately.
Identity 5 has the same proof.
Discussion: If Theorem 4.3 really holds, then the “reverses” of all the identities above should always hold. Let’s consider an example in detail. The reverse of Identity 2 above is
for all strings . However, taking , the weight parsings of are
-
•
.
-
•
.
As usual, our convention is to leave off the words on both sides. Adding up the coefficients, we see that . At the same time, the only weight parsing of is
-
•
Hence . This accords with our supposed equality, but the parsings in the one case don’t obviously cancel out the parsings in the other. In Lemma 4.4, the various parsings matched up and cancelled each other in an obvious way. However, this does not happen for the reverse identities. Nonetheless, we will prove the reverse identities of Lemma 4.4.
Lemma 4.5 (Left Identities)
The following identities hold.
-
1.
.
-
2.
.
-
3.
.
-
4.
.
-
5.
.
Proof: We will prove Identity 5. The other identities have the same proof. First of all, we check computationally that Identity 5 holds (say) for all strings having length at most .
Suppose now that is a shortest word for which we don’t know the result of this lemma. We know that has length at least , so we can write , where has length and ends in . Consider the case when .
By induction, we have
(8) |
(9) |
Using Identity 4 of Lemma 4.4 (three times) we have
(10) |
(11) |
(12) |
When we add together the right hand sides of Equations 10, 11, 12, we get , thanks to Equations 8 and 9. Hence, when we add the left hand sides of Equations 10, 11, 12, we also get . But this last sum is exactly the identity we wanted to prove.
A similar argument works when is any of the other -letter strings
that appear in Lemma 4.4. The only case we haven’t considered
is the case when , but these strings are forbidden.
Now that we have Lemma 4.4 and Lemma 4.5, our proof of Theorem 4.3 goes by induction. First of all, we check Theorem 4.3 for all strings having span at most . Suppose then that is the shortest open sentence for which we do not know Theorem 4.3. We can write where is some string of length that ends in .
Let’s consider the case when . Then we have
The second equality is Identity 2 of Lemma 4.4. The third equality is the induction assumption. The fourth equality is Identity 2 of Lemma 4.5. A similar argument works when is any of the letter strings in Lemma 4.4. The final case, , is forbidden.
This completes the proof of Theorem 4.3.
4.4 The Cyclic Case
We need to mention another convention before we launch into the cyclic case. Besides cyclic and open sentences, there is one more case we can consider. We introduce the notation to denote an open word whose parsings cannot be created by padding s onto the left and right of . We will illustrate what we have in mind by way of example. Setting , the parsings of the open string are
However, the second parsing involves two s that have been padded onto the left of . Only the first parsing of is also a parsing of the locked string . We let be the polynomial that encodes the weights and coefficients of all the parsings of .
Lemma 4.6
Theorem 4.1 holds for any cyclic word with no in it.
Proof: To avoid some messy notation, we will consider an example. The example is sufficiently complex that it should illustrate the general proof. Suppose that
(13) |
Here, for instance . Any parsing of must have the breaks
The point is that we must have a break after each . From this, we see that
To get the list of exponents on the right hand side of this product,
we simply decrement each exponent in Equation 13 by one.
But we would get the same list of exponents (perhaps in a different
order) when considering the reverse word .
Below we prove the following result.
Lemma 4.7
The relation
holds for all open words .
4.5 Some Auxilliary Relations
It remains only to prove Lemma 4.7. We will establish some auxilliary relations in this section, and then use them in the next section to prove Lemma 4.7.
Lemma 4.8
For any string ,
Proof: This is a tautology.
Lemma 4.9
For any string ,
Proof: The proof makes use of the right identities from Lemma 4.4 and is similar to the proof of Theorem 4.3. One first checks the statement for all strings of span three, and then argues inductively on the span. The induction step is proved using the right identities from Lemma 4.4 that hold verbatim for locked strings as well.
To illustrate the idea, we assume that ends with , that is, . Then, by Identity 4 of Lemma 4.4,
and
By the induction assumption,
and the result follows for .
Lemma 4.10
For any string ,
Proof: Either a parsing of the cyclic word contains the
string , or there is a break after the first in . The
former case corresponds to the first term, ,
and the latter case to the second, .
Lemma 4.11
For any string ,
Proof: There must be a break after second in , and this
provides a one-to-one correspondence between the parsings of and .
4.6 Proof of Lemma 4.7
Lemma 4.12
Lemma 4.7 holds if does not start and end with .
Proof: This is a case-by-case analysis. Suppose that for some word . Then . Since we must have a break between and , it follows that
(the second equality holds by Lemma 4.8). On the other hand,
since the combination is prohibited. The claim of Lemma 4.7 follows.
Similarly, if for some word then , and the same argument applies.
If then then each term in the equality of
Lemma 4.7 vanishes. The same holds if .
Finally, if ends in , all terms in Lemma 4.7 are trivial.
The only remaining case is when for some word . What we need to prove is
(14) |
By Lemma 4.10,
By Lemma 4.11,
By Lemma 4.9,
Therefore, the left hand side of (14) equals
(15) |
By Identity 5 of Lemma 4.4 for locked words,
Finally,
since a parsing of must start with . It follows that (15) equals
as needed.
This completes the proof of Lemma 4.7.
5 References
[M] Th. Motskin, The Pentagon in the projective plane, and a
comment on Napier’s rule, Bull. Amer. Math. Soc. 52 (1945), 985-989.
[OST1] V. Ovsienko, R. E. Schwartz, S. Tabachnikov,
The Pentagram Map: A Discrete Integrable System,
Comm. Math. Phys. 2010 (to appear),
math arXiv:0810.5605.
[OST2] V. Ovsienko, R. E. Schwartz, S. Tabachnikov,
Quasiperiodic motion for the pentagram map,
Electr. Res. Announ. Math.
16 (2009), 1–8.
[S1] R. Schwartz, The Pentagram Map
J. Experiment. Math. 1 (1992), 71–81.
[S2] R. Schwartz, Recurrence of the Pentagram Map,
J. Experiment. Math. 110 (2001), 519–528.
[S3] R. Schwartz, Discrete Monodromy, Pentagrams,
and the Method of Condensation,
J. Fixed Point Theory Appl. 3 (2008), 379–409.
[ST] R. Schwartz, S. Tabachnikov,
Elementary Surprises in Projective Geometry,
Math. Intelligencer, 2010 (to appear), math arXiv:0910.1952.