Introduction to weak Pinsker filtrations
Abstract
We introduce the so-called weak Pinsker dynamical filtrations, whose existence in any ergodic system follows from the universality of the weak Pinsker property, recently proved by Austin [1]. These dynamical filtrations appear as a potential tool to describe and classify positive entropy systems. We explore the links between the asymptotic structure of weak Pinsker filtrations and the properties of the underlying dynamical system. A central question is whether, on a given system, the structure of weak Pinsker filtrations is unique up to isomorphism. We give a partial answer, in the case where the underlying system is Bernoulli. We conclude our work by giving two explicit examples of weak Pinsker filtrations.
1 Introduction
In 1958, Kolmogorov and Sinaï introduced the notion of entropy in ergodic theory: the Kolmogorov-Sinaï entropy (or KS-entropy). The same year, Kolmogorov introduced another important notion: K-systems. He defined a K-system as a dynamical system on which there is a generator whose tail -algebra is trivial. There is an equivalent definition that is more intrinsic to the system: is a K-system if, and only if, every non-trivial observable satisfies (a proof of this equivalence, and a more complete presentation of this notion can be found in [4]). It is also equivalent to assume that the Pinsker factor of the system is trivial, the Pinsker factor being the -algebra
The Pinsker factor is simply the largest factor of that is of entropy . Therefore, a K-system has no non-trivial factor of entropy : it is entirely non-deterministic. For example, the most elementary K-systems are the Bernoulli shifts. They are K-systems because i.i.d. processes satisfy Kolmogorov’s - law.
Entropy is an invariant that quantifies the “chaos” of a dynamical system, or more precisely its unpredictability, and many of the questions that arose after its discovery were aimed at understanding the structure of this “chaos”. The first question, which Kolmogorov asked after proving that Bernoulli shifts are K-systems, was whether all K-systems are Bernoulli shifts, which would imply that these chaotic systems have a very simple structure. More general questions then emerged, and we will return to them in the following paragraphs.
The discovery of entropy first led to non-isomorphism results, particularly for Bernoulli shifts: two isomorphic Bernoulli shifts must have the same entropy. The converse of this result, shown by Ornstein [11, 12], is one of the most notable successes of the KS-entropy. But the ramifications of Ornstein’s theory go far beyond Bernoulli shifts, and have had a profound impact on the evolution of ergodic theory. We will confine ourselves here to telling the story of the weak Pinsker property.
In the early 1960s, Pinsker, then working in Moscow with Kolmogorov, showed that any K-factor of is independent of the Pinsker factor (see [19], but this reference is in Russian). Following this result, although the existence of any specific K-factor had not yet been proved, he issued a conjecture (later called the “Pinsker conjecture”): any system of non-zero entropy is isomorphic to the direct product of its Pinsker factor and a K-system. A few years later, Sinai published [23] which seemed to confirm this conjecture: he proved the existence of a factor of isomorphic to a Bernoulli shift of the same entropy as . Given Pinsker’s independence result, it would have been sufficient to prove that the factor constructed by Sinaï and the Pinsker factor generate the entire -algebra to obtain a result even stronger than Pinsker’s conjecture: would then be isomorphic to the direct product of its Pinsker factor and a Bernoulli shift. This “strong Pinsker conjecture” would also have proved that any K-system is isomorphic to a Bernoulli shift.
But this conjecture turned out to be false: Ornstein published a first example of a non-Bernoulli K-system [15] which contradicts the strong Pinsker conjecture. Following that, many other counterexamples were built, showing that the family of all K-systems is very broad, leaving little hope for a complete classification of those systems. Among all these counterexamples, we can find a construction by Ornstein [13] that can be used to contradict Pinsker’s conjecture. Furthermore, he then refined this result by constructing a mixing system that does not verify Pinsker’s conjecture [14]. Thus, all the conjectures formulated in the early years of the study of KS-entropy were wrong, revealing a wide variety of possible phenomena.
One of the ramifications of Ornstein’s theory can be found in the work of Thouvenot, who, starting in 1975, became interested in relatively Bernoulli systems and developed a “relative” version of Ornstein’s theory. Following this work, in his 1977 paper [24], he introduced the weak Pinsker property: for any , is isomorphic to the direct product of a Bernoulli shift and a system of entropy at most :
| (1) |
For four decades, however, it was unclear whether all systems verified the weak Pinsker property. But in 2018, Austin published a paper on the subject [1] in which he proved that all ergodic systems satisfy the weak Pinsker property.
We can then iterate this splitting operation: take an increasing sequence of positive numbers such that , and start by splitting into
then split into
and so on. This yields a sequence of systems such that, for every , is a factor of . By composing the factor maps, it means that each is a factor of , and therefore generates a -invariant -algebra . Because of our iterating construction, we see that , so the sequence is a filtration. This is what we call a weak Pinsker filtration on (see Definition 2.20).
The purpose of this paper is to discuss the links between weak Pinsker filtrations and the structure of dynamical systems with positive entropy. Weak Pinsker filtrations fall into the category of dynamical filtrations, i.e. filtrations on a dynamical system for which each -algebra is -invariant. A framework for the study of such filtrations was introduced in [10], and this will guide our approach of weak Pinsker filtrations. In Section 2.2, we introduce the necessary concepts from the theory of dynamical filtrations. This framework is focused on the various possible structures of filtrations whose tail -algebra is trivial, which is the type of weak Pinsker filtrations that appear on K-systems (see Theorem 2.23). Therefore, the study of weak Pinsker filtrations we suggest would mainly be aimed at classifying K-systems, and in particular non-Bernoulli K-systems.
In Section 2, we give an overview of the results and open questions that arise from the study of the properties of weak Pinsker filtrations, and their relation to the structure of the underlying dynamical system. One of those questions concerns the uniqueness, up to isomorphism, of weak Pinsker filtrations. In Section 3, we give a partial answer to this uniqueness problem in the case of Bernoulli systems. That section is based on ideas suggested to us by Thouvenot. The main result of this paper is Theorem 3.1. Finally, in Section 4, we give explicit examples of weak Pinsker filtrations, in order to give a more concrete meaning to all of those abstract notions.
2 Weak Pinsker filtrations and related questions
In this section, we introduce the notion of weak Pinsker filtrations, the tools necessary to study them and state some of the main questions concerning those filtrations. Since weak Pinsker filtrations are dynamical filtrations, we will use the framework for classifying dynamical filtrations presented in the previous section.
2.1 Basic notation
A dynamical system is a quadruple such that is a Lebesgue probability space, and is an invertible measure-preserving transformation.
Let be sub--algebras. We write mod , if for every , there exists such that . Then, mod if mod and mod . We denote the smallest -algebra that contains and . We say that is a factor -algebra (or -invariant -algebra) if mod . Let and be sub--algebras of . We say that and are relatively independent over if for any -measurable bounded function and -measurable bounded function
In this case, we write . If is trivial, and are independent, which we denote .
If we have two systems and , a factor map is a measurable map such that and , -almost surely. If such a map exists, we say that is a factor of and we denote the -algebra generated by . Conversely, we also say that is an extension of or that is embedded in . Moreover, if there exist invariant sets and of full measure such that is a bijection, then is an isomorphism and we write .
For a given factor -algebra , in general, the quadruple is not a dynamical system since need not separate points on , and in this case is not a Lebesgue probability space. However, there always exist a dynamical system and a factor map such that mod . Moreover, this representation is not unique, but for a given factor , all such systems are isomorphic and there is a canonical construction to get a system and a factor map such that mod (see [6, Chapter 2, Section 2]).
2.2 Dynamical filtrations
Let . A dynamical filtration is a pair such that is a filtration in discrete negative time (i.e. ) on and each is -invariant. The theory of dynamical filtrations was initiated by Paul Lanthier in [9, 10]. For our present work, the primary notion we need is a precise definition of what it means for two dynamical filtrations to be isomorphic:
Definition 2.1.
Let be a dynamical filtration on and a dynamical filtration on . We say that and are isomorphic if there is an isomorphism such that, for all , mod .
We will discuss two specific families of filtrations:
Definition 2.2.
Let be a dynamical filtration on . It is of product type if there is a sequence of mutually independent factor -algebras such that
Definition 2.3.
Let be a dynamical filtration on . It is Kolmogorovian if its tail -algebra is trivial, i.e. if mod .
In particular, because of Kolmogorov’s law, product type filtrations are Kolmogorovian.
In the theory of dynamical filtrations, additional properties, such as standardness and I-cosiness, appear naturally (see [10]). However, for now, we are not able to find relevant applications of those notions in the study of weak Pinsker filtrations, so we dot not discuss them in this paper. That being said, standardness and I-cosiness being looser than “product-type”, a first step in investigating further the examples of Section 4 could be to determine whether they are standard, I-cosy or neither.
2.3 Reminders on KS-entropy
The notion of entropy first appeared in mathematics in 1948, introduced by Shannon in his foundational work on information theory [21]. It is defined as follows:
Definition 2.4 (Shannon entropy).
Let be a probability space and a random variable, with finite or countable. The Shannon entropy of is
The number gives the average amount of information given by the random variable . If we have a probability measure defined directly on , we can also define the entropy of that measure
One can also define conditional entropy: for and be two random variables, we define
where . This quantifies the missing information required to determine once is known. We refer to [3, Chapter 2, Section 6] for the basic properties of this notion. In the present work, conditional entropy will only serve as a computational tool, via Fano’s inequality. See [3, Theorem 6.2] for a proof.
Lemma 2.5 (Fano’s inequality).
Let and be two -valued random variables. Set . We have
In particular, for , if is an -valued random variable such that there exists satisfying , then
In 1958, Kolmogorov and Sinaï used entropy to quantify the randomness of measure preserving dynamical systems.
Let be a dynamical system. To any random variable , with finite, we associate the corresponding -process
Also, for , set .
The Kolmogorov-Sinaï entropy (or KS-entropy) of a dynamical system is:
Definition 2.6 (Kolmogorov-Sinaï entropy).
Let be a dynamical system. For a finite valued random variable , define
For a -invariant -algebra , define
Finally, set
The KS-entropy satisfies the following continuity result:
Lemma 2.7.
Let be a dynamical system and a random variable , with finite. For , there exists such that, for any random variable such that , we have
Proof.
In this proof, we will use Fano’s inequality (Lemma 2.5). Specifically, we compute:
where . And, since is continuous, there exists such that, if , we have
By switching and and doing the same reasoning, we end the proof. ∎
It is useful to locate the deterministic aspects of a dynamical system. We do that by considering the Pinsker factor of a system. For any factor -algebra , we define
The Pinsker factor of the system is then defined as . We will use the following basic result, which can be found in [18, Theorem 14]:
Lemma 2.8.
Let be a dynamical system and and be independent factor -algebras. We have
To be able to compute the entropy of a system, the following result proves to be most useful.
Theorem 2.9 (Kolmogorov-Sinaï).
Let be a dynamical system. Consider a finite valued random variable and the corresponding -process . Then we have
In particular, if is a generator of (i.e. mod ), then .
2.4 Ornstein’s theory and its relative version
From the definition, one easily sees that KS-entropy is invariant under isomorphism of dynamical systems, which makes it a useful tool in the classification of measure preserving dynamical systems. The most remarkable classification results concern Bernoulli and relatively Bernoulli systems:
Definition 2.10 (Bernoulli and relatively Bernoulli).
Let be a dynamical system.
We say that (or ) is Bernoulli if there exists a random variable such that the corresponding -process is i.i.d. and generates , i.e. we have mod .
Let be a factor -algebra. We say that (or ) is relatively Bernoulli over if there is an i.i.d. process of the form such that is independent of and mod .
Those two definitions coincide when is the trivial factor -algebra: is relatively Bernoulli over if and only if is Bernoulli.
Remark 2.11.
We can consider another approach to define Bernoulli systems: take a finite or countable set and a probability measure on . On the product probability space , consider the transformation
The map is called the shift on . One can easily check that is -invariant. Therefore, this yields a measure preserving dynamical system
| (2) |
which is called a Bernoulli shift. Then a system is Bernoulli if and only if it is isomorphic to a Bernoulli shift. Similarly, we can see that a system is relatively Bernoulli over a factor -algebra if and only if is isomorphic to a system of the form via a factor map such that mod (where is the projection of onto ).
Using Theorem 2.9, it is easy to compute the entropy of a Bernoulli system. Let be an process on that generates . We then have
In particular, if is isomorphic to a system of the form (2), we get
Since, to be isomorphic, two systems need to have the same entropy, this computation enables us to get a non-isomorphism result between any two Bernoulli systems of different entropy. Remarkably, Ornstein proved that the converse is also true:
This means that the KS-entropy gives a complete classification of Bernoulli systems. An outstanding result that emerged from Ornstein’s theory was a criterion to characterize Bernoulli systems: finite determination. However, although this notion is useful in proving abstract results, when studying a given system, it is not easy to know whether or not it is finitely determined. Because of that, another criterion called very weak Bernoullicity was developed (see [16, Section 7]). This is the criterion we are interested in.
For the remainder of this section, we assume that the processes are defined on finite alphabets. We first need a technical definition. Given a finite alphabet , an integer and two words of length on , we define the normalized Hamming metric between and as:
where and . We then consider the corresponding transportation metric on :
Then a process is said to be very weak Bernoulli if, for some , the conditional law of given the past of is close enough to the law of in the metric. More formally, we state:
Definition 2.13 (Very weak Bernoulli).
Let be an ergodic dynamical system, equipped with a -process taking values in a finite alphabet. We say that is very weak Bernoulli if, for every , there exists such that for every , we have
where is the law of and, for , is the conditional law of given that equals .
If , we say that (or ) is very weak Bernoulli.
The fact that very weak Bernoullicity characterizes Bernoulli systems can be stated as follows:
Theorem 2.14 (see [16, 17]).
Let be a dynamical system. A -process on is very weak Bernoulli if and only if is Bernoulli.
Following the work of Ornstein, Thouvenot studied relatively Bernoulli systems and adapted the definitions of finite determination and very weak Bernoullicity to get criteria that characterize relatively Bernoulli systems. Here we give his adaptation of very weak Bernoullicity:
Definition 2.15 (Relatively very weak Bernoulli).
Let be an ergodic dynamical system, equipped with two -processes and with finite alphabets. We say that is relatively very weak Bernoulli over if, for every , there exists such that for every and for all large enough, we have
where is the law of and, for , is the conditional law of given that equals and that equals .
If and , we say that (or ) is relatively very weak Bernoulli over .
Many early results from Thouvenot’s theory were stated for relatively finitely determined systems. However, for our work, relative very weak Bernoullicity is a more convenient notion. Fortunately, we have the following equivalence, which enables us to apply to relatively very weak Bernoulli processes results originally stated for relatively finitely determined processes:
Theorem 2.16 (see [20]).
Let be an ergodic system and and be -processes with finite alphabets defined on . Then is relatively very weak Bernoulli over if and only if it is relatively finitely determined over .
We give a summary of the results we will use:
Lemma 2.17.
Let be a finite entropy dynamical system and a factor -algebra. Let and be -processes with finite alphabets defined on such that and . If is relatively very weak Bernoulli over , then
-
(i)
is relatively Bernoulli over ,
-
(ii)
any -process on is relatively very weak Bernoulli over ,
-
(iii)
for any factor -algebra , is relatively very weak Bernoulli over ,
-
(iv)
any factor -algebra that is independent from is Bernoulli.
Proof.
We prove the lemma mainly by referring to the literature. The statement (i) follows from [26, Proposition 5] and Theorem 2.16. Then (ii) follows from [26, Proposition 4] and Theorem 2.16, and (iii) follows from (ii). Let us prove (iv): take a process on such that mod . From (ii), we know that is relatively very weak Bernoulli over . However, since is independent of , is independent of . One can then notice that if we add this independence in the definition of relative very weak Bernoullicity, we end up with the fact that is very weak Bernoulli. Finally, Theorem 2.14 tells us that is Bernoulli. ∎
We have just given many definitions and results concerning processes with finite alphabets, and the -algebras they generate. The following result from Krieger tells that it is applicable on any finite entropy system:
Theorem 2.18 (See [8]).
Let be an ergodic dynamical system and be a factor -algebra. If , there exists a finite alphabet and a random variable such that
We say that is a finite generator of .
2.5 Positive entropy systems and weak Pinsker filtrations
In 2018, Austin proved the following:
Theorem 2.19 (Austin, 2018, [1]).
Let be an ergodic dynamical system. For every there exists a factor -algebra such that:
-
•
,
-
•
is relatively Bernoulli over .
In other words, has the weak Pinsker property (as in (1)).
Definition 2.20.
Let be a dynamical system and a dynamical filtration on such that . We say that is a weak Pinsker filtration if
-
•
for every , is relatively Bernoulli over ,
-
•
and
Then, by iterating Austin’s theorem, we see that we can obtain weak Pinsker filtrations on any ergodic system:
Proposition 2.21.
Let be a dynamical system. If is ergodic, there exists a weak Pinsker filtration on . More specifically, for every increasing sequence such that and , there exists a weak Pinsker filtration such that , .
This simply tells us that weak Pinsker filtrations exist, but gives no explicit description. To start understanding those filtrations better, we can first link them to the Pinsker factor of the system:
Proposition 2.22.
Let be a dynamical system and a weak Pinsker filtration on . Then the tail -algebra is the Pinsker factor of .
Proof.
Let be a weak Pinsker filtration on . Since, for , , it follows that . Then, by taking , this yields . Therefore, .
Conversely, let us show that, for every , . Since is a weak Pinsker filtration, we can choose a Bernoulli factor -algebra such that
Then we use Lemma 2.8:
because, being Bernoulli, its Pinsker factor is trivial. ∎
Weak Pinsker filtrations are dynamical filtrations, and in Section 2.2, we introduced tools to classify dynamical filtrations, which we use here. While trying to connect the properties of a weak Pinsker filtration with the properties of the underlying system, we get the following simple results:
Theorem 2.23.
Let be a dynamical system and be a weak Pinsker filtration on . Then
-
(i)
is a K-system if and only if is Kolmogorovian, i.e. mod .
-
(ii)
If the filtration is of product-type, then is Bernoulli.
Proof.
We know that a system is K if and only if its Pinsker factor is trivial. Then the equivalence in (i) follows from Proposition 2.22.
We now prove (ii). Assume that is a weak Pinsker filtration of product type. This means that there exists a sequence of mutually independent factor -algebras such that . Let . We know that is relatively Bernoulli over and that is independent of . So, Lemma 2.17 tells us that is Bernoulli. Therefore, we have , which shows that we can write as a product of mutually independent Bernoulli factors. Hence, is Bernoulli. ∎
However, this result leaves many open questions. First, we can ask if the converse of (ii) is true. Since we remark at the end of Section 2.6 that, on a Bernoulli shift, there is at least one weak Pinsker filtration of product type, the converse of (ii) is equivalent to the uniqueness problem given in Question 2.27. Another area that is left open is to consider other properties from the theory of dynamical filtrations, like standardness or I-cosiness, and wonder what it implies of the system if a weak Pinsker filtrations has those properties:
Question 2.24.
What can we say about if there is a weak Pinsker filtration on that is standard ? In that case, is Bernoulli ? And if the weak Pinsker filtration is I-cosy ?
2.6 The uniqueness problem
Let be an ergodic dynamical system. As mentioned in Proposition 2.21, the fact that every ergodic systems satisfies the weak Pinsker property implies that, for any given increasing sequence that goes to in such that , there exits a weak Pinsker filtration on such that . But this filtration is not unique. Indeed, in the splitting result given by the weak Pinsker property (1), the choice of the factor -algebra generated by is not unique. For example, take a system of the form
where is a entropy system and and are Bernoulli shifts of equal entropy. Note that and generate two different factor -algebras on . But they are both factors over which is relatively Bernoulli, and they have the same entropy. However, we can notice in this example that and are isomorphic. This observation hints to a general result:
Theorem 2.25 (From Thouvenot in [25]).
Let and be ergodic dynamical systems and be a Bernoulli shift of finite entropy. If and are isomorphic, then and are isomorphic.
Proof.
This proof relies on the weak Pinsker property of and , and Lemma 2.17. We also use many times that Bernoulli shifts with the same entropy are isomorphic.
Since and are isomorphic, we have:
Set . We can then apply the weak Pinsker property of and to find two systems , and a Bernoulli shift such that
and
This implies
In other words, there is a system and two factor maps and such that is relatively Bernoulli over and relatively Bernoulli over . But then, Lemma 2.17 tells us that the factor -algebra is relatively very weak Bernoulli over and relatively very weak Bernoulli over . Therefore, there exist a Bernoulli shift and two factor maps and such that , and
This implies that
But, since we chose to have , we get
Given a last Bernoulli shift of entropy we get and
∎
As a consequence of this result, we see that if and are two weak Pinsker filtrations on such that, for all , , then we must have that, for each , .
However, this only gives “local isomorphisms”, and it does not necessarily mean that the filtrations and are isomorphic (according to the notion of isomorphism introduced in Definition 2.1). Therefore, the following is still an open question:
Question 2.26.
Let be an ergodic dynamical system. Are all weak Pinsker filtrations on with the same entropy isomorphic ?
This question is what we call the uniqueness problem.
If is a Bernoulli shift, and if we take an increasing sequence such that , we can take Bernoulli shifts such that , and define the system
It is a Bernoulli shift of entropy , so it is isomorphic to . Through this isomorphism, the factors of the form generate a product type weak Pinsker filtration on . Therefore, in the case where is a Bernoulli shift, the uniqueness problem becomes:
Question 2.27.
Let be a Bernoulli shift. Are all weak Pinsker filtrations on of product type ?
3 Uniqueness problem on Bernoulli systems
In this section, we present our efforts to tackle Question 2.27. The ideas developed here come from discussions with Jean-Paul Thouvenot, and we thank him for those insights. Specifically, we are going to show:
Theorem 3.1.
Let be a Bernoulli system and let be a weak Pinsker filtration. There exists some sub-sequence which is a weak Pinsker filtration of product type.
The fact that we are only able to describe the structure of a sub-sequence of , for now, seems to be significant. Indeed, we can compare that result to a well known result from Vershik about static filtrations on a probability space: any filtration whose tail -algebra is trivial has a sub-sequence that is standard (see [5, Theorem 3]). However there are many examples of non-standard filtrations with trivial tail -algebra. Therefore, although the context of Vershik’s result is very different, it emphasizes that Theorem 3.1 does not give a complete answer to Question 2.27.
The main step in proving Theorem 3.1 is contained in the following proposition:
Proposition 3.2.
Let be a Bernoulli system of finite entropy and a finite generator of , i.e. a finite valued random variable such that . For every , there exists such that, if is a factor -algebra such that is relatively Bernoulli over , and if , there is a Bernoulli factor -algebra such that
-
(i)
,
-
(ii)
mod ,
-
(iii)
and .
In this proposition, Krieger’ theorem (Theorem 2.18) ensures the existence of a finite generator since has finite entropy. The notation “”, which we use many times below, means that there exists a -measurable random variable such that .
The existence of a Bernoulli factor satisfying (i) and (ii) is simply the definition of relative Bernoullicity, the important part of this proposition is the ability to build a Bernoulli complement that satisfies (iii). Then iterating this result will yield Theorem 3.1 (see Section 3.3).
3.1 The technical lemma
In this section, we tackle the main technical and constructive part of the proof of Proposition 3.2. It is contained in Lemma 3.7.
In Section 2.4, we introduced the notion of very weak Bernoullicity, which gives a characterization of Bernoulli systems. Here, we use another equivalent notion: extremality, due to Thouvenot (see [27, Definition 6.3]).
Definition 3.3.
Let be an ergodic dynamical system and be a process where takes values in some finite alphabet . We say that is extremal if, for every , there exist and , such that for every and every random variable with , there is a set such that and for , we have:
where is the law of and is the law of given that equals .
In [27, Theorem 6.4], it is shown that extremality is equivalent to very weak Bernoullicity (and hence to Bernoullicity). In particular, we will use the fact that any process defined on a Bernoulli system is extremal.
The proof of Lemma 3.7 uses many methods that are usual in Ornstein’s theory of Bernoulli shifts (a presentation can be found in [16] or [22]). Therefore, we need to introduce some commonly used notions and results from that theory. The following combinatorial result is frequently used in Ornstein’s theory:
Lemma 3.4 (Hall’s marriage lemma [7]).
Let and be finite sets, and be a family of subsets of : . There exists an injective map such that if, and only if for every , we have
The main way in which the entropy of the processes is used in our arguments comes from the Shannon-McMillan-Breiman Theorem (see [3, Theorem 13.1]):
Theorem 3.5.
Let be an ergodic dynamical system and . For , define
We have
In particular, we also have the convergence in probability: for every , there exists such that for every , there exists a set such that and for every ,
We also need to introduce another tool that is commonly used in Ornstein’s theory: Rokhlin towers. On a dynamical system , to get a tower of height , we need a set such that the sets , for are disjoint. Then the family is what we call a Rokhlin tower, or, in short, a tower. However, we will also refer to the set as a tower. In particular, many times, we will write for . The following result guaranties that Rokhlin of arbitrary height and total measure almost exist under quite general conditions:
Proposition 3.6 (See [22]).
Let be an ergodic dynamical system and a finite valued random variable. Assume that is non-atomic. For all and , there exists a measurable set such that the sets , for , are disjoint, and .
The set is called the base of the tower and the sets are the levels. For any set , the family
is a tower, and we say that it is a column of . If is a random variable, we will be interested in the columns defined by sets of the form with . We say that is the -name of the column . The columns give a partition of the levels of . Now, conversely, assume that we have a partition of given by sets , then the columns give a partition of the levels of . If, moreover, we associate to each column a name of length , we can define a random variable on the levels of so that, for every , we have . We obtain this random variable simply by setting, for
This is the framework we will use to construct our random variables. We are now ready to turn our attention to the following:
Lemma 3.7.
Let be a Bernoulli system of finite entropy and a finite generator of . For every , there exists satisfying the following:
-
•
if is a finite valued random variable such that ,
-
•
and is a -valued (for some finite set ) i.i.d. process independent from such that mod ,
then for any , there exists a process such that
-
(i)
,
-
(ii)
,
-
(iii)
and .
The proof of the lemma being quite intricate, we start by giving a sketch of the proof. First, we will need a Rokhlin tower of very large height . This tower is then divided into the columns (see (10)) generated by . Each of those columns is then divided into sub-columns (see (14)) generated by . Because generates , we can approach by some random variable depending on finitely many coordinates of . It enables us to associate to each a word which gives a good approximation of on the levels of . We will define by giving a new -name, to replace . Our goal is to choose those names so that we can get a good approximation of by simply knowing the -name of , regardless of . To do that, we fix a column and use it as a “model” for the other columns. Then the extremality of comes into play: it tells us, for most choices of , the families and are quite similar. More specifically, we show that, for most , there are names such that is small. Those names are then suitable -names for . However, when we choose among those suitable names, we need to make sure that we are not giving the same name to too many columns, otherwise we might loose to much information, and we could not get (ii). This is done using Hall’s marriage lemma.
Proof of Lemma 3.7.
In this proof, we use many parameters, which we introduce below in a specific order to highlight the way they depend on each other:
-
(a)
Let . This parameter is chosen first, as it appears in the statement of the lemma. Then we choose and , as the numbers associated to in the definition of extremality of . We assume that .
-
(b)
Let . This is another arbitrarily small parameter that appears in the statement of the lemma. It does not depend on nor .
-
(c)
Next, we introduce , which must be small relative to and for (ii) and (iii) to hold. Specifically, we require that , and that the bound in Lemma 2.7 holds with error , whenever , for any random variables and .
-
(d)
Then we take , which is our most used parameter. We set satisfying the following:
Once is fixed, we choose such that .
-
(e)
Finally, we choose an integer , which will be the height of the Rokhlin tower. It is chosen larger than . We also need it to be large enough for us to apply the Shannon-McMillan-Breiman theorem, as well as Birkhoff’s ergodic theorem. As appears in many estimates, it needs to be large enough depending on , , , and . It would be quite tedious to give an explicit lower bound for , so, since all other parameters are now fixed and do not depend on , we simply point out throughout the proof the estimates where needs to be large.
Having now established the parameters, we begin the proof.
Step 1: The setup of the tower
As mentioned in (e), we choose so that we can apply the Shannon-McMillan-Breiman theorem (i.e. Theorem 3.5) and Birkhoff’s ergodic theorem to know that there exist two sets and such that
| (3) |
on which the estimates (5), (6), (8) and (15) hold. Latter in the proof, we will see that we can take and subsets such that
| (4) |
on which we also have (12) and (16). The fact that (12) holds for appears in Step 2 and the fact that (16) holds for appears in Step 3. Until then, we only use (4). For now, we present some of the estimates we have announced.
The first estimates given by the Shannon-McMillan-Breiman theorem are:
| (5) | |||
| (6) |
For any sequence and any element , denote the frequency at which the element appears in the sequence . This can also be defined as follows:
| (7) |
From this definition of , it becomes clear that, as announced earlier, the estimate given by Birkhoff’s ergodic theorem is:
| (8) |
Since generates , as said in (d), we can find so that . This means that there exists a -measurable random variable such that .
By making use of Proposition 3.6, we can build a set such that is disjoint from and is the base of a tower such that and
| (9) |
The set will be useful later to code the entrance of the tower. We slightly reduce the tower by setting and . One can then use (9) with our previous estimates to see that (by making sure that ).
We then split into -columns: for , we define
| (10) |
so that (we mean that the levels of are disjoint unions of the levels of ). For each , we say that is the column of -name . We also denote by the base of .
Step 2: Using the extremality of
We plan on modifying into a process so that the joint law of is almost the same in most of the columns . We start by using the fact that is Bernoulli to see that the law of is almost the same on each column . Indeed, since is Bernoulli, is extremal, and we fixed and as the numbers associated to in the definition of extremality and assume that (see (a)). On the other hand, from (5), we deduce that
Next we define the partition
In particular, we know that . Moreover, the number of values taken by is bounded by
since . Therefore the extremality of tells us that, since , there exists a subset such that
| (11) |
and for , we have
As mentioned at the start of the proof, the set is chosen so that (4) holds and we have
| (12) |
This is possible because . Therefore, using that with (12), this yields, for :
| (13) |
Step 3: Framework for the construction of
We start the construction of by setting on , where represents a symbol that does not belong to . Later in the proof, this will allow us to detect the entrance into from the value of the process . Then define to take any value in on the rest of . For , on , we set . We are left with defining our new random variable on the columns , with . We start by fixing , and the column will serve as a “model” for the other columns.
Next we fix an . We define sub-columns of : for ,
| (14) |
We say that the -name of . Because of our definition of and (9), the set gives us exactly the -names of all the sub-columns in . We will then give each sub-column a new word and define the random variable on as the only variable such that is the -name of . This means that to conclude the construction of on , we simply need to build a map and the properties we will obtain on will follow from our choice for .
In order to give us some additional leeway, we use the parameter introduced at the start of the proof: we define , and for , we denote by the truncated sub-sequence of of length . Conversely, for , define
and
We will obtain the map by building an injective map and setting . We start be recalling that we chose and so that the estimate given by the Shannon-McMillan-Breiman Theorem, i.e. (6), still holds when replacing by . More precisely, we mean that, for
| (15) |
Moreover, we stated at the start of the proof that is chosen such that, for
| (16) |
We need to prove that statement. We do this by considering the set
From the definition of , we get
Then, we define as , and easily get that . Next, because the set we removed from is measurable with respect to the truncated sequences of length , for , we get
so (16) follows from the definition of .
Step 4: Estimates for Hall’s marriage lemma
From (6) and (15), we can tell that
since and we can choose large enough. This inequality is also clearly true from the definition of , but we include this computation, as a similar one will be essential later in the proof. That being said, this inequality means that it is possible to find an injective map from to , but we want to be more specific about which injective map we choose. To that end, we will make use of Hall’s marriage lemma. To do that, for each , we need to specify which elements of we consider suitable -names for the columns .
We recall that is the integer chosen so that is -measurable. Define and the length of . Because is -measurable, is -measurable. So, for fixed, for each , on the set , there can be only one value of , which we denote .
For , the suitable corresponding -names will be the elements for which . More formally, we set
and we want to build so that we have
| (18) |
for as many as possible.
From (13), it follows that
Therefore:
by choosing large enough. So there exists a coupling of and such that
Denote by and the marginals of , i.e. and . We are interested in the set defined by
The following gives an estimate on the measure of :
so . In other words, if we set
we have . The set is the set on which we want (18) to hold. Hall’s marriage lemma tells us that there exists an injective map for which (18) is true if we have the following:
| (19) |
Let . Consider and note that
Taking that into account, we have
| (20) |
making sure again that is large enough. Moreover, using (17), we get
by definition of . Together with (20), it yields
since . Therefore there exists an injective map for which (18) holds. As we noted that , can then be extended to an injective map defined on (still taking values in ). We recall that, with built, we set .
As we announced at the start of our reasoning, we define on the levels of so that the -name of each sub-column is . Since this construction can be done with every (with the map depending on ), we have completed the construction of . We now need to check that satisfies the conditions (i), (ii) and (iii) of our lemma.
Step 5: Proving that satsifies (i), (ii) and (iii)
We start by estimating the law of . Since , we have
We recall that is the frequency at which the element appears in the sequence (see (7)). Moreover, one can see that, since is the -name of and all the levels of have the same measure, we have
Therefore, because takes values in , (8) yields:
This means that (using (d)).
We now turn our attention to the entropy of . The -name of a column is , and since is invective, we can deduce from the -name of . This means that, on the levels of the truncated tower , is -measurable. Indeed, if is in and the sequence is known, the sequence must contain a “”, which indicates the moment the past orbit of passes trough before entering . So the position of “” in tells us the index of the level of the point is on, which we call . In other words, we mean that . Then, gives the -name of the column is on, from which we deduce the truncated -name of length of the column. Finally, the -th letter of that name gives us .
Therefore, if we combine the previous paragraph with the fact that , there exists a -measurable random variable such that (because ). So, by the choice of made in (c), we can apply Lemma 2.7 and conclude that
Because generates , we also have the converse inequality
so we have proved that satisfies condition (ii) of our lemma.
We are now left with proving (iii). If we consider that is known, we deduce that, if the symbol “” appears in , then is in and the position of “” tells us the index of the level of the point is on. Then, using the notation introduced in our construction above, we can look at the random variable . It is -measurable and we are going to show that it satisfies
| (21) |
We start by looking at a column for some . We then split it into sub-columns . If , we are going to use (18). First, we need to remember that if is in , then gives the -name of the column . But, by construction, that name is , and, because we are looking at the case where and , (18) holds. So we have
We recall that and is its length. By definition of , we know that the number of levels on which we have is greater than . Moreover, by construction, for , on the -th level of , we have . Finally, since , we have
since and we can assume that is large enough so that . Moreover, the fact that implies that , and, combining it with (9), we can see that
Therefore
Next
Finally, was chosen so that , since , we have proven (21), and therefore, up to replacing by , we have shown that
∎
3.2 Application of the technical lemma
We are now left with proving Proposition 3.2 using Lemma 3.7. This is done using some abstract results from Thouvenot [26, Proposition 2’, Proposition 3]. We start by rewriting those results with our notation. We give a slight simplification, adapted to our setup.
First, [26, Proposition 2’] tells us that a process close enough to an i.i.d. process independent from in law and entropy can be turned into an i.i.d. process independent from .
Proposition 3.8.
Let be an ergodic system of finite entropy. Let be a finite valued process defined on and be a probability measure on a finite alphabet . For every , there exist such that if a random variable satisfies
-
(i)
,
-
(ii)
and ,
then there exists a random variable of law such that the process is i.i.d., independent from and we have
Next, [26, Proposition 3] tells us that on a system that is relatively Bernoulli over a factor , any i.i.d. process independent from with the right entropy can be turned into an independent complement of :
Proposition 3.9.
Let be an ergodic system, a finite valued process and a finite valued i.i.d. process independent from such that mod . For any and any i.i.d. process independent from such that , there exists such that , mod and
We are now fully equipped to end the proof of Proposition 3.2:
Proof of Proposition 3.2.
Let be a Bernoulli shift of finite entropy and a finite valued random variable such that . As we consider a factor -algebra of , it has finite entropy, therefore there exists a finite valued random variable such that the process generates . Lastly, we take an i.i.d. process independent from such that mod . Let .
Now, Lemma 3.7 tells us that there is for which, if , then for any , there is a random variable such that
-
(i)
,
-
(ii)
,
-
(iii)
and .
Denote a -measurable random variable such that . We can find an integer for which and set . If is chosen small enough, then Proposition 3.8 tells us that there is a random variable such that the process is i.i.d., independent from and we have . Finally, Proposition 3.9 tells us that we can then find a random variable for which the process is still i.i.d., independent from , but we also have that mod and . So we have .
Combining that with the fact that , we get that , so
Setting , we get the Bernoulli factor desired to prove our proposition. ∎
3.3 Proof of Theorem 3.1
In the previous section, we managed to conclude the proof of Proposition 3.2. We now see how Theorem 3.1 follows from that proposition:
Proof of Theorem 3.1.
Let be a Bernoulli system and be a weak Pinsker filtration. Since is a weak Pinsker filtration, if is of product type, so is . Therefore, up to replacing by the factor generated by , we can assume that has finite entropy. Thanks to Theorem 2.18, this means that we can set a finite alphabet and a random variable such that the corresponding process generates , i.e. mod . Let be a decreasing sequence of positive numbers such that .
We need to build a strictly increasing sequence such that is of product type. We start by setting . Since , we can choose large enough (in absolute value), so that is small enough for Proposition 3.2 to enable us to build a Bernoulli factor -algebra that is an independent complement of such that .
Now take and assume that we have built such that they are mutually independent Bernoulli factors such that for , is independent from , and we have
| (22) |
By construction of the , we know that is measurable with respect to . Moreover, using again Theorem 2.18, there is a random variable such that the process generates . So there exists an integer such that
| (23) |
Then set . As we did above, we choose large enough in absolute value so that is small enough for us to apply Proposition 3.2 to find a Bernoulli factor such that , and
| (24) |
Putting (23) and (24) together, we get
Iterating this for every ends our construction of and . Therefore (22) holds for every . It follows then that is measurable with respect to
Since the are factor -algebras, the full process is also -measurable. Finally, generates , so
Let , and set and . By construction, we have
We use this to see that if is -measurable, we have
which proves that
∎
4 Examples of weak Pinsker filtrations generated by a cellular automaton
Up to this point, we have discussed the existence and abstract properties of weak Pinsker filtrations. Now we want to give explicit examples to get a more concrete idea of what those objects can look like. We take inspiration from [10] and use cellular automata to generate our filtrations. We describe in the following paragraphs how this is done.
Let be a finite alphabet. A cellular automaton (or, more precisely, a deterministic cellular automaton) maps onto itself as follows: take finite, which we call a neighborhood, and a local map . Then define
Here, we will only consider examples in which . Therefore, our automata will be determined by a local map of the form . One can note that, by construction, cellular automata commute with the shift transformation
So we can consider a dynamical system of the form where is a -invariant measure, and note that the -algebra generated by is a factor -algebra. We can do better and iterate to generate a filtration:
In that case, each is a factor -algebra of , and therefore is a dynamical filtration. So, we see that cellular automata give a natural way to construct dynamical filtrations.
In fact, the theory of dynamical filtrations we presented in Section 2.2 was initiated in [10] in the setting of filtrations generated by cellular automata. However, the automata studied there preserve the product measure, and therefore the entropy of the associated factor -algebras will be the same for every . This prevents the filtration from being weak Pinsker.
Here, we will consider a different automaton: take a finite alphabet and assume that one element of is labeled <<>>. Then define the following local map
| (25) |
The associated automaton will eliminate isolated elements, replacing them with , and a maximal string of the form is replaced with . For example, if , this gives:
![[Uncaptioned image]](https://cdn.awesomepapers.org/papers/aea25f6d-8c5f-4c95-893d-784b2843ddad/x1.png)
Therefore, as we iterate the automaton, the proportion of <<>> increases as all other elements are gradually replaced by <<>>. Heuristically, this indicates that the entropy of the factor -algebras will go to zero as goes to infinity. But to state this rigorously, one need to specify the system on which we define . More accurately, it is the alphabet and the measure that need to be specified. However, the entropy goes to regardless of the choice of and :
Proposition 4.1.
Let , where is a -invariant measure and let be the coordinate process on . For every , we have
Proof.
Let be the set of values taken by . We know that
Because of the structure of , in , for , any run of <<>> is placed in between two runs of <<>> of length at least . Therefore, is either a sequence of <<>> or composed of one run of <<>> (with ) in between runs of <<>>. So
In conclusion
∎
In Section 4.1, we deal with the case where is a Bernoulli shift, and in Section 4.2, we deal with the case where is Ornstein’s example of a non-Bernoulli K-system from [15]. In both cases, by Proposition 4.1, the entropy of the filtration generated by the cellular automaton goes to zero. Then we look at each example separately to show the more involved result: each is relatively Bernoulli over . Therefore, we get two examples of weak Pinsker filtrations.
It is interesting to note that those two filtrations are very similar in their construction, but the filtration (or any sub-sequence) on Ornstein’s K-system cannot be of product type (otherwise, the system would be Bernoulli), we know from Theorem 3.1 that the latter has a sub-sequence that is of product type. It shows that there can be subtle differences in the asymptotic structure of weak Pinsker filtrations.
4.1 A cellular automaton on a Bernoulli shift
Here, we consider a Bernoulli shift where is a product measure. To avoid unnecessarily complicated notations, we will also assume that and . Therefore, the local function (25) becomes:
And we study the corresponding automaton:
The automaton replaces an isolated <<>> with a <<>> and reduces sequences of <<>> by replacing the final one by a <<>>.
Theorem 4.2.
On the system , the filtration given by is a weak Pinsker filtration. That is, for every , is relatively Bernoulli over and we have
| (26) |
The convergence of the entropy follows from Proposition 4.1. However, when is a Bernoulli shift, we can compute a better bound, as stated in Proposition 4.3.
Proposition 4.3.
Let denote the coordinate process on . For every , we have
Proof.
Let . One can see that is at if and only if is over the entire segment , as shown below:
![[Uncaptioned image]](https://cdn.awesomepapers.org/papers/aea25f6d-8c5f-4c95-893d-784b2843ddad/x2.png)
We set , and we remark that
Then, combining this with Fano’s inequality (Lemma 2.5) we get
and we can conclude for the KS-entropy:
∎
In addition, we give the following simple lemma on conditional independence:
Lemma 4.4.
Let be a probability space and a sub--algebra. Let , , and be random variables such that
Then we have
Proof.
It follows from the fact that if , , and are respectively , , and -measurable random variables:
∎
Proposition 4.5.
Let be the coordinate process on . For every , is relatively very weak Bernoulli over .
Proof.
Set . Relative very weak Bernoullicity was defined in Definition 2.15. We recall some notation: take to be the law of , and for and , is the conditional law of given that and .
Let . We need to show that there exists such that for every and for large enough, we have
| (27) |
Let . We start by noting that there must be some <<>> that appears in : indeed, the law of large numbers tells us that there exists such that
| (28) |
We then set . Next, we take so that determines entirely .
We fix . First, we note that, as we can see on the following image
![[Uncaptioned image]](https://cdn.awesomepapers.org/papers/aea25f6d-8c5f-4c95-893d-784b2843ddad/x3.png)
if , then is -measurable and is -measurable. So, since the variables are independent, given the variables and are independent. Finally, using Lemma 4.4, for such that , we get:
Therefore, if is chosen so that there exists such that , we see that and are independent given .
We are now ready to prove (27). For any and any such that there exists such that , the fact that and are relatively independent given implies that the measures and have the same marginal on the coordinates of . So the relative product of those measures over is a coupling under which the copies of coincide. It follows that
| (29) |
Proof of Theorem 4.2.
4.2 A cellular automaton on Ornstein’s K-process
Here, we consider the non-Bernoulli K-system introduced by Ornstein in [15]. A more detailed presentation of this system is given in [16, Part III], but we give a sketch of the construction for completeness. It is a process defined on the alphabet . We set , and to be integers depending on used in the construction of the process. For , an -block is a random sequence of length on the alphabet , whose law we define inductively.
To get a -block, take chosen uniformly at random, and consider a sequence that starts with a string of <<>>, followed by a string of <<>>, and ends with a string of <<>>:
![]()
This construction implies that .
To get an -block, take chosen uniformly at random, and i.i.d. random variables such that each is an -block. The -block is then built as follows:
![]()
So an -block starts with a string of <<>>, and ends with a string of <<>>. In between, we put all the -blocks separated by strings of <<>> so that each is placed in between two strings of <<>> of respective lengths and . In particular, is entirely determined by , and .
Ornstein’s K-system is then built by constructing an increasing sequence of towers such that . A tower is given by its base for which the sets are disjoint and
Through a cutting and stacking method, Ornstein builds in [15] the towers along with a process so that the law of given is the law of an -block. In other words, this means that the columns of the form
partition according to the law of an -block. Denote the resulting dynamical system. A proper choice of , and assures that this construction gives a finite measure. Then is a factor map onto the system
where is the law of .
Since is a process on the alphabet , the local function (25) becomes:
From now on, denotes the corresponding cellular automaton. Similarly to what we did in Section 4.1, we prove
Theorem 4.6.
On the system , the filtration given by is a weak Pinsker filtration. That is, for every , is relatively Bernoulli over and we have
| (30) |
The overall structure of the proof will resemble Section 4.1, but the details are adapted to the specific structure of Ornstein’s process. First, the convergence to of the entropy follows from Proposition 4.1. We could also adapt the proof of Proposition 4.3 to get that convergence, but it does not give a better rate of convergence than Proposition 4.1, so we do not give any details.
Proposition 4.7.
If is the process defined above, then for every , is relatively very weak Bernoulli over .
Proof.
We set . Let . Once again, we need to show that there exists such that for every and for large enough, we have
where is the law of and, for , is the conditional law of given that equals and that equals .
Let . We choose so that . By construction of , we know that for any -block in , there exists such that the said -block will come after a string of <<>> and be followed by a string of <<>>. Therefore, by knowing the positions of all the strings of <<>> longer that , we know the position of every -block.
However, since we chose to have , we can say that, for , we have if and only if . This means that the positions of the -blocks contained on a segment are -measurable, for large enough (for example ).
By choosing large enough, we can also assume that . Using Birkhoff’s ergodic theorem, for large enough, the set
satisfies .
In other words, for , the number of elements in the sequence that are part of an -block is greater than . However, among the intervals on which those -blocks are supported, two of them may not be included in , and can intersect . But, if , there are at most elements in those two intervals. To sum up, we get that the number of elements in the sequence that are part of an -block, and for which the position of that -block is contained on the segment , is greater than . Then, we choose so that the positions of the -blocks contained in are -measurable (in particular, is -measurable). So we have the following configuration for :
![[Uncaptioned image]](https://cdn.awesomepapers.org/papers/aea25f6d-8c5f-4c95-893d-784b2843ddad/x6.png)
where the are the positions of the -blocks supported on , and we have shown that
Denote by the random variable that gives the positions of the -blocks on the segment . By construction of , we know that, given , for any -block , the variables and are independent. Moreover, we know that any -block is between two strings of at least <<>>. Therefore, we see that if is fixed, for any -block , is -measurable and is -measurable.
Let us give details on the proof of that last claim: we write as the union of and , the infinite intervals that come before and after respectively. Given the structure of our automaton, it is always true that is -measurable. At the boundary between and , we have the following configuration:
![[Uncaptioned image]](https://cdn.awesomepapers.org/papers/aea25f6d-8c5f-4c95-893d-784b2843ddad/x7.png)
Indeed, in the construction of the blocks, we see that must put an <<>> in the first box of . Therefore, we must have <<>> in the red boxes. So, the values that takes on the boxes preceding are determined. For the rest of the boxes of , it comes from the structure of that the values of are determined by since we are at a distance from . So we have shown that is -measurable. A similar reasoning at the boundary between and shows that is -measurable. And since it is always true that is -measurable, we have proven that is -measurable and is -measurable.
But, we also know from the structure of that, given , and are independent. The previous paragraph enables us to use Lemma 4.4 to extend that to: given , and are independent. Finally, since is -measurable, this yields that and are relatively independent given .
This independence tells us that, for every sequences and , and have the same marginals on the coordinates of the -blocks contained in . Moreover, if is chosen so that is a subset of , we know that the positions of the -blocks cover at least elements in . Then, by considering the relative product of and over , we get:
Finally, since , this yields
∎
Remark 4.8.
We see that the proofs of Theorem 4.6 and Theorem 4.2 are very similar. In both cases, we have a process , whose conditional law given is made of random blocks separated by deterministic blocks, and the random blocks are filled independently from each other. The main difference that prevents Ornstein’s K-process from being Bernoulli is that the position of -blocks is determined by the long sequences of <<>>, and this creates correlations over long distances. But once we condition by , those sequences of <<>> are entirely determined. Therefore we are left with filling independently all the -blocks, and the past has no longer a significant influence on the future.
In that sense, when we look at the relative structure of Ornstein’s K-process over , the non-Bernoulli aspects disappear. However, when we look at the asymptotic properties of the weak Pinsker filtration obtained by applying , whether we start with a Bernoulli process or with a non-Bernoulli K-process, we get different results. Therefore, getting a better understanding of the classification of the various properties of weak Pinsker filtrations could help to develop a new classification of non-Bernoulli K-systems.
Acknowledgements
The author thanks Jean-Paul Thouvenot for the fruitful discussions, the insights and the ongoing exchanges regarding the present work.
References
- [1] Tim Austin. Measure concentration and the weak Pinsker property. Publ. Math., Inst. Hautes Étud. Sci., 128:1–119, 2018.
- [2] Séverin Benzoni. Confined extensions and non-standard dynamical filtrations. Stud. Math., 276(3):233–270, 2024.
- [3] P. Billingsley. Ergodic theory and information. John Wiley & Sons, Hoboken, NJ, 1965.
- [4] I. P. Cornfeld, S. V. Fomin, and Ya. G. Sinai. Ergodic theory. Transl. from the Russian by A. B. Sossinskii., volume 245. Springer, Berlin, 1982.
- [5] M. Émery and W. Schachermayer. On Vershik’s standardness criterion and Tsirelson’s notion of cosiness. In Séminaire de Probabilités XXXV, pages 265–305. Berlin: Springer, 2001.
- [6] Eli Glasner. Ergodic theory via joinings, volume 101. Providence, RI: American Mathematical Society (AMS), 2003.
- [7] Marshall Hall, Jr. Combinatorial theory. Wiley Classics Library. John Wiley & Sons, Inc., New York, second edition, 1998. A Wiley-Interscience Publication.
- [8] W. Krieger. On entropy and generators of measure-preserving transformations. Trans. Am. Math. Soc., 149:453–464, 1970.
- [9] Paul Lanthier. Aspects ergodiques et algébriques des automates cellulaires. PhD thesis, Université de Rouen Normandie, 2020.
- [10] Paul Lanthier and Thierry De La Rue. Classification of backward filtrations and factor filtrations: examples from cellular automata. Ergodic Theory and Dynamical Systems, 42(9):2890–2922, 2022.
- [11] D. Ornstein. Bernoulli shifts with the same entropy are isomorphic. Adv. Math., 4:337–352, 1970.
- [12] Donald Ornstein. Two Bernoulli shifts with infinite entropy are isomorphic. Adv. Math., 5:339–348, 1971.
- [13] Donald S. Ornstein. A K-automorphism with no square root and Pinsker’s conjecture. Adv. Math., 10:89–102, 1973.
- [14] Donald S. Ornstein. A mixing transformation for which Pinsker’s conjecture fails. Adv. Math., 10:103–123, 1973.
- [15] Donald S. Ornstein. An example of a Kolmogorov automorphism that is not a Bernoulli shift. Adv. Math., 10:49–62, 1973.
- [16] Donald S. Ornstein. Ergodic theory, randomness, and dynamical systems. Yale Mathematical Monographs. 5. New Haven - London: Yale University Press. VII, 141 p. £2.50 (1974)., 1974.
- [17] Donald S. Ornstein and Benjamin Weiss. Finitely determined implies very weak Bernoulli. Isr. J. Math., 17:94–104, 1974.
- [18] William Parry. Topics in ergodic theory. Reprint of the 1981 original, volume 75 of Camb. Tracts Math. Cambridge, MA: Cambridge University Press, 2004.
- [19] M. S. Pinsker. Dynamical systems with completely positive or zero entropy. Sov. Math., Dokl., 1:937–938, 1960.
- [20] M. Rahe. Relatively finitely determined implies relatively very weak Bernoulli. Can. J. Math., 30:531–548, 1978.
- [21] C. E. Shannon. A mathematical theory of communication. Bell System Tech. J., 27:379–423, 623–656, 1948.
- [22] Paul Shields. The theory of Bernoulli shifts. Chicago Lectures in Mathematics. Chicago - London: The University of Chicago Press. X, 118 p. £1.35; cloth £3.55 (1973)., 1973.
- [23] Ja. G. Sinaĭ. On a weak isomorphism of transformations with invariant measure. Mat. Sb. (N.S.), 63 (105):23–42, 1964.
- [24] J.-P. Thouvenot. On the stability of the weak Pinsker property. Isr. J. Math., 27:150–162, 1977.
- [25] J.-P. Thouvenot. Two facts concerning the transformations which satisfy the weak Pinsker property. Ergodic Theory Dyn. Syst., 28(2):689–695, 2008.
- [26] Jean-Paul Thouvenot. Quelques propriétés des systèmes dynamiques qui se decomposent en un produit de deux systèmes dont l’un est un schema de Bernoulli. Isr. J. Math., 21:177–207, 1975.
- [27] Jean-Paul Thouvenot. Entropy, isomorphism and equivalence in ergodic theory. In Handbook of dynamical systems. Volume 1A, pages 205–238. Amsterdam: North-Holland, 2002.