Consensus as a Nash Equilibrium of a stochastic differential game
Abstract
In this paper a consensus has been constructed in a social network which is modeled by a stochastic differential game played by agents of that network. Each agent independently minimizes a cost function which represents their motives. A conditionally expected integral cost function has been considered under an agent’s opinion filtration. The dynamic cost functional is minimized subject to a stochastic differential opinion dynamics. As opinion dynamics represents an agent’s differences of opinion from the others as well as from their previous opinions, random influences and stubbornness make it more volatile. An agent uses their rate of change of opinion at certain time point as a control input. This turns out to be a non-cooperative stochastic differential game which have a feedback Nash equilibrium. A Feynman-type path integral approach has been used to determine an optimal feedback opinion and control. This is a new approach in this literature. Later in this paper an explicit solution of a feedback Nash equilibrium opinion is determined.
keywords:
[class=MSC]keywords:
1 Introduction
Social networks influence a lot of behavioral activities including educational achievements (Calvó-Armengol, Patacchini and Zenou, 2009), employment (Calvo-Armengol and Jackson, 2004), technology adoption (Conley and Udry, 2010), consumption (Moretti, 2011) and smoking (Nakajima, 2007; Sheng, 2020). As social networks are the result of individual decisions, consensus takes an important role to understand the formation of networks. Although a lot of theoretical works on social networks have been done (Jackson, 2010; Goyal, 2012; Sheng, 2020), work on consensus as a Nash equilibrium under a stochastic network is very insignificant (Niazi, Özgüler and Yildiz, 2016). Sheng (2020) formalizes network as simultaneous-move game, where social links based on decisions are based on utility externalities from indirect friends. Sheng (2020) proposes a computationally feasible partial identification approach for large social networks. The statistical analysis of network formation dates back to the seminal work by Erdös and Rényi (1959) where a random graph is based on independent links with a fixed probability (Sheng, 2020). Beyond Erdös-Rényi model, many methods have been designed to simulate graphs with characteristics like degree distributions, small world, and Markov type properties (Polansky and Pramanik, 2021). A model based method is useful if this model can be fit successfully and if it is a relatively simple to simulate realizations (Pramanik, 2016; Polansky and Pramanik, 2021). The most frequently used general model for random graphs is the exponential random graph model (ERGM) (Snijders, 2002; Hua, Polansky and Pramanik, 2019; Polansky and Pramanik, 2021) because, this model fits well with the observed network statistics (Sheng, 2020). This ERGM model lacks microfoundations which are important for counterfactual analyses and furthermore, economists view network analysis as the optimal choices of agents who maximizes their utilities (Sheng, 2020). Network evolves as the result of a stochastic process is another popular framework where network may be observed, but it is the parameters of the stochastic process that are of interest, and the observed network is a single realization of the stochastic process (Polansky and Pramanik, 2021).
At birth, humans already posses different types of skills like breathing, digest foods and motor actions which make a human body to behave like an automaton (Kappen, 2007a). Furthermore, like other animals humans acquire skills through learning. Different person has different abilities to acquire a new information in order to get an idea about pleasure, danger or food (Kappen, 2007a). Humans are the most complicated species on earth because, their decisions are not linear and they can learn difficult skills through transitional signals of their complex constellations of sensory patterns. For example, if food is kept in front of a hamster, it would eat immediately. On the other hand, if a plate of food is kept in front of a human, they might not eat because, variety of factors such as the texture, smell, amount of it, their sociocultural background, religion and ethnicity take place before even they think about to tastes it. In order to make this decision, a lot of complex neural activities take place inside a person’s brain. Action of two main parts of a human brain, frontal and occipital lobes, makes them decide what they should do after seeing an object. In this case the occipital lobe sends information of an object through the synaptic systems to frontal lobe, which is known by their previous experiences and knowledge. As for humans one has to consider so many other possibilities compared to a hamster, such that they can choose any of the all available information with some probabilities and make decisions based on it.
This type of human behavior is a feedback circuit where the learning algorithm is determined by a synaptic motor command, more time with an object not only leads to get more information but also the knowledge to adapt with it in the long run and get more intelligence. For example, as humans grow older, more intelligent they become and reflects their genotype closely. On the other hand, environment influences a certain type of decision more with older ages which comes through a process called Hebbian learning (Hebb, 2005; Kappen, 2007a). Ancestors gather more information about an object or circumstance and transfer it to their off-springs in order to help them survive easily and make decisions rationally Kappen (2007a). For example, without having a prior knowledge one does not know how to get a certain type restaurant and which lead them explore their surroundings. If that person finds out a restaurant, they survive for that day. On the next day, they might not be completely sure about full availability of food in that restaurant because of sudden environmental degradation after his previous visit such as flash flood, tornado, an avalanche, landslide, earthquake, other activities like closure due to burglary, fire or some gun related activities so on. Even if that person is sure about the availability of food, they might not go because of other socioeconomic behaviors at the back of their mind. Hence, more information might not lead them react rationally. These types of activities occurs when an event is more uncertain. Consider person is selling their million-dollar car to another person by just . The rationality assumption suggests person to go for this offer but, might think why is giving this offer and might be suspicious about the quality of that car and rejects it.
Therefore, subjective probabilities take an important role to make these types of decisions based on individual judgments such as success in a new job, outcome of an election, state of an economy, difference in learning a new complex topic among students, spreading gossips in small communities (Kahneman and Tversky, 1972; Niazi, Özgüler and Yildiz, 2016; Tversky and Kahneman, 1971). People follow representativeness in judging the likelihood of uncertain events where the probability is defined by the similarities in essential properties to its parent population and reflect the salient features of the process by which it is generated (Kahneman and Tversky, 1972), which makes opinion dynamics of a person to follow a stochastic differential equation. Furthermore, an individual minimizes its cost of foraging for food where finding food can be termed as a reward to them and they want to find their reward with minimal cost. Assume an agent discounts more to the recent future than farther future represented as feedback motor control reinforcement learning problem Kappen (2007a). In an environment of very complex opinion dynamics each agent minimizes their integral cost function subject to a stochastic differential opinion dynamics based on all above cases. This paper considers two environments first, all the agents have same opinion power and second, agents with a leader, where the leader has more power in opinion than others and determines their opinion first based on their own cost minimization mechanism. A feedback Nash equilibrium of opinion is determined by a Feynman-type path integral approach which so far from my knowledge is new (Feynman, 1949; Pramanik and Polansky, 2021). Furthermore, this approach can be used to obtain a solution for stability of an economy after pandemic crisis (Ahamed, 2021a), determine an optimal bank profitability (Hossain and Ahamed, 2015; Ahamed, 2021b).
As each agent’s opinion in a society is assumed to be a quantum particle, I introduce an alternative method based on Feynman-type path integral to solve this stochastic opinion dynamics problem based on Feynman-type path integrals instead of traditional Pontryagin Maximum Principle. If the objective function is quadratic and the differential equations are linear, then solution is given in terms of a number of Ricatti equations which can be solved efficiently (Kappen, 2007b). But the opinion dynamics is more complicated than just an ordinary linear differential equation and non-linear stochastic feature gives the optimal solution a weighted mixture of suboptimal solutions, unlikely in the cases of deterministic or linear optimal control where a unique global optimal solution exists (Kappen, 2007b). In the presence of Wiener noise, Pontryagin Maximum Principle, a variational principle, that leads to a coupled system of stochastic differential equations with initial and terminal conditions, gives a generalized solution (Kappen, 2007b; Øksendal and Sulem, 2019). Although incorporate randomness with its Hamiltonian-Jacobi-Bellman (HJB) equation is straight forward but difficulties come due to dimensionality when a numerical solution is calculated for both of deterministic or stochastic HJB (Kappen, 2007b). General stochastic control problem is intractable to solve computationally as it requires an exponential amount of memory and computational time because, the state space needs to be discretized and hence, becomes exponentially large in the number of dimensions (Theodorou, Buchli and Schaal, 2010; Theodorou, 2011; Yang et al., 2014). Therefore, in order to calculate the expected values it is necessary to visit all states which leads to the summations of exponentially large sums (Kappen, 2007b; Yang et al., 2014). Kappen (2005a) and Kappen (2005b) say that a class of continuous non-linear stochastic finite time horizon control problems can be solved more efficiently than Pontryagin’s Maximum Principle. These control problems reduce to computation of path integrals interpreted as free energy because, of their various statistical mechanics forms such as Laplace approximations, Monte Carlo sampling, mean field approximations or belief propagation (Kappen, 2005a, b, 2007b; Van Den Broek, Wiegerinck and Kappen, 2008). According to Kappen (2007b) these approximate computations are really fast.
Furthermore, one can transform a class of non-linear HJB equations into linear equations by doing a logarithmic transformation. This transformation stems back to the early days of quantum mechanics which was first used by Schrödinger to relate HJB equation to the Schrödinger equation (Kappen, 2007b). Because of this linear feature, backward integration of HJB equation over time can be replaced by computing expectation values under a forward diffusion process which requires a stochastic integration over trajectories that can be described by a path integral (Kappen, 2007b). Furthermore, in more generalized case like Merton-Garman-Hamiltonian system, getting a solution through Pontryagin Maximum principle is impossible and Feynman path integral method gives a solution (Baaquie, 1997; Pramanik, 2020, 2021a). Previous works using Feynman path integral method has been done in motor control theory by Kappen (2005b), Theodorou, Buchli and Schaal (2010) and Theodorou (2011). Applications of Feynman path integral in finance has been discussed rigorously in Baaquie (2007). In Pramanik (2020) a Feynman-type path integral has been introduced to determine a feedback stochastic control. This methods works in both linear and non-linear stochastic differential equations and a Fourier transformation has been used to find out solution of Wick-rotated Schrödinger type equation (Pramanik, 2020; Pramanik and Polansky, 2020a, b; Pramanik, 2021a; Pramanik and Polansky, 2021; Pramanik, 2021b). A more generalized Nash equilibrium on tensor field has been discussed in Pramanik and Polansky (2019).
2 A stochastic differential game of opinion dynamics
Following Niazi, Özgüler and Yildiz (2016) consider a social network of agents by a weighted directed graph , where is the set of all agents. Suppose, is the set of all ordered pairs of all connected agents and, is the influence of agent on agent for all . There are usually two types of connections, one sided or two sided. For the principle-agent problem the connection is one sided (i.e. Stackelberg model) and agent-agent problem it is two sided (i.e. Cournot model). Suppose be the opinion of agent at time with their initial opinion . Then has been normalized into where stands for a strong disagreement and represents strong agreement and all other agreements stays in between. Consider be the opinion profile vector of -agents at time where ‘prime’ represents the transpose. Following Niazi, Özgüler and Yildiz (2016) consider a cost function of agent as
(1) |
where is a parameter which weighs the susceptibility of agent to influence agent , is agent ’s stubbornness, is the control variable of agent and set of all agents with whom interacts is and defined as . The cost function is twice differentiable with respect to time in order to satisfy Wick rotation, is continuously differentiable with respect to agent’s control , non-decreasing in opinion , non-increasing in , and convex and continuous in all opinions and controls (Mas-Colell et al., 1995; Pramanik and Polansky, 2020b). The opinion dynamics of agent follows a stochastic differential equation
(2) |
with the initial condition , where and are the drift and diffusion component of agent with is the Brownian motion. The reason behind incorporating Brownian motion in agent ’s opinion dynamics is because of Hebbian Learning which states that, neurons increase the synaptic connection strength between them when they are active together simultaneously and this behavior in probabilistic in the sense that, resource availability from a particular place is random (Hebb, 2005; Kappen, 2007a). For example, for a given stubbornness, and influence from agent , agent ’s opinion dynamics has some randomness in opinion. Suppose, from other resources agent knows that, the information provided by agent ’s influence is misleading. Apart from that after considering humans as automatons, motor control and foraging for food becomes a big examples of minimization of costs (or the expected return) Kappen (2007a). As control problems like motor controls are stochastic in nature because there is a noise in the relation between the muscle contraction and the actual displacement with joints with the change of the information environment over time, we consider Feynman path integral approach to calculate the stochastic control after assuming the opinion dynamics Equation (2) (Feynman (1949),Fujiwara (2017)). The coefficient of the control term in Equation (1) is normalized to , without loss of generality. The cost functional represented in the Equation (1) is viewed as a model of the motive of agent towards a prevailing social issue Niazi, Özgüler and Yildiz (2016). In this dynamic social network problem agent ’s objective is to subject to the Equation (2), where represents the expectation on at time subject to agent ’s opinion filtration starting at the initial time . A solution of this problem is a feedback Nash equilibrium as the control of agent is updated based on the opinion at the same time .
3 Definitions and Assumptions
Assumption 1.
For and , let and be some measurable function and, for some constant and, for opinion the linear growth of agent ’s control as
such that, there exists another constant and for a different such that the Lipschitz conditions,
and
hold.
Assumption 2.
Agent faces a probability space with sample space , -adaptive filtration at time of opinion as , a probability measure and -dimensional Brownian motion where the control of agent is an adapted process such that Assumption 1 holds, for the feedback control measure of agents in a society there exists a measurable function such that for which such that Equation (2) has a strong unique solution.
Assumption 3.
(i). such that agent cannot go beyond set because of their limitations of acquiring knowledge from their society at a given time. This immediately implies set is different for different agents. If the agent is young , they would have less limitation to acquire new information and make opinions on it.
(ii). The function . Therefore, all agents in a society at the beginning of have the cost function such that in functional spaces and both of them are concave which is equivalent to Slater condition (Marcet and Marimon, 2019). Possibility of giving a partial opinion has been omitted in this paper.
(iii). There exists an with for all and such that
The opinion dynamics of agent is continuous and it is mapped from an interval to a space of continuous functions with initial and terminal time points. Suppose, at time , represents an opinion dynamics of agent with initial and terminal points and respectively, such that, the line path integral is , where is derivative with respect to . In this paper I consider functional path integrals where the domain of the integral is the space of functions (Pramanik, 2020, 2021a; Pramanik and Polansky, 2021). Functional path integrals are very popular in probability theory and quantum mechanics. In Feynman (1948) theoretical physicist Richard Feynman introduced Feynman path integral and popularized it in quantum mechanics. Furthermore, mathematicians develop the measurability of this functional integral and in recent years it has become popular in probability theory (Fujiwara, 2017). In quantum mechanics, when a particle moves from one point to another, between those points it chooses the shortest path out of infinitely many paths such that some of them touch the edge of the universe. After introducing equal length small time interval with such that and using Riemann–Lebesgue lemma if at time one particle touches the end of the universe, then at a later time point it would come back and go to the opposite side of the previous direction to make the path integral a measurable function (Bochner et al., 1949). Similarly, agent has infinitely opinions, out of them they choose the opnion corresponding to least cost given by the constraint explained in Equation (2). Furthermore, the advantage of Feynman approach is that, it can be used in both in linear and non-linear stochastic differential equation systems where constructing an HJB equation is almost impossible (Baaquie, 2007).
Definition 1.
Suppose, for a particle be the Lagrangian in classical sense in generalized coordinate with mass where and are kinetic and potential energies respectively. The transition function of Feynman path integral corresponding to the classical action function
is defined as , where and is an approximated Riemann measure which represents the positions of the particle at different time points in (Pramanik, 2020).
Here agent’s objective is to minimize Equation (1) subject to Equations (2). Following Definition 1 the quantum Lagrangian at time of is
(3) |
where is a time independent quantum Lagrangian multiplier (one can think of as a penalization constant of agent ). As at the beginning of the small time interval , agent does not have any future information, they make expectations based on their opinion . For another normalizing constant and for time interval such that define a transition function from to as
(4) |
where is the value of the transition function based on opinion at time with the initial condition . Therefore, the action function of agent is,
where such that,
Here the action function has the notation which means within the action of agent depends on their opinion and furthermore, I assume this system has a feedback structure. Therefore, the opinion of agent also depends on the strategy as well as the rest of the school. Same argument goes to the transition function .
Definition 2.
For agent optimal opinion and their continuous optimal strategy constitute a dynamic stochastic Equilibrium such that for all the conditional expectation of the cost function is
with the opinion dynamics explained in Equation (2), where is the optimal filtration starting at time such that, .
4 Main results
Suppose, for the opinion space and agent ’s strategy space there exists a permissible strategy and for all define the integrand of the cost function as
Proposition 1.
For stochastic dynamic game of -agents of time interval , let for agent
(i) the feedback control is a continuously differentiable function,
(ii) The cost integrand is a function on for all .
If is a feedback Nash equilibrium and is the opinion trajectory, then there exists Lagrangian multipliers with initial condition such that, for a Lagrangian
with its Euclidean action function
the following conditions hold: (a) , and (b) with . Under this case, the optimal feedback control will be the solution of the following equation
where for a function
Proof.
Equation (2) implies
(5) |
Following Chow (1996) we get our Euclidean action function as
where is the conditional expectation on opinion at the beginning of time . Now, for a small change in time , and for agent ’s normalizing constant , define a transitional wave function in small time interval as
(6) |
for and is the value of the transition function at time and opinion with the initial condition for all .
For the small time interval where the Lagrangian can be represented as,
(7) |
with the initial condition . This conditional expectation is valid when the control of agent ’s opinion dynamics is determined at time and the opinions of all -agents is given (Chow, 1996). The evolution of a process takes place as if the action function is stationary. Therefore, the conditional expectation with respect to time only depends on the expectation of initial time point of interval .
Define , then Fubini’s Theorem implies,
(8) |
By Itô’s Theorem there exists a function such that where is an Itô process (Øksendal, 2003). After assuming
Equation (4) becomes,
(9) |
For a very small interval around time point with , and Itô’s Lemma yields,
(10) |
where , and , and we use the condition with . We use Itô’s Lemma and a similar approximation to approximate the integral. With , dividing throughout by and taking the conditional expectation we get,
(11) |
as and as with the initial condition . For the transition function at is for all . Hence, using Equation (6), the transition function for is
(12) |
As , first order Taylor series expansion on the left hand side of Equation (12) gives
(13) |
For fixed and let so that . When is not around zero, for a positive number we assume so that for , takes even smaller values and agent ’s opinion . Therefore,
Before solving for Gaussian integral of the each term of the right hand side of the above Equation define a function
where is a vector of all -agents’ ’s. Hence,
(14) |
After taking , and a Taylor series expansion with respect to of gives,
Define so that . First integral on the right hand side of Equation (4) becomes,
(15) |
Assuming and the argument of the exponential function in Equation (4) becomes,
(16) |
Therefore,
(17) |
and
(18) |
Substituting second integrand of the right hand side of Equation (4) yields,
(19) |
Substituting in Equation (4) we get,
(20) |
Hence,
(21) |
Using results of Equations (4), and (4) into Equation (4) we get,
(22) |
As is in Schwartz space, derivatives are rapidly falling and assuming , and we get,
such that
Therefore, Wick-rotated Schrödinger type Equation for agent is,
(23) |
Differentiating the Equation (23) with respect to gives us optimal control of agent under this stochastic opinion dynamics which is
(24) |
where , , and . Therefore, optimal feedback control of agent in stochastic opinion dynamics is represented as and is found by setting Equation (24) equal to zero. Hence, is the solution of the following Equation
(25) |
∎
Proposition 2.
For the initial condition the Wick-rotated Schrödinger-type equation of agent
where and , has a unique solution
The optimal opinion can be found after solving the following equation,
(26) |
and corresponding feedback control Nash equilibrium is .
Proof.
Let for three variables and generalized Wick-rotated Schrödinger type equation for agent is,
(27) |
with the initial condition . As agent ’s wave function is a function of opinion for fixed control , the solution to Equation (27) is found by assuming and vary according to the movement of ’s only. Define , and . Hence,
(28) |
For a , the Fourier transformation of is,
(29) |
As then assuming as , Equation (29) gives, . Therefore, and, . Rearranging terms in Equation (28) and Fourier transformation with above conditions give,
(30) |
Let us assume an integrating factor which can be written as Therefore,
or equivalently
so that
(31) |
Integrating both sides of Equation (31) yields,
(32) |
Applying the Fourier transformation on the initial condition yields, which implies . Using this condition Equation (4) gives,
(33) |
where for all . Fourier Inversion Theorem yields,
(34) |
As the Fourier transformation is the product of two Fourier transformations, therefore the Convolution Theorem implies that for and ,
and
for all . Hence, a solution to the Equation (27) is,
(35) |
If one compares Wick-rotated Schrödinger type Equation (23) with (27) we find out and other terms vanishes. Therefore, Equation (34) becomes
(36) |
where is the Dirac -function of the opinion of agent . Now,
Suppose, such that for all . Then
Therefore, the solution to Equation (23) is where . After using this solution to the wave function into Wick-rotated Schrödinger type Equation (23) we get,
and differentiating with respect to gives
(37) |
Optimal opinion of agent , can be found after solving the Equation (37) and an optimal feedback control is obtained. ∎
Corollary 1.
Define for all . As each player has an optimal opinion , is an optimal opinion vector. Furthermore,
is an optimal control vector of feedback Nash equilibrium.
After combining the opinion state variables and the Lagrangian multipliers, the following equation is obtained
where
where is the identity matrix of size , ,
, is an vector, is an -dimensional matrix is an -dimensional Brownian motion corresponding to opinion and is an dimensional Brownian motion of the Lagrangian multiplier. Following Niazi, Özgüler and
Yildiz (2016)
with . is a Laplacian-like matrix of a weighted directed gaph (Niazi, Özgüler and Yildiz, 2016) where element in the off-diagonal shows the weight of the edge directed from to . Define , , , and . Then we get the following equation
(38) |
Following Øksendal (2003) we get a unique solution of the stochastic differential equation expressed in Equation (38) as
(39) |
5 Stochastic differential games with an explicit feedback Nash equilibrium
Propositions 1 and 2 states that, for agent and given one can get a optimal Nash feedback control and for a unique solution of the transition wave function the unique opinion dynamics is . In this section I am considering two main consensus: full consensus or complete information and consensus under a leader who can influence other agents’ opinions.
First, consider the consensus under complete information. Let there be a network where all agents are connected with each other or . As every agent has equal power to influence others, in the long run a consensus will be eventually reached. As some agents are stubborn, their opinions might not be influenced by others and a full consensus is not reached. Following Niazi, Özgüler and Yildiz (2016) assume all the parameters of agent ’s cost function are equal or , for all and where agent ’s stochastic opinion dynamics is represented by
(40) |
where , and is a constant diffusion component. In Equation (40) is the optimal opinion of agent according to agent because, under complete information agent has the information of all possible reaction functions of agent but does not know what reaction function agent will play. Therefore, agent assumes agent is rational and calculates optimal opinion . Opinion trajectory explained in Equation (40) has drift part and a diffusion part. The drift part has three components, the first component is the average of optimal opinions of all the agents in the network, the second term depends on the difference between the opinion of agent at time and the average and the third component is the control of agent . As control is the cost of agent in the opinion dynamics, it comes with a negative sign at the front. I do not consider other agents’ controls in Equation (40) because, I assume all of the agents’ control in this network are independent to each other.
Proposition 3.
Suppose agent minimizes the objective cost function
subject to the stochastic opinion dynamics expressed in Equation (40). For , define .
(i) Then for
a feedback Nash Equilibrium control of opinion dynamics
(41) |
where
, , , , , , and .
(ii) For a unique solution of the wave function as expressed in Proposition 2 and is a function with respect to , an optimal opinion is obtained by solving following equation
(42) |
which is
(43) |
where
and
(iii) The opinion difference between agents and at time is
where , and .
Proof.
(i). Let , for a finite and with , and . Hence, Proposition 1 implies,
(44) |
Now
(45) |
and,
(46) |
Therefore, Equation (25) implies
and we get a cubic polynomial with respect to control
(47) |
where , , , , , , and . Therefore, Equation (47) gives feedback Nash equilibrium of control
(48) |
where
and
(ii). In order to prove the second part let us use Proposition 2. The right hand side of Equation (26) becomes,
(49) |
the left hand side implies
and
(50) |
Matching Equations (5) and (5) we get,
or,
(51) |
As , for , fixed and a very small value of it can be approximated as where assume .
Therefore,
(52) |
After rearranging terms of Equation (5) we get a cubic polynomial opinion of agent
(53) |
where
and
After solving Equation (53) we get a set of optimal opinions for agent
(54) |
where
and
(iii). The integral forms of opinions of agents and obtained from the Equation (40) are
and
where and are the initial opinions of agents and . As agents and comes from the same population hence, . Subtracting from gives,
and taking absolute value on both sides and using triangle inequality we get,
where , and . ∎
Consider the consensus with a leader (agent ) under complete information. It might be a network where agent , the political analyst who can influence the decision of the rest of the agents but not the other way. Furthermore, I also assume that, before a game starts the leader makes their optimal opinion based on the history of the network and their perspective of opinion performance of other agents. Once agent optimizes their opinion at the beginning of the game, they never change their mind and influences in other agents’ decisions. Therefore, leader’s cost functional is defined as,
(55) |
where is a parameter assigned by agent to weight the susceptibility of agent to influence them before the game starts, is a finite positive constant which measures the stubbornness of the leader, is the opinion control and be the fixed opinion values of the other agents according to agent . The reason behind the assumption is that, the leader is a rational person and they want to get more return out of this network than any other agent and assigns an opinion which is less than agent ’s optimal opinion before a game starts. Opinion dynamics of the leader (agent ) is
(56) |
where , and is a constant diffusion component of the leader. Therefore a leader’s problem is to minimize the expected cost functional with respect to their control and opinion subject to the Equation (56). Proposition 3 implies,
Corollary 2.
Suppose the leader (agent ) has the objective cost function
subject to the stochastic opinion dynamics expressed in Equation (56). For , define .
(i) Then for
an optimal control of the leader
(57) |
where
, , , , , , and .
(ii) For a unique solution of the leader’s wave function and is a function with respect to , a leader’s optimal opinion is obtained by solving following equation
which is
(58) |
where
and
As in Corollary 2 optimal opinion of agent is a solution of a cubic equation takes three values and because of rationality he chooses that which has the maximum value. If then optimal opinion of the leader is . Under complete information all the other agents has the information about before a game starts and adjusts their opinions on it. The network is represented by a direct graph with edges directed from all the agents towards the leader. Thus , (Niazi, Özgüler and Yildiz, 2016). Each of other agents represented by minimizes the expectation of his cost functional expressed in Equation (1) where if , subject to his stochastic opinion dynamics
(59) |
where for all , , , is the control of opinion, is a constant diffusion component and the Brownian motion of agent . In this framework we assume that, apart from the leader other agents have very small influence in agent’s opinion.
Proposition 4.
Suppose, there is a network where all agents are unilaterally connected to their leader. Let agent minimizes his objective cost function
(60) |
subject to the stochastic opinion dynamics expressed in Equation (59). For , define .
(i) Then for
we have a feedback Nash Equilibrium control of opinion dynamics
(61) |
where
, , , , , , and .
(ii) For a unique solution of the wave function as expressed in Proposition 2 and is a function with respect to , an optimal opinion is obtained by solving following equation
which is
(62) |
where
and
Proof.
(i). For , let , , and . Hence, Proposition 1 implies,
Now
and,
Therefore, Equation (25) implies
and the cubic polynomial of agent with respect to control under the presence of a leader is
where , , , , , , and . Therefore, feedback Nash equilibrium control under the presence of a leader is
where
(ii). Using Proposition 2 the right hand side of Equation (26) becomes,
(63) |
the left hand side implies
and
(64) |
Comparing Equations (5) and 5 we get,
The polynomial of the opinion dynamics is
or,
(65) |
As in Equation (5) , for , fixed and a very small value of it can be approximated as where assume .
Therefore, we get a cubic equation expressed as,
(66) |
where
and
Solving Equation (5) gives agent ’s optimal opinion
(67) |
where
and
∎
6 Discussion
This paper shows consensus as a feedback Nash equilibrium from a stochastic differential game. The same integral cost function has been used as in Niazi, Özgüler and Yildiz (2016) subject to a stochastic opinion dynamics. A Feynman-type path integral approach has been used to construct a Wick-rotated Schrödinger type equation (i.e a Fokker-Plank diffusion equation). Finally, optimal opinion and control have been determined after solving the first order condition of the Wick-rotated Schrödinger equation. So far from my knowledge, this is a new approach. As different people have different opinions, an opinion changes over time and stubbornness and influence from others have some effects on individual decisions under the assumption that human body is a automaton. The fundamental assumption of this paper is opinion dynamics is stochastic in nature which is another contribution of this paper. Furthermore, results of this paper give more generalized solution of opinion dynamics than (Niazi, Özgüler and Yildiz, 2016).
References
- Ahamed (2021a) {barticle}[author] \bauthor\bsnmAhamed, \bfnmFaruque\binitsF. (\byear2021a). \btitleMacroeconomic Impact of Covid-19: A case study on Bangladesh. \bjournalIOSR Journal of Economics and Finance (IOSR-JEF) \bvolume12 \bpages2021. \endbibitem
- Ahamed (2021b) {barticle}[author] \bauthor\bsnmAhamed, \bfnmFaruque\binitsF. (\byear2021b). \btitleDeterminants of Liquidity Risk in the Commercial Banks in Bangladesh. \bjournalEuropean Journal of Business and Management Research \bvolume6 \bpages164–169. \endbibitem
- Baaquie (1997) {barticle}[author] \bauthor\bsnmBaaquie, \bfnmBelal E\binitsB. E. (\byear1997). \btitleA path integral approach to option pricing with stochastic volatility: some exact results. \bjournalJournal de Physique I \bvolume7 \bpages1733–1753. \endbibitem
- Baaquie (2007) {bbook}[author] \bauthor\bsnmBaaquie, \bfnmBelal E\binitsB. E. (\byear2007). \btitleQuantum finance: Path integrals and Hamiltonians for options and interest rates. \bpublisherCambridge University Press. \endbibitem
- Bochner et al. (1949) {bbook}[author] \bauthor\bsnmBochner, \bfnmSalomon\binitsS., \bauthor\bsnmChandrasekharan, \bfnmKomaravolu\binitsK. \betalet al. (\byear1949). \btitleFourier transforms \bvolume19. \bpublisherPrinceton University Press. \endbibitem
- Calvo-Armengol and Jackson (2004) {barticle}[author] \bauthor\bsnmCalvo-Armengol, \bfnmAntoni\binitsA. and \bauthor\bsnmJackson, \bfnmMatthew O\binitsM. O. (\byear2004). \btitleThe effects of social networks on employment and inequality. \bjournalAmerican economic review \bvolume94 \bpages426–454. \endbibitem
- Calvó-Armengol, Patacchini and Zenou (2009) {barticle}[author] \bauthor\bsnmCalvó-Armengol, \bfnmAntoni\binitsA., \bauthor\bsnmPatacchini, \bfnmEleonora\binitsE. and \bauthor\bsnmZenou, \bfnmYves\binitsY. (\byear2009). \btitlePeer effects and social networks in education. \bjournalThe Review of Economic Studies \bvolume76 \bpages1239–1267. \endbibitem
- Chow (1996) {barticle}[author] \bauthor\bsnmChow, \bfnmGregory C\binitsG. C. (\byear1996). \btitleThe Lagrange method of optimization with applications to portfolio and investment decisions. \bjournalJournal of Economic Dynamics and Control \bvolume20 \bpages1–18. \endbibitem
- Conley and Udry (2010) {barticle}[author] \bauthor\bsnmConley, \bfnmTimothy G\binitsT. G. and \bauthor\bsnmUdry, \bfnmChristopher R\binitsC. R. (\byear2010). \btitleLearning about a new technology: Pineapple in Ghana. \bjournalAmerican economic review \bvolume100 \bpages35–69. \endbibitem
- Erdös and Rényi (1959) {barticle}[author] \bauthor\bsnmErdös, \bfnmPaul\binitsP. and \bauthor\bsnmRényi, \bfnmA\binitsA. (\byear1959). \btitle“On random graphs,”. \bjournalPublicationes Mathematicae \bvolume6 \bpages290–297. \endbibitem
- Feynman (1948) {barticle}[author] \bauthor\bsnmFeynman, \bfnmRichard Phillips\binitsR. P. (\byear1948). \btitleSpace-time approach to non-relativistic quantum mechanics. \bjournalReviews of Modern Physics \bvolume20 \bpages367. \endbibitem
- Feynman (1949) {barticle}[author] \bauthor\bsnmFeynman, \bfnmRichard Phillips\binitsR. P. (\byear1949). \btitleSpace-time approach to quantum electrodynamics. \bjournalPhysical Review \bvolume76 \bpages769. \endbibitem
- Fujiwara (2017) {bbook}[author] \bauthor\bsnmFujiwara, \bfnmDaisuke\binitsD. (\byear2017). \btitleRigorous time slicing approach to Feynman path integrals. \bpublisherSpringer. \endbibitem
- Goyal (2012) {bbook}[author] \bauthor\bsnmGoyal, \bfnmSanjeev\binitsS. (\byear2012). \btitleConnections: an introduction to the economics of networks. \bpublisherPrinceton University Press. \endbibitem
- Hebb (2005) {bbook}[author] \bauthor\bsnmHebb, \bfnmDonald Olding\binitsD. O. (\byear2005). \btitleThe organization of behavior: A neuropsychological theory. \bpublisherPsychology Press. \endbibitem
- Hossain and Ahamed (2015) {barticle}[author] \bauthor\bsnmHossain, \bfnmMd Saimum\binitsM. S. and \bauthor\bsnmAhamed, \bfnmFaruque\binitsF. (\byear2015). \btitleDeterminants of bank profitability: A study on the banking sector of Bangladesh. \bjournalJournal of Finance and Banking \bvolume13 \bpages43–57. \endbibitem
- Hua, Polansky and Pramanik (2019) {barticle}[author] \bauthor\bsnmHua, \bfnmLei\binitsL., \bauthor\bsnmPolansky, \bfnmAlan\binitsA. and \bauthor\bsnmPramanik, \bfnmParamahansa\binitsP. (\byear2019). \btitleAssessing bivariate tail non-exchangeable dependence. \bjournalStatistics & Probability Letters \bpages108556. \endbibitem
- Jackson (2010) {bbook}[author] \bauthor\bsnmJackson, \bfnmMatthew O\binitsM. O. (\byear2010). \btitleSocial and economic networks. \bpublisherPrinceton university press. \endbibitem
- Kahneman and Tversky (1972) {barticle}[author] \bauthor\bsnmKahneman, \bfnmDaniel\binitsD. and \bauthor\bsnmTversky, \bfnmAmos\binitsA. (\byear1972). \btitleSubjective probability: A judgment of representativeness. \bjournalCognitive psychology \bvolume3 \bpages430–454. \endbibitem
- Kappen (2005a) {barticle}[author] \bauthor\bsnmKappen, \bfnmHilbert J\binitsH. J. (\byear2005a). \btitleLinear theory for control of nonlinear stochastic systems. \bjournalPhysical review letters \bvolume95 \bpages200201. \endbibitem
- Kappen (2005b) {barticle}[author] \bauthor\bsnmKappen, \bfnmHilbert J\binitsH. J. (\byear2005b). \btitlePath integrals and symmetry breaking for optimal control theory. \bjournalJournal of statistical mechanics: theory and experiment \bvolume2005 \bpagesP11011. \endbibitem
- Kappen (2007a) {binproceedings}[author] \bauthor\bsnmKappen, \bfnmHilbert J\binitsH. J. (\byear2007a). \btitleAn introduction to stochastic control theory, path integrals and reinforcement learning. In \bbooktitleAIP conference proceedings \bvolume887 \bpages149–181. \bpublisherAIP. \endbibitem
- Kappen (2007b) {binproceedings}[author] \bauthor\bsnmKappen, \bfnmHilbert J\binitsH. J. (\byear2007b). \btitleAn introduction to stochastic control theory, path integrals and reinforcement learning. In \bbooktitleAIP conference proceedings \bvolume887 \bpages149–181. \bpublisherAmerican Institute of Physics. \endbibitem
- Marcet and Marimon (2019) {barticle}[author] \bauthor\bsnmMarcet, \bfnmAlbert\binitsA. and \bauthor\bsnmMarimon, \bfnmRamon\binitsR. (\byear2019). \btitleRecursive contracts. \bjournalEconometrica \bvolume87 \bpages1589–1631. \endbibitem
- Mas-Colell et al. (1995) {bbook}[author] \bauthor\bsnmMas-Colell, \bfnmAndreu\binitsA., \bauthor\bsnmWhinston, \bfnmMichael Dennis\binitsM. D., \bauthor\bsnmGreen, \bfnmJerry R\binitsJ. R. \betalet al. (\byear1995). \btitleMicroeconomic theory \bvolume1. \bpublisherOxford university press New York. \endbibitem
- Moretti (2011) {barticle}[author] \bauthor\bsnmMoretti, \bfnmEnrico\binitsE. (\byear2011). \btitleSocial learning and peer effects in consumption: Evidence from movie sales. \bjournalThe Review of Economic Studies \bvolume78 \bpages356–393. \endbibitem
- Nakajima (2007) {barticle}[author] \bauthor\bsnmNakajima, \bfnmRyo\binitsR. (\byear2007). \btitleMeasuring peer effects on youth smoking behaviour. \bjournalThe Review of Economic Studies \bvolume74 \bpages897–935. \endbibitem
- Niazi, Özgüler and Yildiz (2016) {binproceedings}[author] \bauthor\bsnmNiazi, \bfnmMuhammad Umar B\binitsM. U. B., \bauthor\bsnmÖzgüler, \bfnmArif Bülent\binitsA. B. and \bauthor\bsnmYildiz, \bfnmAykut\binitsA. (\byear2016). \btitleConsensus as a Nash equilibrium of a dynamic game. In \bbooktitle2016 12th International Conference on Signal-Image Technology & Internet-Based Systems (SITIS) \bpages365–372. \bpublisherIEEE. \endbibitem
- Øksendal (2003) {bincollection}[author] \bauthor\bsnmØksendal, \bfnmBernt\binitsB. (\byear2003). \btitleStochastic differential equations. In \bbooktitleStochastic differential equations \bpages65–84. \bpublisherSpringer. \endbibitem
- Øksendal and Sulem (2019) {bmisc}[author] \bauthor\bsnmØksendal, \bfnmBernt\binitsB. and \bauthor\bsnmSulem, \bfnmAgnès\binitsA. (\byear2019). \btitleApplied Stochastic Control of Jump Diffusions. \endbibitem
- Polansky and Pramanik (2021) {barticle}[author] \bauthor\bsnmPolansky, \bfnmAlan M\binitsA. M. and \bauthor\bsnmPramanik, \bfnmParamahansa\binitsP. (\byear2021). \btitleA motif building process for simulating random networks. \bjournalComputational Statistics & Data Analysis \bpages107263. \endbibitem
- Pramanik (2016) {bbook}[author] \bauthor\bsnmPramanik, \bfnmParamahansa\binitsP. (\byear2016). \btitleTail non-exchangeability. \bpublisherNorthern Illinois University. \endbibitem
- Pramanik (2020) {binproceedings}[author] \bauthor\bsnmPramanik, \bfnmParamahansa\binitsP. (\byear2020). \btitleOptimization of Market Stochastic Dynamics. In \bbooktitleSN Operations Research Forum \bvolume1 \bpages1–17. \bpublisherSpringer. \endbibitem
- Pramanik (2021a) {barticle}[author] \bauthor\bsnmPramanik, \bfnmParamahansa\binitsP. (\byear2021a). \btitleEffects of water currents on fish migration through a Feynman-type path integral approach under Liouville-like quantum gravity surfaces. \bjournalTheory in Biosciences \bvolume140 \bpages205-223. \endbibitem
- Pramanik (2021b) {barticle}[author] \bauthor\bsnmPramanik, \bfnmParamahansa\binitsP. (\byear2021b). \btitleEffects of water currents on fish migration through a Feynman-type path integral approach under 8/3 Liouville-like quantum gravity surfaces. \bjournalTheory in Biosciences \bvolume140 \bpages205-223. \endbibitem
- Pramanik and Polansky (2019) {barticle}[author] \bauthor\bsnmPramanik, \bfnmParamahansa\binitsP. and \bauthor\bsnmPolansky, \bfnmAlan M\binitsA. M. (\byear2019). \btitleSemicooperation under curved strategy spacetime. \bjournalarXiv preprint arXiv:1912.12146. \endbibitem
- Pramanik and Polansky (2020a) {barticle}[author] \bauthor\bsnmPramanik, \bfnmParamahansa\binitsP. and \bauthor\bsnmPolansky, \bfnmAlan M\binitsA. M. (\byear2020a). \btitleMotivation to Run in One-Day Cricket. \bjournalarXiv preprint arXiv:2001.11099. \endbibitem
- Pramanik and Polansky (2020b) {barticle}[author] \bauthor\bsnmPramanik, \bfnmParamahansa\binitsP. and \bauthor\bsnmPolansky, \bfnmAlan M\binitsA. M. (\byear2020b). \btitleOptimization of a Dynamic Profit Function using Euclidean Path Integral. \bjournalarXiv preprint arXiv:2002.09394. \endbibitem
- Pramanik and Polansky (2021) {barticle}[author] \bauthor\bsnmPramanik, \bfnmParamahansa\binitsP. and \bauthor\bsnmPolansky, \bfnmAlan M\binitsA. M. (\byear2021). \btitleOptimal Estimation of Brownian Penalized Regression Coefficients. \bjournalarXiv preprint arXiv:2107.02291. \endbibitem
- Sheng (2020) {barticle}[author] \bauthor\bsnmSheng, \bfnmShuyang\binitsS. (\byear2020). \btitleA structural econometric analysis of network formation games through subnetworks. \bjournalEconometrica \bvolume88 \bpages1829–1858. \endbibitem
- Snijders (2002) {barticle}[author] \bauthor\bsnmSnijders, \bfnmTom AB\binitsT. A. (\byear2002). \btitleMarkov chain Monte Carlo estimation of exponential random graph models. \bjournalJournal of Social Structure \bvolume3 \bpages1–40. \endbibitem
- Theodorou (2011) {bbook}[author] \bauthor\bsnmTheodorou, \bfnmEvangelos A\binitsE. A. (\byear2011). \btitleIterative path integral stochastic optimal control: Theory and applications to motor control. \bpublisherUniversity of Southern California. \endbibitem
- Theodorou, Buchli and Schaal (2010) {binproceedings}[author] \bauthor\bsnmTheodorou, \bfnmEvangelos\binitsE., \bauthor\bsnmBuchli, \bfnmJonas\binitsJ. and \bauthor\bsnmSchaal, \bfnmStefan\binitsS. (\byear2010). \btitleReinforcement learning of motor skills in high dimensions: A path integral approach. In \bbooktitleRobotics and Automation (ICRA), 2010 IEEE International Conference on \bpages2397–2403. \bpublisherIEEE. \endbibitem
- Tversky and Kahneman (1971) {barticle}[author] \bauthor\bsnmTversky, \bfnmAmos\binitsA. and \bauthor\bsnmKahneman, \bfnmDaniel\binitsD. (\byear1971). \btitleBelief in the law of small numbers. \bjournalPsychological bulletin \bvolume76 \bpages105. \endbibitem
- Van Den Broek, Wiegerinck and Kappen (2008) {barticle}[author] \bauthor\bsnmVan Den Broek, \bfnmBart\binitsB., \bauthor\bsnmWiegerinck, \bfnmWim\binitsW. and \bauthor\bsnmKappen, \bfnmBert\binitsB. (\byear2008). \btitleGraphical model inference in optimal control of stochastic multi-agent systems. \bjournalJournal of Artificial Intelligence Research \bvolume32 \bpages95–122. \endbibitem
- Yang et al. (2014) {barticle}[author] \bauthor\bsnmYang, \bfnmInsoon\binitsI., \bauthor\bsnmMorzfeld, \bfnmMatthias\binitsM., \bauthor\bsnmTomlin, \bfnmClaire J\binitsC. J. and \bauthor\bsnmChorin, \bfnmAlexandre J\binitsA. J. (\byear2014). \btitlePath integral formulation of stochastic optimal control with generalized costs. \bjournalIFAC Proceedings Volumes \bvolume47 \bpages6994–7000. \endbibitem