This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Analytical Solution for the Size of the Minimum Dominating Set
in Complex Networks

Jose C. Nacher nacher@is.sci.toho-u.ac.jp Department of Information Science, Faculty of Science, Toho University, Miyama 2-2-1, Funabashi, Chiba 274-8510, Japan    Tomoshiro Ochiai ochiai@otsuma.ac.jp Faculty of Social Information Studies, Otsuma Women’s University, 2-7-1 Karakida, Tama-shi, Tokyo 206-8540, Japan
(July 29, 2025)
Abstract

Domination is the fastest-growing field within graph theory with a profound diversity and impact in real-world applications, such as the recent breakthrough approach that identifies optimized subsets of proteins enriched with cancer-related genes. Despite its conceptual simplicity, domination is a classical NP-complete decision problem which makes analytical solutions elusive and poses difficulties to design optimization algorithms for finding a dominating set of minimum cardinality in a large network. Here we derive for the first time an approximate analytical solution for the density of the minimum dominating set (MDS) by using a combination of cavity method and Ultra-Discretization (UD) procedure. The derived equation allows us to compute the size of MDS by only using as an input the information of the degree distribution of a given network.

pacs:

I Introduction

The research on complex networks in diverse fields newman ; caldarelli , based on applied graph theory combined with computational and statistical physics methods, has experienced a spectacular growth in recent years and has led to the discovery of ubiquitous patterns called scale-free networks barabasi , unexpected dynamic behavior vespignani , robustness and vulnerability features havlin1 ; reka ; motter , and applications in natural and social complex systems newman ; caldarelli . On the other hand, domination is an important problem in graph theory which has rich variants, such as independence, covering and matching teresa . The mathematical and computational studies on domination have led to abundant applications in disparate fields such as mobile computing and computer communication networks sto , design of large parallel and distributed systems swapped , analysis of large social networks social1 ; social2 ; social3 , computational biology and biomedical analysis vazquez and discrete algorithms research teresa .

More recently, the Minimum Dominating Set (MDS) has drawn the attention researchers to controllability in complex networks nacher1 ; nacher2 ; nacher3 , to investigate observability in power-grid power and to identify an optimized subset of proteins enriched with essential, cancer-related and virus-targeted genes in protein networks wuchty . The size of MDS was also investigated by extensively analysing several types of artificial scale-free networks using a greedy algorithm in molnar . The problem to design complex networks that are structurally robust has also recently been investigated using the MDS approach nacher4 ; molnar2 .

Despite its conceptual simplicity, the MDS is a classical NP-complete decision problem in computational complexity theory johnson . Therefore, it is believed that there is no a theoretically efficient (i.e., polynomial time) algorithm that finds an exact smallest dominating set for a given graph. It is worth noticing that although it is an NP-hard problem, recent results have shown that we can use Integer Linear Programming (ILP) to find optimal solution nacher1 ; wuchty ; nacher4 . Moreover, for specific types of graph such a tree (i.e., no loops) and even partial-kk-tree, it can be solved using Dynamic Programing (DP) in polynomial time akutsu .

Especially, the density (or fraction) of MDS, defined as the size of MDS versus the total number of nodes in a network, is important to estimate the cost-efficient network deployment of controllers. It is less expensive to have ten power plants operating as controllers in a large network of 1,000 sites than one hundred. Similarly, acting on two protein targets via drug interactions is always better than acting on ten protein targets to minimize adverse side effects. Although we can use an ILP method for obtaining an MDS and the density of MDS, we do not have an analytic solution for the density of MDS. Note that even though optimal solution for MDS can be found using ILP method, this kind of technique is very generic and operates as a black box so that its systematic application does not provide us any knowledge about the details of the particular problem under study. Moreover, from a physical point of view, analytic solutions always enable us to have a deeper understanding of the problem by examining dependences with other variables. Here, we derive for the first time an analytical solution for an MDS by using cavity method.

Cavity method is a well-known methodology developed by physicists working on statistical mechanics (e.g. spin glasses spin ). However, this technique has also been extended and applied to non-physical systems, including network theory in which nodes can be abstracted to some extent. For example, applications include the analysis of random combinatorial problems such as weighted matching mezard1 , the vertex cover hart and the travelling salesman problem mezard2 . Moreover, recently cavity method was used to determine the number of matchings in random graphs mezard_last , the maximal independent sets dalla , the random set packing indian and the minimum weight Steiner tree steiner and controllability using maximum matching liu , extending the work done in mezard_last for undirected networks to directed networks.111Note added: After finishing our paper, we became aware of a recent paper done independently by Zhao et al.Zhao , in which the statistical properties of the MDS were studied. Although the starting point of the cavity analysis is similar, in their results they estimate ths size of MDS by population dynamics simulations. In contrast, as shown later, we derive analytical solution for the size of the MDS. It is also well-known that cavity method computed at zero temperature limit has been applied to networks and used in many works, which have led to derive elegant analytical formulas new1 ; new2 ; new3 .

On the other hand, the Ultra Discretization (UD) procedure has been developed mainly in the field of soliton theory and cellular automaton (CA) theory UD1 ; UD2 ; UD3 ; UD4 ; UDours . Although the common discretization process discretizes independent variables, the UD procedure can discretize dependent variables. As a result, both independent and dependent variables become discrete variables. In other words, UD transforms discretized equations into their corresponding UD equations. In soliton theory, we can obtain the cellular automaton from the corresponding discretized soliton theory after UD procedure. In UD4 , it is shown that the corresponding integrable CA is obtained from Korteweg–de Vries (KdV) soliton equation via Lotka-Volterra equation, using UD procedure. The key point of UD is that the UD procedure can be applied only in the case that the original discretized equation does not have any minus operators.

By combining the cavity method and UD procedure, we solved for the first time the density of MDS problem analytically for any complex network. By cavity method and UD, we derived a combinatorial expression that allows us to identify the density of MDS by only using the degree distribution of the network, which is our main result. We then compare the results of the derived expression for density of MDS with the ILP solutions for regular, random and scale-free networks showing a fair agreement. Moreover, a simple formula is derived for random networks at the large degree limit.

II Theoretical results

II.1 The Hamiltonian of a dominating set

A graph G=(V,E)G=(V,E) is a set of nodes VV and edges EE. A Dominant Set (DS) SS is defined to be a subset of VV, where each node iVi\in V belongs to SS or adjacent to an element of SS. For each node iVi\in V, we define a binary function σi\sigma_{i} to be +1+1 if iSi\in S, otherwise 0. We call the node ii occupied (resp. empty) if σi\sigma_{i} is +1+1 (resp. 0). The set of the adjacent nodes to node ii is denoted by i\partial i. We can then see that the constraint to define a DS is as follows:

σi+jiσj1.\displaystyle\sigma_{i}+\sum_{j\in\partial i}\sigma_{j}\geq 1. (1)

For each node iVi\in V, we define a binary function IiI_{i} to be +1+1 if and only if σi+jiσj1\sigma_{i}+\sum_{j\in\partial i}\sigma_{j}\geq 1, otherwise 0. A set SS is an Minimum Dominating Set (MDS) if the size is smallest among all dominating sets (see Fig. 1).

Let us consider the following Hamiltonian function:

H(σ)=iVσi,\displaystyle H(\sigma)=\sum_{i\in V}\sigma_{i}, (2)

where σ={σi|iV}\sigma=\{\sigma_{i}|i\in V\}. Then the partition function is given by

Z(β)=σ(iVIi)exp(βH(σ)),\displaystyle Z(\beta)=\sum_{\sigma}(\prod_{i\in V}I_{i})\exp(-\beta H(\sigma)), (3)

where the summation is taken over all configurations of σ\sigma and β\beta is the inverse temperature.

II.2 Cavity method analysis

In what follows, we apply the cavity method to derive the analytic formula for the density of MDS in complex networks. Note that the cavity approach has been used in a similar way to solve the maximal independent set problem dalla . First, we assume the graph GG has tree structure and let i\j\partial i\backslash j be the set of i\partial i except for node jj (see Fig. 2). Note also that the minimum degree of the graph should be two, otherwise some of i\j\partial i\backslash j do not exist. Then, we can write σij={σk|ki\j}\sigma_{i\to j}=\{\sigma_{k}|k\in\partial i\backslash j\}. Next, let νij(σi,σij)\nu_{i\to j}(\sigma_{i},\sigma_{i\to j}) be the probability that the node ii and i\j\partial i\backslash j take value σi\sigma_{i} and σij\sigma_{i\to j}, respectively, and constraints IiI_{i} and IjI_{j} are not included:

νij(σi,σij)=1Zijσ(αi,jIα)exp(βH(σ)),\displaystyle\nu_{i\to j}(\sigma_{i},\sigma_{i\to j})=\frac{1}{Z_{i\to j}}\sum_{\sigma^{\prime}}(\prod_{\alpha\neq i,j}I_{\alpha})\exp(-\beta H(\sigma)), (4)

where the first summation is taken over all configurations of σ={σβ|(βi)(βi\j)(βV)}\sigma^{\prime}=\{\sigma_{\beta}|(\beta\neq i)\wedge(\beta\notin\partial i\backslash j)\wedge(\beta\in V)\}. The product for the constraint IαI_{\alpha} is taken over all nodes of VV except for ii, jj (i.e. αV\alpha\in V and αi,j\alpha\neq i,j) and ZijZ_{i\to j} is the normalization constant given by

Zij\displaystyle Z_{i\to j} =\displaystyle= σiσijσ(αi,jIα)exp(βH(σ))\displaystyle\sum_{\sigma_{i}}\sum_{\sigma_{i\to j}}\sum_{\sigma^{\prime}}(\prod_{\alpha\neq i,j}I_{\alpha})\exp(-\beta H(\sigma)) (5)
=\displaystyle= σ(αi,jIα)exp(βH(σ)).\displaystyle\sum_{\sigma}(\prod_{\alpha\neq i,j}I_{\alpha})\exp(-\beta H(\sigma)).

Eq. (4) can be written in a different way, which will be easier to compute. As shown in Fig. 3, we divide a graph into two subgraphs (AA and BB) by cutting the edge (ii, jj) between node ii and jj. Let AA be the set of all the nodes which belongs to the subgraph including ii, and BB be the set of all the nodes which belongs to the subgraph including jj. Then, Eq. (4) can be transformed into

νij(σi,σij)=1ZijσA(αiIα)exp(βγAσγ),\displaystyle\nu_{i\to j}(\sigma_{i},\sigma_{i\to j})=\frac{1}{Z^{\prime}_{i\to j}}\sum_{\sigma^{\prime}_{A}}(\prod_{\alpha\neq i}I_{\alpha})\exp(-\beta\sum_{\gamma\in A}\sigma_{\gamma}), (6)

where the first summation is taken over all configurations of σA={σβ|(βi)(βi\j)(βA)}\sigma^{\prime}_{A}=\{\sigma_{\beta}|(\beta\neq i)\wedge(\beta\notin\partial i\backslash j)\wedge(\beta\in A)\}. The product for the constraint IαI_{\alpha} is taken over all nodes of AA except ii (i.e. αA\alpha\in A and αi\alpha\neq i) and ZijZ^{\prime}_{i\to j} is a normalization constant given by

Zij\displaystyle Z^{\prime}_{i\to j} =\displaystyle= σiσijσA(αiIα)exp(βγAσγ)\displaystyle\sum_{\sigma_{i}}\sum_{\sigma_{i\to j}}\sum_{\sigma^{\prime}_{A}}(\prod_{\alpha\neq i}I_{\alpha})\exp(-\beta\sum_{\gamma\in A}\sigma_{\gamma}) (7)
=\displaystyle= σA(αiIα)exp(βγAσγ),\displaystyle\sum_{\sigma_{A}}(\prod_{\alpha\neq i}I_{\alpha})\exp(-\beta\sum_{\gamma\in A}\sigma_{\gamma}),

where the last summation is taken over all configurations of σA={σβ|(βA)}\sigma_{A}=\{\sigma_{\beta}|(\beta\in A)\}.

Then the following exact recursive equation can be derived:

νij(σi,σij)eβσiki\jσkiIkνki(σk,σki).\displaystyle\nu_{i\to j}(\sigma_{i},\sigma_{i\to j})\propto e^{-\beta\sigma_{i}}\prod_{k\in\partial i\backslash j}\sum_{\sigma_{k\to i}}I_{k}\nu_{k\to i}(\sigma_{k},\sigma_{k\to i}). (8)

This equation can be iteratively solved with its corresponding normalization constant.

Let ν¯(σi,m)\bar{\nu}(\sigma_{i},m) be the summation of νij(σi,σij)\nu_{i\to j}(\sigma_{i},\sigma_{i\to j}), where the number of occupied neighbors σij\sigma_{i\to j} is mm:

ν¯ij(σi,m)=αi\jσα=mνij(σi,σij).\displaystyle\bar{\nu}_{i\to j}(\sigma_{i},m)=\sum_{\sum_{\alpha\in\partial i\backslash j}\sigma_{\alpha}=m}\nu_{i\to j}(\sigma_{i},\sigma_{i\to j}). (9)

Then, let r1ijr_{1}^{i\to j} be the probability that the node ii is occupied in the cavity graph. Let r00ijr_{00}^{i\to j} be the probability that the node ii and all the neighbors σij\sigma_{i\to j} are empty. Let r0ijr_{0}^{i\to j} be the probability that the node ii is empty and at least one of the neighbors σij\sigma_{i\to j} are occupied. Then, we can write:

r1ij\displaystyle r_{1}^{i\to j} =\displaystyle= m=0ki1ν¯ij(1,m),\displaystyle\sum_{m=0}^{k_{i}-1}\bar{\nu}_{i\to j}(1,m), (10)
r00ij\displaystyle r_{00}^{i\to j} =\displaystyle= ν¯ij(0,0),\displaystyle\bar{\nu}_{i\to j}(0,0), (11)
r0ij\displaystyle r_{0}^{i\to j} =\displaystyle= m=1ki1ν¯ij(0,m),\displaystyle\sum_{m=1}^{k_{i}-1}\bar{\nu}_{i\to j}(0,m), (12)

where kik_{i} is the degree of node ii. Here we note that r1ij+r00ij+r0ij=1r_{1}^{i\to j}+r_{00}^{i\to j}+r_{0}^{i\to j}=1, because of probability normalization.

The above definitions lead to the following iterative equations:

r1ij\displaystyle r_{1}^{i\to j} =\displaystyle= 1Neβ,\displaystyle\frac{1}{N}e^{-\beta}, (13)
r00ij\displaystyle r_{00}^{i\to j} =\displaystyle= 1Nki\jr0ki,\displaystyle\frac{1}{N}\prod_{k\in\partial i\backslash j}r_{0}^{k\to i}, (14)
r0ij\displaystyle r_{0}^{i\to j} =\displaystyle= 1N{ki\j(1r00ki)ki\jr0ki},\displaystyle\frac{1}{N}\Bigl{\{}\prod_{k\in\partial i\backslash j}(1-r_{00}^{k\to i})-\prod_{k\in\partial i\backslash j}r_{0}^{k\to i}\Bigr{\}}, (15)

where NN is the normalization constant given by

N\displaystyle N =\displaystyle= eβ+ki\j(1r00ki).\displaystyle e^{-\beta}+\prod_{k\in\partial i\backslash j}(1-r_{00}^{k\to i}). (16)

II.3 Ultra-Discretization (UD) procedure

We note that until now we are considering the problem of an DS, which is treated as finite temperature problem (β\beta is finite) in the context of statistical mechanics. To address the minimum dominating set (MDS) problem, we have to consider the zero-temperature limit (β\beta\to\infty) which gives the ground energy state of the Hamiltonian shown in Eq. (2). To solve the equations associated to the zero-temperature limit, we can use an Ultra-Discretization (UD) procedure UD4 .

Equation independent variable dependent variable
(ii) Continuous equation for x(t)x(t) tt: real value x(t)x(t): real value
(iiii) Discretized equation for xnx_{n} nn: integer xnx_{n}: real value
(iiiiii) Ultra- Discretized (UD) equation for XnX_{n} nn: integer XnX_{n}: integer
Table 1: Discretization procedure changes the equation type from (ii) to (iiii) and Ultra-Discretization procedure transforms the equation type from (iiii) to (iiiiii).

Discretization is a well-known method to discretize independent variable in continuous theory such as differential equation (see Table 1 from (ii) to (iiii)). UD can go further and aims to discretize the dependent variable in the discretized equation (from (iiii) to (iiiiii)). As a result, in the transformed UD equation (iiiiii), both independent and dependent variables are discretized (i.e. both are integer variables). The key formula of UD, which transforms the equation type from (iiii) to (iiiiii) is

limβ1βlog(eβX+eβY)=max(X,Y).\displaystyle\lim_{\beta\to\infty}\frac{1}{\beta}\log(e^{\beta X}+e^{\beta Y})=\max(X,Y). (17)

When the discretized equation (iiii) has no subtraction operator, we can transform it into the corresponding UD equation (iiiiii) by replacing x=eβXx=e^{\beta X} and y=eβYy=e^{\beta Y} and taking limit β\beta\to\infty. Here, xx and yy belong to discretized equation (iiii) and XX and YY belong to UD equation (iiiiii). The operator in the original discretized equation (iiii) is transformed in the UD equation (iiiiii) as follows:

x×yX+Y,\displaystyle x\times y\to X+Y, (18)
x/yXY,\displaystyle x/y\to X-Y, (19)
x+ymax(X,Y).\displaystyle x+y\to\max(X,Y). (20)

Here we remark that in order to obtain the UD equation (iiiiii), it is necessary to be able to remove any minus operator in the original discretized equation (iiii).

In order to ultra-discretize the previous Eqs. (13), (14) and (15), we make them consist of only plus, multiplication and division operators, avoiding minus operators. By simple computation, we have

ki\j(1r00ki)=ki\j(r0ki+r1ki)\displaystyle\prod_{k\in\partial i\backslash j}(1-r_{00}^{k\to i})=\prod_{k\in\partial i\backslash j}(r_{0}^{k\to i}+r_{1}^{k\to i}) (21)

and

ki\j(1r00ki)ki\jr0ki=1m1++mnnrm1k1irm2k2irmnkni,\displaystyle\prod_{k\in\partial i\backslash j}(1-r_{00}^{k\to i})-\prod_{k\in\partial i\backslash j}r_{0}^{k\to i}=\sum_{1\leq m_{1}+\cdots+m_{n}\leq n}r_{m_{1}}^{k_{1}\to i}r_{m_{2}}^{k_{2}\to i}\cdots r_{m_{n}}^{k_{n}\to i}, (22)

where nn is the number of elements of i\j\partial i\backslash j, and mp=0or1m_{p}=0~\mbox{or}~1 (p=1,2,3,,n)(p=1,2,3,\cdots,n).

By inserting Eqs. (21) and (22) into Eqs. (13), (14) and (15), we have

r1ij\displaystyle r_{1}^{i\to j} =\displaystyle= eβeβ+ki\j(r0ki+r1ki),\displaystyle\frac{e^{-\beta}}{e^{-\beta}+\prod_{k\in\partial i\backslash j}(r_{0}^{k\to i}+r_{1}^{k\to i})}, (23)
r00ij\displaystyle r_{00}^{i\to j} =\displaystyle= ki\jr0kieβ+ki\j(r0ki+r1ki),\displaystyle\frac{\prod_{k\in\partial i\backslash j}r_{0}^{k\to i}}{e^{-\beta}+\prod_{k\in\partial i\backslash j}(r_{0}^{k\to i}+r_{1}^{k\to i})}, (24)
r0ij\displaystyle r_{0}^{i\to j} =\displaystyle= 1m1++mnnrm1k1irm2k2irmnknieβ+ki\j(r0ki+r1ki).\displaystyle\frac{\sum_{1\leq m_{1}+\cdots+m_{n}\leq n}r_{m_{1}}^{k_{1}\to i}r_{m_{2}}^{k_{2}\to i}\cdots r_{m_{n}}^{k_{n}\to i}}{e^{-\beta}+\prod_{k\in\partial i\backslash j}(r_{0}^{k\to i}+r_{1}^{k\to i})}. (25)

Here we remark that the above derived equations consist only of three operators (plus, multiplication and division), which is suitable for the ultra-discretization.

Here, we then replace the following three variables as follows:

r1ij\displaystyle r_{1}^{i\to j} =\displaystyle= eβR1ij,\displaystyle e^{\beta R_{1}^{i\to j}}, (26)
r00ij\displaystyle r_{00}^{i\to j} =\displaystyle= eβR00ij,\displaystyle e^{\beta R_{00}^{i\to j}}, (27)
r0ij\displaystyle r_{0}^{i\to j} =\displaystyle= eβR0ij,\displaystyle e^{\beta R_{0}^{i\to j}}, (28)

where R1ijR_{1}^{i\to j}, R00ijR_{00}^{i\to j} and R0ijR_{0}^{i\to j} are variables in UD system.

After inserting (26), (27), (28) into (23), (24), (25) and taking zero-temperature limit β\beta\to\infty, Eqs. (23), (24) and (25) are transformed into

R1ij\displaystyle R_{1}^{i\to j} =\displaystyle= 1max(1,ki\jmax(R0ki,R1ki)),\displaystyle-1-\max(-1,\sum_{k\in\partial i\backslash j}\max(R_{0}^{k\to i},R_{1}^{k\to i})), (29)
R00ij\displaystyle R_{00}^{i\to j} =\displaystyle= ki/jR0kimax(1,ki\jmax(R0ki,R1ki)),\displaystyle\sum_{k\in\partial i/j}R_{0}^{k\to i}-\max(-1,\sum_{k\in\partial i\backslash j}\max(R_{0}^{k\to i},R_{1}^{k\to i})), (30)
R0ij\displaystyle R_{0}^{i\to j} =\displaystyle= max1m1++mnn(Rm1k1i+Rm2k2i++Rmnkni)max(1,ki\jmax(R0ki,R1ki)).\displaystyle\max_{1\leq m_{1}+\cdots+m_{n}\leq n}(R_{m_{1}}^{k_{1}\to i}+R_{m_{2}}^{k_{2}\to i}+\cdots+R_{m_{n}}^{k_{n}\to i})-\max(-1,\sum_{k\in\partial i\backslash j}\max(R_{0}^{k\to i},R_{1}^{k\to i})).

Eq. (23), (24), (25) (or Eq. (29), (30), (LABEL:eqn:UD_equation_3)) are very difficult to solve, because every edge direction has three variables and three associated equations. Therefore, in order to avoid this high complexity, we use a coarse-grained method.

Let pkp_{k} be the degree distribution of the network which gives the probability to find nodes with degree kk and k=kkpk\langle k\rangle=\sum_{k}kp_{k} be the average degree of the network. The excess degree distribution is given by qk=(k+1)pk+1/kq_{k}=(k+1)p_{k+1}/{\langle k\rangle}, that is the probability to find that a neighbor node has degree k.

Let us assume that the network is enough large. Each r1ijr_{1}^{i\to j} and r0ijr_{0}^{i\to j} values are assigned to the edge direction from ii to jj. The opposite edge direction gets the values from r1jir_{1}^{j\to i} and r0jir_{0}^{j\to i}. r00ijr_{00}^{i\to j} is not independent variable because of the normalization. Let P(r1,r0)P(r_{1},r_{0}) be the probability density function of r1ijr_{1}^{i\to j}, r0ijr_{0}^{i\to j}.

Then we have the coarse-grained equation (cavity mean field equation) for (23) and (25) as follows:

P(r1,r0)\displaystyle P(r_{1},r_{0}) =\displaystyle= k=1qkl=1kdr1ldr0lP(r1l,r0l)\displaystyle\sum_{k=1}^{\infty}q_{k}\int\prod_{l=1}^{k}dr_{1}^{l}dr_{0}^{l}P(r_{1}^{l},r_{0}^{l}) (32)
×δ(r1eβeβ+l=1k(r0l+r1l))\displaystyle\times\delta(r_{1}-\frac{e^{-\beta}}{e^{-\beta}+\prod_{l=1}^{k}(r_{0}^{l}+r_{1}^{l})})
×δ(r01m1++mkkrm11rm22rmkkeβ+l=1k(r0l+r1l)).\displaystyle\times\delta(r_{0}-\frac{\sum_{1\leq m_{1}+\cdots+m_{k}\leq k}r_{m_{1}}^{1}r_{m_{2}}^{2}\cdots r_{m_{k}}^{k}}{e^{-\beta}+\prod_{l=1}^{k}(r_{0}^{l}+r_{1}^{l})}).

We transform the probability density function by variables transformation (26) and (28) as follows:

P(r1,r0)dr1dr0=P¯(R1,R0)dR1dR0.\displaystyle P(r_{1},r_{0})dr_{1}dr_{0}=\bar{P}(R_{1},R_{0})dR_{1}dR_{0}. (33)

Then, we have

P¯(R1,R0)\displaystyle\bar{P}(R_{1},R_{0}) =\displaystyle= k=1qkl=1kdR1ldR0lP¯(R1l,R0l)\displaystyle\sum_{k=1}^{\infty}q_{k}\int\prod_{l=1}^{k}dR_{1}^{l}dR_{0}^{l}\bar{P}(R_{1}^{l},R_{0}^{l}) (34)
×δ(r11βlogeβeβ+l=1k(eβR0l+eβR1l))\displaystyle\times\delta(r_{1}-\frac{1}{\beta}\log\frac{e^{-\beta}}{e^{-\beta}+\prod_{l=1}^{k}(e^{\beta R_{0}^{l}}+e^{\beta R_{1}^{l}})})
×δ(r01βlog1m1++mkkeβRm11eβRm22eβRmkkeβ+l=1k(eβR0l+eβR1l)).\displaystyle\times\delta(r_{0}-\frac{1}{\beta}\log\frac{\sum_{1\leq m_{1}+\cdots+m_{k}\leq k}e^{\beta R_{m_{1}}^{1}}e^{\beta R_{m_{2}}^{2}}\cdots e^{\beta R_{m_{k}}^{k}}}{e^{-\beta}+\prod_{l=1}^{k}(e^{\beta R_{0}^{l}}+e^{\beta R_{1}^{l}})}).

Taking UD limit (zero temperature) β\beta\to\infty, we have the ultra-discretization version of cavity equation:

P¯(R1,R0)\displaystyle\bar{P}(R_{1},R_{0}) =\displaystyle= k=1qkl=1kdR1ldR0lP¯(R1l,R0l)\displaystyle\sum_{k=1}^{\infty}q_{k}\int\prod_{l=1}^{k}dR_{1}^{l}dR_{0}^{l}\bar{P}(R_{1}^{l},R_{0}^{l})
×δ(R1+1+max(1,l=1kmax(R0l,R1l)))\displaystyle\times\delta(R_{1}+1+\max(-1,\sum_{l=1}^{k}\max(R_{0}^{l},R_{1}^{l})))
×δ(R0max1m1++mkk(Rm11+Rm22++Rmkk)+max(1,l=1kmax(R0l,R1l))).\displaystyle\times\delta(R_{0}-\max_{1\leq m_{1}+\cdots+m_{k}\leq k}(R_{m_{1}}^{1}+R_{m_{2}}^{2}+\cdots+R_{m_{k}}^{k})+\max(-1,\sum_{l=1}^{k}\max(R_{0}^{l},R_{1}^{l}))).

Next, we will solve this equation. Eqs. (29), (30), (LABEL:eqn:UD_equation_3) imply that R1R_{1} and R0R_{0} takes integer values, since R1R_{1} and R0R_{0} at the boundary of network takes integer values. Furthermore, considering the probability conservation, Eqs. (26), (28) imply that R1R_{1} and R0R_{0} takes value 0 or negative integer. By considering Eq. (29), firstly, we can see that R1ijR_{1}^{i\to j} takes value only 0 or 1-1. Secondly, if R1ijR_{1}^{i\to j} is 1-1, then for each ki\jk\in\partial i\backslash j, one of R0kiR_{0}^{k\to i} or R1kiR_{1}^{k\to i} takes value 0. In this case (R1ij=1R_{1}^{i\to j}=-1), from Eq. (LABEL:eqn:UD_equation_3), R0ijR_{0}^{i\to j} takes value 0 or 1-1. Therefore, we can set the distribution P¯(R1,R0)\bar{P}(R_{1},R_{0}) as follows:

P¯(R1,R0)=aδ(R1+1)δ(R0)+bδ(R1+1)δ(R0+1)+n=0cnδ(R1)δ(R0+n),\displaystyle\bar{P}(R_{1},R_{0})=a\delta(R_{1}+1)\delta(R_{0})+b\delta(R_{1}+1)\delta(R_{0}+1)+\sum_{n=0}^{\infty}c_{n}\delta(R_{1})\delta(R_{0}+n), (36)

where a+b+n=0cn=1a+b+\sum_{n=0}^{\infty}c_{n}=1.

Inserting (36) into the cavity equation (LABEL:eqn:_cavity_equation_UD), we have

a\displaystyle a =\displaystyle= k=1qk((1b)kak),\displaystyle\sum_{k=1}^{\infty}q_{k}((1-b)^{k}-a^{k}), (37)
b\displaystyle b =\displaystyle= k=1qkak,\displaystyle\sum_{k=1}^{\infty}q_{k}a^{k}, (38)
cm1\displaystyle c_{m-1} =\displaystyle= k=mqk(Cmk)bm(1b)km(m=1,2,3,).\displaystyle\sum_{k=m}^{\infty}q_{k}({}_{k}C_{m})b^{m}(1-b)^{k-m}~~~~~(m=1,2,3,\cdots). (39)

Once the degree distribution pkp_{k} is given, we can determine aa, bb, cnc_{n} (n=0,1,2,)(n=0,1,2,\cdots).

Here, we need the average energy of the Hamiltonian (2), which can be identified as the average number of all DS configurations. After taking zero temperature limit β\beta\to\infty, we will obtain the analytical formula for the density of MDS.

The average energy can be computed by <H>=βlogZ<H>=-\frac{\partial}{\partial\beta}\log Z, where the partition function is shown in Supplementary Information. A generic expression for density is given by

ρ\displaystyle\rho =\displaystyle= <H>|V|\displaystyle\frac{<H>}{|V|} (40)
=\displaystyle= k=2pkl=1kdr1ldr0lP(r1l,r0l)eβeβ+l=1k(r0l+r00l)l=1kr0l.\displaystyle\sum_{k=2}^{\infty}p_{k}\int\prod_{l=1}^{k}dr_{1}^{l}dr_{0}^{l}P(r_{1}^{l},r_{0}^{l})\frac{e^{-\beta}}{e^{-\beta}+\prod_{l=1}^{k}(r_{0}^{l}+r_{00}^{l})-\prod_{l=1}^{k}r_{0}^{l}}.

This equation, however, cannot be ultra-discretized because of the minus operator, and without being ultra-discretized, analytical solution cannot be obtained. Therefore, we avoid minus operator by inserting Eqs. (21) and (22), which leads to the following result:

ρ\displaystyle\rho =\displaystyle= k=2pkl=1kdr1ldr0lP(r1l,r0l)eβeβ+1m1++mkk(rm11rmkk).\displaystyle\sum_{k=2}^{\infty}p_{k}\int\prod_{l=1}^{k}dr_{1}^{l}dr_{0}^{l}P(r_{1}^{l},r_{0}^{l})\frac{e^{-\beta}}{e^{-\beta}+\sum_{1\leq m_{1}+\cdots+m_{k}\leq k}(r_{m_{1}}^{1}\cdots r_{m_{k}}^{k})}. (41)

After taking zero temperature limit β\beta\to\infty, we obtain

ρ\displaystyle\rho =\displaystyle= limβk=2pkl=1kdR1ldR0lP(R1l,R0l)eβeβ+1m1++mkkeβ(Rm11+Rm22++Rmkk)\displaystyle\lim_{\beta\to\infty}\sum_{k=2}^{\infty}p_{k}\int\prod_{l=1}^{k}dR_{1}^{l}dR_{0}^{l}P(R_{1}^{l},R_{0}^{l})\frac{e^{-\beta}}{e^{-\beta}+\sum_{1\leq m_{1}+\cdots+m_{k}\leq k}e^{\beta(R_{m_{1}}^{1}+R_{m_{2}}^{2}+\cdots+R_{m_{k}}^{k})}} (42)

By inserting (36) into this equation, we finally obtain

ρ\displaystyle\rho =\displaystyle= k=2pk[1(1b)kkb(1b)k1+ak1k+1\displaystyle\sum_{k=2}^{\infty}p_{k}\Bigl{[}1-(1-b)^{k}-kb(1-b)^{k-1}+a^{k}\frac{1}{k+1} (43)
+p+q+r=k1k!p!r!s!apb(c0)r(1bc0)s11+2r+1\displaystyle+\sum_{p+q+r=k-1}\frac{k!}{p!r!s!}a^{p}b(c_{0})^{r}(1-b-c_{0})^{s}\frac{1}{1+2^{r+1}}
+p+r=k1k!p!r!apb(c0)r(11+2r+112r+1)].\displaystyle+\sum_{p+r=k-1}\frac{k!}{p!r!}a^{p}b(c_{0})^{r}(\frac{1}{1+2^{r+1}}-\frac{1}{2^{r+1}})\Bigr{]}.

More simply, we transform the previous equation into

ρ\displaystyle\rho =\displaystyle= k=2pk[1(1b)kkb(1b)k1+ak1k+1\displaystyle\sum_{k=2}^{\infty}p_{k}\Bigl{[}1-(1-b)^{k}-kb(1-b)^{k-1}+a^{k}\frac{1}{k+1}
+r=0k1k!(kr1)!r!bc0r×{(1bc0)kr111+2r+1akr1(11+2r+112r+1}].\displaystyle+\sum_{r=0}^{k-1}\frac{k!}{(k-r-1)!r!}bc_{0}^{r}\times\bigl{\{}(1-b-c_{0})^{k-r-1}\frac{1}{1+2^{r+1}}-a^{k-r-1}(\frac{1}{1+2^{r+1}}-\frac{1}{2^{r+1}}\bigr{\}}\Bigr{]}.

Eq. (LABEL:eqn:_analytic_formula2) is our main result. Using this equation, we can analytically compute the density of MDS from any degree distribution pkp_{k}. More concretely, we can summarize the procedure as follows. Once the degree density function pkp_{k} is given, we can easily compute the excess degree density function qkq_{k}. Then, by solving (37), (38) and (39), we obtain aa, bb, cnc_{n} (n=0,1,2,)(n=0,1,2,\cdots). Finally, inserting the value aa, bb, cnc_{n} (n=0,1,2,)(n=0,1,2,\cdots) into the above equation (LABEL:eqn:_analytic_formula2), we obtain the density of MDS.

We note that when we consider random network with average degree zz as a special case, we can derive the simple expression ρ=1/z\rho=1/z by using large average degree limit approximation (zz\to\infty) (See supplementary Information).

III Computational results

Here we performed computer simulations to examine the theoretical results (LABEL:eqn:_analytic_formula2) obtained using cavity method. First, we consider regular and random networks constructed with a variety of average degree values (see Fig. 4 (a) and (b)). The results show that cavity method predictions are in excellent agreement with ILP solutions.

Next, we examine the case of scale-free networks. We first generated samples of synthetic scale-free networks with a variety of scaling exponent γ\gamma and average degree <k><k> using the Havel-Hakimi algorithm with random edge swaps (HMC). All samples were generated with a size of N=5000N=5000 nodes. We investigated two different cases. We constructed a set of scale-free network samples with natural cut-off kc=N1k_{c}=N-1 and another set with structural cut-off kc=<k>Nk_{c}=\sqrt{<k>N}. The minimum degree is kmin=2k_{min}=2 in both cases. Fig. 5 shows the results for natural cut-off (i.e. no structural cut-off is considered). The dependence of the MDS density ρ\rho as a function of the average degree <k><k> shows a good agreement with the cavity method analytical predictions (LABEL:eqn:_analytic_formula2), in particular when γ\gamma increases (see Figs. 5 (c)-(e)). By increasing the average degree, the MDS density ρ\rho decreases. For small values of γ\gamma, the predictions of the cavity method deviates from the ILP solutions when average degree increases (see Fig. 5(b)). The reason is because in this case, the network tends to have hubs with very high degree. While these hubs are still visible to ILP method, they are not observed by the cavity method. It is well-known that cavity method addresses better homogeneous networks than extremely inhomogeneous networks. On the other hand, by examining the function of MDS density ρ\rho versus γ\gamma, we clearly observe the influence of the average degree (see Fig. 5(a)). Moreover, in absence of structural cut-off, for high average degree networks, the MDS density ρ\rho for ILP decreases faster than the solution for cavity method when γ\gamma decreases.
We have also considered the case of structural cut-off when constructing finite artificial scale-free networks to address the finite-size effect and eliminate degree correlations. The computer simulations on scale-free networks with structural cut-off shows a different picture (see Fig. 6(a)-(e)). The MDS density ρ\rho does not significantly changes when γ\gamma decreases and remains constant along all the range of γ\gamma values (see Fig. 6(a)). Moreover, when structural cut-off is considered, the agreement between cavity method and ILP results becomes more evident for any value of γ\gamma and average degree (see Fig. 6 (b)-(e)). The reason is because the network tends to be more homogeneous when some hubs with extremely high-degree are knocked out by the structural cut-off.

IV Conclusion

Domination is not only one of the most active research areas in graph theory but also has found abundant real-world applications in many different fields, from engineering to social and natural sciences. With the recent years expansion of networks as data representation framework, domination techniques may provide a rich set of tools to face current network problems in society and nature.

In this work, by using the cavity method and the ultra-discretization procedure, we solved for the first time the MDS problem analytically and derived a combinatorial equation whose computation is easier than that of ILP. By only using the degree distribution of a network as an input information, we can compute the density of an MDS and investigate any dependence with respect to other network variables without using assistance of any complex optimization methods such as ILP or DP.

The present analysis may allow a variety of rich extensions such as computing the corresponding analytical expression for MDS density in directed and bipartite networks.

V Acknowledgements

We thank Prof. Tatsuya Akutsu for insighful comments. J.C.N. was partially supported by MEXT, Japan (Grant-in-Aid No. 25330351), and T.O. was partially supported by JSPS Grants-in-Aid for Scientific Research (Grant Number 15K01200) and Otsuma Grant-in Aid for Individual Exploratory Research (Grant Number S2609).

Refer to caption
Figure 1: Illustration of an MDS. Filled nodes belong to the MDS.
Refer to caption
Figure 2: An example of a tree graph. The highlighted nodes i\j\partial i\backslash j indicate the set of nodes adjacent to ii except for node jj.
Refer to caption
Figure 3: Two subgraphs (AA and BB) are obtained by cutting an edge (ii, jj). Note that subgraph AA and BB includes nodes ii and jj, respectively.
Refer to caption
Figure 4: Comparison of the cavity method and ILP results for the MDS density ρ\rho in regular (a) and random (b) networks constructed with a variety of average degree values. The results are averaged using five network samples. For cavity method the standard deviation (SD) is smaller than symbol. For ILP the SD is shown for each symbol.
Refer to caption
Figure 5: Comparison of the cavity method and ILP results for the MDS density ρ\rho in samples of synthetic scale-free networks generated without structural cut-off kc=N1k_{c}=N-1. ρ\rho as a function of the scaling exponent γ\gamma (a) and average degree <k><k> (b-d) are shown in figure. The results are averaged using five network samples with a size of N=5000N=5000 nodes. For cavity method the standard deviation (SD) is smaller than symbol. For ILP the SD is shown for each symbol.
Refer to caption
Figure 6: Comparison of the cavity method and ILP results for the MDS density ρ\rho in samples of synthetic scale-free networks generated with structural cut-off kc=<k>Nk_{c}=\sqrt{<k>N}. ρ\rho as a function of the scaling exponent γ\gamma (a) and average degree <k><k> (b-d) are shown in figure. The results are averaged using five network samples with a size of N=5000N=5000 nodes. For cavity method the standard deviation (SD) is smaller than symbol. For ILP the SD is shown for each symbol.

References

  • (1) M.E.J. Newman, Networks: An Introduction. (Oxford University Press, New York, 2010)
  • (2) G. Caldarelli, Scale-Free Networks: Complex Webs in Nature and Technology (Oxford Univ. Press, Oxford, 2007).
  • (3) A.-L. Barabási, R. Albert, Science 286, 509 (1999).
  • (4) R. Pastor-Satorras, A. Vespignani, Phys. Rev. Lett. 86, 3200 (2001).
  • (5) R. Cohen, K. Erez, D. ben-Avraham and S. Havlin, Phys. Rev. Lett. 85, 4626 (2000).
  • (6) R. Albert, H. Jeong, A.-L. Barabasi, Nature 406, 378 (2000).
  • (7) A. E. Motter, Phys. Rev. Lett. 93, 098701 (2004).
  • (8) T.W. Haynes, S.T. Hedetniemi and P.J. Slater, Fundamentals of Domination in Graphs (Pure Applied Mathematics, Chapman and Hall/CRC, New York, 1998).
  • (9) I. Stojmenovic, M. Seddigh and J. Zunic, IEEE Trans. Parallel Distributed Systems 13, 14 (2002).
  • (10) W. Chena, Z. Lub and W. Wub, Information Sciences 269, 286 (2014).
  • (11) L. Kelleher and M. Cozzens, Graphs. Math. Soc. Sciences 16, 267 (1988).
  • (12) F. Wang, et al. , Theo. Comp. Sci. 412, 265 (2011).
  • (13) S. Eubank, V.S. Anil Kumar, M.V. Marathe, A. Srinivasan and N. Wang, in SODA ’04 Proceedings of the fifteenth annual ACM-SIAM symposium on Discrete algorithms pp. 718-727 (2004).
  • (14) A. Vazquez, BMC Systems Biology 3, 81 (2009).
  • (15) J.C. Nacher and T. Akutsu, New J. Phys. 14, 073005 (2012).
  • (16) J.C. Nacher and T. Akutsu, Journal of Physics: Conf. Ser. 410, 012104 (2013).
  • (17) J.C. Nacher and T. Akutsu, Scientific Reports 3, 1647 (2013).
  • (18) Y. Yang, J. Wang and A.E. Motter, Phys. Rev. Lett. 109, 258701 (2012).
  • (19) S. Wuchty, Proc. Natl. Acad. Sci. USA 111, 7156 (2014).
  • (20) F. Molnár, S. Sreenivasan, B.K. Szymanski and G. Korniss, Scientific Reports 3, 1736 (2013).
  • (21) J.C. Nacher and T. Akutsu, Phys. Rev. E 91, 012826 (2015).
  • (22) F. Molnár, N. Derzsy, B.K. Szymanski and G. Korniss, Scientific Reports 5, 8321 (2015).
  • (23) M. Garey, and D.S. Johnson, Computers and Intractability: A Guide to the Theory of NP-Completeness, (W. H. Freeman, 1979).
  • (24) C. Cooper and M. Zito, Discrete Applied Mathematics 157, 2010 (2009).
  • (25) M. Mezard, G. Parisi G and M.A. Virasoro, Spin-Glass Theory and Beyond (Lecture notes in Physics vol. 9) (World Scientific, Singapore, 1987)
  • (26) M. Mezard and G. Parisi, J. Phys. Lett. 46 L771, (1985).
  • (27) M. Weigt and A.K. Hartmann, Phys. Rev. Lett. 84 6118 (2000).
  • (28) M. Mezard and G. Parisi, J. Physique 47 1285, (1986).
  • (29) L. Zdeborova and M. Mezard, J. Stat. Mech. P05003 (2006).
  • (30) L. Dall’Asta, P. Pin, and A. Ramezanpour, Phys. Rev. E 80, 061136 (2009).
  • (31) C. Lucibello and F. Ricci-Tersenghi, Int. J. Statistical Mechanics vol. 2014, Article ID 136829, (2014)
  • (32) M. Bayati, C. Borgs, A. Braunstein, J. Chayes, A. Ramezanpour, and R. Zecchina, Phys. Rev. Lett. 101, 037208 (2008).
  • (33) Y.-Y. Liu, J.-J. Slotine, A.-L. Barabási, Nature 473, 167 (2011).
  • (34) M. Mezard, G. Parisi, Journal of Statistical Physics 111, 1 (2003)
  • (35) M. Mezard, F. Ricci-Tersenghi, R. Zecchina, Journal of Statistical Physics, 111, 505 (2003)
  • (36) O. Rivoire, G. Biroli, O.C. Martin, M. Mezard, Eur. Phys. J. B, 37, 55 (2004)
  • (37) D. Takahashi and J. Satsuma, J. Phys. Soc. Jpn. 59, 3514 (1990).
  • (38) D. Takahashi, T. Tokihiro, B.Grammaticos, Y. Ohta, A Ramani, J. Phys. A 30 7953 (1997).
  • (39) T. Tokihiro, D. Takahashi, J. Matsukidaira, J. Phys. A 33 607 (2000).
  • (40) T. Tokihiro, D. Takahashi, J. Matsukidaira and J. Satsuma, Phys. Rev. Lett. 76 3247 (1996).
  • (41) T. Ochiai, J.C. Nacher, Journal of Mathematical Physics 46 063507 (2005).
  • (42) M. Bayati, C. Nair, http://arxiv.org/abs/cond-mat/0607290
  • (43) J.H. Zhao, Y. Habibulla, H.J. Zhou Journal of Statistical Physics159 1154 (2015).

V.1

V.1.1