This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Complementary Vanishing Graphs

Craig Erickson Saint Paul, MN, USA (cerickson.phd@gmail.com).    Luyining Gan Department of Mathematics and Statistics, University of Nevada Reno, Reno, NV 89557, USA (lgan@unr.edu).    Jürgen Kritschgau Department of Mathematics, Carnegie Mellon University, Pittsburgh, PA 15213, USA (jkritsch@andrew.cmu.edu).    Jephian C.-H. Lin Department of Applied Mathematics, National Sun Yat-sen University, Kaohsiung 80424, Taiwan (jephianlin@gmail.com)    Sam Spiro Department of Mathematics, UC San Diego, La Jolla, CA 92093, USA (sspiro@ucsd.edu).
Abstract

Given a graph GG with vertices {v1,,vn}\{v_{1},\ldots,v_{n}\}, we define 𝒮(G)\mathcal{S}(G) to be the set of symmetric matrices A=[ai,j]A=[a_{i,j}] such that for iji\neq j we have ai,j0a_{i,j}\neq 0 if and only if vivjE(G)v_{i}v_{j}\in E(G). Motivated by the Graph Complement Conjecture, we say that a graph GG is complementary vanishing if there exist matrices A𝒮(G)A\in\mathcal{S}(G) and B𝒮(G¯)B\in\mathcal{S}(\overline{G}) such that AB=OAB=O. We provide combinatorial conditions for when a graph is or is not complementary vanishing, and we characterize which graphs are complementary vanishing in terms of certain minimal complementary vanishing graphs. In addition to this, we determine which graphs on at most 88 vertices are complementary vanishing.

Keywords: Graph Complement Conjecture, Minimum rank, Maximum nullity, Inverse eigenvalue problem for graphs

AMS subject classifications: 05C50, 15A18, 15B57, 65F18.

1 Introduction

Let GG be a simple graph with vertex set {v1,v2,,vn}\{v_{1},v_{2},\dots,v_{n}\}. The set of symmetric matrices described by GG, 𝒮(G)\mathcal{S}(G), is the set of all real symmetric n×nn\times n matrices A=[ai,j]A=\begin{bmatrix}a_{i,j}\end{bmatrix} such that for iji\neq j we have ai,j0a_{i,j}\neq 0 if and only if vivjE(G)v_{i}v_{j}\in E(G). Note that no restrictions are placed on the diagonal entries of A𝒮(G)A\in\mathcal{S}(G). Given a graph GG, the inverse eigenvalue problem of a graph (IEP-G for short) refers to the problem of determining the spectra of matrices in 𝒮(G)\mathcal{S}(G). A large amount of work has been done in this area, see for example [12, 11, 5, 1, 7, 6].

Specifying exactly which graphs can achieve a particular spectrum is generally hard, and because of this, much of the work in the IEP-G considers parameters that measure more general spectral properties. For example, the minimum rank, mr(G)\text{mr}(G), of a graph GG is defined as the minimum rank of a matrix A𝒮(G)A\in\mathcal{S}(G). The maximum nullity, M(G)M(G), of a graph GG is defined as the maximum nullity of a matrix A𝒮(G)A\in\mathcal{S}(G). Since 𝒮(G)\mathcal{S}(G) is closed under translations by rInrI_{n} for real values rr, the parameter M(G)M(G) is also equal to the largest multiplicity of an eigenvalue of A𝒮(G)A\in\mathcal{S}(G). It follows from the definitions that mr(G)+M(G)=n\text{mr}(G)+M(G)=n whenever GG is an nn-vertex graph. The minimum rank and maximum nullity of graphs was one of the main subjects of the 2006 American Institute of Mathematics workshop [3]. This workshop was a catalyst for research on the IEP-G, and there the following conjecture was posed.

Conjecture 1.1 (Graph Complement Conjecture).

For any graph GG of order nn,

mr(G)+mr(G¯)n+2,\text{mr}(G)+\text{mr}(\overline{G})\leq n+2,

where G¯\overline{G} denotes the complement of GG. Equivalently,

M(G)+M(G¯)n2.M(G)+M(\overline{G})\geq n-2.

Conjecture 1.1 is an example of a Nordhaus–Gaddum problem, which are problems where one tries to prove bounds for f(G)+f(G¯)f(G)+f(\overline{G}) where ff is some graph-theoretic function. An analogous Nordhaus–Gaddum conjecture for the Colin de Verdière number μ(G)\mu(G) (defined in [8]) was made by Kotlov et al. [14]. Levene, Oblak and Šmigoc verified a Nordhaus–Gaddum conjecture for the minimum number of distinct eigenvalues for several graph families [15].

Conjecture 1.1 is open in general and is evidently difficult. Some results related to Conjecture 1.1 for joins of graphs were studied and established by Barioli et al. [4], and the authors also showed that Conjecture 1.1 is true for graphs on up to 1010 vertices. Li, Nathanson, and Phillips [16] showed that it suffices to prove Conjecture 1.1 for a certain class of graphs called complement critical graphs. Theorem 3.16 in [3] implies that Conjecture 1.1 is true for trees. On the other hand, it was shown in [2] that the zero forcing number Z(G)Z(G) is an upper bound of M(G)M(G), and it is known Z(G)+Z(G¯)n2Z(G)+Z(\overline{G})\geq n-2 for any graph on nn vertices [10, 13].

With Conjecture 1.1 in mind, we make the following definition.

Definition 1.2.

A graph GG is said to be complementary vanishing if there are matrices A𝒮(G)A\in\mathcal{S}(G) and B𝒮(G¯)B\in\mathcal{S}(\overline{G}) such that AB=OAB=O (the zero matrix).

The motivation for this definition is the following observation.

Proposition 1.3.

If GG is an nn-vertex graph which is complementary vanishing, then

M(G)+M(G¯)n.M(G)+M(\overline{G})\geq n.

In particular, Conjecture 1.1 holds for GG.

Proof.

By assumption there exist A𝒮(G),B𝒮(G¯)A\in\mathcal{S}(G),B\in\mathcal{S}(\overline{G}) such that AB=OAB=O. Thus the sum of the nullities of AA and BB is at least nn. This in turn implies that M(G)+M(G¯)nM(G)+M(\overline{G})\geq n as desired. ∎

The goal of this paper is to establish necessary and sufficient conditions for a graph to be complementary vanishing. Our main result characterizes which graphs are complementary vanishing in terms of a class of “minimal” complementary vanishing graphs \mathcal{M}.

Definition 1.4.
  • Let \mathcal{M} denote the set of complementary vanishing graphs GG such that both GG and G¯\overline{G} are connected.

  • Let \mathcal{R} denote the smallest set of graphs containing \mathcal{M} which is closed under taking disjoint unions, joins, and complements111We will use GHG\sqcup H to denote the disjoint union of two graphs G,HG,H. We recall that the join GHG\vee H of two graphs G,HG,H is obtained by taking the disjoint union of GG and HH and then adding all possible edges between vertices of GG to vertices of HH..

  • Let 𝒞\mathcal{C} denote all the graphs GG such that either in GG or G¯\overline{G} there exist distinct vertices u,v,wu,v,w with v,wN(u)v,w\notin N(u), |N(u)N(w)|=1|N(u)\setminus N(w)|=1, and |N(u)N(v)|=0|N(u)\setminus N(v)|=0.

Theorem 1.5.

A graph GG is complementary vanishing if and only if G𝒞G\in\mathcal{R}\setminus\mathcal{C}.

Roughly speaking, Theorem 1.5 states that a graph GG is complementary vanishing if and only if GG can be constructed out of minimal complementary vanishing graphs \mathcal{M} while avoiding a particular combinatorial structure.

Example 1.6.

Let GG be the graph obtained by adding a perfect matching to K4,4K_{4,4}. We will use Theorem 1.5 to determine whether GG is complementary vanishing or not. It is easy to check that G𝒞G\notin\mathcal{C}, so GG will be complementary vanishing if and only if GG\in\mathcal{R}. Observe that

G=(K2K2)(K2K2)=((K1K1)(K1K1))((K1K1)(K1K1)).G=(K_{2}\sqcup K_{2})\vee(K_{2}\sqcup K_{2})=((K_{1}\vee K_{1})\sqcup(K_{1}\vee K_{1}))\vee((K_{1}\vee K_{1})\sqcup(K_{1}\vee K_{1})).

Thus, to have GG\in\mathcal{R}, it would suffice to have K1K_{1}\in\mathcal{M}, and it is easy to verify that this is the case. This proves that GG is complementary vanishing.

One can also use Theorem 1.5 to determine which trees are complementary vanishing.

Corollary 1.7.

A tree is complementary vanishing if and only if it is a star.

Proof.

If TT is a tree which is not a star, then there exists an induced path uxvwuxvw in TT where uu is a leaf. Note that v,wN(u),|N(u)N(w)|=1v,w\notin N(u),\ |N(u)\setminus N(w)|=1, and |N(u)N(v)|=0|N(u)\setminus N(v)|=0. We conclude that T𝒞T\in\mathcal{C}, and hence TT is not complementary vanishing by Theorem 1.5.

If TT is a star, then TT\in\mathcal{R} (since TT is the disjoint union of K1K_{1}’s joined to a K1K_{1}) and it is not difficult to show that T𝒞T\notin\mathcal{C}, so Theorem 1.5 implies that TT is complementary vanishing. Alternatively, one can show that TT is complementary vanishing directly by taking A𝒮(T)A\in\mathcal{S}(T) to be the adjacency matrix of TT and B𝒮(T¯)B\in\mathcal{S}(\overline{T}) to be the Laplacian matrix of the complement of TT, since in this case AB=OAB=O. ∎

By Theorem 1.5, it suffices to characterize the set of graphs \mathcal{M} in order to characterize the set of complementary vanishing graphs. To this end, we determine exactly which graphs on at most 8 vertices are in \mathcal{M}.

Theorem 1.8.

A graph GG on n8n\leq 8 vertices is in \mathcal{M} if and only if it is one of the graphs listed in Appendix A.

A precise break down of the number of pairs {G,G¯}\{G,\overline{G}\} of connected graphs on n8n\leq 8 vertices which are complementary vanishing is given in Table 1.

nn not complementary vanishing complementary vanishing total pairs
1 0 1 1
2 0 0 0
3 0 0 0
4 1 0 1
5 5 0 5
6 34 0 34
7 327 4 331
8 4917 32 4949
Table 1: Each row breaks down the number of pairs {G,G¯}\{G,\overline{G}\} which are complementary vanishing, where GG and G¯\overline{G} are connected graphs on nn vertices.

Organization and Notation. The rest of the paper is organized as follows. In Section 2, we give combinatorial conditions for when a graph is complementary vanishing or not. In Section 3, we prove Theorem 1.5 by studying a slightly stronger notion of being complementary vanishing. In Section 4, we prove Theorem 1.8 and describe several algorithms which can be used to test whether a graph is complementary vanishing or not. We close with a few open problems in Section 5.

Throughout we use standard notation from graph theory and linear algebra. Given a graph GG, we define the open neighborhood NG(v)N_{G}(v) to be the set of vertices adjacent to vv in GG, and the closed neighborhood NG[v]=NG(v){v}N_{G}[v]=N_{G}(v)\cup\{v\}. We will drop the subscript on NN when the graph GG is clear from context.

We write Om×nO_{m\times n} for the m×nm\times n zero matrix, which we denote simply by OO whenever the dimensions are clear from context. We use bold lower case letters to denote vectors, and in particular we write 𝟎{\bf 0} and 𝟏{\bf 1} to denote the all 0’s and all 1’s vectors.

2 Combinatorial Conditions for Complementary Vanishing Graphs

The results in this section use structural properties of graphs to determine whether it is possible for a graph to be complementary vanishing. Most of our conditions will show that a graph cannot be complementary vanishing since the diagonal entries for AA and BB are over-constrained. The following observation leads to our first condition for the diagonal entries of AA and BB. Here and throughout the paper, if SS is a set of vertices of a graph GG then S¯:=V(G)S\overline{S}:=V(G)\setminus S.

Observation 2.1.

If vV(G)v\in V(G), then NG¯[v]=NG(v)¯N_{\overline{G}}[v]=\overline{N_{G}(v)} and NG¯(v)=NG[v]¯N_{\overline{G}}(v)=\overline{N_{G}[v]}.

Lemma 2.2.

Let A𝒮(G)A\in\mathcal{S}(G), B𝒮(G¯)B\in\mathcal{S}(\overline{G}). If AB=OAB=O, then ai,ia_{i,i} or bi,ib_{i,i} is 0 for each 1in1\leq i\leq n.

Proof.

Because AB=OAB=O, taking the dot product of the ithi\text{th} row of AA and the ithi\text{th} column of BB gives

0=1knai,kbk,i=ai,ibi,i+kNG(i)NG¯(i)ai,kbk,i=ai,ibi,i+kNG(i)NG[i]¯ai,kbk,i,0=\sum_{1\leq k\leq n}a_{i,k}b_{k,i}=a_{i,i}b_{i,i}+\sum_{k\in N_{G}(i)\cap N_{\overline{G}}(i)}a_{i,k}b_{k,i}=a_{i,i}b_{i,i}+\sum_{k\in N_{G}(i)\cap\overline{N_{G}[i]}}a_{i,k}b_{k,i},

where the second equality used, for example, that ai,k=0a_{i,k}=0 if kNG[i]k\notin N_{G}[i] by definition of 𝒮(G)\mathcal{S}(G), and the third used Observation 2.1. Since NG(i)NG[i]¯=N_{G}(i)\cap\overline{N_{G}[i]}=\emptyset, this reduces to

0=ai,ibi,i.0=a_{i,i}b_{i,i}.

This implies ai,i=0a_{i,i}=0 or bi,i=0b_{i,i}=0. ∎

The key fact in proving Lemma 2.2 is that NG(i)N_{G}(i) is contained in NG[i]N_{G}[i] (since this made the final sum equal to 0). This step in the proof can be generalized to pairs of vertices iji\neq j when the closed neighborhood of jj contains the neighborhood of ii.

Lemma 2.3.

Let A𝒮(G)A\in\mathcal{S}(G) and B𝒮(G¯)B\in\mathcal{S}(\overline{G}) be such that AB=OAB=O. For distinct vertices i,jV(G)i,j\in V(G), if NG(i)NG[j]N_{G}(i)\subseteq N_{G}[j], then we have the following.

  1. (1)

    If iGji\sim_{G}j, then bj,j=0b_{j,j}=0.

  2. (2)

    If i≁Gji\not\sim_{G}j, then ai,i=0a_{i,i}=0.

Proof.

Let i,ji,j be distinct vertices with NG(i)NG[j]N_{G}(i)\subseteq N_{G}[j]. Then we have NG(i)NG[j]¯=N_{G}(i)\cap\overline{N_{G}[j]}=\emptyset. Taking the dot product of the ithi\text{th} row of AA and jthj\text{th} column of BB gives

0\displaystyle 0 =ai,ibi,j+ai,jbj,j+kNG(i)NG[j]¯ai,kbk,j\displaystyle=a_{i,i}b_{i,j}+a_{i,j}b_{j,j}+\sum_{k\in N_{G}(i)\cap\overline{N_{G}[j]}}a_{i,k}b_{k,j}
=ai,ibi,j+ai,jbj,j.\displaystyle=a_{i,i}b_{i,j}+a_{i,j}b_{j,j}.

Suppose iGji\sim_{G}j. Then ai,j0a_{i,j}\neq 0 and bi,j=0b_{i,j}=0 (since i≁G¯ji\not\sim_{\overline{G}}j). Therefore, ai,jbj,j=0a_{i,j}b_{j,j}=0, which implies that bj,j=0b_{j,j}=0. A completely symmetric argument gives the result when i≁Gji\not\sim_{G}j. ∎

We can also extract information from the dot product of the ithi\text{th} row of AA and the jjth column of BB if we “almost” have NG(i)NG[j]N_{G}(i)\subseteq N_{G}[j].

Lemma 2.4.

Let A𝒮(G)A\in\mathcal{S}(G) and B𝒮(G¯)B\in\mathcal{S}(\overline{G}) be such that AB=OAB=O. For distinct vertices i,jV(G)i,j\in V(G), if NG(i)NG[j]={v}N_{G}(i)\setminus N_{G}[j]=\{v\} for some vV(G)v\in V(G), then we have the following.

  1. (1)

    If iGji\sim_{G}j, then bj,j0b_{j,j}\neq 0 and aj,j=0a_{j,j}=0.

  2. (2)

    If i≁Gji\not\sim_{G}j, then ai,i0a_{i,i}\neq 0 and bi,i=0b_{i,i}=0.

Proof.

Let NG(i)NG[j]={v}N_{G}(i)\setminus N_{G}[j]=\{v\}. Then NG(i)NG[j]¯={v}N_{G}(i)\cap\overline{N_{G}[j]}=\{v\} and taking the dot product of the ithi\text{th} row of AA and the jthj\text{th} column of BB gives

0\displaystyle 0 =ai,ibi,j+ai,jbj,j+kNG(i)NG[j]¯ai,kbk,j\displaystyle=a_{i,i}b_{i,j}+a_{i,j}b_{j,j}+\sum_{k\in N_{G}(i)\cap\overline{N_{G}[j]}}a_{i,k}b_{k,j}
=ai,ibi,j+ai,jbj,j+ai,vbv,j\displaystyle=a_{i,i}b_{i,j}+a_{i,j}b_{j,j}+a_{i,v}b_{v,j}

Suppose iGji\sim_{G}j. Then we have that ai,j0a_{i,j}\neq 0 and bi,j=0b_{i,j}=0 (since i≁G¯ji\not\sim_{\overline{G}}j). This implies that ai,jbj,j+ai,vbv,j=0a_{i,j}b_{j,j}+a_{i,v}b_{v,j}=0 where ai,j0a_{i,j}\neq 0, ai,v0a_{i,v}\neq 0, and bv,j0b_{v,j}\neq 0. Solving for bj,jb_{j,j} shows that bj,j0b_{j,j}\neq 0. Moreover, aj,j=0a_{j,j}=0 by Lemma 2.2. A completely symmetric argument gives the result when i≁Gji\not\sim_{G}j. ∎

Applying both Lemmas 2.3 and 2.4 can lead to contradictions along the diagonal of AA. This gives the following corollary, which helps motivate the set 𝒞\mathcal{C} defined in the introduction.

Corollary 2.5.

Let GG be a graph such that there exist distinct vertices u,v,wu,v,w with v,wN(u)v,w\notin N(u), |N(u)N(w)|=1|N(u)\setminus N(w)|=1, and |N(u)N(v)|=0|N(u)\setminus N(v)|=0. Then GG is not complementary vanishing.

Proof.

For the sake of contradiction suppose that GG is complementary vanishing with matrices A𝒮(G)A\in\mathcal{S}(G) and B𝒮(G¯).B\in\mathcal{S}(\overline{G}). Notice that |N(u)N(w)|=1|N(u)\setminus N(w)|=1 and u≁wu\not\sim w implies au,u0a_{u,u}\neq 0 by Lemma 2.4. Furthermore, N(u)N(v)N(u)\subseteq N(v) and u≁vu\not\sim v implies au,u=0a_{u,u}=0 by Lemma 2.3. This is a contradiction, so it follows that GG is not complementary vanishing. ∎

Lemma 2.2, Lemma 2.3, and Lemma 2.4 are the elementary calculations we use to show that particular graphs are not complementary vanishing. One improvement on these elementary calculations is to look for induced cycles with particular neighborhood properties.

Lemma 2.6.

Let GG be a graph that has an induced odd cycle CC on vertices 1,2,,2k+11,2,\ldots,2k+1 and a vertex vv such that 1,2,,2k+1N(v)1,2,\ldots,2k+1\notin N(v) and

1i2k+1N(i){1,,2k+1}N(v).\bigcup_{1\leq i\leq 2k+1}N(i)\setminus\{1,\ldots,2k+1\}\subseteq N(v).

If A𝒮(G)A\in\mathcal{S}(G) and B𝒮(G¯)B\in\mathcal{S}(\overline{G}) satisfy AB=OAB=O, then ai,i0a_{i,i}\neq 0 for some i{1,2,,2k+1}i\in\{1,2,\ldots,2k+1\}.

Note that N(i){1,,2k+1}N(v)\bigcup N(i)\setminus\{1,\ldots,2k+1\}\subseteq N(v) is equivalent to saying that vv is adjacent to the entire “external neighborhood” of {1,,2k+1}\{1,\ldots,2k+1\}.

Proof.

Assume for the sake of contradiction that there exist such A,BA,B with ai,i=0a_{i,i}=0 for all i{1,,2k+1}i\in\{1,\dots,2k+1\}. By taking the dot product of the ithi\text{th} row of AA with the vthv\text{th} column of BB, we obtain (writing indices mod 2k+12k+1 and using the symmetry of AA and BB)

ai1,ibi1,v+ai,i+1bi+1,v=0,a_{i-1,i}b_{i-1,v}+a_{i,i+1}b_{i+1,v}=0,

where we used that ai,i=0a_{i,i}=0 and the fact that the neighborhood of vv contains the external neighborhood of CC. Also note that each term in this sum is nonzero since implicitly we have assumed v{1,,2k+1}v\notin\{1,\ldots,2k+1\}.

For convenience, define c2i1=ai,i+1c_{2i-1}=a_{i,i+1} and c2i=bi+1,vc_{2i}=b_{i+1,v} for all 1i2k+11\leq i\leq 2k+1. Notice that cjc_{j} is an entry from BB if jj is even, and cjc_{j} is an entry from AA when jj is odd. The relationship above can be rewritten as c2i3c2i4+c2i1c2i=0c_{2i-3}c_{2i-4}+c_{2i-1}c_{2i}=0, where the indices of cc are written mod 4k+24k+2. In particular, this implies that amongst all of the terms cjcj+1c_{j}c_{j+1}, exactly 2k+12k+1 of them are negative. Thus,

1=sgn(jcj2)=jsgn(cjcj+1)=(1)2k+1=1,1=\operatorname{sgn}\left(\prod_{j}c_{j}^{2}\right)=\prod_{j}\operatorname{sgn}(c_{j}c_{j+1})=(-1)^{2k+1}=-1,

which is a contradiction. ∎

The following example shows the power of Lemma 2.6.

11223355667744
Figure 1: A graph that is not complementary vanishing due to Lemma 2.6. Notice that Corollary 2.5 does not apply for this graph.
Example 2.7.

Let GG be the graph in Figure 1. Suppose A=[ai,j]𝒮(G)A=\begin{bmatrix}a_{i,j}\end{bmatrix}\in\mathcal{S}(G) and B=[bi,j]𝒮(G¯)B=\begin{bmatrix}b_{i,j}\end{bmatrix}\in\mathcal{S}(\overline{G}) are such that AB=OAB=O. By applying the latter case of Lemma 2.4 with u=1u=1 and w=5w=5, we know b1,10b_{1,1}\neq 0 and a1,1=0a_{1,1}=0. By symmetry, a2,2=a3,3=0a_{2,2}=a_{3,3}=0. However, by Lemma 2.6 with the odd cycle on {1,2,3}\{1,2,3\} and v=4v=4, at least one of a1,1a_{1,1}, a2,2a_{2,2}, and a3,3a_{3,3} is nonzero, which is a contradiction. Thus, GG is not complementary vanishing.

We end the section with some ways to construct complementary vanishing graphs. We say that two vertices u,vu,v are twins in a graph GG if N(u){v}=N(v){u}N(u)\setminus\{v\}=N(v)\setminus\{u\}. Note that we allow for twins to either be adjacent or nonadjacent.

Proposition 2.8.

If GG is a graph such that every vertex has at least one twin, then GG is complementary vanishing.

Proof.

Partition the vertex set of GG into equivalence classes of twins so that V(G)=V=V1VkV(G)=V=V_{1}\cup\cdots\cup V_{k} where every u,vViu,v\in V_{i} are twins. By the hypothesis, |Vi|2|V_{i}|\geq 2 for i=1,,ki=1,\dots,k. Let 𝐱(i),𝐲(i)|V|{\bf x}^{(i)},{\bf y}^{(i)}\in\mathbb{R}^{|V|} be orthogonal vectors that are nonzero on entries indexed by ViV_{i} and zero otherwise. Since |Vi|2|V_{i}|\geq 2 for 1ik1\leq i\leq k, we know that the vectors 𝐱(i),𝐲(i){\bf x}^{(i)},{\bf y}^{(i)} exist for 1ik1\leq i\leq k. Notice that (𝐱(i))𝐲(j)=0({\bf x}^{(i)})^{\top}{\bf y}^{(j)}=0 for all 1i,jk1\leq i,j\leq k.

Let P={(i,j):uvE(G),uVi,vVj}P=\{(i,j):uv\in E(G),u\in V_{i},v\in V_{j}\} and P¯={(i,j):uvE(G),uVi,vVj}\overline{P}=\{(i,j):uv\notin E(G),u\in V_{i},v\in V_{j}\}. Notice that every pair (i,j)(i,j) with i,j{1,,k}i,j\in\{1,\ldots,k\} distinct appears in exactly one of the sets PP or P¯\overline{P} since every two vertices in ViV_{i} and VjV_{j} are twins. Let

A=(i,j)P𝐱(i)(𝐱(j))A=\sum_{(i,j)\in P}{\bf x}^{(i)}({\bf x}^{(j)})^{\top}

and

B=(i,j)P¯𝐲(i)(𝐲(j)),B=\sum_{(i,j)\in\overline{P}}{\bf y}^{(i)}({\bf y}^{(j)})^{\top},

where each term is a square matrix indexed by VV resulting from an outer product. Notice that A𝒮(G)A\in\mathcal{S}(G), B𝒮(G¯)B\in\mathcal{S}(\overline{G}), and that

AB=(i,j)P(,m)P¯𝐱(i)(𝐱(j))𝐲()(𝐲(m))=O.AB=\sum_{(i,j)\in P}\sum_{(\ell,m)\in\overline{P}}{\bf x}^{(i)}({\bf x}^{(j)})^{\top}{\bf y}^{(\ell)}({\bf y}^{(m)})^{\top}=O.

This completes the proof. ∎

Proposition 2.8 is tight in the sense that there are graphs GG which are not complementary vanishing where all but one vertex in GG has a twin. Indeed, consider P3P2P_{3}\sqcup P_{2} with uxvuxv the copy of P3P_{3} and ywyw the copy of P2P_{2}. Then xx does not have a twin, |N(u)N(w)|=1|N(u)\setminus N(w)|=1, |N(u)N(v)|=0|N(u)\setminus N(v)|=0 and v,wN(u)v,w\notin N(u). Thus, P3P2P_{3}\sqcup P_{2} is not complementary vanishing by Corollary 2.5.

Lemma 2.9.

Let GG be a graph with A𝒮(G)A\in\mathcal{S}(G) and B𝒮(G¯)B\in\mathcal{S}(\overline{G}) such that AB=OAB=O. If ai,i=0a_{i,i}=0, then the graph HH obtained by adding a new vertex jj with NH(j)=NG(i)N_{H}(j)=N_{G}(i) is complementary vanishing.

Proof.

Let n=|V(G)|n=|V(G)|. For convenience of illustrating the matrices, we assume i=ni=n and write

A=[C𝐮𝐮0] and B=[D𝐯𝐯b].A=\begin{bmatrix}C&{\bf u}\\ {\bf u}^{\top}&0\\ \end{bmatrix}\text{ and }B=\begin{bmatrix}D&{\bf v}\\ {\bf v}^{\top}&b\\ \end{bmatrix}.

Thus, AB=OAB=O can be expanded as

CD+𝐮𝐯\displaystyle CD+{\bf u}{\bf v}^{\top} =O,\displaystyle=O,
C𝐯+b𝐮\displaystyle C{\bf v}+b{\bf u} =𝟎,\displaystyle={\bf 0},
𝐮D\displaystyle{\bf u}^{\top}D =𝟎,\displaystyle={\bf 0}^{\top},
𝐮𝐯\displaystyle{\bf u}^{\top}{\bf v} =0.\displaystyle=0.

Let j=n+1j=n+1 and ζ=12𝟏\zeta=\frac{1}{\sqrt{2}}{\bf 1}, where 𝟏{\bf 1} is the all 11’s vector in 2\mathbb{R}^{2}. Then we may construct the following matrices

A=[C𝐮ζζ𝐮O2,2] and B=[D𝐯ζζ𝐯E],A^{\prime}=\begin{bmatrix}C&{\bf u}\zeta^{\top}\\ \zeta{\bf u}^{\top}&O_{2,2}\\ \end{bmatrix}\text{ and }B^{\prime}=\begin{bmatrix}D&{\bf v}\zeta^{\top}\\ \zeta{\bf v}^{\top}&E\\ \end{bmatrix},

where EE is either

b2[1111] or [1111]\frac{b}{2}\begin{bmatrix}1&1\\ 1&1\end{bmatrix}\text{ or }\begin{bmatrix}1&-1\\ -1&1\end{bmatrix}

depending on if b0b\neq 0 or b=0b=0. Note that with this we always have ζE=bζ\zeta^{\top}E=b\zeta^{\top}.

By direct computation of each block of ABA^{\prime}B^{\prime}, we have

CD+𝐮ζζ𝐯\displaystyle CD+{\bf u}\zeta^{\top}\zeta{\bf v}^{\top} =CD+𝐮𝐯=O,\displaystyle=CD+{\bf u}{\bf v}^{\top}=O,
C𝐯ζ+𝐮ζE\displaystyle C{\bf v}\zeta^{\top}+{\bf u}\zeta^{\top}E =(C𝐯+b𝐮)ζ=𝟎ζ=O,\displaystyle=(C{\bf v}+b{\bf u})\zeta^{\top}={\bf 0}\zeta^{\top}=O,
ζ𝐮D\displaystyle\zeta{\bf u}^{\top}D =ζ𝟎=O,\displaystyle=\zeta{\bf 0}^{\top}=O,
ζ𝐮𝐯ζ\displaystyle\zeta{\bf u}^{\top}{\bf v}\zeta^{\top} =0ζζ=O.\displaystyle=0\zeta\zeta^{\top}=O.

Along with the fact that A𝒮(H)A^{\prime}\in\mathcal{S}(H) and B𝒮(H¯)B^{\prime}\in\mathcal{S}(\overline{H}), we have that HH and H¯\overline{H} are both complementary vanishing. ∎

Note that the above proof easily generalizes to duplicating a vertex any number of times or duplicating multiple vertices.

3 Proof of Theorem 1.5

In order to prove Theorem 1.5, we will need to understand when the disjoint union and join of two graphs are complementary vanishing, and to do this we need consider graphs which are in some sense “robust” with respect to being complementary vanishing.

3.1 Robust Graphs

A graph GG is said to be α\alpha-robust if there are matrices A𝒮(G)A\in\mathcal{S}(G) and B𝒮(G¯)B\in\mathcal{S}(\overline{G}) such that AB=OAB=O and ker(A)\operatorname{ker}(A) contains a nowhere-zero vector. Similarly we say that GG is β\beta-robust if there are matrices A𝒮(G)A\in\mathcal{S}(G) and B𝒮(G¯)B\in\mathcal{S}(\overline{G}) such that AB=OAB=O and ker(B)\operatorname{ker}(B) contains a nowhere-zero vector. Note that α\alpha-robust and β\beta-robust graphs are in particular complementary vanishing.

Proposition 3.1.

Let GG and HH be two graphs. Then the following are equivalent:

  • (1a)

    The graphs GG and HH are both α\alpha-robust.

  • (1b)

    The disjoint union GHG\sqcup H is α\alpha-robust.

  • (1c)

    The disjoint union GHG\sqcup H is complementary vanishing.

Similarly, the following are equivalent:

  • (2a)

    The graphs GG and HH are both β\beta-robust.

  • (2b)

    The join GHG\vee H is β\beta-robust.

  • (2c)

    The join GHG\vee H is complementary vanishing.

Proof.

Suppose GG and HH are both α\alpha-robust. By definition, there exist matrices AG𝒮(G)A_{G}\in\mathcal{S}(G), AH𝒮(H)A_{H}\in\mathcal{S}(H), BG𝒮(G¯)B_{G}\in\mathcal{S}(\overline{G}), BH𝒮(H¯)B_{H}\in\mathcal{S}(\overline{H}) such that AGBG=OA_{G}B_{G}=O and AHBH=OA_{H}B_{H}=O. Moreover, ker(AG)\operatorname{ker}(A_{G}) and ker(AH)\operatorname{ker}(A_{H}) contain nowhere-zero vectors 𝐯G{\bf v}_{G} and 𝐯H{\bf v}_{H}, respectively. Thus, we have

A=[AGOOAH]𝒮(GH),B=[BG𝐯G𝐯H𝐯H𝐯GBH]𝒮(GH¯),A=\begin{bmatrix}A_{G}&O\\ O&A_{H}\end{bmatrix}\in\mathcal{S}(G\sqcup H),\ B=\begin{bmatrix}B_{G}&{\bf v}_{G}{\bf v}_{H}^{\top}\\ {\bf v}_{H}{\bf v}_{G}^{\top}&B_{H}\end{bmatrix}\in\mathcal{S}(\overline{G\sqcup H}),

and AB=OAB=O. Moreover, the vector 𝐯=[𝐯G𝐯H]{\bf v}=\begin{bmatrix}{\bf v}_{G}\\ {\bf v}_{H}\end{bmatrix} is a nowhere-zero vector in ker(A)\operatorname{ker}(A). Thus GHG\sqcup H is α\alpha-robust, showing that (1a) implies (1b).

A graph being α\alpha-robust immediately implies that it is complementary vanishing, so (1b) implies (1c). To show (1c) implies (1a), suppose GHG\sqcup H is complementary vanishing. Then there are matrices

A=[AGOOAH]𝒮(GH) and B=[BGCCBH]𝒮(GH¯)A=\begin{bmatrix}A_{G}&O\\ O&A_{H}\end{bmatrix}\in\mathcal{S}(G\sqcup H)\text{ and }B=\begin{bmatrix}B_{G}&C\\ C^{\top}&B_{H}\end{bmatrix}\in\mathcal{S}(\overline{G\sqcup H})

such that AB=OAB=O for some matrices AG𝒮(G)A_{G}\in\mathcal{S}(G), AH𝒮(H)A_{H}\in\mathcal{S}(H), BG𝒮(G¯)B_{G}\in\mathcal{S}(\overline{G}), BH𝒮(H¯)B_{H}\in\mathcal{S}(\overline{H}), and some CC which is nowhere-zero. Thus, we have AGBG=OA_{G}B_{G}=O and AGC=OA_{G}C=O. Since any column of CC is a nowhere-zero vector in ker(AG)\operatorname{ker}(A_{G}), we know GG is α\alpha-robust. Similarly, HH is also α\alpha-robust, proving that (1c) implies (1a).

Notice that G,HG,H satisfying one of (2a), (2b), or (2c) is equivalent to G¯,H¯\overline{G},\overline{H} satisfying one of (1a), (1b), or (1c). Therefore, the equivalence of (2a), (2b), and (2c) follows from the equivalence of (1a), (1b), and (1c). ∎

The next lemma gives an effective way for finding nowhere-zero vectors 𝐱{\bf x} such that B𝐱B{\bf x} is also nowhere-zero.

Lemma 3.2.

If BB is a matrix which does not contain a row of zeros, then there exists a nowhere-zero vector 𝐱{\bf x} such that B𝐱B{\bf x} is nowhere-zero.

Proof.

Suppose that 𝐯{\bf v} is a vector in the column space of BB with the fewest 0 entries. For the sake of contradiction, suppose that 𝐯i=0{\bf v}_{i}=0 for some ii. Since BB does not contain a row of zeros, there exists some column 𝐜{\bf c} of BB such that 𝐜i0{\bf c}_{i}\neq 0. Choose ε>0\varepsilon>0 so that

εmaxj{|𝐜j|}<minj𝐯j0{|𝐯j|}.\varepsilon\max_{j}\{|{\bf c}_{j}|\}<\min_{\begin{subarray}{c}j\\ {\bf v}_{j}\neq 0\end{subarray}}\{|{\bf v}_{j}|\}.

Notice that 𝐯+ε𝐜{\bf v}+\varepsilon{\bf c} has fewer 0 entries than 𝐯{\bf v} by construction. This contradiction leads to the conclusion that 𝐯{\bf v} is a nowhere-zero vector.

Suppose that 𝐱{\bf x} is a vector with the minimum number of 0 entries such that B𝐱B{\bf x} is nowhere-zero (this vector exists by the previous paragraph). For the sake of contradiction, suppose that 𝐱i=0{\bf x}_{i}=0. Let 𝐜{\bf c} be the ithi\text{th} column of BB. Choose ε>0\varepsilon>0 such that

εmaxj{|𝐜j|}<minj{|(B𝐱)j|}.\varepsilon\max_{j}\{|{\bf c}_{j}|\}<\min_{j}\{|(B{\bf x})_{j}|\}.

Recall that minj{|(B𝐱)j|}\min_{j}\{|(B{\bf x})_{j}|\} is positive since B𝐱B{\bf x} is nowhere-zero. Notice that 𝐱+ε𝐞i{\bf x}+\varepsilon{\bf e}_{i} has exactly one less 0 entry than 𝐱{\bf x} and B(𝐱+ε𝐞i)=B𝐱+ε𝐜B({\bf x}+\varepsilon{\bf e}_{i})=B{\bf x}+\varepsilon{\bf c} is nowhere-zero by construction. This contradiction leads to the conclusion that 𝐱{\bf x} is a nowhere-zero vector such that B𝐱B{\bf x} is also nowhere-zero. ∎

By using this lemma, we give some sufficient conditions for a complementary vanishing graph to be α\alpha- or β\beta-robust.

Lemma 3.3.

Let GG be a complementary vanishing graph.

  • (1)

    If GG has no dominating vertices, then GG is α\alpha-robust.

  • (2)

    If GG has no isolated vertices, then GG is β\beta-robust.

Proof.

Since GG is complementary vanishing, there are matrices A𝒮(G)A\in\mathcal{S}(G) and B𝒮(G¯)B\in\mathcal{S}(\overline{G}) such that AB=OAB=O. First assume that GG has no dominating vertices. This means G¯\overline{G} has no isolated vertices, and thus the column space of BB contains a nowhere-zero vector 𝐯{\bf v} by Lemma 3.2. Having AB=OAB=O implies A𝐯=𝟎A{\bf v}={\bf 0}, so 𝐯{\bf v} is a nowhere-zero vector in ker(A)\operatorname{ker}(A). Therefore, GG is α\alpha-robust. An identical argument shows that if GG has no isolated vertices, then there is some nowhere-zero 𝐱{\bf x} in the row space of AA. This implies that GG is β\beta-robust. ∎

Lemma 3.4.

Let GG be a graph. If there exist matrices A𝒮(G),B𝒮(G¯)A\in\mathcal{S}(G),B\in\mathcal{S}(\overline{G}), and a nowhere-zero vector 𝐱{\bf x} such that AB=OAB=O and B𝐱B{\bf x} is nowhere-zero, then GK1G\sqcup K_{1} is β\beta-robust.

Proof.

Assume there exist matrices and vectors as in the hypothesis of the lemma. Let 𝐯=B𝐱{\bf v}=B{\bf x}, which is nowhere-zero by assumption. Define

A=[AOO0],B=[B𝐯𝐯𝐱𝐯].A^{\prime}=\begin{bmatrix}A&O\\ O&0\end{bmatrix},\,B^{\prime}=\begin{bmatrix}B&{\bf v}\\ {\bf v}^{\top}&{\bf x}^{\top}{\bf v}\end{bmatrix}.

Observe that A𝒮(GK1),B𝒮(GK1¯)A^{\prime}\in\mathcal{S}(G\sqcup K_{1}),B^{\prime}\in\mathcal{S}(\overline{G\sqcup K_{1}}). Furthermore, AB=OA^{\prime}B^{\prime}=O since AB=OAB=O and A𝐯=AB𝐱=𝟎A{\bf v}=AB{\bf x}={\bf 0}. If 𝐱=[𝐱1]{\bf x}^{\prime}=\begin{bmatrix}-{\bf x}\\ 1\end{bmatrix}, then 𝐱{\bf x}^{\prime} is nowhere-zero and

B𝐱=[B𝐱+𝐯𝐯𝐱+𝐱𝐯]=𝟎.B^{\prime}{\bf x}^{\prime}=\begin{bmatrix}-B{\bf x}+{\bf v}\\ -{{\bf v}}^{\top}{\bf x}+{\bf x}^{\top}{\bf v}\end{bmatrix}={\bf 0}.

Thus, GK1G\sqcup K_{1} is β\beta-robust.∎

Corollary 3.5.

If GG is complementary vanishing and does not have a dominating vertex, then GK1G\sqcup K_{1} is β\beta-robust.

Proof.

By assumption there exist A𝒮(G),B𝒮(G¯)A\in\mathcal{S}(G),B\in\mathcal{S}(\overline{G}) with AB=OAB=O. Since GG has no dominating vertices, G¯\overline{G} has no isolated vertices and BB does not have a row of zeros. By Lemma 3.2, there exists nowhere-zero vector 𝐱{\bf x} such that B𝐱B{\bf x} is also nowhere-zero. Therefore, GK1G\sqcup K_{1} is β\beta-robust by Lemma 3.4. ∎

3.2 Completing the Proof of Theorem 1.5

We first recall the statement of Theorem 1.5. Let \mathcal{M} denote the set of graphs such that GG is complementary vanishing, GG is connected, and G¯\overline{G} is connected. One can check that K1K_{1}\in\mathcal{M}. Let \mathcal{R} denote the smallest set of graphs which contains \mathcal{M} and which is closed under taking disjoint unions, joins, and complements. For example, having K1K_{1}\in\mathcal{M} implies that \mathcal{R} contains every complete multipartite graph and every threshold graph. Let 𝒞\mathcal{C} denote all the graphs GG such that either in GG or G¯\overline{G} there exist distinct vertices u,v,wu,v,w with v,wN(u)v,w\notin N(u), |N(u)N(w)|=1|N(u)\setminus N(w)|=1, and |N(u)N(v)|=0|N(u)\setminus N(v)|=0. Notice that graphs in 𝒞\mathcal{C} are not complementary vanishing by Corollary 2.5. Our aim is to prove the following.

Theorem 1.5.

A graph GG is complementary vanishing if and only if G𝒞G\in\mathcal{R}\setminus\mathcal{C}.

Part of Theorem 1.5 can be proven immediately.

Proposition 3.6.

If GG\notin\mathcal{R}, then GG is not complementary vanishing.

Proof.

We will prove the proposition by minimal counterexample. Suppose that GG is complementary vanishing and a vertex minimal graph not in .\mathcal{R}. This implies that either GG or G¯\overline{G} is not connected as otherwise G.G\in\mathcal{M}\subseteq\mathcal{R}. This implies that either G=HKG=H\sqcup K or G=HKG=H\vee K where H,KH,K are non-empty complementary vanishing graphs by Proposition 3.1. Since GG is assumed to be a vertex minimal counterexample, it follows that H,KH,K\in\mathcal{R}. However, this implies that GG\in\mathcal{R}; which is a contradiction. ∎

In order to prove Theorem 1.5, we need to keep track of the graphs in \mathcal{R} that are close to falling within 𝒞\mathcal{C}. Let 𝒟\mathcal{D}\subseteq\mathcal{R} denote the set of graphs GG that contain a leaf which is adjacent to a vertex of degree at least 22. For example, stars on at least three vertices are in 𝒟\mathcal{D}, as is a triangle with a pendant (notice that both of these graphs are in \mathcal{R}). The main observation regarding 𝒟\mathcal{D} is the following.

Lemma 3.7.

If D𝒟D\in\mathcal{D} and HH is a graph on at least one vertex, then DH𝒞D\sqcup H\in\mathcal{C}. Furthermore, DHD\sqcup H is not complementary vanishing.

Proof.

By definition of 𝒟\mathcal{D}, the graph DD contains a leaf uu adjacent to xx such that there exists vN(x){u}.v\in N(x)\setminus\{u\}. Notice that |N(u)N(v)|=0|N(u)\setminus N(v)|=0. Since HH is not the empty graph, there exists a vertex ww in HH such that |N(u)N(w)|=1|N(u)\setminus N(w)|=1. By Corollary 2.5, DH𝒞D\sqcup H\in\mathcal{C} is not complementary vanishing. ∎

To prove Theorem 1.5, we prove the following stronger version to aid with an inductive proof.

Theorem 3.8.

Let GG\in\mathcal{R}.

  • (1)

    If G𝒞G\in\mathcal{C}, then GG is not complementary vanishing.

  • (2)

    If G𝒞G\notin\mathcal{C} and G𝒟G\in\mathcal{D}, then GG is β\beta-robust but not α\alpha-robust.

  • (3)

    If G𝒞G\notin\mathcal{C} and G¯𝒟\overline{G}\in\mathcal{D}, then GG is α\alpha-robust but not β\beta-robust.

  • (4)

    If G𝒞G\notin\mathcal{C} and G,G¯𝒟G,\overline{G}\notin\mathcal{D}, then GG is both α\alpha and β\beta-robust.

Proof.

Statement (1): This follows immediately from Corollary 2.5.

Statement (2): Suppose that G(𝒟𝒞)G\in(\mathcal{D}\setminus\mathcal{C})\cap\mathcal{R}. Since G𝒟G\in\mathcal{D}, it follows that GK1G\sqcup K_{1} is not complementary vanishing by Lemma 3.7. By Proposition 3.1 and the fact that K1K_{1} is α\alpha-robust, it follows that GG is not α\alpha-robust.

For the sake of contradiction, suppose that GG is disconnected. This implies that G=HKG=H\sqcup K, and without loss of generality we can assume HH contains a leaf with a neighbor whose degree is at least 22 since G𝒟.G\in\mathcal{D}. In particular, H𝒟H\in\mathcal{D}, and therefore, G𝒞G\in\mathcal{C} by Lemma 3.7. This is a contradiction. Thus, we can assume that GG is connected.

For the sake of contradiction, suppose that GG is a counterexample to the statement. If GG\in\mathcal{M}, then GG is complementary vanishing by definition, and it is β\beta-robust by Lemma 3.3 since GG is connected (and since GK1G\neq K_{1} because K1𝒟K_{1}\notin\mathcal{D}). Therefore, GG\notin\mathcal{M}. Thus, there exists non-empty H,KH,K\in\mathcal{R} such that G=HKG=H\vee K or G=HKG=H\sqcup K, and we must have G=HKG=H\vee K since GG is connected. However, G𝒟G\in\mathcal{D} must have a leaf, which is only possible if H=K=K1H=K=K_{1}. In this case, GG does not have a vertex of degree at least 22, which is a contradiction to G𝒟G\in\mathcal{D}.

Statement (3): Suppose G¯𝒟\overline{G}\in\mathcal{D} and G𝒞G\notin\mathcal{C}. Since 𝒞\mathcal{C} is closed under complements, it follows that G¯𝒞\overline{G}\notin\mathcal{C}. By statement (2), we see that G¯\overline{G} is β\beta-robust but not α\alpha-robust. Thus, GG is α\alpha-robust but not β\beta-robust.

Statement (4): Suppose G𝒞G\notin\mathcal{C} and G,G¯𝒟G,\overline{G}\notin\mathcal{D} is a vertex minimal counterexample. It is easy to see that G=K1G=K_{1} is both α\alpha and β\beta-robust, so this is not a counterexample. Observe that graphs in {K1}\mathcal{M}\setminus\{K_{1}\} are complementary vanishing with no isolated vertices or dominating vertices. Therefore, graphs in \mathcal{M} are both α\alpha and β\beta-robust by Lemma 3.3. Thus we can assume that GG\notin\mathcal{M}. In particular, G=HKG=H\sqcup K or G=HKG=H\vee K for some non-empty H,KH,K\in\mathcal{R}. Since 𝒞\mathcal{C} is closed under complements, we can assume that G=HKG=H\sqcup K without loss of generality.

Since G𝒞G\notin\mathcal{C}, it follows from Lemma 3.7 that neither HH nor KK is in 𝒟𝒞\mathcal{D}\cup\mathcal{C}. If H¯𝒟\overline{H}\in\mathcal{D}, then HH is α\alpha-robust by statement (3), and if H¯𝒟\overline{H}\notin\mathcal{D} then HH is α\alpha-robust since GG is a vertex minimal counterexample to statement (4). Similarly we conclude that KK is α\alpha-robust, and hence by Proposition 3.1 we see that GG is complementary vanishing and α\alpha-robust.

If GG does not have an isolated vertex, then we are done by Lemma 3.3. Therefore, we can assume that K=K1K=K_{1}. Notice that if H=K1H=K_{1}, then it is easy to show that GG is not a counterexample. Furthermore, if HH does not have a dominating vertex, then GG is β\beta-robust by Corollary 3.5. Thus, we can assume HH contains at least two vertices and a dominating vertex. However, this implies G¯\overline{G} has a leaf which is adjacent to a vertex of degree at least two. This is a contradiction, since we assume G¯𝒟\overline{G}\notin\mathcal{D}. ∎

We can now prove our main result.

Proof of Theorem 1.5.

Suppose that G𝒞G\notin\mathcal{R}\setminus\mathcal{C}. If G𝒞G\in\mathcal{C}, then GG is not complementary vanishing by Corollary 2.5. If GG\notin\mathcal{R}, then GG is not complementary vanishing by Proposition 3.6.

Suppose that G𝒞G\in\mathcal{R}\setminus\mathcal{C}. There are three cases which exhaust all possibilities: either G𝒟G\in\mathcal{D}, G¯𝒟\overline{G}\in\mathcal{D}, or G,G¯𝒟G,\overline{G}\notin\mathcal{D}. In any case, GG is complementary vanishing by Theorem 3.8. ∎

4 Algorithmic approaches and the proof of Theorem 1.8

In this section we introduce two algorithmic methods. The first uses Gröbner basis arguments to conclude that a graph is not complementary vanishing. The second is an algorithm that takes a matrix A𝒮(G)A\in\mathcal{S}(G) and checks for the existence of a matrix B𝒮(G¯)B\in\mathcal{S}(\overline{G}) such that AB=OAB=O. Using this second algorithm together with a randomly selected A𝒮(G)A\in\mathcal{S}(G) (or B𝒮(G¯)B\in\mathcal{S}(\overline{G})) will allow us to find certificates for a graph being complementary vanishing.

4.1 Gröbner basis

12345678H1:G@BMP{H_{1}:\texttt{G{@}BMP\{} 12345678H¯1:G}{pm?\overline{H}_{1}:\texttt{G\}\{pm?} 12345678H2:GU`}WH_{2}:\texttt{G$\sim$U\`{}\}W} 12345678H¯2:G?h]@c\overline{H}_{2}:\texttt{G?h]{@}c}
12345678H3:G@hkacH_{3}:\texttt{G{@}hkac} 12345678H¯3:G}UR\W\overline{H}_{3}:\texttt{G\}UR\textbackslash W} 12345678H4:G@h^UgH_{4}:\texttt{G{@}h\^{}Ug} 12345678H¯4:G}U_hS\overline{H}_{4}:\texttt{G\}U\_hS}
Figure 2: Graphs that are not complementary vanishing due to the Gröbner basis argument. Each graph is labelled by its graph6 string, as used by the nauty package.

Let GG be a graph. Define AA to be the variable matrix whose i,ji,j-entry is a variable ai,j=aj,ia_{i,j}=a_{j,i} if {i,j}\{i,j\} is an edge or i=ji=j, and otherwise we set the entry to be 0. Similarly, we define a variable matrix BB with variables bi,jb_{i,j} for G¯\overline{G}. Thus, AB=OAB=O is equivalent to a system of n2n^{2} polynomial equations fi,j=0f_{i,j}=0 for all 1in1\leq i\leq n and 1jn1\leq j\leq n, where

fi,j=kNG[i]NG¯[j]ai,kbk,j.f_{i,j}=\sum_{k\in N_{G}[i]\cap N_{\overline{G}}[j]}a_{i,k}b_{k,j}.

With this in mind, we let IA,B=fi,jI_{A,B}=\langle f_{i,j}\rangle be the ideal generated by all fi,jf_{i,j} in the polynomial ring over \mathbb{R} (for computational purposes, we sometimes consider the polynomial ring over \mathbb{Q}). Recall that the Gröbner basis of an ideal is a particularly nice generating set for an ideal, see [9] for background on this. The only fact about Gröbner bases that we need is that there exist computer algebra systems, such as SageMath, that can be used to calculate the Gröbner basis 𝐁\mathbf{B} of IA,BI_{A,B} so that 𝐁=IA,B\langle\mathbf{B}\rangle=I_{A,B}.

Suppose 𝐁={1}\mathbf{B}=\{1\}. Then we know a combination of fi,jf_{i,j} generates 11, which is impossible when we assume fi,j=0f_{i,j}=0 for all i,ji,j. Therefore, if 𝐁={1}\mathbf{B}=\{1\}, then GG is not complementary vanishing. More formally, this argument implies the following.

Proposition 4.1 (Gröbner basis argument).

Let GG be a graph on nn vertices. Let A𝒮(G)A\in\mathcal{S}(G) and B𝒮(G¯)B\in\mathcal{S}(\overline{G}) be the corresponding variable matrices. If the Gröbner basis of IA,BI_{A,B} contains 11, then GG is not complementary vanishing.

The calculation of a Gröbner basis can be expensive. There are a few ways to reduce the number of variables:

  1. 1.

    Lemmas 2.2, 2.3 and 2.4 imply that certain diagonal entries in AA and BB are either 0 or nonzero.

  2. 2.

    By replacing AA by kAkA, we may assume one of the nonzero variables in AA is 11. (This variable can be an off-diagonal entry, or a diagonal entry that is guaranteed to be nonzero by one of our lemmas). The same argument applies simultaneously for the matrix BB.

  3. 3.

    Pick a spanning forest TT of GG. By replacing AA by ±DAD\pm DAD and BB by D1BD1D^{-1}BD^{-1} for an appropriate invertible diagonal matrix DD, we may assume the entries in AA corresponding to the edges of TT along with some (arbitrary) nonzero diagonal entry are 11. We may switch the role of AA and BB, but we cannot apply this to AA and BB simultaneously.

Note that this last technique relies on the choice of the spanning tree, and that the Gröbner basis depends on the choice of the monomial order, which often depends on the order of the variables. In most of the cases we have tried, we used the degree reverse lexicographic order of the monomials (which is the default setting for SageMath) with the off-diagonal variables preferred over the diagonal variables. We use this method to show the graphs in Figure 2 are not complementary vanishing.

Example 4.2.

Let H2H_{2} and H¯2\overline{H}_{2} be the graphs in Figure 2. Assume A=[ai,j]𝒮(H2)A=\begin{bmatrix}a_{i,j}\end{bmatrix}\in\mathcal{S}(H_{2}) and B=[bi,j]𝒮(H¯2)B=\begin{bmatrix}b_{i,j}\end{bmatrix}\in\mathcal{S}(\overline{H}_{2}) satisfy AB=OAB=O. According to Lemmas 2.3 and 2.4, we may assume

ai,i=0\displaystyle a_{i,i}=0 for i{1,2,3,4},\displaystyle\text{ for }i\in\{1,2,3,4\},
ai,i0\displaystyle a_{i,i}\neq 0 for i{5,6,7,8},\displaystyle\text{ for }i\in\{5,6,7,8\},
bi,i0\displaystyle b_{i,i}\neq 0 for i{1,2,3,4},\displaystyle\text{ for }i\in\{1,2,3,4\},
bi,i=0\displaystyle b_{i,i}=0 for i{5,6,7,8}.\displaystyle\text{ for }i\in\{5,6,7,8\}.

Moreover, we may pick a spanning tree TT of H2H_{2} using the edges

E(T)={{1,8},{2,8},{3,7},{4,7},{5,8},{6,7},{6,8}}.E(T)=\{\{1,8\},\{2,8\},\{3,7\},\{4,7\},\{5,8\},\{6,7\},\{6,8\}\}.

By replacing AA with ±DAD\pm DAD and BB with D1BD1D^{-1}BD^{-1} for some appropriate diagonal matrix DD, we may assume the entries in AA corresponding to E(T)E(T) along with a5,5a_{5,5} are 11. Moreover, by replacing BB with kBkB for some nonzero kk, we may further assume b1,5=1b_{1,5}=1. In conclusion, we have

A=[0a1,2a1,3a1,40a1,601a1,20a2,3a2,4a2,5001a1,3a2,30a3,40a3,610a1,4a2,4a3,40a4,50100a2,50a4,510a5,71a1,60a3,600a6,6110011a5,71a7,701100110a8,8]A=\begin{bmatrix}0&a_{1,2}&a_{1,3}&a_{1,4}&0&a_{1,6}&0&1\\ a_{1,2}&0&a_{2,3}&a_{2,4}&a_{2,5}&0&0&1\\ a_{1,3}&a_{2,3}&0&a_{3,4}&0&a_{3,6}&1&0\\ a_{1,4}&a_{2,4}&a_{3,4}&0&a_{4,5}&0&1&0\\ 0&a_{2,5}&0&a_{4,5}&1&0&a_{5,7}&1\\ a_{1,6}&0&a_{3,6}&0&0&a_{6,6}&1&1\\ 0&0&1&1&a_{5,7}&1&a_{7,7}&0\\ 1&1&0&0&1&1&0&a_{8,8}\end{bmatrix}

and

B=[b1,100010b1,700b2,3000b2,6b2,7000b3,30b3,500b3,8000b4,40b4,60b4,810b3,500b5,6000b2,60b4,6b5,6000b1,7b2,700000b7,800b3,8b4,800b7,80].B=\begin{bmatrix}b_{1,1}&0&0&0&1&0&b_{1,7}&0\\ 0&b_{2,3}&0&0&0&b_{2,6}&b_{2,7}&0\\ 0&0&b_{3,3}&0&b_{3,5}&0&0&b_{3,8}\\ 0&0&0&b_{4,4}&0&b_{4,6}&0&b_{4,8}\\ 1&0&b_{3,5}&0&0&b_{5,6}&0&0\\ 0&b_{2,6}&0&b_{4,6}&b_{5,6}&0&0&0\\ b_{1,7}&b_{2,7}&0&0&0&0&0&b_{7,8}\\ 0&0&b_{3,8}&b_{4,8}&0&0&b_{7,8}&0\end{bmatrix}.

By treating the remaining ai,ja_{i,j} and bi,jb_{i,j} as variables, the 828^{2} polynomial entries of ABAB generate the ideal IA,BI_{A,B}.

We then find the Gröbner basis of IA,BI_{A,B} under the degree reverse lexicographic order, which requires an ordering of the variables. Order the variables by

ai,j(ij)bi,j(ij)ai,ibi,ia_{i,j}\ (i\neq j)\succ b_{i,j}\ (i\neq j)\succ a_{i,i}\succ b_{i,i}

(and order the off-diagonal entries by lexicographic order). This means that the algorithm prioritizes elimination of ai,ja_{i,j} first, and then bi,jb_{i,j}, and so on. With these settings, the Gröbner basis of IA,BI_{A,B} is {1}\{1\}. This means some polynomials in IA,BI_{A,B} generates 11, which is impossible if they are all zero. Therefore, H2H_{2} is not complementary vanishing.

Remark 4.3.

The same technique applies to H3H_{3} and H4H_{4} in Figure 2, so they are not complementary vanishing. For H1H_{1}, one may choose a spanning tree in H¯1\overline{H}_{1} instead. In this case the Gröbner basis will contain a polynomial of a single nonzero variable, e.g., a5,8a_{5,8}. Therefore, it again shows H1H_{1} is not complementary vanishing.

4.2 Solving for BB

Let GG be a graph and A𝒮(G)A\in\mathcal{S}(G). We present an algorithm for checking if there is a matrix B𝒮(G¯)B\in\mathcal{S}(\overline{G}) such that AB=OAB=O.

Let GG be a graph with nn vertices and mm edges. We define 𝒮cl(G)\mathcal{S}^{\rm cl}(G) as the topological closure of 𝒮(G)\mathcal{S}(G); that is, 𝒮cl(G)\mathcal{S}^{\rm cl}(G) consists of all real symmetric matrices whose i,ji,j-entry can be nonzero only when {i,j}\{i,j\} is an edge or i=ji=j. Thus, 𝒮cl(G)\mathcal{S}^{\rm cl}(G) is a vector subspace of dimension n+mn+m in the space of real symmetric matrices of order nn.

When A𝒮(G)A\in\mathcal{S}(G) is given, define

𝒲={B𝒮cl(G¯):AB=O},\mathcal{W}=\{B\in\mathcal{S}^{\rm cl}(\overline{G}):AB=O\},

which is a vector space. Our task is to determine whether the vector space 𝒲\mathcal{W} contains a matrix B𝒮(G¯)B\in\mathcal{S}(\overline{G}). To do so, we need to check the existence of a matrix B𝒲B\in\mathcal{W} that is nonzero on the edges of G¯\overline{G}.

Let 𝐁\mathbf{B} be a basis of 𝒲\mathcal{W}. Then every matrix in 𝒲\mathcal{W} is zero at a given entry if and only if every matrix in 𝐁\mathbf{B} is zero at the given entry. Moreover, when SS is a subset of indices, 𝒲\mathcal{W} has a matrix that is nowhere-zero on SS if and only if for each index {i,j}S\{i,j\}\in S, there is a vector in 𝐁\mathbf{B} that is not zero at index {i,j}\{i,j\}. With these observations, whether there is a matrix B𝒮(G¯)B\in\mathcal{S}(\overline{G}) with AB=OAB=O can be determined through standard linear algebra techniques. This idea is summarized in Proposition 4.4.

Proposition 4.4.

Let GG be a graph, A𝒮(G)A\in\mathcal{S}(G), and let 𝒲\mathcal{W} be as defined above with 𝐁\mathbf{B} a basis of 𝒲\mathcal{W}. Then a matrix B𝒮(G¯)B\in\mathcal{S}(\overline{G}) exists with AB=OAB=O if and only if for every edge {i,j}E(G¯)\{i,j\}\in E(\overline{G}), there is a matrix C𝐁C\in\mathbf{B} whose i,ji,j-entry is nonzero. Moreover, if the latter condition holds, then a random linear combination of 𝐁\mathbf{B} with each coefficient in Gaussian distribution gives a matrix BB in 𝒮(G¯)\mathcal{S}(\overline{G}) with probability 11.

4.3 Proof of Theorem 1.8

We summarize some necessary and sufficient conditions for a graph to be complementary vanishing:

  • Methods which can verify a graph is not complementary vanishing:

    • Diagonal lemmas: Lemma 2.2 and Corollary 2.5

    • Odd cycle lemma: Lemma 2.6 in conjunction with Lemmas 2.3 and 2.4

    • Gröbner basis arguments: Proposition 4.1 and its modifications

  • Methods to verify a graph is complementary vanishing:

    • Twin proposition: Proposition 2.8

    • Duplication lemma: Lemma 2.9

    • Random trial: Randomly pick matrices A𝒮(G)A\in\mathcal{S}(G) and apply Proposition 4.4 to see if B𝒮(G¯)B\in\mathcal{S}(\overline{G}) with AB=OAB=O exists; or do the same for G¯\overline{G}

    • Manual construction: Judiciously construct matrices A𝒮(G)A\in\mathcal{S}(G) and B𝒮(G¯)B\in\mathcal{S}(\overline{G}) with AB=OAB=O

Not complementary vanishing Complementary vanishing
nn diag oc grob twin dup certificate found total
1 1 1
2 0
3 0
4 1 1
5 5 5
6 34 34
7 326 1 4 331
8 4905 8 4 6 14 12 4949
Table 2: The number of pairs {G,G¯}\{G,\overline{G}\} verified using each method, where GG and G¯\overline{G} are connected graphs on nn vertices. The columns diag, oc, grob counts the graph pairs that are not complementary vanishing due to the diagonal lemmas, the odd-cycle lemma, and Gröbner basis arguments. The columns twin, dup, and certificate found count the graph pairs that are complementary vanishing because of the twin proposition, the duplication lemma, or because the certificate is found by random trial or manual construction. The methods were tested in order from left to right.

It turns out that these techniques are enough for us to classify all pairs {G,G¯}\{G,\overline{G}\} with GG and G¯\overline{G} connected on 88 or few vertices. The results are shown in Table 2. Therefore, we have the following lemma, which one can check by computer is equivalent to Theorem 1.8.

Lemma 4.5.

Let GG be a graph on n8n\leq 8 vertices such that both GG and G¯\overline{G} are connected. Then GG is complementary vanishing if and only if GG and G¯\overline{G} do not satisfy the hypotheses of Lemmas 2.2, 2.6, Corollary 2.5, and are not one of the four pairs in Figure 2.

We note that for one of the pairs in Appendix A, namely G8G_{8}, we were unable to find a certificate using random trials; the only construction we know of for this graph was found manually.

We note that in principle, one could try to apply these techniques to determine which pairs of graphs on 99 vertices are complementary vanishing. However, applying the diagonal conditions to graphs on 99 vertices leaves to many possibilities for random trails to effectively resolve.

5 Conclusion

In this paper we began the study of complementary vanishing graphs. In view of Theorem 1.5, we have essentially reduced the problem to graphs GG such that both GG and G¯\overline{G} are connected. In Theorem 1.8 we completely solved this problem for graphs on at most 8 vertices.

As things currently stand, there does not seem to be a simple description which categorizes graphs that are complementary vanishing. However, there are certain classes of graphs where one might be able to determine the answer. For example, it is natural to ask what happens for random graphs.

Question 5.1.

If GG is chosen uniformly at random amongst all nn-vertex graphs, does the probability that GG is complementary vanishing tend to 1 as nn tends towards infinity?

If the answer to this question is “yes”, then it is likely that this will be difficult to prove. Indeed, this would imply that the random graph satisfies the Graph Complement Conjecture asymptotically almost surely, which is an open problem.

Recall that there are four connected graphs GG on at most 88 vertices such that their complements are also connected and such that these GG were shown not to be complementary vanishing using a Gröbner Basis argument. It would be nice if one could prove that these four graphs are not complementary vanishing by utilizing combinatorial techniques.

Question 5.2.

Can one use a combinatorial argument to prove that the graphs in Figure 2 are not complementary vanishing?

Acknowledgments. The authors would like to express their sincere gratitude to the organizers of the Mathematics Research Community on Finding Needles in Haystacks: Approaches to Inverse Problems Using Combinatorics and Linear Algebra, and the AMS for funding the program through NSF grant 1916439. J. C.-H. Lin was supported by the Young Scholar Fellowship Program (grant no. MOST-110-2628-M-110-003) from the Ministry of Science and Technology of Taiwan. S. Spiro was supported by the National Science Foundation Graduate Research Fellowship under Grant No. DGE-1650112.

References

  • [1] B. Ahmadi, F. Alinaghipour, M. S. Cavers, S. Fallat, K. Meagher, and S. Nasserasr. Minimum number of distinct eigenvalues of graphs. Electron. J. Linear Algebra, 26:673–691, 2013.
  • [2] AIM Minimum Rank-Special Graphs Work Group. Zero forcing sets and the minimum rank of graphs. Linear Algebra Appl., 428(7):1628–1648, 2008.
  • [3] American Institute of Mathematics Workshop. Spectra of families of matrices described by graphs, digraphs, and sign patterns. http://aimath.org/pastworkshops/matrixspectrum.html, 2006.
  • [4] F. Barioli, W. Barrett, S. M. Fallat, H. T. Hall, L. Hogben, and H. van der Holst. On the graph complement conjecture for minimum rank. Linear Algebra Appl., 436(12):4373–4391, 2012.
  • [5] F. Barioli and S. M. Fallat. On two conjectures regarding an inverse eigenvalue problem for acyclic symmetric matrices. Electron. J. Linear Algebra, 11:41–50, 2004.
  • [6] W. Barrett, S. Butler, S. M. Fallat, H. T. Hall, L. Hogben, J. C.-H. Lin, B. L. Shader, and M. Young. The inverse eigenvalue problem of a graph: multiplicities and minors. J. Combin. Theory Ser. B, 142:276–306, 2020.
  • [7] W. Barrett, S. Fallat, H. T. Hall, L. Hogben, J. C.-H. Lin, and B. L. Shader. Generalizations of the strong Arnold property and the minimum number of distinct eigenvalues of a graph. Electron. J. Combin., 24(2):Paper No. 2.40, 28, 2017.
  • [8] Y. Colin de Verdière. On a new graph invariant and a criterion for planarity. In Graph structure theory (Seattle, WA, 1991), volume 147 of Contemp. Math., pages 137–147. Amer. Math. Soc., Providence, RI, 1993.
  • [9] D. A. Cox, J. Little, and D. O’Shea. Ideals, varieties, and algorithms: An introduction to computational algebraic geometry and commutative algebra. Springer, fourth edition, 2015.
  • [10] J. Ekstrand, C. Erickson, H. T. Hall, D. Hay, L. Hogben, R. Johnson, N. Kingsley, S. Osborne, T. Peters, J. Roat, A. Ross, D. D. Row, N. Warnberg, and M. Young. Positive semidefinite zero forcing. Linear Algebra Appl., 439(7):1862–1874, 2013.
  • [11] S. M. Fallat and L. Hogben. The minimum rank of symmetric matrices described by a graph: a survey. Linear Algebra Appl., 426(2-3):558–582, 2007.
  • [12] F. R. Gantmacher and M. G. Krein. Oszillationslnatrizen, Oszillationskerne und kleine Schwingungen nzechanischer Systenze. Translated by Alfred Stohr, Berlin, Akademie-Verlag, 1960.
  • [13] G. Joret and D. R. Wood. Nordhaus-Gaddum for treewidth. European J. Combin., 33(4):488–490, 2012.
  • [14] A. Kotlov, L. Lovász, and S. Vempala. The Colin de Verdière number and sphere representations of a graph. Combinatorica, 17(4):483–521, 1997.
  • [15] R. H. Levene, P. Oblak, and H. Šmigoc. A Nordhaus-Gaddum conjecture for the minimum number of distinct eigenvalues of a graph. Linear Algebra Appl., 564:236–263, 2019.
  • [16] X. Li, M. Nathanson, and R. Phillips. Minimum vector rank and complement critical graphs. Electron. J. Linear Algebra, 27:100–123, 2014.

See pages 1- of appendix.pdf