Hubbry Logo
GraphonGraphonMain
Open search
Graphon
Community hub
Graphon
logo
8 pages, 0 posts
0 subscribers
Be the first to start a discussion here.
Be the first to start a discussion here.
Graphon
Graphon
from Wikipedia
A realization of an exchangeable random graph defined by a graphon. The graphon is shown as a magenta heatmap (lower right). A random graph of size is generated by independently assigning to each vertex a latent random variable (values along vertical axis) and including each edge independently with probability . For example, edge (green, dotted) is present with probability ; the green boxes in the right square represent the values of and . The upper left panel shows the graph realization as an adjacency matrix.

In graph theory and statistics, a graphon (also known as a graph limit) is a symmetric measurable function , that is important in the study of dense graphs. Graphons arise both as a natural notion for the limit of a sequence of dense graphs, and as the fundamental defining objects of exchangeable random graph models. Graphons are tied to dense graphs by the following pair of observations: the random graph models defined by graphons give rise to dense graphs almost surely, and, by the regularity lemma, graphons capture the structure of arbitrary large dense graphs.

Statistical formulation

[edit]

A graphon is a symmetric measurable function . Usually a graphon is understood as defining an exchangeable random graph model according to the following scheme:

  1. Each vertex of the graph is assigned an independent random value
  2. Edge is independently included in the graph with probability .

A random graph model is an exchangeable random graph model if and only if it can be defined in terms of a (possibly random) graphon in this way. The model based on a fixed graphon is sometimes denoted , by analogy with the Erdős–Rényi model of random graphs. A graph generated from a graphon in this way is called a -random graph.

It follows from this definition and the law of large numbers that, if , exchangeable random graph models are dense almost surely.[1]

Examples

[edit]

The simplest example of a graphon is for some constant . In this case the associated exchangeable random graph model is the Erdős–Rényi model that includes each edge independently with probability .

If we instead start with a graphon that is piecewise constant by:

  1. dividing the unit square into blocks, and
  2. setting equal to on the block,

the resulting exchangeable random graph model is the community stochastic block model, a generalization of the Erdős–Rényi model. We can interpret this as a random graph model consisting of distinct Erdős–Rényi graphs with parameters respectively, with bigraphs between them where each possible edge between blocks and is included independently with probability .

Many other popular random graph models can be understood as exchangeable random graph models defined by some graphon, a detailed survey is included in Orbanz and Roy.[1]

Jointly exchangeable adjacency matrices

[edit]

A random graph of size can be represented as a random adjacency matrix. In order to impose consistency (in the sense of projectivity) between random graphs of different sizes it is natural to study the sequence of adjacency matrices arising as the upper-left sub-matrices of some infinite array of random variables; this allows us to generate by adding a node to and sampling the edges for . With this perspective, random graphs are defined as random infinite symmetric arrays .

Following the fundamental importance of exchangeable sequences in classical probability, it is natural to look for an analogous notion in the random graph setting. One such notion is given by jointly exchangeable matrices; i.e. random matrices satisfying

for all permutations of the natural numbers, where means equal in distribution. Intuitively, this condition means that the distribution of the random graph is unchanged by a relabeling of its vertices: that is, the labels of the vertices carry no information.

There is a representation theorem for jointly exchangeable random adjacency matrices, analogous to de Finetti’s representation theorem for exchangeable sequences. This is a special case of the Aldous–Hoover theorem for jointly exchangeable arrays and, in this setting, asserts that the random matrix is generated by:

  1. Sample independently
  2. independently at random with probability

where is a (possibly random) graphon. That is, a random graph model has a jointly exchangeable adjacency matrix if and only if it is a jointly exchangeable random graph model defined in terms of some graphon.

Graphon estimation

[edit]

Due to identifiability issues, it is impossible to estimate either the graphon function or the node latent positions and there are two main directions of graphon estimation. One direction aims at estimating up to an equivalence class,[2][3] or estimate the probability matrix induced by .[4][5]

Analytic formulation

[edit]

Any graph on vertices can be identified with its adjacency matrix . This matrix corresponds to a step function , defined by partitioning into intervals such that has interior and for each , setting equal to the entry of . This function is the associated graphon of the graph .

In general, if we have a sequence of graphs where the number of vertices of goes to infinity, we can analyze the limiting behavior of the sequence by considering the limiting behavior of the functions . If these graphs converge (according to some suitable definition of convergence), then we expect the limit of these graphs to correspond to the limit of these associated functions.

This motivates the definition of a graphon (short for "graph function") as a symmetric measurable function which captures the notion of a limit of a sequence of graphs. It turns out that for sequences of dense graphs, several apparently distinct notions of convergence are equivalent and under all of them the natural limit object is a graphon.[6]

Examples

[edit]

Constant graphon

[edit]

Take a sequence of Erdős–Rényi random graphs with some fixed parameter . Intuitively, as tends to infinity, the limit of this sequence of graphs is determined solely by edge density of these graphs. In the space of graphons, it turns out that such a sequence converges almost surely to the constant , which captures the above intuition.

Half graphon

[edit]

Take the sequence of half-graphs, defined by taking to be the bipartite graph on vertices and such that is adjacent to precisely when . If the vertices are listed in the presented order, then the adjacency matrix has two corners of "half square" block matrices filled with ones, with the rest of the entries equal to zero. For example, the adjacency matrix of is given by

As gets large, these corners of ones "smooth" out. Matching this intuition, the sequence converges to the half-graphon defined by when and otherwise.

Complete bipartite graphon

[edit]

Take the sequence of complete bipartite graphs with equal sized parts. If we order the vertices by placing all vertices in one part at the beginning and placing the vertices of the other part at the end, the adjacency matrix of looks like a block off-diagonal matrix, with two blocks of ones and two blocks of zeros. For example, the adjacency matrix of is given by

As gets larger, this block structure of the adjacency matrix remains constant, so that this sequence of graphs converges to a "complete bipartite" graphon defined by whenever and , and setting otherwise.

If we instead order the vertices of by alternating between parts, the adjacency matrix has a chessboard structure of zeros and ones. For example, under this ordering, the adjacency matrix of is given by

As gets larger, the adjacency matrices become a finer and finer chessboard. Despite this behavior, we still want the limit of to be unique and result in the graphon from example 3. This means that when we formally define convergence for a sequence of graphs, the definition of a limit should be agnostic to relabelings of the vertices.

Limit of W-random graphs

[edit]

Take a random sequence of -random graphs by drawing for some fixed graphon . Then just like in the first example from this section, it turns out that converges to almost surely.

Recovering graph parameters from graphons

[edit]

Given graph with associated graphon , we can recover graph theoretic properties and parameters of by integrating transformations of . For example, the edge density (i.e. average degree divided by number of vertices) of is given by the integral This is because is -valued, and each edge in corresponds to a region of area where equals .

Similar reasoning shows that the triangle density in is equal to

Notions of convergence

[edit]

There are many different ways to measure the distance between two graphs. If we are interested in metrics that "preserve" extremal properties of graphs, then we should restrict our attention to metrics that identify random graphs as similar. For example, if we randomly draw two graphs independently from an Erdős–Rényi model for some fixed , the distance between these two graphs under a "reasonable" metric should be close to zero with high probability for large .

Naively, given two graphs on the same vertex set, one might define their distance as the number of edges that must be added or removed to get from one graph to the other, i.e. their edit distance. However, the edit distance does not identify random graphs as similar; in fact, two graphs drawn independently from have an expected (normalized) edit distance of .

There are two natural metrics that behave well on dense random graphs in the sense that we want. The first is a sampling metric, which says that two graphs are close if their distributions of subgraphs are close. The second is an edge discrepancy metric, which says two graphs are close when their edge densities are close on all their corresponding subsets of vertices.

Miraculously, a sequence of graphs converges with respect to one metric precisely when it converges with respect to the other. Moreover, the limit objects under both metrics turn out to be graphons. The equivalence of these two notions of convergence mirrors how various notions of quasirandom graphs are equivalent.[7]

Homomorphism densities

[edit]

One way to measure the distance between two graphs and is to compare their relative subgraph counts. That is, for each graph we can compare the number of copies of in and in . If these numbers are close for every graph , then intuitively and are similar looking graphs. Rather than dealing directly with subgraphs, however, it turns out to be easier to work with graph homomorphisms. This is fine when dealing with large, dense graphs, since in this scenario the number of subgraphs and the number of graph homomorphisms from a fixed graph are asymptotically equal.

Given two graphs and , the homomorphism density of in is defined to be the number of graph homomorphisms from to . In other words, is the probability a randomly chosen map from the vertices of to the vertices of sends adjacent vertices in to adjacent vertices in .

Graphons offer a simple way to compute homomorphism densities. Indeed, given a graph with associated graphon and another , we have

where the integral is multidimensional, taken over the unit hypercube . This follows from the definition of an associated graphon, by considering when the above integrand is equal to . We can then extend the definition of homomorphism density to arbitrary graphons , by using the same integral and defining

for any graph .

Given this setup, we say a sequence of graphs is left-convergent if for every fixed graph , the sequence of homomorphism densities converges. Although not evident from the definition alone, if converges in this sense, then there always exists a graphon such that for every graph , we have simultaneously.

Cut distance

[edit]

Take two graphs and on the same vertex set. Because these graphs share the same vertices, one way to measure their distance is to restrict to subsets of the vertex set, and for each such pair of subsets compare the number of edges from to in to the number of edges between and in . If these numbers are similar for every pair of subsets (relative to the total number of vertices), then that suggests and are similar graphs.

As a preliminary formalization of this notion of distance, for any pair of graphs and on the same vertex set of size , define the labeled cut distance between and to be

In other words, the labeled cut distance encodes the maximum discrepancy of the edge densities between and . We can generalize this concept to graphons by expressing the edge density in terms of the associated graphon , giving the equality

where are unions of intervals corresponding to the vertices in and . Note that this definition can still be used even when the graphs being compared do not share a vertex set. This motivates the following more general definition.

Definition 1. For any symmetric, measurable function , define the cut norm of to be the quantity

taken over all measurable subsets of the unit interval.[6]

This captures our earlier notion of labeled cut distance, as we have the equality .

This distance measure still has one major limitation: it can assign nonzero distance to two isomorphic graphs. To make sure isomorphic graphs have distance zero, we should compute the minimum cut norm over all possible "relabellings" of the vertices. This motivates the following definition of the cut distance.

Definition 2. For any pair of graphons and , define their cut distance to be

where is the composition of with the map , and the infimum is taken over all measure-preserving bijections from the unit interval to itself.[8]

The cut distance between two graphs is defined to be the cut distance between their associated graphons.

We now say that a sequence of graphs is convergent under the cut distance if it is a Cauchy sequence under the cut distance . Although not a direct consequence of the definition, if such a sequence of graphs is Cauchy, then it always converges to some graphon .

Equivalence of convergence

[edit]

As it turns out, for any sequence of graphs , left-convergence is equivalent to convergence under the cut distance, and furthermore, the limit graphon is the same. We can also consider convergence of graphons themselves using the same definitions, and the same equivalence is true. In fact, both notions of convergence are related more strongly through what are called counting lemmas.[6]

Counting Lemma. For any pair of graphons and , we have

for all graphs .

The name "counting lemma" comes from the bounds that this lemma gives on homomorphism densities , which are analogous to subgraph counts of graphs. This lemma is a generalization of the graph counting lemma that appears in the field of regularity partitions, and it immediately shows that convergence under the cut distance implies left-convergence.

Inverse Counting Lemma. For every real number , there exist a real number and a positive integer such that for any pair of graphons and with

for all graphs satisfying , we must have .

This lemma shows that left-convergence implies convergence under the cut distance.

The space of graphons

[edit]

We can make the cut-distance into a metric by taking the set of all graphons and identifying two graphons whenever . The resulting space of graphons is denoted , and together with forms a metric space.

This space turns out to be compact. Moreover, it contains the set of all finite graphs, represented by their associated graphons, as a dense subset. These observations show that the space of graphons is a completion of the space of graphs with respect to the cut distance. One immediate consequence of this is the following.

Corollary 1. For every real number , there is an integer such that for every graphon , there is a graph with at most vertices such that .

To see why, let be the set of graphs. Consider for each graph the open ball containing all graphons such that . The set of open balls for all graphs covers , so compactness implies that there is a finite subcover for some finite subset . We can now take to be the largest number of vertices among the graphs in .

Applications

[edit]

Regularity lemma

[edit]

Compactness of the space of graphons can be thought of as an analytic formulation of Szemerédi's regularity lemma; in fact, a stronger result than the original lemma.[9] Szemeredi's regularity lemma can be translated into the language of graphons as follows. Define a step function to be a graphon that is piecewise constant, i.e. for some partition of , is constant on for all . The statement that a graph has a regularity partition is equivalent to saying that its associated graphon is close to a step function.

The proof of compactness requires only the weak regularity lemma:

Weak Regularity Lemma for Graphons. For every graphon and , there is a step function with at most steps such that .

but it can be used to prove stronger regularity results, such as the strong regularity lemma:

Strong Regularity Lemma for Graphons. For every sequence of positive real numbers, there is a positive integer such that for every graphon , there is a graphon and a step function with steps such that and

The proof of the strong regularity lemma is similar in concept to Corollary 1 above. It turns out that every graphon can be approximated with a step function in the norm, showing that the set of balls cover . These sets are not open in the metric, but they can be enlarged slightly to be open. Now, we can take a finite subcover, and one can show that the desired condition follows.

Sidorenko's conjecture

[edit]

The analytic nature of graphons allows greater flexibility in attacking inequalities related to homomorphisms.

For example, Sidorenko's conjecture is a major open problem in extremal graph theory, which asserts that for any graph on vertices with average degree (for some ) and bipartite graph on vertices and edges, the number of homomorphisms from to is at least .[10] Since this quantity is the expected number of labeled subgraphs of in a random graph , the conjecture can be interpreted as the claim that for any bipartite graph , the random graph achieves (in expectation) the minimum number of copies of over all graphs with some fixed edge density.

Many approaches to Sidorenko's conjecture formulate the problem as an integral inequality on graphons, which then allows the problem to be attacked using other analytical approaches. [11]

Generalizations

[edit]

Graphons are naturally associated with dense simple graphs. There are extensions of this model to dense directed weighted graphs, often referred to as decorated graphons.[12] There are also recent extensions to the sparse graph regime, from both the perspective of random graph models [13] and graph limit theory.[14][15]

References

[edit]
Revisions and contributorsEdit on WikipediaRead on Wikipedia
from Grokipedia
A graphon is a symmetric, Lebesgue-measurable function W:[0,1]2[0,1]W: [0,1]^2 \to [0,1], which serves as the canonical limit object for convergent sequences of dense finite graphs under the cut metric, enabling the study of asymptotic properties through continuous analogs of discrete graph structures. Introduced by László Lovász and Balázs Szegedy in their 2006 paper "Limits of Dense Graph Sequences," graphons formalize the convergence of graph sequences where homomorphism densities—quantities measuring the relative frequency of subgraphs—tend to limits, bridging combinatorial graph theory with functional analysis. Graphons generalize finite graphs by representing edge probabilities between points in a continuous vertex set, with the value W(x,y)W(x,y) indicating the probability of an edge between vertices labeled by xx and yy. Key properties include weak , where two graphons are equivalent up to measure-preserving relabelings, and the of the space of graphons under the cut distance metric δ\delta_\square, which ensures every sequence of dense graphs has a convergent . This framework, building on Szemerédi's regularity lemma, allows graphons to model limits via integrals for subgraph densities, such as the edge density t(K2,W)=0101W(x,y)dxdyt(K_2, W) = \int_0^1 \int_0^1 W(x,y) \, dx \, dy. The theory of graphons has profound implications in extremal graph theory, where it resolves optimization problems like those in Turán's theorem by identifying extremal graphons that maximize or minimize subgraph densities. For instance, the Turán graphon for KrK_r-free graphs achieves the maximum edge density without rr-cliques. In property testing and algorithmic graph theory, graphons facilitate efficient estimation of graph parameters and testing for properties like quasirandomness, with applications to the removal lemma and counting lemmas for subgraphs. Additionally, graphons underpin models of random graphs, such as WW-random graphs generated by sampling vertices from [0,1] and edges according to WW, and extend to sparse graphs via graphings for bounded-degree sequences. Their influence extends to spectral graph theory, where operator norms of integral operators associated with graphons converge to those of graph sequences, and to open conjectures like Sidorenko's on bipartite subgraph densities. Overall, graphons provide a robust mathematical foundation for analyzing large-scale networks in combinatorics, probability, and computer science.

Introduction

Definition and Motivation

A graphon is formally defined as a symmetric measurable function W:[0,1]×[0,1][0,1]W: [0,1] \times [0,1] \to [0,1], considered up to measure-preserving reparameterizations of the domain, which represent the limiting objects of convergent sequences of finite dense graphs. This continuous kernel encodes the asymptotic edge probabilities between vertices, generalizing the adjacency matrix of a finite graph to a probabilistic continuum where vertices are points in the unit interval. Graphons arise in theory as a natural framework to address the limitations of studying sequences of graphs with diverging vertex counts, where traditional convergence in graph properties like edge densities fails due to non-uniform sizes. By capturing the limits of such sequences, graphons enable the rigorous analysis of asymptotic behaviors, particularly densities—the relative frequencies of subgraphs—which converge pointwise for every fixed graph under appropriate conditions. This allows for the extension of results, such as bounds on subgraph counts, to infinite or growing structures without relying on finite approximations. The motivation for graphons originated from the in theory, where models like the Erdős–Rényi process required tools to describe limiting behaviors, and from , which sought to formalize "graph limits" for optimizing properties like Turán numbers in asymptotic regimes. These foundations, developed through early works on quasirandom graphs and regularity lemmas, culminated in a unified theory for handling sequences that do not converge in simpler metrics but do so in terms of subgraph densities. In the context of sampling from a graphon, expectations of linear statistics can be represented via the kernel integral 0101W(x,y)f(x)g(y)dxdy\int_0^1 \int_0^1 W(x,y) f(x) g(y) \, dx \, dy, where ff and gg are measurable functions on [0,1][0,1] approximating vertex labels or indicators, providing a continuous analog to sums over adjacency matrices in finite graphs.

Historical Background

The development of graphon theory traces its roots to foundational results in extremal graph theory from the 1970s, particularly Endre Szemerédi's regularity lemma, which established that large graphs can be partitioned into a bounded number of relatively uniform bipartite subgraphs, providing a structural approximation for dense graphs. This lemma laid the groundwork for analyzing asymptotic behaviors in graph sequences, influencing later limit theories by offering a tool to control irregularities in large structures. In the 1980s, the Aldous-Hoover theorem provided a representation for exchangeable random arrays, characterizing infinite symmetric random matrices as mixtures of deterministic functions over random variables, which implicitly connected to graph-like structures through exchangeability. This probabilistic framework, developed by David Aldous and Douglas N. Hoover, offered a for random graphs that would later inform the sampling properties of graphons. The modern theory of graphons emerged in the mid-2000s through work on limits of sequences, initiated by and Balázs Szegedy in their 2006 paper, which showed that sequences of graphs where homomorphism densities converge admit a limit object definable via these densities, bridging combinatorial and analytic perspectives. Building on this, Christian Borgs, Jennifer Chayes, Lovász, Vera T. Sós, and Katalin Vesztergombi formalized the graphon as the canonical limit representation in their 2008 paper, establishing metric properties and convergence criteria that integrated the earlier approach with probabilistic sampling from graphons. These contributions around 2006–2008 by Lovász, Szegedy, Borgs, Chayes, Sós, and Vesztergombi solidified graphons as a versatile tool for studying graph limits, evolving from the exchangeable array representations of the 1980s and the structural decompositions of the 1970s.

Mathematical Foundations

Analytic Formulation

In the analytic formulation, a graphon is a symmetric measurable function W:[0,1]2[0,1]W: [0,1]^2 \to [0,1] that belongs to the space L([0,1]2)L^\infty([0,1]^2), consisting of essentially bounded functions with respect to the Lebesgue measure on the unit square. The symmetry condition W(x,y)=W(y,x)W(x,y) = W(y,x) holds for almost every x,y[0,1]x,y \in [0,1], reflecting the undirected nature of the graphs it models. Graphons are identified up to equivalence under measure-preserving bijections ϕ:[0,1][0,1]\phi: [0,1] \to [0,1], meaning two graphons WW and WW' represent the same object if W(x,y)=W(ϕ(x),ϕ(y))W'(x,y) = W(\phi(x), \phi(y)) almost everywhere, forming a quotient space that accounts for relabeling invariance. This continuous object serves as the limit for sequences of dense graphs: a sequence of graphs (Gn)(G_n) with V(Gn)=n|V(G_n)| = n converges to the graphon WW if the homomorphism densities of every fixed simple graph FF in GnG_n approach the corresponding integrals defined by WW. Specifically, the homomorphism density t(F,Gn)t(F, G_n) converges to t(F,W)t(F, W), where t(F,W)=[0,1]V(F)e=uvE(F)W(xu,xv)vV(F)dxv.t(F, W) = \int_{[0,1]^{|V(F)|}} \prod_{e=uv \in E(F)} W(x_u, x_v) \prod_{v \in V(F)} \, dx_v. This equation expresses the density of homomorphisms from FF into the graphon as a multivariate over the unit , with the product taken over the edges of FF. The functional of graphons stem from their membership in L([0,1]2)L^\infty([0,1]^2), which guarantees essential boundedness ( 0W(x,y)10 \leq W(x,y) \leq 1) and thus integrability over the compact domain, ensuring that subgraph integrals and related functionals are well-defined and finite. A central norm in this space is the cut norm, given by W=supS,T[0,1]STW(x,y)dxdy,\|W\|_\square = \sup_{S,T \subseteq [0,1]} \left| \int_S \int_T W(x,y) \, dx \, dy \right|,
Add your contribution
Related Hubs
User Avatar
No comments yet.