Hubbry Logo
search
logo

Quadratic unconstrained binary optimization

logo
Community Hub0 Subscribers
Read side by side
from Wikipedia

Quadratic unconstrained binary optimization (QUBO), also known as unconstrained binary quadratic programming (UBQP), is a combinatorial optimization problem with a wide range of applications from finance and economics to machine learning.[1] QUBO is an NP hard problem, and for many classical problems from theoretical computer science, like maximum cut, graph coloring and the partition problem, embeddings into QUBO have been formulated.[2][3] Embeddings for machine learning models include support-vector machines, clustering and probabilistic graphical models.[4] Moreover, due to its close connection to Ising models, QUBO constitutes a central problem class for adiabatic quantum computation, where it is solved through a physical process called quantum annealing.[5]

Definition

[edit]

Let the set of binary digits (or bits), then is the set of binary vectors of fixed length . Given a symmetric or upper triangular matrix , whose entries define a weight for each pair of indices , we can define the function that assigns a value to each binary vector through

Alternatively, the linear and quadratic parts can be separated as

where and . This is equivalent to the previous definition through using the diag operator, exploiting that for all binary values .

Intuitively, the weight is added if both and . The QUBO problem consists of finding a binary vector that minimizes , i.e., .

In general, is not unique, meaning there may be a set of minimizing vectors with equal value w.r.t. . The complexity of QUBO arises from the number of candidate binary vectors to be evaluated, as grows exponentially in .

Sometimes, QUBO is defined as the problem of maximizing , which is equivalent to minimizing .

Properties

[edit]

QUBO is scale invariant for positive factors , which leave the optimum unchanged:

.

In its general form, QUBO is NP-hard and cannot be solved efficiently by any polynomial-time algorithm.[6] However, there are polynomially-solvable special cases, where has certain properties,[7] for example:

  • If all coefficients are positive, the optimum is trivially . Similarly, if all coefficients are negative, the optimum is .
  • If is diagonal, the bits can be optimized independently, and the problem is solvable in . The optimal variable assignments are simply if , and otherwise.
  • If all off-diagonal elements of are non-positive, the corresponding QUBO problem is solvable in polynomial time.[8]

QUBO can be solved using integer linear programming solvers like CPLEX or Gurobi Optimizer. This is possible since QUBO can be reformulated as a linear constrained binary optimization problem. To achieve this, substitute the product by an additional binary variable and add the constraints , and . Note that can also be relaxed to continuous variables within the bounds zero and one.

Applications

[edit]

QUBO is a structurally simple, yet computationally hard optimization problem. It can be used to encode a wide range of optimization problems from various scientific areas.[9]

Maximum Cut

[edit]

Given a graph with vertex set and edges , the maximum cut (max-cut) problem consists of finding two subsets with , such that the number of edges between and is maximized.

The more general weighted max-cut problem assumes edge weights , with , and asks for a partition that maximizes the sum of edge weights between and , i.e.,

By setting for all this becomes equivalent to the original max-cut problem above, which is why we focus on this more general form in the following.

For every vertex in we introduce a binary variable with the interpretation if and if . As , every is in exactly one set, meaning there is a 1:1 correspondence between binary vectors and partitions of into two subsets.

We observe that, for any , the expression evaluates to 1 if and only if and are in different subsets, equivalent to logical XOR. Let with . By extending above expression to matrix-vector form we find that

is the sum of weights of all edges between and , where . As this is a quadratic function over , it is a QUBO problem whose parameter matrix we can read from above expression as

after flipping the sign to make it a minimization problem.

Cluster Analysis

[edit]
Binary Clustering with QUBO
20 points with random cluster assignment
A bad cluster assignment.
20 points with sensible cluster assignment
A good cluster assignment.
Visual representation of a clustering problem with 20 points: Circles of the same color belong to the same cluster. Each circle can be understood as a binary variable in the corresponding QUBO problem.

Next, we consider the problem of cluster analysis, where we are given a set of points in -dimensional space and want to assign each point to one of two classes or clusters, such that points in the same cluster are similar to each other. For this example we set and . The data is given as a matrix , where each row contains two cartesian coordinates. For two clusters, we can assign a binary variable to the point corresponding to the -th row in , indicating whether it belongs to the first () or second cluster (). Consequently, we have 20 binary variables, which form a binary vector that corresponds to a cluster assignment of all points (see figure).

One way to derive a clustering is to consider the pairwise distances between points. Given a cluster assignment , the expression evaluates to 1 if points and are in the same cluster. Similarly, indicates that they are in different clusters. Let denote the Euclidean distance between the points and , i.e.,

,

where is the -th row of .

In order to define a cost function to minimize, when points and are in the same cluster we add their positive distance , and subtract it when they are in different clusters. This way, an optimal solution tends to place points which are far apart into different clusters, and points that are close into the same cluster.

Let with for all . Given an assignment , such a cost function is given by

where .

From the second line we can see that this expression can be re-arranged to a QUBO problem by defining

and ignoring the constant term . Using these parameters, a binary vector minimizing this QUBO instance will correspond to an optimal cluster assignment w.r.t. above cost function.

Connection to Ising models

[edit]

QUBO is very closely related and computationally equivalent to the Ising model, whose Hamiltonian function is defined as

with real-valued parameters for all . The spin variables are binary with values from instead of . Note that this formulation is simplified, since, in a physics context, are typically Pauli operators, which are complex-valued matrices of size , whereas here we treat them as binary variables. Many formulations of the Ising model Hamiltonian further assume that the variables are arranged in a lattice, where only neighboring pairs of variables can have non-zero coefficients; here, we simply assume that if and are not neighbors.

Applying the identity yields an equivalent QUBO problem [10]

whose weight matrix is given by

again ignoring the constant term, which does not affect the minization. Using the identity , a QUBO problem with matrix can be converted to an equivalent Ising model using the same technique, yielding

and a constant offset of .[10]

References

[edit]
[edit]
Revisions and contributorsEdit on WikipediaRead on Wikipedia
from Grokipedia
Quadratic unconstrained binary optimization (QUBO) is a combinatorial optimization problem that seeks to minimize a quadratic objective function over binary variables, formulated as finding the vector $ x \in {0,1}^n $ that minimizes $ x^T Q x + c^T x $, where $ Q $ is an $ n \times n $ symmetric matrix and $ c $ is a vector in $ \mathbb{R}^n $.[1] This model unifies a wide range of NP-hard problems, including maximum cut, stable set, and graph partitioning, by encoding constraints implicitly through penalty terms in the objective function.[2] The origins of QUBO trace back to the late 1950s and early 1960s, with foundational work on pseudo-Boolean functions by researchers such as Fortet (1959–1960) and Hammer and Rudeanu (1968), who formalized quadratic forms in binary variables for operations research applications.[1] Subsequent developments in the 1970s and 1980s, including contributions from Hammer, Hansen, and others, advanced algorithms for quadratic 0-1 optimization and established its NP-hardness.[1] QUBO gained renewed prominence in the 2000s due to its equivalence to the Ising model in statistical physics, enabling applications in quantum annealing hardware like D-Wave systems and neuromorphic computing platforms.[2][3] Key features of QUBO include its flexibility in modeling diverse problems through binary decision variables and quadratic interactions, as well as its amenability to both classical solvers (e.g., semidefinite programming relaxations, tabu search metaheuristics) and quantum-inspired methods.[1] While general QUBO is intractable, special cases—such as those on planar graphs or with low-rank matrices—admit polynomial-time solutions, and approximation algorithms achieve ratios like $ O(1/\log n) $ via SDP-based rounding.[1] Notable applications span machine learning (e.g., clustering), computational biology (e.g., RNA folding, protein design), logistics (e.g., vehicle routing), and computer vision (e.g., object detection via energy minimization).[1][2] This versatility positions QUBO as a cornerstone for advancing optimization in both theoretical and practical domains, particularly with emerging hardware accelerations.[2]

Definition and Formulation

Problem Statement

Quadratic unconstrained binary optimization (QUBO) provides a unified framework for modeling and solving a broad class of NP-hard combinatorial optimization problems through the use of binary decision variables xi{0,1}x_i \in \{0,1\}. This approach represents decision choices as binary states, with interactions between variables captured in a quadratic objective function, enabling the encoding of diverse discrete problems such as graph partitioning and facility location without additional constraints beyond the binary requirement. By reducing complex constrained formulations to an unconstrained quadratic form via penalty terms, QUBO simplifies the structure while preserving the problem's essential features.[2] The origins of QUBO trace back to the late 1950s and early 1960s, when studies on pseudo-Boolean functions and binary quadratic optimization emerged in operations research.[1] Seminal work by Hammer and Rudeanu in 1968 formalized these concepts in the context of Boolean methods for optimization, laying the groundwork for unconstrained quadratic binary programming. Subsequent contributions by Fred Glover in the 1970s advanced binary optimization techniques, further developing QUBO's applications in areas like project selection and clustering.[1] QUBO's value stems from its versatility in transforming varied combinatorial problems into a standard quadratic format, which supports efficient solving with metaheuristic algorithms like tabu search and specialized hardware such as quantum annealers from D-Wave Systems.[2][4] This encoding facilitates exploration by classical and quantum methods alike, often outperforming problem-specific solvers for large instances. For example, a basic two-variable QUBO instance seeks to minimize the expression involving the product of the variables plus each variable individually, yielding the optimal binary assignment of zeros for both to achieve the minimum value.[2]

Mathematical Expression

The Quadratic Unconstrained Binary Optimization (QUBO) problem seeks to find a binary vector $ x \in {0,1}^n $ that minimizes the objective function $ x^T Q x + c^T x $, where $ Q $ is an $ n \times n $ symmetric matrix and $ c $ is a vector in $ \mathbb{R}^n $, both containing real-valued entries.[2] Linear terms from an original problem formulation can be absorbed into this quadratic objective by adding the components of $ c $ to the diagonal entries of $ Q $, yielding the compact form $ x^T Q x $ (leveraging the binary property $ x_i^2 = x_i $). This quadratic form naturally incorporates both linear and quadratic contributions: the diagonal elements $ q_{ii} $ (including absorbed linear biases) capture the linear terms $ q_{ii} x_i $, while the off-diagonal elements $ q_{ij} $ (for $ i \neq j $) encode the quadratic interactions. In the common QUBO convention, the objective expands to
xTQx+cTx=i=1n(qii+ci)xi+1i<jnqijxixj, x^T Q x + c^T x = \sum_{i=1}^n (q_{ii} + c_i) x_i + \sum_{1 \leq i < j \leq n} q_{ij} x_i x_j,
where for a symmetric $ Q $, the off-diagonal $ q_{ij} $ typically represents half the pairwise interaction strength to avoid double-counting in the matrix form.[2] In standard notation, $ Q $ serves as the interaction matrix, with diagonal elements $ q_{ii} $ termed biases (representing unary influences on each variable) and off-diagonal elements $ q_{ij} $ termed couplings (representing pairwise interactions between variables).[4] This convention facilitates formulation across diverse optimization contexts, including those amenable to quantum annealing hardware.[2]

Properties

Computational Complexity

The quadratic unconstrained binary optimization (QUBO) problem is NP-hard. This hardness follows from a polynomial-time reduction from the maximum cut problem, which is a well-known NP-hard combinatorial optimization task that can be directly formulated as a QUBO instance by setting the off-diagonal elements of the matrix $ Q $ to negative weights corresponding to the graph edges and adjusting the diagonal for normalization. The decision version of QUBO—determining whether there exists a binary vector $ \mathbf{x} \in {0,1}^n $ such that $ \mathbf{x}^T Q \mathbf{x} \leq k $ for a given threshold $ k $—is NP-complete. Membership in NP is straightforward, as a proposed solution $ \mathbf{x} $ can be verified in polynomial time by computing the quadratic form. Completeness arises from the equivalence between QUBO and the Ising spin glass model, whose decision problem (e.g., whether the ground-state energy is at most zero) is NP-complete via reductions from problems like 3-SAT. Certain special cases of QUBO are solvable in polynomial time. If $ Q $ is diagonal, the objective decouples into $ n $ independent linear terms over single binary variables, allowing the minimum to be found by evaluating each term separately in $ O(n) $ time. If $ Q $ is positive semidefinite, the quadratic form is convex, and its minimum over the hypercube $ [0,1]^n $ occurs at a binary vertex; this can be solved exactly via a semidefinite programming relaxation or interior-point methods for convex quadratic programming, both polynomial-time algorithms. QUBO remains NP-hard in parameterized settings with sparsity constraints on $ Q $. Specifically, the problem is NP-hard even when $ Q $ has only $ O(n) $ nonzero entries, as the reduction from maximum cut preserves sparsity: maximum cut is NP-hard on cubic graphs (maximum degree 3), which yield QUBO instances with $ O(n) $ nonzeros corresponding to the $ O(n) $ edges.

Mathematical Characteristics

Quadratic unconstrained binary optimization (QUBO) constitutes a specialized instance of quadratic programming, wherein the decision variables are confined to the vertices of the binary hypercube {0,1}n\{0,1\}^n. The objective is to minimize f(x)=xTQx+cTxf(x) = x^T Q x + c^T x, with QQ a symmetric n×nn \times n matrix and cRnc \in \mathbb{R}^n. This formulation inherits the combinatorial structure of the hypercube, rendering the feasible set discrete and finite, while the quadratic nature introduces potential non-convexity depending on the spectral properties of QQ.[5] The matrix QQ fundamentally shapes the optimization landscape. As a symmetric matrix, its definiteness is determined by its eigenvalues: if QQ is positive semidefinite (all eigenvalues non-negative), f(x)f(x) is convex over the continuous relaxation, facilitating easier analysis. However, in typical QUBO instances, QQ is indefinite, featuring both positive and negative eigenvalues, which induces non-convexity and results in a multimodal objective function replete with local minima over the binary domain. This indefiniteness arises frequently in applications, as off-diagonal elements of QQ encode interactions that can be attractive or repulsive, mirroring spin systems without explicit constraints.[1] The eigenvalue spectrum of QQ further elucidates the problem's intrinsic challenges. The presence of negative eigenvalues signals non-convexity, with the count of such eigenvalues correlating to the extent of the ruggedness in the energy landscape; more negative eigenvalues generally amplify the number and depth of local minima, thereby increasing the likelihood of suboptimal solutions in local search methods. For instance, even low-rank indefinite QQ (e.g., rank 2) can harbor exponentially many local optima, underscoring the deceptive nature of the discrete quadratic form.[1] To derive bounds, QUBO problems are often relaxed to continuous quadratic optimization over [0,1]n[0,1]^n, which preserves the objective but expands the domain. When QQ is indefinite, this relaxation remains non-convex, but semidefinite programming (SDP) provides a convex outer approximation yielding lower bounds on the minimum value. A standard SDP relaxation minimizes Q,X\langle Q, X \rangle subject to
(1\diag(X)T\diag(X)X)0, \begin{pmatrix} 1 & \diag(X)^T \\ \diag(X) & X \end{pmatrix} \succeq 0,
with \diag(X)=1\diag(X) = \mathbf{1} (typically after linear transformation to the equivalent Ising model over {1,1}\{-1,1\} variables), solvable in polynomial time via interior-point methods and offering tightness for certain structured instances.[6][1]

Equivalences and Mappings

Relation to Ising Model

The quadratic unconstrained binary optimization (QUBO) problem is mathematically equivalent to the classical Ising model from statistical mechanics, allowing for a direct transformation between the two formulations.[4] In the QUBO framework, the objective is to minimize $ H_{\text{QUBO}} = \sum_{i} Q_{i,i} x_i + \sum_{i < j} Q_{i,j} x_i x_j $, where $ x_i \in {0, 1} $ are binary variables and $ Q $ is an upper-triangular matrix of coefficients.[4] The Ising model, conversely, minimizes the Hamiltonian $ H_{\text{Ising}} = \sum_{i} h_i s_i + \sum_{i < j} J_{i,j} s_i s_j $, where $ s_i \in {-1, +1} $ represent spin variables, $ h_i $ are local magnetic fields, and $ J_{i,j} $ are coupling strengths between spins.[7] This equivalence arises because both models are binary quadratic optimization problems, differing only in the representation of the binary domain.[2] The transformation between QUBO and the Ising model is achieved via a simple affine mapping of variables. To convert from QUBO to Ising, substitute $ s_i = 2x_i - 1 $, which maps the binary $ x_i $ to spins $ s_i $. Substituting this into the QUBO objective yields the Ising Hamiltonian:
HIsing=ihisi+i<jJi,jsisj+C, H_{\text{Ising}} = \sum_i h_i s_i + \sum_{i<j} J_{i,j} s_i s_j + C,
where the constant $ C = \sum_i \frac{Q_{i,i}}{2} + \sum_{i<j} \frac{Q_{i,j}}{4} $ does not affect minimization, the linear terms become magnetic fields $ h_i = \frac{Q_{i,i}}{2} + \frac{1}{4} \sum_{j \neq i} Q_{\min(i,j),\max(i,j)} $, and the quadratic terms map to couplings $ J_{i,j} = \frac{Q_{i,j}}{4} $.[4] The reverse mapping from Ising to QUBO uses $ x_i = \frac{s_i + 1}{2} $, transforming the fields and couplings back to QUBO coefficients accordingly.[7] This bijection preserves the structure and optimal solutions of the problems, with linear terms in QUBO corresponding to external fields that bias spins, and quadratic terms to interactions that favor alignments or anti-alignments.[2] This equivalence gained prominence in the 2010s through the development of quantum annealing hardware by D-Wave Systems, which natively solves Ising models on superconducting quantum processors. Earlier formulations of the Ising model date to 1925 in statistical physics for magnetism studies, but its optimization applications were revitalized by D-Wave's 2007 prototype and subsequent systems, prompting widespread reformulation of combinatorial problems as QUBO for compatibility.[2] Seminal works, such as Lucas's 2014 survey, demonstrated how numerous NP-hard problems could be mapped to Ising form, bridging computer science and physics. The mapping is significant because it enables the application of physics-inspired algorithms and hardware designed for Ising problems to solve QUBO instances efficiently.[4] For instance, the large effective fields in QUBO-derived Ising models can simplify spin configurations by coercing many variables to fixed states, potentially reducing computational complexity in certain solvers.[7] This interchangeability has facilitated interdisciplinary advances, allowing optimization researchers to leverage tools from statistical mechanics while adapting Ising solvers for binary decision problems in fields like operations research.[2]

Reductions to Other Problems

Quadratic unconstrained binary optimization (QUBO) provides a versatile framework for solving NP-hard combinatorial problems, as numerous such problems can be polynomially reduced to QUBO by encoding binary decision variables and their interactions directly into the Q matrix. The diagonal elements of Q represent linear costs or rewards associated with individual binary variables, while off-diagonal elements capture pairwise interactions, allowing constraints and objectives to be integrated without explicit enforcement mechanisms. This reduction strategy leverages penalty functions to handle constraints implicitly, transforming constrained problems into unconstrained quadratic forms over binary variables.[2] A prominent example is the reduction of the minimum vertex cover problem, where binary variables xj{0,1}x_j \in \{0,1\} indicate whether vertex jj is included in the cover (xj=1x_j = 1) or not. The objective minimizes the cover size via linear terms jVxj\sum_{j \in V} x_j, while ensuring every edge is covered adds penalty terms P(i,j)E(1xixj+xixj)P \sum_{(i,j) \in E} (1 - x_i - x_j + x_i x_j), with P>0P > 0 a sufficiently large scalar to enforce feasibility. The resulting QUBO formulation is minxTQx\min \mathbf{x}^T Q \mathbf{x}, where uncovered vertices contribute negatively to the diagonal through the expanded penalty but are outweighed by the inclusion cost, prioritizing minimal covers.[2] The maximum independent set problem similarly reduces to QUBO, using binary variables xi{0,1}x_i \in \{0,1\} to denote inclusion in the set. To maximize the set size while prohibiting adjacent vertices, the formulation minimizes H=iVxi+2(i,j)ExixjH = -\sum_{i \in V} x_i + 2 \sum_{(i,j) \in E} x_i x_j, with negative diagonal entries in Q encouraging vertex selection and positive off-diagonal entries for edges penalizing violations. This setup ensures the optimal solution corresponds to the largest non-adjacent vertex subset.[8] QUBO is synonymous with unconstrained quadratic binary programming, serving as a canonical representation that facilitates these reductions across diverse domains. In modeling constraint satisfaction problems, QUBO incorporates restrictions—such as equalities Ax=bA\mathbf{x} = b or inequalities—via quadratic penalty terms like P(Axb)2P (A\mathbf{x} - b)^2, embedding them into the objective to maintain the pure unconstrained binary quadratic structure without separate constraint sets.[2] Conversely, QUBO itself, being NP-hard, can be reduced to other NP-hard problems to leverage specialized solvers; for instance, any QUBO instance can be transformed into an equivalent maximum cut problem by constructing a graph where edge weights derive from the off-diagonal Q elements and additional self-loops or gadgets account for diagonal terms, allowing MaxCut algorithms to solve the original QUBO. Similar graph-based transformations extend this to problems like the traveling salesman problem via embeddings that map QUBO interactions onto TSP tour constraints.[9]

Applications

Maximum Cut Problem

The maximum cut problem seeks to partition the vertices of an undirected graph into two disjoint subsets such that the total weight of edges with endpoints in different subsets is maximized. For an unweighted graph, this corresponds to maximizing the number of crossing edges; for a weighted graph with edge weights wij>0w_{ij} > 0, it maximizes (i,j)Ewij(xi(1xj)+xj(1xi))\sum_{(i,j) \in E} w_{ij} (x_i (1 - x_j) + x_j (1 - x_i)), where xi{0,1}x_i \in \{0,1\} assigns vertex ii to one subset if xi=1x_i = 1 and the other if xi=0x_i = 0. This objective simplifies to ixi(jiwij)2(i,j)Ewijxixj\sum_i x_i \left( \sum_{j \sim i} w_{ij} \right) - 2 \sum_{(i,j) \in E} w_{ij} x_i x_j, which is a quadratic form $ \mathbf{x}^T Q \mathbf{x} $ with $ Q_{ii} = \sum_{j \sim i} w_{ij} $ (the weighted degree of ii) and $ Q_{ij} = Q_{ji} = -w_{ij} $ if (i,j)E(i,j) \in E, or 00 otherwise. Maximizing this form yields the maximum cut value, making Max-Cut a direct QUBO instance; equivalently, minimizing xTQx-\mathbf{x}^T Q \mathbf{x} aligns with standard QUBO minimization frameworks. For example, consider an unweighted graph with 5 vertices and edges {(1,2),(1,3),(2,4),(3,4),(3,5),(4,5)}\{ (1,2), (1,3), (2,4), (3,4), (3,5), (4,5) \}. The corresponding QQ matrix is
Q=(2110012010103110113100112). Q = \begin{pmatrix} 2 & -1 & -1 & 0 & 0 \\ -1 & 2 & 0 & -1 & 0 \\ -1 & 0 & 3 & -1 & -1 \\ 0 & -1 & -1 & 3 & -1 \\ 0 & 0 & -1 & -1 & 2 \end{pmatrix}.
The optimal assignment x=(0,1,1,0,0)\mathbf{x} = (0,1,1,0,0) places vertices 2 and 3 in one subset and 1, 4, 5 in the other, yielding xTQx=5\mathbf{x}^T Q \mathbf{x} = 5 and a maximum cut of 5 edges (specifically, the crossing edges are (1,2), (1,3), (2,4), (3,4), (3,5)). Max-Cut has served as a canonical QUBO application in combinatorial optimization since the 1970s, highlighted by seminal approximation algorithms such as the Goemans-Williamson method, which achieves a 0.878-approximation ratio via semidefinite programming relaxation.

Graph Partitioning and Clustering

Graph partitioning problems involve dividing the vertices of an undirected graph into two subsets of approximately equal size while minimizing the number of edges that connect vertices across the subsets, known as the cut size. This can be formulated as a QUBO problem by assigning binary variables xi{0,1}x_i \in \{0,1\} to each vertex ii, where xi=1x_i = 1 indicates assignment to one subset and 00 to the other. The objective function minimizes the cut size through terms in the Q matrix that penalize differing assignments for connected vertices, such as Qij=γQ_{ij} = -\gamma for edges (i,j)E(i,j) \in E, where γ>0\gamma > 0 is a scaling factor derived from the adjacency matrix. Balance constraints are encoded via quadratic penalties in QQ, for example, by adding terms like λ(ixin/2)2\lambda \left( \sum_i x_i - n/2 \right)^2, where λ>0\lambda > 0 enforces equal partition sizes by penalizing deviations from half the total vertices nn; this expands to diagonal and off-diagonal entries in QQ that discourage imbalance.[10] Such QUBO formulations for graph partitioning have been applied in VLSI design since the 1990s to optimize circuit placement and routing by minimizing wire lengths modeled as cut edges, with early quadratic binary models paving the way for modern unconstrained approaches. In social network analysis, these methods facilitate balanced divisions for studying network structures, emerging prominently in the 2010s with quantum-enhanced solvers. Unlike the maximum cut problem, which seeks unbalanced partitions to maximize crossing edges, graph partitioning prioritizes size equality for practical load balancing.[10] For clustering tasks, a k-means-like binary clustering (with k=2k=2) can be expressed in QUBO form by setting the Q matrix to reward intra-cluster similarity, such as minimizing squared distances between points assigned to the same cluster via terms Qij=d(pi,pj)Q_{ij} = d(p_i, p_j) for data points pi,pjp_i, p_j, where positive entries penalize (discourage) co-assignment of dissimilar points. Imbalance is penalized quadratically, similar to partitioning, ensuring each cluster has roughly half the points; for instance, the full objective becomes minxxTQx+λ(ixiN/2)2\min_x x^T Q x + \lambda \left( \sum_i x_i - N/2 \right)^2, where NN is the number of points. This approach targets global optima unlike iterative classical k-means, with empirical tests showing competitive results on small datasets.[11] A specific QUBO formulation arises in modularity maximization for community detection, where the goal is to partition graphs into communities with dense internal connections. The modularity matrix Bij=Aijγdidj2mB_{ij} = A_{ij} - \gamma \frac{d_i d_j}{2m} defines Q entries, with AijA_{ij} the adjacency, did_i degrees, mm total edges, and γ\gamma a resolution parameter; positive BijB_{ij} rewards co-assignment in the same community, while negative values penalize it based on expected edge densities under a null model. The QUBO objective is then minxxT(B)x+balance terms\min_x x^T (-B) x + \text{balance terms}, maximizing modularity Q14mijBijsisjQ \approx \frac{1}{4m} \sum_{ij} B_{ij} s_i s_j (with Ising spins si=2xi1s_i = 2x_i - 1) to identify communities in complex networks like social graphs. This has been implemented on quantum annealers for scalable detection.[12]

Machine Learning and Other Uses

Quadratic unconstrained binary optimization (QUBO) formulations have found significant applications in machine learning, particularly for feature selection tasks, where the objective is to identify a subset of relevant features from high-dimensional data while minimizing redundancy. In this context, the QUBO problem is constructed by encoding feature importance scores in the linear terms of the objective function and penalizing pairwise redundancies through the quadratic matrix $ Q $, allowing binary variables to represent inclusion or exclusion of features. [13] This approach balances model interpretability and performance, as demonstrated in quantum annealing implementations that outperform classical greedy methods on datasets with thousands of features. [14] For sparse principal component analysis (PCA), QUBO models incorporate quadratic regularization terms to enforce sparsity in the loading vectors, enabling the identification of key principal components with minimal non-zero entries while preserving variance explanation. In finance, QUBO is widely used for portfolio optimization, where binary variables indicate asset inclusion, and the quadratic terms in $ Q $ capture risk-return tradeoffs via covariance matrices and expected returns. [15] This formulation addresses the Markowitz mean-variance problem in a discrete setting, often solved with quantum-inspired algorithms to handle combinatorial explosion in large asset universes. [16] Empirical studies show that such QUBO-based portfolios achieve Sharpe ratios comparable to continuous optimizations but with enhanced diversification through binary constraints. [17] In computational biology, QUBO models extend to RNA secondary structure prediction, where binary variables represent base-pairing decisions, and quadratic terms encode thermodynamic stability and pseudoknot constraints to minimize free energy.[18] Protein design uses QUBO to optimize amino acid sequences for target folds, penalizing unfavorable interactions via the Q matrix to achieve desired structures.[19] Protein folding, as in the hydrophobic-polar (HP) lattice model, employs binary variables for residue placements to minimize energy through quadratic interactions; pioneering QUBO formulations in the 2010s enabled quantum solver applications on small peptides. [20] [21] In logistics, QUBO formulations solve vehicle routing problems (VRP), such as the capacitated VRP, by using binary variables for route assignments and quadratic penalties for capacity and distance constraints, facilitating efficient delivery scheduling.[22] For computer vision, QUBO minimizes energy functions in tasks like object detection and segmentation, where binary labels indicate object presence, and quadratic terms model spatial and appearance consistencies.[23] In scheduling problems, QUBO encodes resource allocation constraints quadratically, as seen in NASA's Deep Space Network antenna assignments, where it optimizes track durations while avoiding overlaps. Applications in recommender systems, emerging around 2020–2024, use QUBO for binary user-item interactions or feature selection in collaborative filtering, optimizing over sparse matrices; advancements include counterfactual analysis for improved personalization on e-commerce datasets. [24] [25]

Solution Methods

Exact Algorithms

Exact algorithms for solving quadratic unconstrained binary optimization (QUBO) problems guarantee finding the global optimum by systematically exploring the 2n2^n possible binary solutions, where nn is the number of variables, while employing bounding techniques to prune the search tree and avoid exhaustive enumeration. These methods are essential due to the NP-hardness of QUBO, which arises from reductions to problems like maximum cut. Branch-and-bound approaches dominate, often incorporating Lagrangian relaxation to derive tight upper bounds on the objective function, enabling efficient pruning of suboptimal branches.[26] In branch-and-bound for QUBO, the search tree is constructed by sequentially fixing binary variables to 0 or 1, partitioning the problem into subproblems at each node. Lagrangian relaxation is applied to the quadratic form xTQx\mathbf{x}^T Q \mathbf{x}, where QQ is the coefficient matrix, by dualizing the unfixed variables to obtain a relaxed bound; specifically, the dual function provides an upper bound that is compared against the current best solution to fathom branches where no improvement is possible. Seminal implementations, such as those by Pardalos and Rodgers, use dynamic preprocessing and parallelization to solve small instances up to around 30 variables, with extensions allowing larger specific cases, though general performance is limited by density. Enhancements like roof duality can further tighten bounds by identifying persistent variables—those fixed to optimal values across the roof dual—reducing the effective search space.[27] For sparse QUBO instances, where the matrix QQ has limited nonzero entries corresponding to a graph with low density, dynamic programming exploits structural properties to enumerate states efficiently. Roof duality plays a key role here by computing the convex roof of the quadratic function, allowing variable fixing and decomposition into smaller subproblems solvable via dynamic programming over the variable dependencies, such as in chain-like or tree-structured graphs. This approach reduces the state space from exponential in nn to polynomial in the sparsity degree, as demonstrated in methods for pseudo-Boolean optimization that extend to sparse QUBO. For example, in instances with bounded treewidth, dynamic programming tables track partial solutions over bags of variables, achieving exact solutions for graphs with hundreds of nodes if sparsity is high.[9] QUBO can also be reformulated as a mixed-integer quadratic program (MIQP) by enforcing binarity through linear constraints, such as xi{0,1}x_i \in \{0,1\} or big-M formulations for quadratic terms if needed, though modern solvers handle general quadratic objectives directly. Commercial solvers like Gurobi and CPLEX employ branch-and-bound internally with advanced presolving, cutting planes, and Lagrangian heuristics to solve the MIQP exactly, making this a practical exact method despite deviating from pure QUBO. These tools integrate seamlessly with QUBO via simple variable declarations and objective specification. Performance of exact solvers typically limits to n50n \leq 50 for dense random instances, with solution times ranging from seconds to days depending on structure; sparse cases extend to n>1000n > 1000 using graph-based reductions. Benchmarks from QUBO workshops since 2010, such as those comparing SCIP-based branch-and-cut against BiqCrunch, highlight that Lagrangian-enhanced branch-and-bound solves 80-90% of n=30n=30 dense instances within hours, underscoring scalability challenges for larger nn.

Heuristic and Approximation Techniques

Heuristic and approximation techniques for quadratic unconstrained binary optimization (QUBO) focus on efficiently finding high-quality solutions for large-scale instances where exact methods are computationally prohibitive. These approaches trade optimality for speed, leveraging local improvements, population-based searches, and relaxation-based rounding to navigate the combinatorial landscape of binary variables. Seminal developments include local search methods that iteratively flip variables to reduce the objective function while incorporating mechanisms to avoid stagnation, as well as metaheuristics inspired by natural processes. One prominent local search heuristic is tabu search, introduced by Fred Glover for binary quadratic programs, including QUBO formulations. In this method, starting from an initial binary solution, the algorithm evaluates neighborhood moves by flipping a single variable and selects the one yielding the largest improvement in the objective function. To escape local minima, it maintains a tabu list of recently flipped variables, temporarily forbidding their reversal to promote diversification, while allowing aspiration criteria to override tabus if a superior solution is found. Glover's adaptive memory variant enhances this by storing elite solutions and guiding future searches toward promising regions, demonstrating superior performance over prior heuristics on challenging instances up to n=100 variables by finding optimal or near-optimal solutions faster than exact branch-and-bound methods.[28] Metaheuristics extend local search through broader exploration strategies. Genetic algorithms treat binary strings as chromosomes representing QUBO solutions and evolve populations via selection, crossover, and mutation operators to minimize the quadratic objective. For instance, uniform crossover exchanges bits between parent solutions to generate offspring, while mutation randomly flips bits to introduce diversity; hybrid variants integrate local search for intensification, outperforming standalone tabu search and simulated annealing on large QUBO instances with n=200–2500 by discovering new best-known solutions for dense problems. Simulated annealing, another metaheuristic, draws brief inspiration from Ising model thermal transitions by probabilistically accepting worsening flips at high "temperatures" to escape local minima, gradually cooling to favor improvements; applied to unconstrained quadratic pseudo-Boolean functions equivalent to QUBO, it achieves efficient solution quality on hundreds of test problems with O(n²) complexity per temperature step.[29] For specific QUBO reductions like the maximum cut problem, approximation algorithms provide worst-case guarantees. The Goemans-Williamson algorithm relaxes the binary quadratic form to a semidefinite program, solving it to obtain vector embeddings, then rounds via random hyperplane cuts to yield a binary partition; this delivers an expected approximation ratio of at least 0.878 to the optimum, marking a significant advance for graph partitioning applications of QUBO.[30] To scale these techniques, hybrid CPU-GPU implementations parallelize metaheuristic components across hardware. The Diverse Adaptive Bulk Search (DABS) framework, for example, uses CPU threads to manage diverse solution pools and GPU kernels to execute bulk neighborhood evaluations and genetic operations simultaneously on multiple devices, enabling effective solving of QUBO instances with n exceeding 1000 variables, such as dense MaxCut problems, while outperforming commercial solvers like Gurobi in solution quality and speed.

Quantum and Hybrid Approaches

Quantum annealing leverages the equivalence between QUBO problems and the Ising model to map optimization objectives onto the Hamiltonian of quantum annealers, enabling hardware-based solving through adiabatic evolution. D-Wave Systems introduced commercial quantum annealers in 2011, with processors like the D-Wave 2000Q (2017), Advantage (2020), and Advantage2 (2023, with over 7,000 qubits and improved connectivity) designed to natively handle Ising formulations derived from QUBO instances. As of November 2025, D-Wave announced a new annealing quantum computer scheduled for release in February 2025. These systems embed QUBO problems by representing binary variables as qubit states and quadratic terms as interactions, allowing the annealer to find low-energy configurations that correspond to near-optimal solutions. Early demonstrations showed applicability to problems like portfolio optimization, where quantum annealing provided solutions competitive with classical methods for instances up to hundreds of variables.[31] The gate-model quantum approach employs the Quantum Approximate Optimization Algorithm (QAOA), a variational hybrid method introduced by Farhi et al. in 2014, which applies alternating layers of cost and mixer Hamiltonians to approximate solutions for combinatorial problems including QUBO.[32] In QAOA, the QUBO objective is encoded as a cost Hamiltonian, and tunable parameters are optimized classically to maximize expectation values on noisy intermediate-scale quantum (NISQ) devices like those from IBM or IonQ. As of 2025, this enables solving QUBO instances with up to hundreds of variables on devices exceeding 100 qubits, with performance improving through deeper circuits and error mitigation, though limited by hardware coherence times.[33] Hybrid solvers combine quantum-inspired classical techniques with quantum hardware or simulations for larger-scale QUBO problems. Fujitsu's Digital Annealer, first detailed in 2019, uses a CMOS-based architecture mimicking quantum annealing dynamics to solve fully connected QUBO models without embedding constraints, achieving near-real-time performance for combinatorial tasks.[34] Post-2020 advancements, including the third-generation unit introduced in 2021 with capacity up to 100,000 bits, integrate software layers for partitioning large problems, demonstrating scalability for dense QUBO instances beyond pure quantum limits as of 2025.[35] These hybrids often outperform standalone quantum annealers in solution quality for industrial applications like scheduling. Recent 2025 benchmarks confirm the Digital Annealer's competitive performance on large-scale Max-Cut problems, a key QUBO application.[36] Current quantum and hybrid methods face significant limitations, including embedding overhead in annealers like D-Wave, where dense QUBO graphs require chaining multiple qubits per variable, increasing problem size by factors of 5-10 and amplifying noise effects.[37] NISQ devices introduce gate and readout errors, degrading QAOA accuracy for depths beyond p=3, with error rates necessitating error mitigation techniques.[38] Benchmarks indicate speedups for sparse QUBO problems with n≈10^4 variables, where D-Wave hybrids solve instances in seconds compared to hours on classical solvers, though dense cases remain challenging due to overhead.[39]

References

User Avatar
No comments yet.