Log in

Bandgap optimization in combinatorial graphs with tailored ground states: application in quantum annealing

  • Research Article
  • Published:
Optimization and Engineering Aims and scope Submit manuscript

Abstract

A mixed-integer linear programming (MILP) formulation is presented for parameter estimation of the Potts model. Two algorithms are developed; the first method estimates the parameters such that the set of ground states replicate the user-prescribed data set; the second method allows the user to prescribe the ground states multiplicity. In both instances, the optimization process ensures that the bandgap is maximized. Consequently, the model parameter efficiently describes the user data for a broad range of temperatures. This is useful in the development of energy-based graph models to be simulated on Quantum annealing hardware where the exact simulation temperature is unknown. Computationally, the memory requirement in this method grows exponentially with the graph size. Therefore, this method can only be practically applied to small graphs. Such applications include learning of small generative classifiers and spin-lattice model with energy described by Ising hamiltonian. Learning large data sets poses no extra cost to this method; however, applications involving the learning of high dimensional data are out of scope.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
EUR 32.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or Ebook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price includes VAT (Germany)

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3

Similar content being viewed by others

Abbreviations

\(\varvec{\theta }\) :

Set of parameters for potts model

\(\varvec{S}\) :

A state of the graph

\(\Delta E\) :

Band gap

\(\eta \) :

Negative Log-likelihood

\(\mathcal {C}\) :

Set of graph’s connection

\(\mathcal {S}\) :

Set of all possible states

\(\mathcal {S}_D\) :

Set of the data states

\(\mathcal {S}_E^{\varvec{\theta }}\) :

Set of all Excited states

\(\mathcal {S}_G^{\varvec{\theta }}\) :

Set of all Ground states

\(\mathcal {V}\) :

Set of graph’s vertices

E :

Potts Energy

\(E_0\) :

Energy of ground state

\(E_1\) :

Energy of \(1^{st}\) excited state

G :

Simple undirected weighted Graph

\(N_{C}\) :

Number of graph connections

\(N_{DS}\) :

Number of data states

\(N_{ES}\) :

Number of excited states

\(N_{GS}\) :

Number of ground states

\(N_{L}\) :

Number of labels

\(N_{TS}\) :

Number of total states

\(N_{V}\) :

Number of graph vertices

\(p(\varvec{S})\) :

Probability of a state, \(\varvec{S}\)

\(s_i\) :

Label of Vertex with index i

\(v_i\) :

Vertex with index i

References

  • Graner François, Glazier James A (1992) Simulation of biological cell sorting using a two-dimensional extended potts model. Phys Rev Lett 69(13):2013

    Article  Google Scholar 

  • Miodownik Mark (2007) Monte carlo potts model. In: Janssens Koenraad G.F., Raabe Dierk, Kozeschnik Ernst, Miodownik Mark A, Nestler Britta (eds) Computational materials engineering. Academic Press, Burlington, pp 47–108

    Chapter  Google Scholar 

  • Boykov Yuri, Veksler Olga, Zabih Ramin (2001) Fast approximate energy minimization via graph cuts. IEEE Trans Pattern Anal Mach Intell 23(11):1222–1239

    Article  Google Scholar 

  • Shai Bagon (2012) Discrete energy minimization, beyond submodularity: Applications and approximations. ar**v preprint ar**v:1210.7362

  • Descombes Xavier, Morris Robin D, Zerubia Josiane, Berthod Marc (1999) Estimation of Markov random field prior parameters using Markov chain monte Carlo maximum likelihood. IEEE Trans Image Process 8(7):954–963

    Article  MathSciNet  MATH  Google Scholar 

  • Hinton Geoffrey E (2002) Training products of experts by minimizing contrastive divergence. Neural Comput 14(8):1771–1800

    Article  MATH  Google Scholar 

  • Asja Fischer and Christian Igel (2012) An introduction to restricted boltzmann machines. In: Iberoamerican Congress on Pattern Recognition, Springer, pp. 14–36

  • Steven H Adachi and Maxwell P Henderson (2015) Application of quantum annealing to training of deep neural networks. ar**v preprint ar**v:1510.06356

  • Siddhartha Srivastava, Veera Sundararaghavan (2020) Machine learning in quantum computers via general boltzmann machines: Generative and discriminative training through annealing. ar**v preprint ar**v:2002.00792

  • Johnson Mark W, Amin Mohammad HS, Gildert Suzanne, Lanting Trevor, Hamze Firas, Dickson Neil, Harris Richard, Berkley Andrew J, Johansson Jan, Bunyk Paul et al (2011) Quantum annealing with manufactured spins. Nature 473(7346):194–198

    Article  Google Scholar 

  • Hassan K. Khalil (2002) Fundamental properties. In: Nonlinear systems, Prentice Hall, pp. 87 – 110.

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Siddhartha Srivastava.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendices

Appendices

The codes are available at https://github.com/sidsriva/PEP.

Appendix 1: Proof of theorem

(a) Since \(\mathcal {S}_{G}(\varvec{\theta },\beta ) = \mathcal {S}_D\), the Negative Log Likelhood, \(\eta (\varvec{\theta }_D,\beta )\), is estimated as:

$$\begin{aligned} \eta (\varvec{\theta }_D,\beta ) = N_{GS} \beta E_0 + N_{GS} \log Z \end{aligned}$$

The derivative is estimated as::

$$\begin{aligned} \frac{d\eta }{d\beta }&= N_{GS} \left( (E_{0} - \mathbb {E}(E) \right) \end{aligned}$$
(10)

where

$$\begin{aligned} \mathbb {E}(E) = \sum _{\varvec{S}\in \mathcal {S}} E(\varvec{S}) p(\varvec{S}|\varvec{\theta }_D,\beta ) \end{aligned}$$

Since \(\Delta E > 0\), the expected energy is strictly bounded below as \(\mathbb {E}(E) > E_0\). Consequently:

$$\begin{aligned} \frac{d\eta }{d\beta } < 0 \end{aligned}$$

In the low temperature limit, Eq. (2) estimates that the probability of all excited states approaches 0 while all ground states are equally likely with probability \((N_{GS})^{-1}\). Therefore, the value of \(\eta \) in this limit is estimated as Eq. (4).

(b) Let \(\varvec{S}_G \in \mathcal {S}_G\) and \(P = p(\varvec{S}_G|\varvec{\theta }_D,\beta )\) so that \(\eta (\varvec{\theta }_D,\beta ) = - N_{GS} \log {P}\). The probability of occurrence of a ground state is given by \(N_{GS}P\) and occurrence of a excited state is given as \(\left( 1-N_{GS}P\right) \). Moreover, for any finite value of \(\beta \) both of these probabilities are finite. Therefore, the expectation of energy, \(\mathbb {E}\), can be bounded as

$$\begin{aligned} \mathbb {E} = N_{GS} P E_0 + \sum _{\varvec{S}\in \mathcal {S}_E} E(\varvec{S}) p(\varvec{S}|\varvec{\theta }_D,\beta ) \le N_{GS} P E_0 + (1- N_{GS} P) E_1 \end{aligned}$$

Substituting in Eq. (10),

$$\begin{aligned} \frac{d \eta }{d\beta } = E_{0} - \mathbb {E}(E)&\le \left( N_{GS}P - 1\right) N_{GS} \Delta E \end{aligned}$$

Substituting \(P = e^{-\eta /N_{GS}}\) gives the following differential inequality

$$\begin{aligned} \frac{d \eta }{d\beta } \le \left( N_{GS} e^{-\eta /N_{GS}} - 1\right) N_{GS} \Delta E \end{aligned}$$
(11)

Consider the differential equation for \(\beta \in [0,\infty )\),

$$\begin{aligned} \frac{d \xi }{d\beta } = \left( N_{GS} e^{-\xi /N_{GS}} - 1\right) N_{GS} \Delta E \end{aligned}$$
(12)

with initial condition \(\xi (\varvec{\theta }_D,0) = \eta (\varvec{\theta }_D,0) = N_{GS}\log N_{TS} \). Noting that \(N_{GS} e^{-\xi /N_{GS}} - 1 = N_{GS}P-1 > 0\), this ODE is integrated to give the following solution:

$$\begin{aligned} \xi (\varvec{\theta }_D,\beta ) = N_{GS} \log { \left( N_{GS} + N_{ES} e^{-\beta \Delta E} \right) } \end{aligned}$$
(13)

Using Comparison Lemma (Khalil 2002), for all \(0<\beta <\infty \),

$$\begin{aligned} \eta (\varvec{\theta }_D,\beta ) \le \xi (\varvec{\theta }_D,\beta ) \end{aligned}$$
(14)

This proves the upper bound. The lower bound is a direct consequence from monotonicity proved in part 1.

(c) For any \(\beta < \infty \)

$$\begin{aligned} \eta (\varvec{\theta }_D,\beta ) - N_{GS} \log {N_{GS}} \le N_{GS} \log \left( 1+ \frac{N_{ES}}{N_{GS}} e^{-\beta \Delta E} \right) \end{aligned}$$

For any \(\epsilon >0\), choose a \(\beta > \beta ^*(\epsilon )\) using Eq. (6) and observe that,

$$\begin{aligned} N_{GS}\log \left( 1+ \frac{N_{ES}}{N_{GS}} e^{-\beta \Delta E} \right) < \epsilon \end{aligned}$$

This proves the third statement.

Appendix 2: Optimized graphs

1.1 Appendix 2.1: K-3 graph

A fully connected 3-noded graph is optimized for 4 data states. The energy of the graph is modeled using Ising model Eq. (9) with \(|H|\le 1\) and \(|J|\le 1\). The optimized parameters using the (1) Minimization of Negative Log-likelihood, and (2) PEPDAS method are presented in Fig. 4

Fig. 4
figure 4

a Training data set of states with green representing a ‘+1’ state and red representing a ‘-1’ state. b Optimized graph using minimization of Negative Log-likelhood at \(\beta =1\) c Optimized graph using PEPDAS method. The field terms are mentioned in blue color and interaction terms are mentioned in red color

1.2 Appendix 2.2: Peterson graph

A Peterson graph is first optimized for upto 3 user prescribed data states using PEPDAS method. Then it is optimized for 3 ground states using PEPGSM method. The energy of the graph is modeled using Ising model Eq. (9) with \(|H|<1\) and \(|J|<1\). The optimized graphs are presented in Fig. 5 and their respective Negative log likelhood is presented in Fig. 6.

Fig. 5
figure 5

Optimal Ising parameters of a Peterson graph found using PEPDAS method (left) and PEPGSM method (right). The ground states are presented as the colored graph in the top right corner of each image. A green label denotes the ‘\(+1\)’ state and the red label denotes the ‘\(-1\)’ state

Fig. 6
figure 6

Normalized Negative log likelihood and their respective bounds for Peterson graphs trained using PEPDAS method

Rights and permissions

Springer Nature or its licensor holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Srivastava, S., Sundararaghavan, V. Bandgap optimization in combinatorial graphs with tailored ground states: application in quantum annealing. Optim Eng 24, 1931–1949 (2023). https://doi.org/10.1007/s11081-022-09758-9

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11081-022-09758-9

Keywords

Navigation