Abstract
A mixed-integer linear programming (MILP) formulation is presented for parameter estimation of the Potts model. Two algorithms are developed; the first method estimates the parameters such that the set of ground states replicate the user-prescribed data set; the second method allows the user to prescribe the ground states multiplicity. In both instances, the optimization process ensures that the bandgap is maximized. Consequently, the model parameter efficiently describes the user data for a broad range of temperatures. This is useful in the development of energy-based graph models to be simulated on Quantum annealing hardware where the exact simulation temperature is unknown. Computationally, the memory requirement in this method grows exponentially with the graph size. Therefore, this method can only be practically applied to small graphs. Such applications include learning of small generative classifiers and spin-lattice model with energy described by Ising hamiltonian. Learning large data sets poses no extra cost to this method; however, applications involving the learning of high dimensional data are out of scope.
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11081-022-09758-9/MediaObjects/11081_2022_9758_Fig1_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11081-022-09758-9/MediaObjects/11081_2022_9758_Fig2_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11081-022-09758-9/MediaObjects/11081_2022_9758_Fig3_HTML.png)
Similar content being viewed by others
Abbreviations
- \(\varvec{\theta }\) :
-
Set of parameters for potts model
- \(\varvec{S}\) :
-
A state of the graph
- \(\Delta E\) :
-
Band gap
- \(\eta \) :
-
Negative Log-likelihood
- \(\mathcal {C}\) :
-
Set of graph’s connection
- \(\mathcal {S}\) :
-
Set of all possible states
- \(\mathcal {S}_D\) :
-
Set of the data states
- \(\mathcal {S}_E^{\varvec{\theta }}\) :
-
Set of all Excited states
- \(\mathcal {S}_G^{\varvec{\theta }}\) :
-
Set of all Ground states
- \(\mathcal {V}\) :
-
Set of graph’s vertices
- E :
-
Potts Energy
- \(E_0\) :
-
Energy of ground state
- \(E_1\) :
-
Energy of \(1^{st}\) excited state
- G :
-
Simple undirected weighted Graph
- \(N_{C}\) :
-
Number of graph connections
- \(N_{DS}\) :
-
Number of data states
- \(N_{ES}\) :
-
Number of excited states
- \(N_{GS}\) :
-
Number of ground states
- \(N_{L}\) :
-
Number of labels
- \(N_{TS}\) :
-
Number of total states
- \(N_{V}\) :
-
Number of graph vertices
- \(p(\varvec{S})\) :
-
Probability of a state, \(\varvec{S}\)
- \(s_i\) :
-
Label of Vertex with index i
- \(v_i\) :
-
Vertex with index i
References
Graner François, Glazier James A (1992) Simulation of biological cell sorting using a two-dimensional extended potts model. Phys Rev Lett 69(13):2013
Miodownik Mark (2007) Monte carlo potts model. In: Janssens Koenraad G.F., Raabe Dierk, Kozeschnik Ernst, Miodownik Mark A, Nestler Britta (eds) Computational materials engineering. Academic Press, Burlington, pp 47–108
Boykov Yuri, Veksler Olga, Zabih Ramin (2001) Fast approximate energy minimization via graph cuts. IEEE Trans Pattern Anal Mach Intell 23(11):1222–1239
Shai Bagon (2012) Discrete energy minimization, beyond submodularity: Applications and approximations. ar**v preprint ar**v:1210.7362
Descombes Xavier, Morris Robin D, Zerubia Josiane, Berthod Marc (1999) Estimation of Markov random field prior parameters using Markov chain monte Carlo maximum likelihood. IEEE Trans Image Process 8(7):954–963
Hinton Geoffrey E (2002) Training products of experts by minimizing contrastive divergence. Neural Comput 14(8):1771–1800
Asja Fischer and Christian Igel (2012) An introduction to restricted boltzmann machines. In: Iberoamerican Congress on Pattern Recognition, Springer, pp. 14–36
Steven H Adachi and Maxwell P Henderson (2015) Application of quantum annealing to training of deep neural networks. ar**v preprint ar**v:1510.06356
Siddhartha Srivastava, Veera Sundararaghavan (2020) Machine learning in quantum computers via general boltzmann machines: Generative and discriminative training through annealing. ar**v preprint ar**v:2002.00792
Johnson Mark W, Amin Mohammad HS, Gildert Suzanne, Lanting Trevor, Hamze Firas, Dickson Neil, Harris Richard, Berkley Andrew J, Johansson Jan, Bunyk Paul et al (2011) Quantum annealing with manufactured spins. Nature 473(7346):194–198
Hassan K. Khalil (2002) Fundamental properties. In: Nonlinear systems, Prentice Hall, pp. 87 – 110.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendices
Appendices
The codes are available at https://github.com/sidsriva/PEP.
Appendix 1: Proof of theorem
(a) Since \(\mathcal {S}_{G}(\varvec{\theta },\beta ) = \mathcal {S}_D\), the Negative Log Likelhood, \(\eta (\varvec{\theta }_D,\beta )\), is estimated as:
The derivative is estimated as::
where
Since \(\Delta E > 0\), the expected energy is strictly bounded below as \(\mathbb {E}(E) > E_0\). Consequently:
In the low temperature limit, Eq. (2) estimates that the probability of all excited states approaches 0 while all ground states are equally likely with probability \((N_{GS})^{-1}\). Therefore, the value of \(\eta \) in this limit is estimated as Eq. (4).
(b) Let \(\varvec{S}_G \in \mathcal {S}_G\) and \(P = p(\varvec{S}_G|\varvec{\theta }_D,\beta )\) so that \(\eta (\varvec{\theta }_D,\beta ) = - N_{GS} \log {P}\). The probability of occurrence of a ground state is given by \(N_{GS}P\) and occurrence of a excited state is given as \(\left( 1-N_{GS}P\right) \). Moreover, for any finite value of \(\beta \) both of these probabilities are finite. Therefore, the expectation of energy, \(\mathbb {E}\), can be bounded as
Substituting in Eq. (10),
Substituting \(P = e^{-\eta /N_{GS}}\) gives the following differential inequality
Consider the differential equation for \(\beta \in [0,\infty )\),
with initial condition \(\xi (\varvec{\theta }_D,0) = \eta (\varvec{\theta }_D,0) = N_{GS}\log N_{TS} \). Noting that \(N_{GS} e^{-\xi /N_{GS}} - 1 = N_{GS}P-1 > 0\), this ODE is integrated to give the following solution:
Using Comparison Lemma (Khalil 2002), for all \(0<\beta <\infty \),
This proves the upper bound. The lower bound is a direct consequence from monotonicity proved in part 1.
(c) For any \(\beta < \infty \)
For any \(\epsilon >0\), choose a \(\beta > \beta ^*(\epsilon )\) using Eq. (6) and observe that,
This proves the third statement.
Appendix 2: Optimized graphs
1.1 Appendix 2.1: K-3 graph
A fully connected 3-noded graph is optimized for 4 data states. The energy of the graph is modeled using Ising model Eq. (9) with \(|H|\le 1\) and \(|J|\le 1\). The optimized parameters using the (1) Minimization of Negative Log-likelihood, and (2) PEPDAS method are presented in Fig. 4
a Training data set of states with green representing a ‘+1’ state and red representing a ‘-1’ state. b Optimized graph using minimization of Negative Log-likelhood at \(\beta =1\) c Optimized graph using PEPDAS method. The field terms are mentioned in blue color and interaction terms are mentioned in red color
1.2 Appendix 2.2: Peterson graph
A Peterson graph is first optimized for upto 3 user prescribed data states using PEPDAS method. Then it is optimized for 3 ground states using PEPGSM method. The energy of the graph is modeled using Ising model Eq. (9) with \(|H|<1\) and \(|J|<1\). The optimized graphs are presented in Fig. 5 and their respective Negative log likelhood is presented in Fig. 6.
Rights and permissions
Springer Nature or its licensor holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Srivastava, S., Sundararaghavan, V. Bandgap optimization in combinatorial graphs with tailored ground states: application in quantum annealing. Optim Eng 24, 1931–1949 (2023). https://doi.org/10.1007/s11081-022-09758-9
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11081-022-09758-9