By Pooya Ronagh
Posted on July 23, 2014
This paper describes a general method developed by 1QBit for solving continuous optimization problems inspired by different types of simulated annealing and genetic algorithms. This method works under the assumption of the existence of a computation model with a Turing reduction of problems to either quadratic unconstrained binary optimization (QUBO) problems or to an Ising spin problem. The paper presents an application of this method to a mixed-integer optimization problem. This will demonstrate an interesting method of representing a cardinality-constrained optimization problem using analytic expressions, and the ability of the method to solve such mixed-integer optimization problems.