A tractable Walsh analysis of SAT and its implications for genetic algorithms
AAAI '98/IAAI '98 Proceedings of the fifteenth national/tenth conference on Artificial intelligence/Innovative applications of artificial intelligence
Introduction to algorithms
Genetic Algorithms in Search, Optimization and Machine Learning
Genetic Algorithms in Search, Optimization and Machine Learning
SIAM Review
Discrete Applied Mathematics
Improved Design Debugging Using Maximum Satisfiability
FMCAD '07 Proceedings of the Formal Methods in Computer Aided Design
Algorithms for maximum satisfiability using unsatisfiable cores
Proceedings of the conference on Design, automation and test in Europe
Backbone guided local search for maximum satisfiability
IJCAI'03 Proceedings of the 18th international joint conference on Artificial intelligence
Integrating systematic and local search paradigms: a new strategy for MaxSAT
IJCAI'09 Proceedings of the 21st international jont conference on Artifical intelligence
Iterated robust tabu search for MAX-SAT
AI'03 Proceedings of the 16th Canadian society for computational studies of intelligence conference on Advances in artificial intelligence
Combining adaptive noise and look-ahead in local search for SAT
SAT'07 Proceedings of the 10th international conference on Theory and applications of satisfiability testing
Learning the large-scale structure of the MAX-SAT landscape using populations
IEEE Transactions on Evolutionary Computation
A new method for solving hard satisfiability problems
AAAI'92 Proceedings of the tenth national conference on Artificial intelligence
Automated design debugging with maximum satisfiability
IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems
UBCSAT: an implementation and experimentation environment for SLS algorithms for SAT and MAX-SAT
SAT'04 Proceedings of the 7th international conference on Theory and Applications of Satisfiability Testing
Constant time steepest descent local search with lookahead for NK-landscapes and MAX-kSAT
Proceedings of the 14th annual conference on Genetic and evolutionary computation
Hi-index | 0.00 |
By converting the MAXSAT problem to Walsh polynomials, we can efficiently and exactly compute the hyperplane averages of fixed order k. We use this fact to construct initial solutions based on variable configurations that maximize the sampling of hyperplanes with good average evaluations. The Walsh coefficients can also be used to implement a constant time neighborhood update which is integral to a fast next descent local search for MAXSAT (and for all bounded pseudo-Boolean optimization problems.) We evaluate the effect of initializing local search with hyperplane averages on both the first local optima found by the search and the final solutions found after a fixed number of bit flips. Hyperplane initialization not only provides better evaluations, but also finds local optima closer to the globally optimal solution in fewer bit flips than search initialized with random solutions. A next descent search initialized with hyperplane averages is able to outperform several state-of-the art stochastic local search algorithms on both random and industrial instances of MAXSAT.