Download PDFOpen PDF in browser

Deep Reinforcement Learning for Synthesizing Functions in Higher-Order Logic

19 pagesPublished: May 27, 2020

Abstract

The paper describes a deep reinforcement learning framework based on self-supervised learning within the proof assistant HOL4. A close interaction between the machine learning modules and the HOL4 library is achieved by the choice of tree neural networks (TNNs) as machine learning models and the internal use of HOL4 terms to represent tree structures of TNNs. Recursive improvement is possible when a task is expressed as a search problem. In this case, a Monte Carlo Tree Search (MCTS) algorithm guided by a TNN can be used to explore the search space and produce better examples for training the next TNN. As an illustration, term synthesis tasks on combinators and Diophantine equations are specified and learned. We achieve a success rate of 65% on combinator synthesis problems outperforming state-of-the-art ATPs run with their best general set of strategies. We set a precedent for statistically guided synthesis of Diophantine equations by solving 78.5% of the generated test problems.

Keyphrases: combinators, Diophantine equations, HOL, Reinforcement Learning, tree neural networks

In: Elvira Albert and Laura Kovács (editors). LPAR23. LPAR-23: 23rd International Conference on Logic for Programming, Artificial Intelligence and Reasoning, vol 73, pages 230--248

Links:
BibTeX entry
@inproceedings{LPAR23:Deep_Reinforcement_Learning_for,
  author    = {Thibault Gauthier},
  title     = {Deep Reinforcement Learning for Synthesizing Functions in Higher-Order  Logic},
  booktitle = {LPAR23. LPAR-23: 23rd International Conference on Logic for Programming, Artificial Intelligence and Reasoning},
  editor    = {Elvira Albert and Laura Kovacs},
  series    = {EPiC Series in Computing},
  volume    = {73},
  pages     = {230--248},
  year      = {2020},
  publisher = {EasyChair},
  bibsource = {EasyChair, https://easychair.org},
  issn      = {2398-7340},
  url       = {https://easychair.org/publications/paper/Tctp},
  doi       = {10.29007/7jmg}}
Download PDFOpen PDF in browser