Deep Network-based Automated Theorem Prover

This is a review of the paper — “Deep network guided proof search”. arXiv preprint arXiv:1701.06972, as given in reference [1].

— Automated first-order theorem provers (ATP) can be used to prove mathematical theorems and these can also be used in applications such as program analysis, theory exploration, and reasoning.

— Higher-order logic, Naive Bayes, techniques are used in prior works to improve search.

— This article provides deep learning-based guidance in proofs of theorems provers inside an ATP.

— Modern ATP does not perform well in general when many facts are present.

— AI can help select lemmas to be tested on ATP for better results and time complexity.

— Mentioned in paper that ATP can find proofs upto 20 human steps [in Mizar Mathematical Library proof sets wherein proofs upto 100 steps are provided]

— Does not rely on Tableau Methods for ATP.

— Experiments on Mizar Mathematical Library converted to First Order Logic (FOL).

— Learns syntax tree of proof clause.

— First-order logic theorem prover E preprocesses the FOL into CNF to do proof search.

— CNF form if the formula is the conjunction of clauses and each clause is a disjunction of literals. A literal is a predicate term.

— Clause Selection Heuristic is the main thing driving this work.

— Unprocessed clauses: those clauses that are not processed.

— Once a clause is viewed its consequences are generated, each of these is added to the unprocessed clause list.

— The viewed clause is shifted to the processed clause list.

— The algorithm used to view clauses determines how long would it take to find the proof.

— ATP uses a heuristic to select a clause search heuristic and sets other parameters and decides the next step.

— As specified in the paper the best-performing heuristics were selected by targeting various criteria in a round-robin manner.

— Other heuristics can be FIFO on shortest unprocessed clause.

— Hybrid approach combining the above two heuristics produced better results.

— Here, a machine learning-based optimal clause selection heuristic given time and computing powers, is proposed in this paper.

— Mizar Mathematical Library (MML) is a collection of formularized proofs

— MML has been translated to FOL (the number goes up to 57,882 Mizar theorems), and this makes it best choice for experiments in CS and ML.

— The FOL proofs used by authors are based out of work by Kaliszyk & Urban (2015).

— Around 91000 proofs were used to train and learn the ATP.

— Network architecture was taken by premise selection work by Alemi et al. (2016).

— Model takes — (a) negated conjecture in CNF form, and, (b) unprocessed clause in CNF form.

— Convert the above two in fixed-sized vectors using an embedding network and combine the two using a combining network and generate a score to select a clause. The same architecture but different weights are used for this. This was tried using (a) a CNN (b) WaveNet, and (c) a RNN.

— Loss is computed towards 1 if the clause selected is used in proof, else it is set to be near 0.

— Combiner use one 1024-unit hidden layer.

— The input layer converts each symbol to an embedding vector of dimension 1024.

— Ranks are produced for the clauses in addition to the given clause selection methods mentioned above.

— Rankings are produced for clauses to be used in ATP.

— Harder and deeper theorems were proved with improvements in strategy with this approach.

— Due to the time taken by neural networks, a hybrid approach is better, for example, one where a combinatorial search follows a deep neural-based layer with a hand-crafted clause selection phase.

— However, the computational time of the method proposed here is high as compared to prior works.

— Experimental proofs are provided by authors that the two-phase process reduces the time to prove a theorem by reducing the steps and hence increasing the number of theorems proved in the given time.

— Results increase the number of theorems with ATP proofs.

References

Loos, S., Irving, G., Szegedy, C., & Kaliszyk, C. (2017). Deep network guided proof search. arXiv preprint arXiv:1701.06972.

Alexander A Alemi, Francois Chollet, Geoffrey Irving, Niklas Een, Christian Szegedy, and Josef Urban. Deepmath-deep sequence models for premise selection. In Advances in Neural Information Processing Systems, pp. 2235–2243, 2016.

Cezary Kaliszyk and Josef Urban. MizAR 40 for Mizar 40. J. Autom. Reasoning, 55(3): 245–256, 2015a. doi: 10.1007/s10817–015–9330–8. URL http://dx.doi.org/10.1007/ s10817–015–9330–8.

Published by Nidhika

Hi, Apart from profession, I have inherent interest in writing especially about Global Issues of Concern, fiction blogs, poems, stories, doing painting, cooking, photography, music to mention a few! And most important on this website you can find my suggestions to latest problems, views and ideas, my poems, stories, novels, some comments, proposals, blogs, personal experiences and occasionally very short glimpses of my research work as well.

Leave a comment