Poster
in
Workshop: MATH-AI: The 4th Workshop on Mathematical Reasoning and AI
Library Learning Doesn’t: The Curious Case of the Single-Use “Library”
Ian Berlot-Attwell · Frank Rudzicz · Xujie Si
Keywords: [ library learning ] [ LEGO-Prover ] [ Self-Correction ] [ tool creation ] [ TroVE ] [ ablation ] [ self-consistency ]
Advances in Large Language Models (LLMs) have spurred a wave of LLM library learning systems for mathematical reasoning. These systems aim to learn a reusable library of tools, such as formal Isabelle lemmas or Python programs that are tailored to a family of tasks. Many of these systems are inspired by the human structuring of knowledge into reusable and extendable concepts, but do current methods actually learn reusable libraries of tools? We study two library learning systems for mathematics which both reported increased accuracy: LEGO-Prover and TroVE. We find that function reuse is extremely infrequent on miniF2F and MATH. Our followup ablation experiments suggest that, rather than reuse, self-correction and self-consistency are the primary drivers of the observed performance gains. Our code and data are available at https://github.com/ikb-a/curious-case.