Get AI summaries of any video or article — Sign up free
Alpha Geometry thumbnail

Alpha Geometry

6 min read

Based on West Coast Machine Learning's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

Alpha Geometry solves a subset of plane-geometry proof tasks by alternating between a symbolic theorem prover (closure computation) and a language model that proposes auxiliary constructions.

Briefing

Alpha Geometry is a system that solves a difficult subset of geometry proofs—specifically “plane geometry” problems—without human demonstrations, by combining a symbolic deduction engine with a small GPT-style language model. The key breakthrough is not that the model performs geometric reasoning end-to-end; instead, it learns to propose promising auxiliary constructions (new points, lines, circles) while the symbolic engine deterministically derives all consequences and checks whether the target statement follows. That division of labor matters because geometry proof datasets are scarce, while the symbolic engine can verify correctness instantly once the right construction is proposed.

The symbolic side works like an open-book theorem prover for geometry. It stores a database of geometric rules expressed as equality-style propositions (for example, angle relationships in triangles, parallel-line angle facts, and other standard inference patterns). Given an initial set of premises—such as “AB is parallel to CD,” “ABC is a triangle,” and the goal statement—this engine repeatedly matches rules to the current facts, adds newly implied facts, and computes the closure: the full set of statements derivable from the premises. For many simpler geometry tasks, this closure computation can reach the answer in seconds.

The hard cases are those where the proof requires auxiliary construction: adding specific new objects like perpendicular feet, angle bisectors, midpoints, or circles inside/outside a configuration. A central example discussed is proving equal areas of two triangles that share a base under a parallel-lines condition. The symbolic engine alone can’t “invent” the needed perpendicular constructions; it can only propagate consequences of what’s already present. The combinatorial explosion is the problem: there are many possible points and constructions one could add, and generic heuristics for choosing them don’t scale.

To address data scarcity, the system generates its own training proofs. It first samples random premise sets, then uses the symbolic engine to grow a directed graph of derivable facts. Each node corresponds to a statement that becomes provable from earlier nodes; by tracing backwards from a randomly chosen node, the system extracts a synthetic proof: a minimal set of premises plus the sequence of deduction steps. From an initial pool of hundreds of millions of synthetic proofs, deduplication leaves about 100 million, and a further backward “hardening” step removes premises that the final conclusion doesn’t depend on—forcing the proof to reintroduce those missing objects via auxiliary construction.

The language model is trained on these synthetic proof strings using next-token prediction from scratch (not pretrained weights), with a custom tokenizer tailored to geometry facts and construction commands. During inference, it does not see the symbolic engine’s entire closure. Instead, it alternates with the symbolic engine: given the current premises and goal, the model proposes one auxiliary construction; the symbolic engine then recomputes closure to see whether the goal is reached. Because the model’s suggestions can be wrong, the system uses beam search—keeping hundreds of candidate construction paths in parallel (reported as beam width 512) and limiting how many constructions each path can add (max depth 16). Only the top candidates survive to the next iteration.

On the benchmark of 30 International Mathematical Olympiad-style geometry problems, the system solves 25, beating prior computer baselines (previous bests were around 10/30) and landing near the average performance of top human gold medalists over the same set. Ablations show performance drops when training data is reduced, when fine-tuning on auxiliary-construction-heavy proofs is removed, or when beam width is lowered—suggesting the search strategy and the synthetic proof curriculum are both crucial. The result is a practical, human-readable proof trace: each step corresponds to an explicit construction, followed by symbolic rule applications that justify the derived facts.

Cornell Notes

Alpha Geometry combines a symbolic deduction engine with a GPT-style language model to solve plane-geometry proof problems. The symbolic engine stores geometric rules and, given premises, computes the full deductive closure to verify whether the target statement follows. The language model’s job is narrower: propose one auxiliary construction at a time (new points, perpendiculars, midpoints, circles, etc.) that the symbolic engine can then use to derive consequences. Because real geometry proofs are scarce, training uses synthetic proofs generated by sampling premises and growing proof graphs with the symbolic engine, then “hardening” proofs so they require auxiliary construction. Inference alternates model suggestions with symbolic verification using beam search (beam width 512, max depth 16), yielding 25/30 solutions on the referenced Olympiad geometry benchmark—close to top human averages.

Why can’t the symbolic deduction engine solve the hardest geometry problems by itself?

It can only derive consequences of what’s already in the premise set. Many Olympiad geometry goals require adding specific auxiliary objects—like perpendicular feet to compute heights, angle bisectors, midpoints, or circles positioned relative to the figure. Without those constructions, the engine’s closure never reaches the needed relationships. The system’s example of equal triangle areas under parallel lines illustrates this: proving equal areas may require introducing perpendiculars from vertices to the shared base, which means adding new points (not just reusing existing ones).

How does the system generate training data when geometry proof datasets are too small?

It synthesizes proofs using the symbolic engine. First, it samples random theorem premises from the space of possible geometric facts. Then it repeatedly applies deduction rules to grow a directed graph of provable statements. By selecting a node (a statement) and tracing backward through the graph, it extracts a proof consisting of the minimal premises needed and the deduction steps used. This produces a large synthetic proof corpus without relying on human-written demonstrations.

What does “hardening” proofs mean, and why does it force auxiliary construction?

Some synthetic proofs may reach the conclusion without ever mentioning certain auxiliary points. The system removes premises that the conclusion doesn’t depend on, then regenerates the proof so that those missing objects must be reintroduced through auxiliary construction. This backward process turns easier proofs into ones that explicitly require adding new points/objects—matching the inference-time role of the language model (suggesting constructions).

What exactly does the language model do during inference, and what does it not do?

During inference, the model suggests one new construction at a time—e.g., “construct point D midpoint of BC” or “construct a point such that … is a parallelogram.” It does not perform the full deductive reasoning itself. The symbolic engine performs the heavy lifting: after each suggested construction, it recomputes closure from the updated premises and checks whether the goal statement is now derivable. A crucial detail is that the language model does not receive the symbolic engine’s entire closure as input in the loop.

How does beam search fit into the proof search process?

Because the model’s next construction proposals can be wrong, beam search keeps many candidate construction sequences alive. With beam width 512, the system explores hundreds of alternative auxiliary constructions in parallel at each iteration, then prunes back to the top candidates based on the model’s token probabilities. It repeats this alternation: propose constructions → symbolic closure check → prune → propose again, up to a max depth of 16 constructions per path.

Why are the reported results meaningful even if the method uses brute-force search?

The symbolic engine guarantees correctness once the right construction is proposed, so the search is about finding useful constructions in a huge space. The system’s performance (25/30 on the benchmark) indicates that the model plus search strategy can repeatedly land on constructions that unlock the symbolic closure needed for the goal. Ablations show that reducing training data or narrowing beam width reduces solved problems, reinforcing that both the synthetic proof curriculum and the breadth of search are central to the outcome.

Review Questions

  1. What is the division of labor between the symbolic deduction engine and the language model, and how does that affect what each component is responsible for during inference?
  2. Describe how synthetic proofs are generated from premise sampling and proof graphs, and explain how “hardening” changes which steps require auxiliary construction.
  3. Why does beam search help in this setting, and what do beam width and max depth control in the construction sequence exploration?

Key Points

  1. 1

    Alpha Geometry solves a subset of plane-geometry proof tasks by alternating between a symbolic theorem prover (closure computation) and a language model that proposes auxiliary constructions.

  2. 2

    The symbolic deduction engine can only derive consequences of given premises; it cannot invent the new points/circles/perpendiculars that many Olympiad proofs require.

  3. 3

    Training data is synthesized: random premise sets are expanded into proof graphs using the symbolic engine, then backward-traced to extract proof sequences.

  4. 4

    A “hardening” step removes premises not needed for the conclusion, forcing later proofs to reintroduce those objects via auxiliary construction.

  5. 5

    Inference uses beam search (beam width 512, max depth 16) to explore many candidate construction paths in parallel, pruning by the language model’s next-token probabilities.

  6. 6

    The language model is trained on full proof strings but, during inference, it does not receive the symbolic engine’s full closure; it only outputs the next construction suggestion.

  7. 7

    On the referenced 30-problem Olympiad geometry benchmark, the system solves 25, outperforming earlier computer baselines and approaching top human averages.

Highlights

The symbolic engine computes the deductive closure from premises; the language model’s job is mainly to suggest which auxiliary construction to add next.
Synthetic proofs are built by sampling premises, growing a proof graph with symbolic rules, and extracting backward-traced proof sequences.
Beam search explores hundreds of construction candidates per iteration (reported beam width 512) while the symbolic engine verifies each candidate’s closure.
The system’s reported benchmark performance is 25/30, close to the average gold-medalist score on the same set.
Proof traces remain human-readable because each step corresponds to explicit constructions followed by symbolic rule applications.

Topics

  • Auxiliary Construction
  • Symbolic Deduction
  • Synthetic Proof Generation
  • Beam Search
  • Plane Geometry Proofs

Mentioned

  • DeepMind
  • LLM
  • GPT
  • MLP
  • Ru
  • GPT-4
  • TST
  • USAMO
  • AIME
  • IMO
  • AR
  • GP4