Boolean Satisfiability via Imitation Learning
Overview
Overall Novelty Assessment
ImitSAT proposes an imitation learning approach to CDCL branching that learns from expert KeyTrace demonstrations—compact sequences of surviving decisions extracted from full solver runs. The paper resides in the 'Imitation Learning and Supervised Approaches' leaf under 'Machine Learning-Based Branching Heuristics', where it appears as the sole occupant among the 50 papers surveyed. This isolation suggests the leaf represents an emerging or underexplored direction within the broader landscape of learned branching policies, contrasting with more populated sibling categories like 'Graph Neural Network-Based RL Branching' containing four papers.
The taxonomy reveals substantial activity in neighboring reinforcement learning branches, particularly GNN-based RL methods that learn through trial-and-error exploration rather than expert demonstration. ImitSAT diverges from these by avoiding exploration entirely, instead extracting supervision directly from CDCL traces. The broader 'Machine Learning-Based Branching Heuristics' category also includes 'Neural Heuristics Without Explicit RL' (three papers) and 'Machine Learning for Solver Component Selection' (two papers), indicating that while neural approaches to branching are well-represented, the specific combination of imitation learning with decision-level supervision from KeyTrace appears less crowded.
Among the nineteen candidates examined through limited semantic search, none clearly refute any of the three core contributions. The 'ImitSAT branching policy' contribution examined three candidates with zero refutations, 'KeyTrace extraction' examined ten candidates with zero refutations, and 'autoregressive sequence modeling formulation' examined six candidates with zero refutations. This absence of overlapping prior work across all contributions suggests that within the examined scope, the combination of KeyTrace-based supervision, prefix-conditioned autoregressive modeling, and direct propagation reduction represents a distinct approach not previously documented in the top-nineteen semantic matches.
The analysis reflects a focused literature search rather than exhaustive coverage, examining nineteen candidates from semantic similarity and citation expansion. While no refutations emerged within this scope, the singleton status in the taxonomy leaf and the limited search scale mean the assessment captures novelty relative to closely related work but cannot rule out relevant contributions outside the examined set. The approach's distinctiveness appears strongest in its KeyTrace formulation and decision-level supervision mechanism.
Taxonomy
Research Landscape Overview
Claimed Contributions
The authors introduce ImitSAT, the first branching policy for CDCL solvers that uses imitation learning instead of reinforcement learning. It learns from expert KeyTrace demonstrations to predict branching decisions, providing dense decision-level supervision and directly reducing propagations.
The authors develop a method to collapse full CDCL solver runs into KeyTrace sequences that contain only surviving decisions by systematically removing backtracked detours. These sequences serve as clean, conflict-free training targets that align naturally with prefix-conditioned autoregressive modeling.
The authors formulate CDCL branching as a prefix-conditioned autoregressive sequence prediction problem. They serialize the CNF formula and KeyTrace prefix into a single sequence and train a Transformer-based model to predict the next branching decision autoregressively.
Core Task Comparisons
Comparisons with papers in the same taxonomy category
Contribution Analysis
Detailed comparisons for each claimed contribution
ImitSAT: Imitation learning-based branching policy for CDCL solvers
The authors introduce ImitSAT, the first branching policy for CDCL solvers that uses imitation learning instead of reinforcement learning. It learns from expert KeyTrace demonstrations to predict branching decisions, providing dense decision-level supervision and directly reducing propagations.
KeyTrace: Compact expert trace extraction from CDCL runs
The authors develop a method to collapse full CDCL solver runs into KeyTrace sequences that contain only surviving decisions by systematically removing backtracked detours. These sequences serve as clean, conflict-free training targets that align naturally with prefix-conditioned autoregressive modeling.
[51] Proof-Driven Clause Learning in Neural Network Verification PDF
[52] Conflict-Driven XOR-Clause Learning (extended version) PDF
[53] Proofs in conflict-driven theory combination PDF
[54] Understanding the Relative Strength of QBF CDCL Solvers and QBF Resolution PDF
[55] Proof Generation for CDCL Solvers Using Gauss-Jordan Elimination PDF
[56] Proofs for Incremental SAT with Inprocessing. PDF
[57] Trimming while checking clausal proofs PDF
[58] Verifying refutations with extended resolution PDF
[59] On the power of clause-learning SAT solvers as resolution engines PDF
[60] Mechanical verification of SAT refutations with extended resolution PDF
Autoregressive sequence modeling formulation for CDCL branching
The authors formulate CDCL branching as a prefix-conditioned autoregressive sequence prediction problem. They serialize the CNF formula and KeyTrace prefix into a single sequence and train a Transformer-based model to predict the next branching decision autoregressively.