Accepted at IJCAI 2026

A Foundation Model for
Zero-Shot Logical Rule Induction

Yin Jun Phua
Institute of Science Tokyo

Trained once on synthetic Boolean formulas, NRI induces interpretable DNF rules zero-shot on any tabular task. No retraining, no fine-tuning, no per-task weights.

Zero-shot
14 UCI tasks, one checkpoint
+17 pts
over RIPPER / DT at 30% label noise
>92%
accuracy with 32 spurious distractors
~12 ms
inference at $N{=}512$ variables
Example rule plas>med ∧ pres>med ∧ skin>med ∧ age>med 68.0% UCI diabetes · produced in a single forward pass, no per-task training

Results

Zero-shot transfer to UCI

NRI is trained only on random synthetic Boolean formulas with $N \in [6, 12]$ variables. It is then applied directly to 14 UCI tabular tasks, with no retraining and no per-task fitting. 12 of those 14 tasks have more features than the training range, so most of this is genuine out-of-distribution transfer. Baselines are trained per dataset; NRI uses a single checkpoint everywhere.

Dataset $N$ EBM (per-task) RIPPER (per-task) NRI (zero-shot)
diabetes866.966.868.0
breast-cancer993.088.388.3
vote3287.491.688.3
mushroom11699.498.587.8
kr-vs-kp7392.589.972.3
credit4682.283.171.5
nursery2791.186.271.3
spambase5791.185.771.9

5-fold stratified CV accuracy (%). NRI reaches 69.7% on average across all 14 tasks, 13 points behind EBM (which is trained per dataset). NRI is strongest on the two in-distribution tasks ($N \leq 12$) and beats every per-task baseline on diabetes. See the paper, Table 1, for the full comparison against XGBoost, LightGBM, EBM, RIPPER, RuleFit, FIGS, DT, and N-DNF.

Robustness to label noise

This is where NRI most clearly outperforms classical rule learners. Symbolic methods fit their training data exactly and pay for it the moment labels are noisy. NRI's statistical encoding acts as implicit regularization, so accuracy degrades smoothly instead of falling off a cliff.

Label noise robustness curve
NRI (blue) maintains accuracy from 92.3% at 0% noise to 87.4% at 30% noise. RIPPER and DT drop from ~99% to ~70% over the same range. NRI overtakes both around 15% noise.

Robustness to spurious variables

Real datasets are full of features that correlate with the label but should not appear in the rule. We append $D$ distractor features that correlate with the label at strength $\rho$ while having no causal relationship. NRI has to learn to ignore them.

Spurious variable robustness heatmap
Accuracy stays above 92% across the entire $(D, \rho)$ grid, including 32 highly-correlated distractors at $\rho{=}0.9$. The counterfactual-necessity loss and the spurious-environment training regime are what make this work.

Rule complexity scaling

We measure exact logical equivalence between predicted and ground-truth rules across rule complexity dimensions $K$ (number of clauses) and $L$ (literals per clause).

Rule complexity scaling heatmap
Exact logical match degrades with complexity, dominated by $K$. Recovery is 99.5% for $K{=}1, L{=}1$ and 24.0% for the hardest $K{=}4, L{=}3$ regime. Prediction accuracy stays in the 85–100% range even when the recovered rule is not literally identical, indicating semantically equivalent or near-equivalent solutions.

What NRI actually outputs

The point of this work is not a number on a leaderboard, it is the rule itself. Each card below shows the DNF that NRI produced for a UCI task in a single forward pass, with no fine-tuning. The rules are short, readable, and align with what a domain expert would write down. This is what zero-shot interpretable induction looks like in practice.

diabetes 68.0%
plas>med ∧ pres>med ∧ skin>med ∧ age>med
Elevated plasma glucose, blood pressure, skin thickness, and older age — a clinically plausible summary of the binarized Pima dataset.
breast-cancer-wisconsin 88.3%
Clump_Thickness>med ∧ Cell_Size_Uniformity>med ∧ Cell_Shape_Uniformity>med ∧ Bare_Nuclei>med
Larger and more irregular cells with prominent bare nuclei indicate malignancy — standard cytological features used by pathologists.
tic-tac-toe 56.6%
¬middle-left_x ∧ ¬middle-middle_o ∧ middle-middle_x
A short rule centered on board control: when the centre is X (and not O) and the middle-left is not X, the model leans toward an X-positive outcome.
vote 88.3%
¬physician-fee-freeze_n ∧ ¬education-spending_n ∧ ¬crime_n ∧ ¬duty-free-exports_y
Recovers a familiar political signature from US congressional voting records, dominated by the physician-fee-freeze indicator.
spambase 71.9%
¬word_freq_hp>med ∧ ¬word_freq_hpl>med ∧ ¬word_freq_george>med ∧ ¬word_freq_meeting>med
Absence of HP-internal vocabulary (the dataset comes from HP employee mail) is a strong indicator of spam, exactly the artefact known to dominate this benchmark.
mushroom 87.8%
(¬odor_n ∧ ¬gill-spacing_w ∧ ¬stalk-root_e ∧ ¬spore-print-color_k) ∨ (¬odor_n ∧ ¬stalk-root_e ∧ ¬stalk-surface-above-ring_s ∧ ¬spore-print-color_k) ∨ ...
A 5-clause DNF centered on odor (the canonical mushroom-edibility predictor) combined with stalk and spore features.
kr-vs-kp (chess endgame) 72.3%
(bkxwp_f ∧ ¬bkxwp_t ∧ bxqsq_f ∧ ¬bxqsq_t) ∨ ... 5 more clauses
A 6-clause rule that exhausts NRI's training-time clause budget; this is one of the cases where the task genuinely needs more clauses than NRI was trained for.
german-credit 59.8%
checking_status_no_checking ∧ (¬purpose_domestic_appliance ∧ ¬employment_1≤X<4 ∧ ¬property_no_known) ∨ ...
Credit risk centered on the checking-account indicator, the strongest single feature in this benchmark.

All rules above are produced by the same checkpoint, with no per-task fine-tuning. The full set of 14 rules (including the multi-class breakdowns for car and nursery) is in Appendix B.

Why this matters

Interpretability has always come at the cost of generality

For thirty years, practitioners faced a forced choice. Want a model you can read? Use a decision tree or an ILP system, and accept that you will retrain it from scratch on every dataset, that it will overfit noisy labels, and that it will not transfer anything it learned on one problem to the next. Want a model that scales? Reach for gradient boosting or a deep network and explain the predictions with post-hoc tools you cannot fully trust.

NRI removes that trade-off. The model is general (one checkpoint, many domains), the output is symbolic (a DNF formula you can paste into a slide), and the reasoning is grounded in the data rather than in language priors.

What this enables

Instant hypothesis generation on new datasets

Run NRI on a fresh tabular dataset and get a candidate rule in a single forward pass. Domain experts can then accept, refute, or refine it, treating the model as a fast first-pass analyst rather than a black box that needs to be re-trained, re-tuned, and re-validated for every new question.

Auditable predictions in regulated domains

Because NRI emits an explicit DNF, the rule itself is the audit trail. There is no need for SHAP, LIME, or counterfactual explainers: the rule that produced the prediction is the explanation, and it is the same rule on every example.

Robust learning with messy labels

Statistical encoding gives NRI implicit regularization. Where symbolic rule learners memorize noise, NRI's accuracy degrades smoothly, making it usable in settings where labels come from human raters, weak supervision, or imperfect logging pipelines.

Resistance to spurious shortcuts

NRI is trained against environment-dependent distractors, so it learns to prefer features whose predictive power survives a change in environment. This is the same property that domain-generalization methods aim for, achieved here as a byproduct of how the model is pretrained.

A foundation for symbolic reasoning at scale

Vision and language each have foundation models. Symbolic reasoning does not. NRI is a step toward that: a single model whose pretraining objective is induction itself, ready to be extended to multi-valued attributes, continuous variables via discretization, and eventually first-order relational rules.

How NRI works

The whole architecture is built around one principle: never look at variable names. Anything tied to specific predicates would prevent transfer to new domains, so NRI replaces identities with statistical signatures and replaces ordered generation with parallel slots.

NRI pipeline diagram
NRI architecture. An episode $(X, Y)$ is converted into per-literal statistics $\phi$. Cross-attention over examples produces literal embeddings, and a slot-based decoder synthesizes $K$ candidate clauses in parallel through learned literal gates $z$ and clause gates $w$. The rule is evaluated with the product T-norm for end-to-end training, then discretized into an interpretable DNF rule.

Statistical encoding

Each literal $l_j$ is described by 18 identity-free statistics: class-conditional truth rates $P(l_j \mid y{=}1)$ and $P(l_j \mid y{=}0)$, marginal entropy $\mathcal{H}(l_j)$, polarity, observation rates, and mean co-occurrence with other literals. Two literals with the same statistics are interchangeable from NRI's point of view, which is exactly the invariance we want for transfer.

Parallel slot decoder

An autoregressive decoder would force a left-to-right order on a disjunction $C_1 \lor \ldots \lor C_K$, which has none. Instead, $K$ learned slot queries decode all clauses in parallel under FiLM modulation, preserving permutation invariance while letting different slots specialize.

T-norm execution

Literal gates $z_{k,j}$ and clause gates $w_k$ define a soft DNF that is executed under the product T-norm. Conjunctions become products, disjunctions become probabilistic OR, and the entire pipeline from raw episode to predicted label is differentiable. Discretizing the gates at inference yields the symbolic rule.

Synthetic-only training

NRI is trained on millions of randomly generated DNFs with up to $K{=}6$ clauses and $L{=}4$ literals. Each episode also injects spurious environment features with opposite correlations across two halves of the data, forcing the model to rely on causal signal rather than marginal correlation. No real-world supervision is used.

BibTeX

TBD — will be updated once the official IJCAI 2026 proceedings citation is available.