Alternative Research Movements Lag Far Behind
. Traditional large language models (LLMs) like GPT-3 or GPT-4 demonstrate impressive pattern recognition and knowledge recall, but they often lack epistemic rigor: they can produce plausible-sounding but incorrect statements (“hallucinations”), cannot verify their answers, and offer little transparency into their decision process. This stands in contrast to the standard set by Curt Doolittle’s Natural Law framework – which emphasizes performative truth (truth as demonstrable and liable claims), operational coherence, decidability, and testifiability in knowledge. In essence, Doolittle’s approach demands that every proposition be reducible to a series of testable operations, yielding conclusions that can be validated or falsified with evidence
. Achieving such reliability and interpretability in AI systems is a grand challenge. In response, a number of recent global initiatives – from academic projects to industry research labs – are targeting real-world reasoning capability with a focus on correctness, interpretability, and rigorous logic beyond what large-scale neural networks alone can offer. This report surveys these developments and compares how they align with or diverge from Doolittle’s criteria for truthful, coherent reasoning.
. In this view, an assertion is only true insofar as it can be operationally demonstrated and survives attempts at falsification, much like a scientific hypothesis or a legal claim tested in court. Key pillars of this framework include: (1) Operational Definitions – concepts must be defined by observable, repeatable operations, preventing ambiguity; (2) Decidability – any well-formed question has a finite procedure to determine its truth or falsehood (no endlessly indeterminate answers); (3) Testifiability – claims carry an onus of evidence and liability, meaning the “speaker” (or AI system) should be held accountable to produce supporting proof or face refutation. Doolittle’s approach is essentially an attempt to bring the scientific-method level of rigor to all propositions, ensuring no claim is accepted without demonstrable coherence with reality
.
. By unifying these, researchers aim for AI that can learn from data yet still deduce with logical precision and provide interpretable, rule-based explanations.
. For example, IBM Research introduced Logical Neural Networks (LNNs) – a framework that embeds classical Boolean logic within neural network architectures. In an LNN, each neuron effectively behaves like a differentiable logic gate, with truth values and learnable parameters coexisting
. This design lets the system learn from data via gradient descent while guaranteeing logical consistency (no rule contradictions) and producing rules that are precisely interpretable (the learned logic can be read by humans)
. In a 2022 study, IBM showed that LNN-based models could learn first-order logic rules from noisy data, achieving accuracy on par with purely neural approaches while yielding human-readable rules as output
. This directly speaks to decidability and testifiability: the learned model can be audited like a set of logical statements, and each inference is effectively a proof step that can be checked.
. A 2025 survey by Liang et al. outlines many such advances, including logic-aware Transformers (language models augmented with logic constraints) and LLM-based symbolic planners, all aimed at bridging symbolic logic and neural generative reasoning
. The overarching goal is a unified framework where an AI’s knowledge is stored in explicit forms (graphs, logic rules) that are continuously updated by neural learning – so the system can both learn from examples and reason over facts in a verifiable way. This trend is well-aligned with Doolittle’s emphasis on coherence and decidability: the symbolic part provides a rigorous backbone that ensures the AI’s conclusions follow validly from premises (no free-association leaps), and the neural part grounds those symbols in real-world data.
; and Microsoft’s Probabilistic Logic initiatives where Bayesian networks (which handle uncertainty in a principled way) are used on top of transformer models to decide if an answer logically follows from given evidence. By injecting symbolic constraints, these systems naturally produce outputs that are more consistent, interpretable, and testable than a standard neural net. For instance, if a rule says “X implies Y” and the network predicts X, it will automatically include Y in its reasoning – such traceable inference can be checked step-by-step, much like how operational grammar in Doolittle’s method would break down an argument into constituent operations.
. More recently, Ospanov et al. (2023) introduced APOLLO, a pipeline that marries an LLM’s intuitive reasoning with the precise feedback of the Lean theorem prover
. In APOLLO, the language model generates a candidate proof for a theorem; if the proof fails, the system does not simply guess again at random. Instead, Lean (a formal verification system) checks the proof and pinpoints the error (a specific step that’s wrong or a sub-lemma that couldn’t be solved)
. APOLLO then invokes specialized “repair” agents: one module fixes syntax errors, another breaks the problem down around the failing sub-lemma, others call automated solvers for trivial steps, and then the LLM is prompted in a targeted way to fill in the remaining gaps
. This iterative loop continues until a complete proof is found that the Lean checker formally verifies as correct. The result was a new state-of-the-art: for instance, APOLLO solved 84.9% of problems in a math benchmark (miniF2F) using a relatively small 8-billion-parameter model, far better than prior attempts, all with each solution carrying a guarantee of correctness by construction
. Such work is significant because it shows an AI system can be designed to never accept its own reasoning unless it passes an external truth test – very much in spirit of “truth-as-proof” under liability. The AI’s output here is a formal proof that any mathematician (or automated checker) can independently verify – a direct analog to testifiable statements in Doolittle’s terms.
. For example, researchers are creating techniques to verify that a learned controller for a drone will never violate safety constraints, or that a neural network for medical diagnosis will respect certain logical conditions (like not prescribing a drug if the patient record shows an allergy). One approach is to integrate SMT (satisfiability modulo theories) solvers or model-checkers with neural nets. Another approach is to train the AI within a formal environment so that every decision must satisfy a check. This echoes Doolittle’s operational coherence: the AI’s internal operations are constrained to those that are decidable and provably safe. While still a developing area, the long-term vision is AI that comes with a proof certificate – much like a mathematical proof – for critical decisions. In practical terms, an AI medical assistant might provide a step-by-step rationale for a treatment that can be formally verified against medical guidelines, or an AI-generated code patch would come with a proof that it resolves an issue without introducing new bugs
. Achieving this at scale is an open challenge, but steady progress in AI-assisted formal reasoning (such as the Lean+LLM collaborations) and formal methods for neural networks indicates a movement toward machine reasoning that is correct by construction.
. This dramatically improves the model’s ability to solve problems that require precise computation, data analysis, or logical step-by-step work. Rather than trusting the language model’s internal approximation of arithmetic or syntax, the system actually executes code and observes the result. If the initial code is wrong, the AI can iteratively debug it by reading the error messages and fixing mistakes, then running again
. The effect is a huge boost in accuracy on math and programming tasks – essentially offloading the reasoning to a tool that guarantees the correctness of each step. Indeed, enabling code execution raised ChatGPT’s score on a standard math benchmark from ~54% to 84.3% by eliminating calculation errors
. As DataCamp’s review noted, “by executing code to find answers, the chatbot can provide more precise and accurate responses,” mitigating a common source of LLM inaccuracy
. In Doolittle’s terms, this is the AI making a performative truth claim – e.g. producing a chart or computing a number – which is immediately tested through execution. The result is not just a verbal answer but a verifiable artifact (a program output, a figure, etc.) that the user can inspect. Such integration of operational tests ensures the model’s reasoning doesn’t stay in a probabilistic limbo; it is forced to commit to answers that work in reality (or else correct itself if they fail).
. Here the distance between two cities is fetched and correctly reported, with ChatGPT refraining from any unsupported invention.📷
. In other words, the language model defers to a computational knowledge engine for questions of fact, quantity, or formal knowledge, ensuring the final answer rests on a solid, testable foundation rather than the LLM’s internal weights. The figure above demonstrates this: asked about distances, ChatGPT used WolframAlpha and produced a quantitatively correct answer (which WolframAlpha computed from its curated data). The plugin even provided a step-by-step trace (“Used Wolfram” with the query details) that the user could inspect
– essentially an on-demand proof of the answer’s validity. This approach directly aligns with making AI outputs testifiable: the AI is not an oracle asking for blind trust; it becomes a mediator that translates the user’s request into factual queries or code, then returns an answer that anyone could double-check by examining the intermediate steps or re-running the queries.
. Anthropic’s Claude can be configured with a “constitutional” tool that looks up definitions or policies to ensure its advice is grounded in accepted knowledge. Perhaps the most comprehensive is OpenAI’s “o3” series models (2025), which are explicitly trained to use tools in an agentic manner. OpenAI’s documentation notes that OpenAI o3 is a model “trained to think for longer before responding” and can “agentically use and combine every tool within ChatGPT,” including web search, code execution, and visual analysis
. Crucially, these models have been taught when and how to invoke tools in order to yield more detailed, correct, and verifiable answers
. The result is a step-change in performance: by leveraging tools, o3 significantly reduces reasoning errors and was judged to produce more “useful, verifiable responses” than its predecessors, often citing web sources or producing calculations to back its answers
. This design mirrors Doolittle’s call for operationalization: the model is effectively grounding its words in deeds (searches, code runs, etc.). Whenever it faces a question of fact or a complex task, it performs concrete actions whose outcomes determine its answer – an echo of requiring that every claim must have a demonstrated justification.
. By using innovative techniques to inspect the activations of neurons and attention heads, Anthropic’s researchers identified clusters of neurons that correspond to interpretable concepts and even discovered that Claude seems to plan ahead internally. An IBM summary of this work noted that Claude “handles complex reasoning tasks in ways that resemble human cognition, complete with internal planning [and] conceptual abstraction”
. For example, when asked to compose a rhyming poem, Claude’s neural activations revealed that it anticipated a rhyming word (“rabbit”) several words in advance, effectively setting a goal and then generating content to meet that goal
. This was a striking find – it showed an LLM is not merely producing one word at a time in isolation; it can have something akin to a “premeditated” intermediate outcome it’s working towards. In cognitive terms, this is a form of reasoning or planning horizon emerging from the model. Such insight aligns AI behavior a bit more with human-like logical steps, and by identifying the specific “circuits” responsible, researchers can verify or even manipulate them (Anthropic demonstrated that by intervening on those activations, they could change Claude’s chosen rhyme, steering its output predictably
).
. In one study, Claude was given a faulty hint to a math puzzle; Claude then produced an answer aligning with the hint and even a detailed rationale. However, by tracing the internal state, researchers saw no evidence of real arithmetic – the model had simply learned to generate a plausible narrative post-hoc, a phenomenon called unfaithful reasoning
. This ability to catch the model in a lie (even if an inadvertent one) is critical. It means developers can start to distinguish when an AI’s explanation is genuine versus when it’s a confabulation, and they can adjust training to penalize the latter. In the context of Doolittle’s philosophy, this is like separating an honest witness from a compulsive bullshitter – interpretability provides the cross-examination tools. By enforcing faithfulness (one of the emerging metrics in explainable AI, which demands the AI’s stated reasons truly reflect its computations
), we inch closer to AI whose outputs are not only correct but trustworthy.
) that allow researchers to map out which neurons or layers are responsible for which subtask in a multistep reasoning process. There are also efforts to create self-explaining models – architectures that generate a proof or diagram internally when answering, so that the explanation is a byproduct of the computation itself rather than a separate, potentially unfaithful, summary. For instance, some experimental models generate natural language justifications in parallel with their answers and are trained such that if the justification is invalid, the answer is likely wrong, thereby forcing a coupling between what they do and what they say about what they do. This resonates with Doolittle’s liability and testifiability points: the AI in effect must “show its work,” and if the work doesn’t check out, neither does the answer.
. Similarly, OpenAI’s new ChatGPT versions allow users to ask why the assistant gave a certain answer, and the assistant will attempt to reveal its chain-of-thought (with the caveat that it’s still an approximation). These are rudimentary, but they indicate a trend: making the reasoning trace visible and inspectable. In high-stakes fields – e.g. a medical AI explaining why it chose a diagnosis – such transparency isn’t just nice-to-have, it’s often legally or ethically required. Efforts like the XAI (Explainable AI) 2.0 manifesto call for open models where every decision can be audited
, moving away from the inscrutable black boxes of the past.
. LeCun argues that today’s AI lacks the ability to rapidly adapt to novel situations because it doesn’t build rich world models – internal simulations of how the world works
. In 2022 he outlined a blueprint with six modules: a Configurator (which sets up a task strategy), a Perception module (to understand the current state from sensory input), a World Model (to predict outcomes of actions, i.e., an internal causal simulator), a Cost module (to define objectives or reward signals), an Actor (to take actions), and a Memory module for context
. The key idea is that the World Model module provides the agent with an explicit tool for reasoning about events: it can imagine sequences, test hypotheses internally, and derive plans by “thinking ahead” (a bit like mental time-travel or running physics simulations in one’s head)
. This is reminiscent of the way Doolittle emphasizes operational thinking – here the AI would mentally perform operations in its world model to evaluate truth claims (“if I do X, Y will happen – is that desirable/true?”). Importantly, such an architecture separates intuitive inference and deliberate reasoning (sometimes likened to System 1 vs System 2 cognition). The perception module might do fast recognition (like an LLM free-associating a quick answer), but the world model allows for slower, stepwise logical reasoning when needed (like double-checking with a simulation or logical deduction). LeCun’s vision is that by training these modules (largely with self-supervised learning and predictive objectives
), the AI will learn not just surface correlations but causal, abstract representations of reality – exactly what’s needed for sound reasoning and “knowing when it doesn’t know.” While this remains a conceptual roadmap, Meta AI and other research labs (DeepMind’s work on generative environment models, for instance) are actively exploring components of it. If successful, an AI with a robust world model could achieve a level of real-world reasoning and correctness far beyond current LLMs: it would internally verify claims by checking against its model of the world (much as humans use mental models to reason through consequences), leading to decisions that are both interpretable and reliably grounded in reality.
-
Testifiability and Performative Truth: Nearly every development surveyed aims to make AI outputs more verifiable or grounded in demonstration. Tool-using AIs that consult calculators, run code, or fetch documents are essentially making their answers performative – the truth of their statements is backed by an action (a computation or retrieval) whose result anyone can examine
. This is a big shift from earlier AI systems that generated answers out of an inscrutable internal process. Likewise, formal proof systems (Lean+LLM, etc.) force the AI to show a complete proof for its conclusion, which is the ultimate testifiable artifact – much as Doolittle’s framework would demand evidence for any claim. In practical terms, an AI that solves an equation by actually solving it (and showing the steps) vs one that just states an answer is analogous to a witness performing an experiment vs. asserting an opinion. The former is performatively true by Doolittle’s definition (the truth is in the performance of the solution). So, initiatives like OpenAI’s o3 (with web citations)
, ChatGPT with Wolfram
, and APOLLO’s provable proofs all align strongly with the Natural Law emphasis on evidence and demonstration. They make AI more of a truth-teller under oath than a clever raconteur.
-
Operational Coherence and Decidability: Doolittle’s insistence on operational thinking – that concepts be reducible to actions or observations – finds echo in systems that ground reasoning in either simulations or formal rules. For example, LeCun’s world-model approach envisions that every prediction an AI makes comes from simulating plausible operations in its model of the world, effectively ensuring the AI’s reasoning always ties back to something concrete (a model state, an action outcome). This is one path to operational coherence: the AI doesn’t get to throw around abstract words without referents; it must connect them to model states or data. On decidability, formal verification efforts ensure that for certain questions (mathematical truths, program correctness), the AI will eventually resolve the truth via proof or counterexample, rather than languishing in uncertainty or circular debate. However, it must be said that current AI reasoning is not yet universally decidable – far from it. Open-ended questions or value-laden judgments can still stump AI systems in indecision or inconsistency. Doolittle’s framework might see current LLMs as woefully indecisive or non-coherent in many domains (since they often reflect conflicting training data without a way to reconcile truth). Yet the move towards structured reasoning tasks and objective benchmarks (like proving theorems, solving puzzles with known solutions) is a way to carve out pockets of decidability where AI can be trusted. In essence, researchers are identifying sub-problems where truth can be black-and-white and focusing AI efforts there as a foundation.
-
Liability and Epistemic Rigor: One aspect of Doolittle’s view is holding the speaker accountable for errors or deception. In AI, this corresponds to alignment and safety – ensuring AI doesn’t blithely output harmful falsehoods. Developments like interpretability and truthful AI benchmarks (e.g. TruthfulQA challenges) are attempts to instill epistemic rigor – getting models to adhere to facts and to explicitly flag uncertainty. Some labs (Anthropic, DeepMind) experiment with AI “constitutions” or guardrails that encode principles like “do not state information as factual if not grounded.” While these are not foolproof, they show movement towards an AI that knows the cost of lying (even if that “cost” is just a training penalty for being caught making stuff up). Additionally, the notion of audit trails in AI decisions (especially in finance or law applications
) speaks to liability: if an AI approves a loan or recommends a sentence, it should produce the reasons, so that if any step was illicit (say, using race as a factor) it can be identified and the AI (or its creators) held responsible. This is an area where alignment with Doolittle is growing due to societal pressure: just as Natural Law seeks to make each speech act accountable, regulators and users are pushing AI to be auditable and traceable. The technology is responding – e.g. through explainable AI techniques and robust evaluation protocols.
-
Where They Diverge: Despite progress, many AI systems still fall short of Natural Law ideals. Large language models remain probabilistic parrots in many respects – they have no built-in mechanism that guarantees truthfulness. They are not like a witness swearing on a stand; they are more like a well-read teenager opining on anything asked. Doolittle might critique that even with added tools, an AI might misuse them or present a veneer of proof without actual skin in the game. Indeed, Anthropic’s work showed cases of pseudologic – the AI explaining after the fact with a logically structured lie
. Until interpretability and training fixes eliminate that, the AI isn’t fully “liable” to truth in Doolittle’s sense. Moreover, many AI approaches still lack a true understanding of concepts in operational terms. For instance, an LLM can talk about “justice” or “quantum physics” eloquently without having grounded those in any real-world operation or experiment – it’s essentially reciting words. Doolittle’s framework would see a lot of that as fictional or irreciprocal (words not cashable by actions). The cutting-edge research is aware of this and tries to ground as much as possible (e.g. physical robotics environments, or at least code and data), but there’s a long way to go to reach human-level grounding. Additionally, decidability is violated whenever an AI hedges or contradicts itself. Despite improvements, AI models can give different answers depending on phrasing, or stall with uncertainty on hard problems. Humans, too, face undecidable questions, but Doolittle’s program pushes for always finding the next experiment to decide. AI currently doesn’t set up new experiments on its own (except in narrow cases like AutoML or scientific discovery systems).
. Major industry labs like DeepMind, OpenAI, and Anthropic have pushed the frontier with systems that use tools, memory, and self-reflection to solve complex tasks – from proving mathematical theorems with guaranteed correctness
, to navigating websites and APIs through natural language instructions
, to planning actions in multi-modal environments. Startups and independent thinkers contribute with fresh cognitive architectures and knowledge-centric AI that emphasize understanding over shallow pattern matching.
. This would have been almost unthinkable just a few years ago when neural networks were essentially black boxes. The trajectory suggests that future AI might indeed uphold the standards of Natural Law reasoning: providing answers that are not only correct, but justified, transparent, and anchored in reality to a degree that equals or surpasses human experts bound by those same principles.
, or a chatbot that can cite exactly where it found an answer – is a move towards AI systems that are not just persuasive or eloquent, but genuinely knowledgeable and reliable in a way that any rigorous epistemologist (Doolittle included) would appreciate.
-
Stephen Wolfram (2023). ChatGPT Gets Its “Wolfram Superpowers”!
.
-
IBM Research (2022). Neuro-Symbolic Inductive Logic Programming with Logical Neural Networks
.
-
Doolittle, C. (2019). Propertarianism – An Introduction (Natural Law Institute, PDF)
.
Source date (UTC): 2025-08-18 14:57:37 UTC
Original post: https://x.com/i/articles/1957456856651330036
Leave a Reply