-
Testifiability and Performative Truth: Nearly every development surveyed aims to make AI outputs more verifiable or grounded in demonstration. Tool-using AIs that consult calculators, run code, or fetch documents are essentially making their answers performative – the truth of their statements is backed by an action (a computation or retrieval) whose result anyone can examine
. This is a big shift from earlier AI systems that generated answers out of an inscrutable internal process. Likewise, formal proof systems (Lean+LLM, etc.) force the AI to show a complete proof for its conclusion, which is the ultimate testifiable artifact – much as Doolittle’s framework would demand evidence for any claim. In practical terms, an AI that solves an equation by actually solving it (and showing the steps) vs one that just states an answer is analogous to a witness performing an experiment vs. asserting an opinion. The former is performatively true by Doolittle’s definition (the truth is in the performance of the solution). So, initiatives like OpenAI’s o3 (with web citations)
, ChatGPT with Wolfram
, and APOLLO’s provable proofs all align strongly with the Natural Law emphasis on evidence and demonstration. They make AI more of a truth-teller under oath than a clever raconteur.
-
Operational Coherence and Decidability: Doolittle’s insistence on operational thinking – that concepts be reducible to actions or observations – finds echo in systems that ground reasoning in either simulations or formal rules. For example, LeCun’s world-model approach envisions that every prediction an AI makes comes from simulating plausible operations in its model of the world, effectively ensuring the AI’s reasoning always ties back to something concrete (a model state, an action outcome). This is one path to operational coherence: the AI doesn’t get to throw around abstract words without referents; it must connect them to model states or data. On decidability, formal verification efforts ensure that for certain questions (mathematical truths, program correctness), the AI will eventually resolve the truth via proof or counterexample, rather than languishing in uncertainty or circular debate. However, it must be said that current AI reasoning is not yet universally decidable – far from it. Open-ended questions or value-laden judgments can still stump AI systems in indecision or inconsistency. Doolittle’s framework might see current LLMs as woefully indecisive or non-coherent in many domains (since they often reflect conflicting training data without a way to reconcile truth). Yet the move towards structured reasoning tasks and objective benchmarks (like proving theorems, solving puzzles with known solutions) is a way to carve out pockets of decidability where AI can be trusted. In essence, researchers are identifying sub-problems where truth can be black-and-white and focusing AI efforts there as a foundation.
-
Liability and Epistemic Rigor: One aspect of Doolittle’s view is holding the speaker accountable for errors or deception. In AI, this corresponds to alignment and safety – ensuring AI doesn’t blithely output harmful falsehoods. Developments like interpretability and truthful AI benchmarks (e.g. TruthfulQA challenges) are attempts to instill epistemic rigor – getting models to adhere to facts and to explicitly flag uncertainty. Some labs (Anthropic, DeepMind) experiment with AI “constitutions” or guardrails that encode principles like “do not state information as factual if not grounded.” While these are not foolproof, they show movement towards an AI that knows the cost of lying (even if that “cost” is just a training penalty for being caught making stuff up). Additionally, the notion of audit trails in AI decisions (especially in finance or law applications
) speaks to liability: if an AI approves a loan or recommends a sentence, it should produce the reasons, so that if any step was illicit (say, using race as a factor) it can be identified and the AI (or its creators) held responsible. This is an area where alignment with Doolittle is growing due to societal pressure: just as Natural Law seeks to make each speech act accountable, regulators and users are pushing AI to be auditable and traceable. The technology is responding – e.g. through explainable AI techniques and robust evaluation protocols.
-
Where They Diverge: Despite progress, many AI systems still fall short of Natural Law ideals. Large language models remain probabilistic parrots in many respects – they have no built-in mechanism that guarantees truthfulness. They are not like a witness swearing on a stand; they are more like a well-read teenager opining on anything asked. Doolittle might critique that even with added tools, an AI might misuse them or present a veneer of proof without actual skin in the game. Indeed, Anthropic’s work showed cases of pseudologic – the AI explaining after the fact with a logically structured lie
. Until interpretability and training fixes eliminate that, the AI isn’t fully “liable” to truth in Doolittle’s sense. Moreover, many AI approaches still lack a true understanding of concepts in operational terms. For instance, an LLM can talk about “justice” or “quantum physics” eloquently without having grounded those in any real-world operation or experiment – it’s essentially reciting words. Doolittle’s framework would see a lot of that as fictional or irreciprocal (words not cashable by actions). The cutting-edge research is aware of this and tries to ground as much as possible (e.g. physical robotics environments, or at least code and data), but there’s a long way to go to reach human-level grounding. Additionally, decidability is violated whenever an AI hedges or contradicts itself. Despite improvements, AI models can give different answers depending on phrasing, or stall with uncertainty on hard problems. Humans, too, face undecidable questions, but Doolittle’s program pushes for always finding the next experiment to decide. AI currently doesn’t set up new experiments on its own (except in narrow cases like AutoML or scientific discovery systems).
-
Liang, B. et al. (2025). “AI Reasoning in Deep Learning Era: From Symbolic AI to Neural–Symbolic AI.” Mathematics 13(11): 1707
.
-
Ospanov, A. et al. (2023). “APOLLO: Automated LLM and Lean Collaboration for Advanced Formal Reasoning.” arXiv preprint
.
-
Stephen Wolfram (2023). ChatGPT Gets Its “Wolfram Superpowers”!
.
-
DataCamp Tutorial (2024). How to Use ChatGPT Code Interpreter
.
-
Google DeepMind (2024). Google DeepMind at ICLR 2024 (blog)
.
-
Anthropic (2024). Tracing the thoughts of a large language model (blog)
.
-
IBM Research (2022). Neuro-Symbolic Inductive Logic Programming with Logical Neural Networks
.
-
OpenAI (2025). Introducing OpenAI o3 and o4-mini (release blog)
.
-
Doolittle, C. (2019). Propertarianism – An Introduction (Natural Law Institute, PDF)
.
-
Cogni Down Under (2024). Inside Logical AI: Explainable Reasoning (Medium)
.