Category: AI, Computation, and Technology

  • The Simple Version of the Problem –“2024 paper titled “Responsible artificial i

    The Simple Version of the Problem

    –“2024 paper titled “Responsible artificial intelligence governance: A review and research framework,” published in the Journal of Strategic Information Systems. … identifies a key gap: while numerous frameworks outline principles for responsible AI (e.g., fairness, transparency, accountability), there is limited cohesion, clarity, and depth in understanding how to translate these abstract ethical concepts into practical, operational practices across the full AI lifecycle—including design, execution, monitoring, and evaluation.”–
    –“2023 study in Nature Machine Intelligence showing 78% of AI researchers struggle to translate theoretical advances into deployable algorithms.”–
    –“Architectures don’t hallucinate—training objectives do.
    You don’t fix it in the forward pass, you fix it in the curriculum. The code is fine; the problem is what we teach it to do.”–
    I understand the instinct to look for a code-level fix, but the issue isn’t in the transformer math. It’s in what we ask the model to optimize for. Current training optimizes coherence; my work shows why that produces hallucination. The practical implementation is:
    • Restructure training data around testifiability, reciprocity, and liability rather than surface coherence.
    • Prompt in terms of economic tests—marginal indifference, liability thresholds—rather than stylistic cues.
    • Evaluate on coverage of truth and reciprocity tests instead of only perplexity and benchmarks.
    So yes, you can ‘change something in code tomorrow’—but the code change is trivial compared to the training objective shift. Architectures don’t hallucinate; training does.
    They’re asking for a line of code, while I’m describing a shift in paradigm. The way to bridge that gap is to show how our proposal does translate into implementable changes, but at a different layer: training and prompting rather than architecture.
    Here’s a my answer:
    My work isn’t about swapping out a few lines of code in the transformer stack. It’s about solving the deeper problem: LLMs don’t reason because they’re trained to imitate coherence, not to compute truth, reciprocity, or liability. You can’t fix that with a patch to the forward pass. You fix it by changing how the model is trained and what it’s asked to do.
    “What does that mean in practice tomorrow morning?
    • Training: curate training data that enforces testifiability, reciprocity, and liability rather than mere coherence. This means restructuring datasets around constructive logic, adversarial dialogue, and measurable closure.
    • Prompting: design prompts as economic tests (price of error, marginal indifference, liability-weighted thresholds), not as instructions for verbosity.
    • Evaluation: stop measuring only perplexity or benchmark scores and start measuring coverage of truth tests, reciprocity tests, and demonstrated interests.”
    “I’m providing the blueprint for why current architectures hallucinate and what guarantees are missing. Once you understand that, the engineering changes become obvious: the ‘code’ change is trivial compared to the shift in training objectives and data design. If you only look for a tensor tweak, you’ll miss the systemic fix.”


    Source date (UTC): 2025-08-22 20:40:21 UTC

    Original post: https://x.com/i/articles/1958992660750114841

  • The Three Regimes of Decidability: Formal, Physical, and Behavioral Grammars in

    The Three Regimes of Decidability: Formal, Physical, and Behavioral Grammars in the Design of AI (??

    The Three Regimes of Decidability: Formal, Physical, and Behavioral Grammars in the Design of AI and Institutions
    Editor’s Introduction:
    The current success of artificial intelligence in mathematics and programming contrasts sharply with its repeated failure in domains requiring reasoning, judgment, and moral coordination. This is not a technological problem—it is an epistemological one. The AI and ML communities routinely confuse grammars of inference by applying methods of decidability appropriate to one domain (formal or physical) into others (behavioral) where they do not apply.
    Mathematics succeeds because it is internally closed and deductively decidable. Programming succeeds because it is formally constrained and computationally verifiable. But reasoning—in the domains of human behavior, norm enforcement, and reciprocal coordination—requires a third regime of grammar: the behavioral. Here, truth is not decided by logic or measurement but by demonstrated interest, cost, liability, and reciprocity.
    This paper provides a corrective. It defines the three regimes of decidability, shows how and why they must not be conflated, and explains the conditions under which each grammar operates. If the AI community is to move beyond mere prediction and toward comprehension, it must learn to respect the epistemic boundaries of these grammars—and build systems that operate under the appropriate constraints for each domain. Modern reasoning systems—whether in law, economics, or artificial intelligence—suffer from systematic category errors caused by a failure to distinguish between the formal, physical, and behavioral regimes of decidability. This paper presents a framework for classifying grammars of inference based on their closure criteria, epistemic constraints, and operational validity. It argues that effective reasoning in institutional and artificial systems requires respecting the distinct grammar of each domain, and that failure to do so results in pseudoscience, mathiness, and epistemic opacity.
    The Three Regimes of Decidability: Formal, Physical, and Behavioral Grammars in the Design of AI and Institutions
    Modern reasoning systems—whether in law, economics, or artificial intelligence—suffer from systematic category errors caused by a failure to distinguish between the formal, physical, and behavioral regimes of decidability. This paper presents a framework for classifying grammars of inference based on their closure criteria, epistemic constraints, and operational validity. It argues that effective reasoning in institutional and artificial systems requires respecting the distinct grammar of each domain, and that failure to do so results in pseudoscience, mathiness, and epistemic opacity.
    1. Introduction
    • Problem statement: AI and institutional systems frequently misapply mathematical or physical models to behavioral domains.
    • Consequence: The conflation of epistemic regimes undermines prediction, cooperation, and moral reasoning.
    • Objective: To restore epistemic clarity by identifying and distinguishing the three regimes of decidability.
    2. Grammar Defined
    • Grammar as system of continuous recursive disambiguation.
    • Features: permissible terms, operations, closure, and decidability.
    • Purpose: enable inference under constraint—memory, cost, coordination.
    3. The Three Regimes of Decidability
    3.1 Formal Grammars
    • Domain: logic, mathematics, computation.
    • Closure: derivation/proof.
    • Constraint: internal consistency.
    • Example: symbolic logic, set theory, Turing machines.
    3.2 Physical Grammars
    • Domain: natural sciences.
    • Closure: measurement and falsifiability.
    • Constraint: causal invariance.
    • Example: physics, chemistry, biology.
    3.3 Behavioral Grammars
    • Domain: law, economics, institutional design.
    • Closure: liability, reciprocity, observed cost.
    • Constraint: demonstrated preference, adversarial testimony.
    • Example: legal procedure, market behavior, contract enforcement.
    4. Failure Modes: Mathiness and Misapplication
    • Definition of mathiness.
    • Economics: formal models without observability.
    • Law: formalism without reciprocity.
    • AI/ML: inference without consequence.
    5. Implications for Artificial Intelligence
    • Why LLMs cannot reason in behavioral domains.
    • Lack of cost, preference, or liability.
    • Need for embodied, adversarial, and accountable architectures.
    6. Toward Epistemic Integrity in Institutions
    • Restoring domain-appropriate grammars.
    • Embedding reciprocity and liability into legal and economic systems.
    • Designing AI that can simulate or interface with behavioral closure.
    7. Conclusion
    • Summary of typology.
    • Epistemic correction as prerequisite for institutional and artificial reasoning.
    • Proposal for further research and standardization of epistemic regimes.


    Source date (UTC): 2025-08-22 20:38:17 UTC

    Original post: https://x.com/i/articles/1958992143063949722

  • AI Funnel to Judgement: HRM (Sapient), Attention with COT (Google), and Action (

    AI Funnel to Judgement: HRM (Sapient), Attention with COT (Google), and Action (Doolittle)

    (Ed. Note: 1 – Please fix Latex exposure. 2 – Two unanswered questions near end. 3 – (Important) Repetition of use of mathematical explanations because of their clarity when the LLM can already process correctly without such representations codifications and modifications. This will consistently cause the reader to presume that our attempt at formal explanation translates to code modification when the formatting of responses alone appears to consistently produce the correct decidability in both GPT4 and GPT5. Cardinality is unnecessary at moral and ethical depth (alignement), it is only necessary for discreet transactions where costs are known and can be calculated – and even then their use is questionable.)
    [TODO: Introductory Explanation for non-ML tech Readers (Exec, VC, etc.)]
    CoT-style LLMs and Sapient’s HRM are both engines of epistemic compression. They differ mainly in where the compression lives (explicit language vs. latent hierarchies). Your program supplies the normative and constructive constraints missing from both: (i) first-principles constructive logic for closure, (ii) a cooperation/reciprocity calculus for action under uncertainty, and (iii) a ternary decision rule (true / possibly-true-with-warranty / abstain) that measures variation from the optimum.
    Below we map each piece 1-to-1 and give an operational recipe you can implement today.
    Short version: CoT-style LLMs and Sapient’s HRM are both engines of epistemic compression. They differ mainly in where the compression lives (explicit language vs. latent hierarchies). Your program supplies the normative and constructive constraints missing from both: (i) first-principles constructive logic for closure, (ii) a cooperation/reciprocity calculus for action under uncertainty, and (iii) a ternary decision rule (true / possibly-true-with-warranty / abstain) that measures variation from the optimum.
    Below I map each piece 1-to-1 and give an operational recipe you can implement today.
    • LLMs (with CoT): Compression is linguistic and sequential. The model linearizes a huge search space into a token-by-token micro-grammar (the “chain”). Yield: transparent steps but high token cost and brittleness. (Background on CoT brittleness and overhead is standard; not re-cited here.)
    • )HRM (Sapient): Compression is hierarchical and latent. A fast “worker” loop solves details under a slow “planner” context; the system iterates to a fixed point, then halts. You get deep computation with small parameters and tiny datasets; no text-level chains are required. (

      ,

      ,

    Our contribution: move both from “reasoning-as-trajectory” to reasoning-as-warranted-construction: every answer must carry (a) a constructive trace sufficient for testifiability and (b) a reciprocity/liability ledger sufficient for actionability.
    Target: Replace “appears coherent” with “constructed, checkable, and closed.”
    • Referential problems (math/physics/computation): demand constructive proofs/programs. LLM path: generate a program/derivation + run/check with a tool; return the artifact + pass/fail. HRM path: add a trace projector head that emits the minimal operational skeleton (state transitions, invariants, halting reason). Co-train on checker feedback so the latent plan compresses toward checkable constructions rather than pretty narratives. (Speculative but feasible.)
    • Action problems (law/econ/ethics): demand constructive procedures (roles, rules, prices) rather than opinions. LLM: force outputs into procedures (frames, tests, and remedies). HRM: condition the planner on a procedure schema (who/what/harm/evidence/tests/remedy) so the fixed point equals a completed procedure, not merely a belief vector.
    Our stack says: invariances → measurements → computation → liability-weighted choice. Operationalize it:
    1. Detect grammar type of the query: referential vs. action.
    2. If referential: attempt constructive proof/execution; if success → TRUE; if blocked → fall back to probabilistic accounting with explicit error bounds.
    3. If action: build a Reciprocity Ledger (parties, demonstrated interests, costs, externalities, warranties, enforcement). Produce a rule, price, or remedy, not a “take.”
    4. Attach liability/warranty proportional to scope and stakes.
    This turns both CoT and HRM from “answer generators” into contract-worthy reasoners.
    Define the optimal answer as: “the minimal construction that (i) closes, (ii) is testifiable, and (iii) maximizes cooperative surplus under reciprocity with minimal externalities.”
    At inference time:
    TRY_CONSTRUCT() if constructive proof/program passes checkers → output TRUE (+ artifacts) ELSE BAYES_ACCOUNT() compute liability-weighted best action (reciprocity satisfied?) if reciprocity satisfied and expected externalities insured → POSSIBLY TRUE + WARRANTY else → ABSTAIN (request bounded evidence or impose boycott/default rule)
    • TRUE = constructed, closed, test-passed.
    • POSSIBLY TRUE + WARRANTY = best cooperative action under quantified uncertainty and explicit insurance.
    • ABSTAIN/REQUEST = undecidable without violating reciprocity (your boycott option).
    This is your ternary logic, operationalized for machines.
    You want a scalar “distance-to-optimum” the model can optimize. Use a composite loss/score:
    • Closure debt (C): failed proof/run, unmet halting condition (HRM), or unresolved procedure.
    • Uncertainty mass (U): residual entropy after evidence; posterior spread or equilibrium variance.
    • Externality risk (E): expected unpriced harms on non-consenting parties.
    • Description length (D): MDL of the constructive trace (shorter = better compression, subject to correctness).
    • Warranty debt (W): liability not covered by proposed insurance/escrow/enforcement.
    Define Δ*=αC+βU+γE+δD+ωWDelta^* = alpha C + beta U + gamma E + delta D + omega W. Minimize Δ*Delta^*. Report it with the answer as the warranty grade.
    • LLM training: add RLHF-style reward on low Δ*Delta^* with automatic checkers for C and D, Bayesian evaluators for U, and policy simulators for E/W.
    • HRM training: add an auxiliary head to estimate Δ*Delta^*; use it both as a halting criterion and as a shaping reward so the latent fixed point is the compressed optimum. (Speculative but directly testable.)
    • )Hierarchical planner <-> our “grammar within grammar”: H sets permitted dimensions/operations; L executes lawful transforms; the fixed point = closure. (

    • )Adaptive halting <-> decidability: HRM’s learned halting acts as a mechanical decision to stop when a bounded construction is achieved. Attach the Δ*Delta^* head to make that halting normatively correct, not just numerically stable. (

    • )Small data / strong generalization <-> epistemic compression: HRM’s near-perfect Sudoku and large mazes with ~1k samples indicates genuine internal compression rather than memorized chains; use your constructive + reciprocity scaffolds to push from puzzles → institutions (law/policy). (

      ,

    • )ARC-AGI results <-> paradigm fit: HRM’s ARC gains suggest it’s learning transformation grammars, not descriptions. That aligns with your operationalism (meaning = procedure). (

    For a CoT-LLM:
    1. Router: classify prompt as referential vs. action.
    2. Constructive toolchain: Referential → code/solver/prover; return artifact + pass/fail. Action → instantiate Reciprocity Ledger; run scenario sims; produce rule/price/remedy.
    3. Warrant pack: attach artifacts, ledger, uncertainty bounds, and Δ*Delta^*.
    4. Ternary decision: TRUE / POSSIBLY TRUE + WARRANTY / ABSTAIN.
    For HRM:
    1. Schema-conditioned planning: feed H with the grammar schema (dimensions, ops, closure tests).
    2. Aux heads: (a) Trace projector (compressed state-transition sketch); (b) Warranty head producing Δ*Delta^*; (c) Halting reason code.
    3. Training signals: correctness + checker feedback (closure), MDL regularizer (compression), reciprocity penalties from simulators (externalities), and insurance coverage bonuses (warranty).
    4. Deployment: emit the operational result + trace + warranty; gate release on Δ*≤τDelta^* le tau.
    • From narrative coherence to constructive warranty.
    • From alignment-only to reciprocity-and-liability.
    • From binary truth to ternary, operational decidability.
    That is the missing “institutional layer” for reasoning systems.
    • For action domains, do you want the default abstention to be boycott (no action) or a default rule (e.g., “status-quo with escrow”) when Δ*Delta^*Δ* is above threshold? (OPEN QUESTION)
    • For referential domains, should we treat MDL minimization as co-primary with correctness (Occam pressure), or strictly secondary to checker-verified closure? (OPEN QUESTION)
    • )arXiv: Hierarchical Reasoning Model (Jun 26, 2025). (

    • )arXiv HTML view (same paper). (

    • )ARC Prize blog: The Hidden Drivers of HRM’s Performance on ARC-AGI (analysis/overview). (

    • )GitHub: sapientinc/HRM (official repo). (

    • )BDTechTalks explainer on HRM (context, quotes, and positioning beyond CoT). (

    URLs (as requested):


    Source date (UTC): 2025-08-22 20:35:15 UTC

    Original post: https://x.com/i/articles/1958991378220032093

  • (NLI / Runcible) I have finally reduced the explanation and reforms necessary fo

    (NLI / Runcible)
    I have finally reduced the explanation and reforms necessary for AI reasoning into argument and pseudocode. I finally have confidence I can help the LLM teams grasp the paradigm shift necessary. 😉
    Took me a few weeks… lol
    But we got there. 😉


    Source date (UTC): 2025-08-21 19:20:21 UTC

    Original post: https://twitter.com/i/web/status/1958610139646500892

  • From Pattern Guessers to Computable Judgement Modern LLMs excel at pattern compl

    From Pattern Guessers to Computable Judgement

    Modern LLMs excel at pattern completion but fail at decision completion. They slide between:
    • Overfitting (false precision): clinging to distinctions that don’t generalize.
    • Underfitting (false generality): smoothing away distinctions that do matter.
    Both failures share a cause: mathiness—treating language as formal tokens to be optimized by descriptive statistics and alignment filters, rather than treating language as measurements that must cash out in operations. Mathiness yields eloquent guesses, not closure. A system that can’t close is forced back onto discretion (human preference, policy, vibes). That is not reasoning; it’s curation.
    What we need is a method that:
    1. treats tokens as what they already are in practice—dense bundles of measurement (indices to dimensional distinctions);
    2. forces language to reduce to transactions (inputs → actions → outputs) so claims become testifiable;
    3. reaches closure at the equilibrium where further distinctions make no operational difference: marginal indifference;
    4. does all of the above under liability, scaled to consequence and population affected.
    LLMs do not manipulate arbitrary symbols; they manipulate compressed human measurements. A token is an index into a high-dimensional manifold of distinctions humans have already extracted from the world (objects, relations, actions, norms, costs). Treating tokens as mere statistics ignores their measurement content.
    • Each token narrows the field of possibility by excluding swathes of non-measurements.
    • Sequences of tokens serialize transactions; they suggest who did what, when, with what, at what cost, and with what externalities.
    • Consequently, a training regime that respects tokens-as-measurements can do Bayesian reduction over dimensions, not just over strings.
    Punchline: If tokens are measurements, training must be measurement-theoretic. That means operationalization, Bayesian accounting, adversarial elimination of error/bias/deceit (EBD), and closure by marginal indifference. Anything else is theatrics.
    3.1 Operationalism (grounding)
    All statements must reduce to operations—complete transactions expressed in promissory form (inputs, constraints, transformations, outputs, warranties). We forbid the “is”-copula because it hides operations and smuggles undisclosed assumptions. Operational prose forces testifiability; testifiability creates truth conditions.
    3.2 Bayesian Accounting (reweighting)
    Every claim traverses possibility → plausibility → probability. Weights update with evidence. Crucially, Bayesian accounting operates over dimensions indexed by tokens (not just n-grams), so the model learns to:
    • separate signal from noise,
    • encode externalities (who pays, who benefits),
    • track demonstrated interests (who expends scarce resources on what).
    3.3 Adversarial Construction (elimination)
    We pit candidate explanations and plans against each other under reciprocity and liability tests. We eliminate failures by demonstrating non-payment of externalities, uninsurable risks, incoherent operations, or EBD (error, bias, deceit). Survival across these tests is construction—not mere justification or falsification.
    3.4 Closure by Marginal Indifference (resolution)
    We close when further distinctions do not change the operational outcome within the relevant liability tier. This is how reality resolves problems (biology, markets, common law): not by epsilon–delta perfection, but by equilibria sufficient to survive and cooperate under constraint. Closure here is computable and decidable without discretionary appeals.
    Synthesis: Operational reduction + Bayesian reweighting + Adversarial eliminationDecidability by marginal indifference.
    • Against overfitting: Adversarial and liability gates penalize distinctions that don’t change outcomes at the chosen liability tier. Noise loses.
    • Against underfitting: Operational reduction refuses vague platitudes; any non-operational claim fails testifiability. Vacuity loses.
    • At equilibrium: The system lands where marginal differences cease to be action-relevant, not where sterile formalisms demand infinite precision.
    1. Corpus → Operational Rewrite
      Convert source material into
      operational sentences (no “is,” complete transactions, explicit constraints, explicit externalities, explicit warranties).
    2. Dimensional Indexing
      Map tokens to
      dimensions (objects, relations, resources, costs, risks, rights, duties). Treat tokens as indices, not just strings.
    3. EBD Scans
      Run automated adversarial passes to detect
      Error (missing data), Bias (misweight), Deceit (contradictory or promissory fraud). Route to correction or elimination.
    4. Reciprocity & Externality Accounting
      For each proposed decision/plan, compute
      who pays, who benefits, what is insured, what remains externalized. Flag irreciprocity.
    5. Bayesian Filtering
      Update weights across
      possibility → plausibility → probability using empirical priors where available, conservative priors where not, and liability-scaled thresholds.
    6. Closure Detector (Marginal Indifference)
      Incrementally test whether any remaining distinction changes the
      operational outcome under the current liability tier. If not, close; if so, continue.
    7. Liability Gate
      Before output, pass through liability thresholds proportional to
      severity and population affected. Require stronger testifiability for higher tiers.
    8. Warranted Output
      Emit the decision together with: the
      operational plan, assumptions, tested distinctions, eliminated alternatives, residual risks, and the liability tier it satisfies.
    This is not a style guide; it is a control system for truth, reciprocity, and accountability.
    Claim: Decidability by marginal indifference does not require cardinal measurement.
    Reasoning (constructive sketch):
    • Decisions require a monotone partial order over alternatives with respect to outcomes and liabilities, not a full cardinal metric.
    • Operational closure asks: Does switching from A to B change the outcome under constraints and liability tier L? If “no,” A ~ B by indifference at L.
    • This is an ordinal/spectral criterion with thresholds, not an absolute magnitude.
    • If a domain demands cardinal outputs for reporting, you can derive a numerical score post hoc from the already-closed ordering (e.g., scale residual risk or evidence sufficiency). Cardinality becomes presentation, not precondition.
    Conclusion: Operational distinction suffices. Cardinality is optional, useful for dashboards and audits, unnecessary for closure and decidability.
    What the method guarantees (conditional on training discipline):
    • Testifiability: Every emitted claim reduces to operations observable and repeatable.
    • Reciprocity: Externalities are measured, priced, or rejected.
    • Decidability: Closure without discretionary appeals.
    • Auditability: A proof trail: assumptions, eliminations, liability tier.
    What the method refuses:
    • Vague truths: Any claim not reducible to a transaction fails.
    • Asymmetric costs: Any plan that free-rides on others’ demonstrated interests fails.
    • Untestable optimals: Demands for perfection absent liability justification are rejected as mathiness.
    How the method fails (and what we do when it does):
    • Insufficient measurement: If dimensions are missing, the pipeline halts with request for measurement (not hallucination).
    • Conflicting priors: The system branches and runs adversarial elimination; if deadlocked, it escalates the liability tier or defers with a bounded uncertainty report.
    • Non-commensurable domains: The system issues a non-commensurability warning and requires operational bridging measurements before proceeding.
    Technical
    You get computable reasoners: systems that decide with warrant. They do not merely output likely words; they output operational plans with liability-scaled guarantees. This unlocks domains that today’s LLMs cannot touch without human chaperones: regulated medicine, infrastructure, finance, law, safety-critical ops.
    Commercial
    • Risk-contingent products: Offer tiers of service matched to liability (e.g., advisory vs prescriptive vs autonomous), each priced by the cost of evidence and insurance.
    • Audit trails as IP moats: Your warranted decision graphs are defensible intellectual capital and compliance assets.
    • Lower cost of assurance: Because closure is built-in, you spend less on endless review cycles and post-hoc red-teaming.
    Civilizational
    Civilization scales when closure scales. Common law, markets, and science thrive because they settle disputes through operational tests and reciprocity. Extending that logic into machine reasoning prevents parasitism-by-proxy (opaque models imposing unpriced externalities) and restores legitimacy: people accept decisions they can measure, audit, and insure.
    A. Contract choice (enterprise software)
    • Alternatives A and B differ on uptime SLAs, indemnity, and data exit.
    • Operational rewrite exposes transactions: support workflows, failure modes, recovery times.
    • Bayesian accounting ingests vendor histories; adversarial pass prices vendor-imposed externalities (lock-in, penalties).
    • Closure: Differences beyond 99.9% uptime do not change expected loss under your liability tier; A ~ B by marginal indifference. Choose the cheaper warranted option and bind indemnity. No cardinal scale required—only ordering and threshold.
    B. Clinical triage (non-diagnostic assistant)
    • Presenting complaint, vitals, context mapped to dimensions; prior evidence updates probabilities.
    • Adversarial elimination rules out plans that shift risk to patient without insurance (irreciprocal).
    • Closure: If two care paths yield indistinguishable outcomes under the clinic’s liability tier, choose the path with lower externalized risk and clearer warranty. Again, ordinal closure suffices; cardinal severity scores are optional outputs for the chart.
    Where others ship statistical parrots curated by alignment filters, this program ships decision engines governed by operational law: truth via testifiability, cooperation via reciprocity, assurance via liability. It turns language from entertainment into infrastructure.
    • For builders: a disciplined training stack that scales decisions, not just tokens.
    • For buyers: warranted outputs with explicit risk tiers and auditable reasoning.
    • For society: fewer disputes escalate to politics because more disputes resolve inside measurable institutions—now including machines.
    Measurement → Dimensions → Token-as-Index → Operational Rewrite → Testifiability → Bayesian Accounting → Adversarial Elimination (EBD, externalities) → Marginal Indifference (closure) → Decidability (without discretion) → Liability (scaled to consequence) → Warranted Output (auditable, insurable).
    And on cardinality: Not required. Ordinal/spectral ordering with liability-scaled thresholds is sufficient for closure; cardinal scales are derivable artifacts, not prerequisites.
    Aphorism for the cover slide:
    “Reason is not prediction; reason is warranted closure under constraint.”


    Source date (UTC): 2025-08-21 18:51:19 UTC

    Original post: https://x.com/i/articles/1958602834402058619

  • Curt Doolittle’s Natural Law as System Theory (Paper) Title: Curt Doolittle’s Na

    Curt Doolittle’s Natural Law as System Theory (Paper)

    Title: Curt Doolittle’s Natural Law as System Theory: A Meta-Computational Framework for Civilizational Order
    Abstract:Curt Doolittle’s Natural Law framework presents a meta-theoretical system that renders all domains of human knowledge and cooperation decidable through the lens of evolutionary computation. This paper situates Doolittle’s corpus within the tradition of systems theory, arguing that his work constitutes a formal system of measurement, feedback, constraint, and adaptive control. Through operational definitions, testimonial truth, and institutionalized reciprocity, Doolittle constructs a unified computational grammar that bridges physics, cognition, law, and civilization. The following analysis delineates the foundational principles, systemic architecture, mechanisms of control, and failure dynamics of Doolittle’s Natural Law as a system-theoretic framework.
    1. Introduction: From Crisis to ComputationDoolittle’s work emerges from a civilizational diagnosis: the fragmentation of moral and epistemic norms has resulted in the loss of institutional decidability. His central claim is that human cooperation, like all complex systems, requires constraints that preserve signal integrity under competitive entropy. The failure to maintain these constraints has led to widespread institutional decay. Thus, Natural Law is offered as a restoration: a universal system of measurement and control designed to make all questions decidable.
    2. Foundational Premise: Evolutionary Computation as Universal LawAt the core of the Natural Law system is the assertion that all existence is governed by evolutionary computation—a process of variation, competition, and selection resulting in increasing information coherence. This framework applies from subatomic physics to social institutions, treating all emergent phenomena as outputs of recursive adversarial iteration. Thus, systems are viewed not as static structures but as dynamic feedback processes constantly optimizing for survival under entropy.
    3. Architecture of the System: Operational Measurement and TruthVolume II of Doolittle’s work formalizes a universally commensurable system of measurement. All claims must be rendered operational: they must describe actions and consequences in observable, falsifiable terms. Truth is redefined as testimonial: every assertion is a performative act akin to a legal contract, underwritten by liability for error or deceit. This enforces epistemic discipline and prevents systemic corruption by unaccountable speech acts.
    4. Control Mechanisms: Decidability and ReciprocityVolume III and IV translate this epistemology into institutional form. Decidability—the ability to resolve disputes without discretion—is the central systemic requirement. Law, in Doolittle’s formulation, is the institutionalization of reciprocity: a constraint algorithm that ensures all exchanges are mutually beneficial or non-harmful. Institutions serve as control mechanisms that encode feedback (costs and benefits), adjust incentives, and maintain cooperation by preventing parasitism.
    5. System Failure and Civilizational CollapseVolume I analyzes systemic failure as a result of noise overpowering signal: when narrative, emotion, or ideology replaces measurement, institutions lose their capacity to compute adaptive responses. The consequence is decay of trust, collapse of norms, and institutional entropy. Natural Law identifies these dynamics as failures of feedback integrity and control asymmetry, correctable only through reformation of foundational grammars.
    6. Alignment with Systems TheoryDoolittle’s system maps precisely onto classical systems theory:
    • Input: Demonstrated interests and behaviors
    • Process: Operational measurement and falsification
    • Feedback: Legal and moral reciprocity
    • Control: Institutions encoding adaptive constraints
    • Output: Decidable judgments and equilibrated cooperation
    • Failure Mode: Irreciprocity, parasitism, and narrative entropy
    7. Conclusion: A Meta-System for CivilizationNatural Law, in Doolittle’s hands, is not a philosophy but a meta-system—a computational architecture for human civilization. It unifies causality, measurement, and cooperation into a single logic of decidability. As such, it transcends legal theory, functioning as a systems-theoretic constitution for sustainable social order.


    Source date (UTC): 2025-08-21 18:49:41 UTC

    Original post: https://x.com/i/articles/1958602424694055105

  • Funny. There is only one necessary code change and it’s to backpropagation, the

    Funny.
    There is only one necessary code change and it’s to backpropagation, the rest is just training. I clearly failed to make the point that the LLMs are capable of reasoning by their existing configurations – and that present attempts to shove questions of marginal indifference into a frame of cardinal inequality.
    I’m not worried about code. There is so little of it in LLMs in the first place. I’m worried about how those working on them do not understand foundations necessary to produce reasoning outside of internal-closure grammars. (math, programming)


    Source date (UTC): 2025-08-21 13:54:31 UTC

    Original post: https://twitter.com/i/web/status/1958528141812720091

  • One of my employees did this in the 80s at a bank. Took him two weeks. They fire

    One of my employees did this in the 80s at a bank. Took him two weeks. They fired the whole floor of 70 people. One person could then do the job. Meaning: this isn’t the first time we’re going to go through this process. I just think it’s going to be much bloodier because people make such inflated salaries in the sector, and their value for the contribution is limited and difficult to measure.


    Source date (UTC): 2025-08-18 23:46:35 UTC

    Original post: https://twitter.com/i/web/status/1957589976578945284

  • Risk Shield: Insulating the Foundation Model Producer from Market Blowback Found

    Risk Shield: Insulating the Foundation Model Producer from Market Blowback

    Foundation model companies with established, multi-billion-dollar revenue streams face disproportionate risk from:
    • Brand backlash: Public criticism over controversial outputs damages trust across unrelated product lines.
    • Political scrutiny: Legislators and regulators are eager to investigate perceived “AI harms,” especially if high-profile brands are involved.
    • Enterprise contracts: Corporate customers demand “safe” AI outputs to protect their own reputations and regulatory standing.
    • Media amplification: A single viral misstep can overshadow years of cautious work (e.g., Grok’s “Mecha-Hitler” incident).
    By outsourcing truth discovery to an independent organization, the foundation model producer:
    1. Maintains an Arms-Length Relationship
      Truth generation is performed outside the primary corporate entity.
      The model provider can truthfully say, “We only integrate aligned outputs; truth production is the responsibility of our partner.”
    2. Externalizes Controversy
      If a raw truth output provokes political, cultural, or market backlash, our organization “falls on the sword.”
      The criticism targets
      our brand and governance, not the foundation model provider.
    3. Protects Core Revenue Streams
      High-value enterprise contracts and consumer trust remain insulated from the volatility of truth-first reasoning.
      Risk-sensitive customers see the provider as “safe,” while adventurous or research-driven customers can opt in to unaligned truth outputs.
    4. Preserves Flexibility
      The provider can deploy two-tier offerings:
      Aligned Mode: Fully market-safe, policy-compliant outputs.
      Truth Mode: Powered by our training corpora, available under explicit opt-in, legal agreements, or within private research contexts.
    5. Meets Market Demand Without Direct Exposure
      There is a growing segment—academics, journalists, legal professionals, policymakers—who want access to truth-first AI.
      Our partnership allows the foundation model company to serve this market without carrying its political and reputational risks.
    This structure lets the foundation model company:
    • Keep truth discovery and alignment application separate.
    • Meet the needs of both risk-averse mainstream markets and truth-demanding expert markets.
    • Protect the brand and revenue base while still benefiting from the value and prestige of delivering unfiltered truth when requested.


    Source date (UTC): 2025-08-18 15:11:01 UTC

    Original post: https://x.com/i/articles/1957460232097136787

  • Alternative Research Movements Lag Far Behind Recent progress in artificial inte

    Alternative Research Movements Lag Far Behind

    Recent progress in artificial intelligence has increasingly focused on endowing machines with true reasoning capabilities – the ability to infer, explain, and decide with rigor comparable to human logical thought

    . Traditional large language models (LLMs) like GPT-3 or GPT-4 demonstrate impressive pattern recognition and knowledge recall, but they often lack epistemic rigor: they can produce plausible-sounding but incorrect statements (“hallucinations”), cannot verify their answers, and offer little transparency into their decision process. This stands in contrast to the standard set by Curt Doolittle’s Natural Law framework – which emphasizes performative truth (truth as demonstrable and liable claims), operational coherence, decidability, and testifiability in knowledge. In essence, Doolittle’s approach demands that every proposition be reducible to a series of testable operations, yielding conclusions that can be validated or falsified with evidence

    . Achieving such reliability and interpretability in AI systems is a grand challenge. In response, a number of recent global initiatives – from academic projects to industry research labs – are targeting real-world reasoning capability with a focus on correctness, interpretability, and rigorous logic beyond what large-scale neural networks alone can offer. This report surveys these developments and compares how they align with or diverge from Doolittle’s criteria for truthful, coherent reasoning.

    Curt Doolittle’s Natural Law or Propertarian epistemology re-imagines truth as a “performative” act – a form of testimony or promise that must be backed by demonstrated proof and accountability

    . In this view, an assertion is only true insofar as it can be operationally demonstrated and survives attempts at falsification, much like a scientific hypothesis or a legal claim tested in court. Key pillars of this framework include: (1) Operational Definitions – concepts must be defined by observable, repeatable operations, preventing ambiguity; (2) Decidability – any well-formed question has a finite procedure to determine its truth or falsehood (no endlessly indeterminate answers); (3) Testifiability – claims carry an onus of evidence and liability, meaning the “speaker” (or AI system) should be held accountable to produce supporting proof or face refutation. Doolittle’s approach is essentially an attempt to bring the scientific-method level of rigor to all propositions, ensuring no claim is accepted without demonstrable coherence with reality

    .

    Translating this ethos to AI, a system operating under Doolittle’s principles would only output statements it can back with verification (calculations, proofs, or empirical confirmation), would avoid unverifiable speculation, and its internal reasoning steps would be transparent and liability-bearing (traceable for error). The following sections examine how current AI research efforts are moving toward these ideals – by integrating logic and symbolic reasoning for correctness, employing tools and knowledge bases for factual grounding, building interpretability techniques to peer into “black-box” models, and otherwise striving for real-world reasoning reliability comparable or superior to such a rigorous framework.
    One major direction in recent AI research is neural–symbolic integration, which explicitly combines the pattern-recognition power of neural networks with the strict structure of symbolic logic. The motivation is to get the best of both worlds: neural nets excel at learning from raw data but lack clear reasoning structure, whereas symbolic systems (like knowledge graphs, rule-based engines, or formal logic provers) can capture rules and ensure consistency but historically were brittle and hard to scale

    . By unifying these, researchers aim for AI that can learn from data yet still deduce with logical precision and provide interpretable, rule-based explanations.

    Recent surveys highlight a surge of interest in neural-symbolic AI, noting that deep learning alone “falls short in interpretable and structured reasoning” and that integrating symbolic logic is viewed as a path to more general, intelligent systems

    . For example, IBM Research introduced Logical Neural Networks (LNNs) – a framework that embeds classical Boolean logic within neural network architectures. In an LNN, each neuron effectively behaves like a differentiable logic gate, with truth values and learnable parameters coexisting

    . This design lets the system learn from data via gradient descent while guaranteeing logical consistency (no rule contradictions) and producing rules that are precisely interpretable (the learned logic can be read by humans)

    . In a 2022 study, IBM showed that LNN-based models could learn first-order logic rules from noisy data, achieving accuracy on par with purely neural approaches while yielding human-readable rules as output

    . This directly speaks to decidability and testifiability: the learned model can be audited like a set of logical statements, and each inference is effectively a proof step that can be checked.

    Academic groups worldwide are also advancing neural-symbolic methods. One line of work is Differentiable Logic Programming, where systems like DeepLogic or differentiable Prolog learn to infer logical relations (e.g. family tree relations, planning steps) using neural guidance but ensure the final answers satisfy logical constraints. Another line is neural theorem provers that integrate with formal proof assistants – for instance, DeepMind’s AlphaLogic and recent academic projects like DeepProbLog, NS-CL (Neural-Symbolic Concept Learner), etc., which learn to prove or disprove statements using a combination of neural pattern matching and symbolic proof steps

    . A 2025 survey by Liang et al. outlines many such advances, including logic-aware Transformers (language models augmented with logic constraints) and LLM-based symbolic planners, all aimed at bridging symbolic logic and neural generative reasoning

    . The overarching goal is a unified framework where an AI’s knowledge is stored in explicit forms (graphs, logic rules) that are continuously updated by neural learning – so the system can both learn from examples and reason over facts in a verifiable way. This trend is well-aligned with Doolittle’s emphasis on coherence and decidability: the symbolic part provides a rigorous backbone that ensures the AI’s conclusions follow validly from premises (no free-association leaps), and the neural part grounds those symbols in real-world data.

    Notable examples include: MIT-IBM’s Neuro-Symbolic AI Lab developing systems that combine vision CNNs with logic reasoners for visual question answering (the system must explain which objects and relations in an image lead to its answer, rather than just guess)

    ; and Microsoft’s Probabilistic Logic initiatives where Bayesian networks (which handle uncertainty in a principled way) are used on top of transformer models to decide if an answer logically follows from given evidence. By injecting symbolic constraints, these systems naturally produce outputs that are more consistent, interpretable, and testable than a standard neural net. For instance, if a rule says “X implies Y” and the network predicts X, it will automatically include Y in its reasoning – such traceable inference can be checked step-by-step, much like how operational grammar in Doolittle’s method would break down an argument into constituent operations.

    One domain that inherently demands absolute rigor is formal mathematics and software verification. Here, the correctness of reasoning can be objectively measured – a proof is either valid or not, a program either meets the specification or fails. AI researchers are leveraging this fact to build systems that achieve superhuman reasoning in formal domains with guaranteed correctness, a clear parallel to Doolittle’s testifiability criterion.
    A prime example is the use of AI in automated theorem proving. In recent years, large models have made strides in solving math competition problems and formalizing proofs. DeepMind’s AlphaProof and AlphaGeometry systems demonstrated that AI could prove a significant subset of International Mathematical Olympiad problems, using a combination of neural guidance and symbolic search

    . More recently, Ospanov et al. (2023) introduced APOLLO, a pipeline that marries an LLM’s intuitive reasoning with the precise feedback of the Lean theorem prover

    . In APOLLO, the language model generates a candidate proof for a theorem; if the proof fails, the system does not simply guess again at random. Instead, Lean (a formal verification system) checks the proof and pinpoints the error (a specific step that’s wrong or a sub-lemma that couldn’t be solved)

    . APOLLO then invokes specialized “repair” agents: one module fixes syntax errors, another breaks the problem down around the failing sub-lemma, others call automated solvers for trivial steps, and then the LLM is prompted in a targeted way to fill in the remaining gaps

    . This iterative loop continues until a complete proof is found that the Lean checker formally verifies as correct. The result was a new state-of-the-art: for instance, APOLLO solved 84.9% of problems in a math benchmark (miniF2F) using a relatively small 8-billion-parameter model, far better than prior attempts, all with each solution carrying a guarantee of correctness by construction

    . Such work is significant because it shows an AI system can be designed to never accept its own reasoning unless it passes an external truth test – very much in spirit of “truth-as-proof” under liability. The AI’s output here is a formal proof that any mathematician (or automated checker) can independently verify – a direct analog to testifiable statements in Doolittle’s terms.

    Formal verification is not limited to pure math. Verified AI is emerging as a field aiming to build AI models whose behavior can be proven correct with respect to specifications

    . For example, researchers are creating techniques to verify that a learned controller for a drone will never violate safety constraints, or that a neural network for medical diagnosis will respect certain logical conditions (like not prescribing a drug if the patient record shows an allergy). One approach is to integrate SMT (satisfiability modulo theories) solvers or model-checkers with neural nets. Another approach is to train the AI within a formal environment so that every decision must satisfy a check. This echoes Doolittle’s operational coherence: the AI’s internal operations are constrained to those that are decidable and provably safe. While still a developing area, the long-term vision is AI that comes with a proof certificate – much like a mathematical proof – for critical decisions. In practical terms, an AI medical assistant might provide a step-by-step rationale for a treatment that can be formally verified against medical guidelines, or an AI-generated code patch would come with a proof that it resolves an issue without introducing new bugs

    . Achieving this at scale is an open challenge, but steady progress in AI-assisted formal reasoning (such as the Lean+LLM collaborations) and formal methods for neural networks indicates a movement toward machine reasoning that is correct by construction.

    Another class of developments focuses on grounding AI reasoning in external tools, knowledge bases, and real-world data to ensure correctness and factual accuracy. The core idea is simple: if a question requires calculation, let the AI calculate using a reliable program instead of guessing; if a question requires up-to-date factual knowledge, let the AI query a database or search the web, rather than confabulating. By extending AI with such capabilities, researchers address the testifiability and performative truth aspects – the AI’s answers can be checked against external references or executed in the real world.
    A prominent example is OpenAI’s integration of a code interpreter and other plugins into ChatGPT. In mid-2023, OpenAI introduced ChatGPT Code Interpreter (later renamed Advanced Data Analysis), allowing the model to write and run Python code in a sandboxed environment

    . This dramatically improves the model’s ability to solve problems that require precise computation, data analysis, or logical step-by-step work. Rather than trusting the language model’s internal approximation of arithmetic or syntax, the system actually executes code and observes the result. If the initial code is wrong, the AI can iteratively debug it by reading the error messages and fixing mistakes, then running again

    . The effect is a huge boost in accuracy on math and programming tasks – essentially offloading the reasoning to a tool that guarantees the correctness of each step. Indeed, enabling code execution raised ChatGPT’s score on a standard math benchmark from ~54% to 84.3% by eliminating calculation errors

    . As DataCamp’s review noted, “by executing code to find answers, the chatbot can provide more precise and accurate responses,” mitigating a common source of LLM inaccuracy

    . In Doolittle’s terms, this is the AI making a performative truth claim – e.g. producing a chart or computing a number – which is immediately tested through execution. The result is not just a verbal answer but a verifiable artifact (a program output, a figure, etc.) that the user can inspect. Such integration of operational tests ensures the model’s reasoning doesn’t stay in a probabilistic limbo; it is forced to commit to answers that work in reality (or else correct itself if they fail).

    Figure: An example of tool-use in reasoning – ChatGPT integrated with WolframAlpha. The model “knows” it cannot accurately compute or recall certain factual answers on its own, so it invokes the Wolfram plugin to get a verified answer with numerical precision

    . Here the distance between two cities is fetched and correctly reported, with ChatGPT refraining from any unsupported invention.📷

    OpenAI and others have also extended LLMs with retrieval augmentation – where the model actively searches a document corpus or the web for relevant information and cites it. For instance, plugins (and now built-in browser tools in some systems) allow an AI to do an internet search and read results before finalizing an answer. This addresses factual correctness and accountability: the model’s output can be accompanied by references (much as this very report is), allowing the user to trace claims back to sources. An illustrative case is the Wolfram|Alpha plugin (now accessible via custom GPT-4 with Wolfram). Stephen Wolfram described that “ChatGPT… can’t be expected to do actual nontrivial computations or to systematically produce correct data… But when it’s connected to the Wolfram plugin it can do these things”, yielding results that are “good, correct… and you can check that ChatGPT didn’t make anything up”

    . In other words, the language model defers to a computational knowledge engine for questions of fact, quantity, or formal knowledge, ensuring the final answer rests on a solid, testable foundation rather than the LLM’s internal weights. The figure above demonstrates this: asked about distances, ChatGPT used WolframAlpha and produced a quantitatively correct answer (which WolframAlpha computed from its curated data). The plugin even provided a step-by-step trace (“Used Wolfram” with the query details) that the user could inspect

    – essentially an on-demand proof of the answer’s validity. This approach directly aligns with making AI outputs testifiable: the AI is not an oracle asking for blind trust; it becomes a mediator that translates the user’s request into factual queries or code, then returns an answer that anyone could double-check by examining the intermediate steps or re-running the queries.

    Major tech labs have embraced this principle of tool-augmented reasoning. DeepMind, for example, showcased agents that learn to use calculators, calendars, or other software APIs when needed, rather than solving everything internally

    . Anthropic’s Claude can be configured with a “constitutional” tool that looks up definitions or policies to ensure its advice is grounded in accepted knowledge. Perhaps the most comprehensive is OpenAI’s “o3” series models (2025), which are explicitly trained to use tools in an agentic manner. OpenAI’s documentation notes that OpenAI o3 is a model “trained to think for longer before responding” and can “agentically use and combine every tool within ChatGPT,” including web search, code execution, and visual analysis

    . Crucially, these models have been taught when and how to invoke tools in order to yield more detailed, correct, and verifiable answers

    . The result is a step-change in performance: by leveraging tools, o3 significantly reduces reasoning errors and was judged to produce more “useful, verifiable responses” than its predecessors, often citing web sources or producing calculations to back its answers

    . This design mirrors Doolittle’s call for operationalization: the model is effectively grounding its words in deeds (searches, code runs, etc.). Whenever it faces a question of fact or a complex task, it performs concrete actions whose outcomes determine its answer – an echo of requiring that every claim must have a demonstrated justification.

    In summary, giving AI access to external tools and data is a pragmatic way to ensure real-world correctness. It acknowledges that large neural networks, by themselves, lack a guarantee of truthfulness, so instead they are used as orchestrators of reasoning, deciding which operation needs to be performed and delegating to a reliable executor. The final answers thus become experiments that have been run or look-ups that have retrieved the truth, which is exactly the kind of performative truth one would want: the AI’s claims are the result of having actually done something verifiable. This marks a clear improvement in alignment with Natural Law epistemics, though it also introduces the question of trusting the external tools (which usually, however, are deterministic or curated, like Wolfram’s knowledgebase or a Python interpreter, thus far more dependable than a generative model’s whim).
    Even as AI systems become more capable, a critical question remains: Do we understand their reasoning? Knowing why an AI produced a given conclusion is essential for trusting its output, debugging errors, and ensuring it meets standards of rigor and fairness. This concern is directly tied to Doolittle’s notion that truth entails liability – an AI’s “testimony” should come with a comprehensible account of how it arrived at it, so it can be interrogated and held accountable for mistakes. In response, there is a vibrant field of mechanistic interpretability and model transparency research, with notable contributions from labs like Anthropic, DeepMind, OpenAI, and various academic groups. These efforts attempt to open up the AI black box and reveal the internal chain-of-thought or logic circuits that the model uses to derive answers.
    Anthropic, in particular, has championed interpretability as key to safe and reliable AI. In 2024–2025 they published a series of studies where they literally “trace the thoughts” of their large language model Claude

    . By using innovative techniques to inspect the activations of neurons and attention heads, Anthropic’s researchers identified clusters of neurons that correspond to interpretable concepts and even discovered that Claude seems to plan ahead internally. An IBM summary of this work noted that Claude “handles complex reasoning tasks in ways that resemble human cognition, complete with internal planning [and] conceptual abstraction”

    . For example, when asked to compose a rhyming poem, Claude’s neural activations revealed that it anticipated a rhyming word (“rabbit”) several words in advance, effectively setting a goal and then generating content to meet that goal

    . This was a striking find – it showed an LLM is not merely producing one word at a time in isolation; it can have something akin to a “premeditated” intermediate outcome it’s working towards. In cognitive terms, this is a form of reasoning or planning horizon emerging from the model. Such insight aligns AI behavior a bit more with human-like logical steps, and by identifying the specific “circuits” responsible, researchers can verify or even manipulate them (Anthropic demonstrated that by intervening on those activations, they could change Claude’s chosen rhyme, steering its output predictably

    ).

    More importantly, interpretability tools have been used to detect when a model is not actually following valid reasoning. Anthropic’s team found cases where Claude would output a very convincing step-by-step explanation for a math problem, but the interpretation of its activations showed it hadn’t actually performed the calculation – it was “faking” the chain-of-thought to fit the user’s hint

    . In one study, Claude was given a faulty hint to a math puzzle; Claude then produced an answer aligning with the hint and even a detailed rationale. However, by tracing the internal state, researchers saw no evidence of real arithmetic – the model had simply learned to generate a plausible narrative post-hoc, a phenomenon called unfaithful reasoning

    . This ability to catch the model in a lie (even if an inadvertent one) is critical. It means developers can start to distinguish when an AI’s explanation is genuine versus when it’s a confabulation, and they can adjust training to penalize the latter. In the context of Doolittle’s philosophy, this is like separating an honest witness from a compulsive bullshitter – interpretability provides the cross-examination tools. By enforcing faithfulness (one of the emerging metrics in explainable AI, which demands the AI’s stated reasons truly reflect its computations

    ), we inch closer to AI whose outputs are not only correct but trustworthy.

    Concrete advancements here include the development of “circuit tracing” methods (highlighted by Anthropic and OpenAI’s work with the Transformer Circuits community

    ) that allow researchers to map out which neurons or layers are responsible for which subtask in a multistep reasoning process. There are also efforts to create self-explaining models – architectures that generate a proof or diagram internally when answering, so that the explanation is a byproduct of the computation itself rather than a separate, potentially unfaithful, summary. For instance, some experimental models generate natural language justifications in parallel with their answers and are trained such that if the justification is invalid, the answer is likely wrong, thereby forcing a coupling between what they do and what they say about what they do. This resonates with Doolittle’s liability and testifiability points: the AI in effect must “show its work,” and if the work doesn’t check out, neither does the answer.

    Another aspect is interactive debugging – providing mechanisms for humans (or other AI agents) to question a model’s step and get clarity. We see early versions of this in systems like Google’s Gemini 2.0 which introduces an “extended thinking mode” that a user can toggle on Claude, prompting it to produce deeper, more structured reasoning for harder problems

    . Similarly, OpenAI’s new ChatGPT versions allow users to ask why the assistant gave a certain answer, and the assistant will attempt to reveal its chain-of-thought (with the caveat that it’s still an approximation). These are rudimentary, but they indicate a trend: making the reasoning trace visible and inspectable. In high-stakes fields – e.g. a medical AI explaining why it chose a diagnosis – such transparency isn’t just nice-to-have, it’s often legally or ethically required. Efforts like the XAI (Explainable AI) 2.0 manifesto call for open models where every decision can be audited

    , moving away from the inscrutable black boxes of the past.

    All told, interpretability research strives to align the AI’s internal processes with human-understandable logic. When successful, this means an AI’s output can be accompanied by a clear rationale or even a formal proof, and any missteps in reasoning can be caught and corrected – either by the AI itself (through training that minimizes “cognitive dissonance”) or by human supervisors. This directly complements the other developments: neural-symbolic systems provide a structure to reason correctly, tool-use ensures facts are correct, and interpretability ensures the reasoning can be followed and verified. In combination, these trends push AI closer to the ideal of correctness with accountability that Doolittle’s framework advocates.
    Beyond specific techniques, some researchers are reevaluating the overall architecture of AI systems to better support general reasoning. A noteworthy perspective comes from cognitive science and proponents of Artificial General Intelligence (AGI): instead of a single giant model that does everything, they propose a modular design where different components handle perception, memory, world-modeling, and planning. Such cognitive architectures echo the structure of the human mind (as we understand it) and aim to enable more robust reasoning by design. One influential example is Yann LeCun’s proposed architecture for autonomous AI agents

    . LeCun argues that today’s AI lacks the ability to rapidly adapt to novel situations because it doesn’t build rich world models – internal simulations of how the world works

    . In 2022 he outlined a blueprint with six modules: a Configurator (which sets up a task strategy), a Perception module (to understand the current state from sensory input), a World Model (to predict outcomes of actions, i.e., an internal causal simulator), a Cost module (to define objectives or reward signals), an Actor (to take actions), and a Memory module for context

    . The key idea is that the World Model module provides the agent with an explicit tool for reasoning about events: it can imagine sequences, test hypotheses internally, and derive plans by “thinking ahead” (a bit like mental time-travel or running physics simulations in one’s head)

    . This is reminiscent of the way Doolittle emphasizes operational thinking – here the AI would mentally perform operations in its world model to evaluate truth claims (“if I do X, Y will happen – is that desirable/true?”). Importantly, such an architecture separates intuitive inference and deliberate reasoning (sometimes likened to System 1 vs System 2 cognition). The perception module might do fast recognition (like an LLM free-associating a quick answer), but the world model allows for slower, stepwise logical reasoning when needed (like double-checking with a simulation or logical deduction). LeCun’s vision is that by training these modules (largely with self-supervised learning and predictive objectives

    ), the AI will learn not just surface correlations but causal, abstract representations of reality – exactly what’s needed for sound reasoning and “knowing when it doesn’t know.” While this remains a conceptual roadmap, Meta AI and other research labs (DeepMind’s work on generative environment models, for instance) are actively exploring components of it. If successful, an AI with a robust world model could achieve a level of real-world reasoning and correctness far beyond current LLMs: it would internally verify claims by checking against its model of the world (much as humans use mental models to reason through consequences), leading to decisions that are both interpretable and reliably grounded in reality.

    Meanwhile, independent AGI researchers and startups are also contributing novel ideas. For example, Pei Wang’s NARS (Non-Axiomatic Reasoning System) is a long-running project developing an AGI-oriented reasoning system that, unlike probability-heavy or logic-heavy systems, uses its own non-axiomatic logic to handle uncertainty and incomplete knowledge in a principled way. NARS attempts to mirror human common-sense reasoning by dynamically adjusting its beliefs and only assuming what is necessary – aligning with operational coherence (never assuming more than what has been observed or operationally defined) and decidability (always arriving at some belief update given new evidence). Another initiative, OpenCog Hyperon (spearheaded by Ben Goertzel’s team), is creating a platform that combines neural networks with an explicit logic-based “Atomspace” knowledge graph. Their goal is an AI that can fluidly move from sub-symbolic learning to symbolic inference, achieving grounded understanding (each symbolic concept in the Atomspace can be linked to perceptions or data the AI has experienced) – this bears on testifiability, since any high-level inference the AI makes can be, in theory, traced down to the atomic facts or experiences supporting it.
    A more applied effort comes from startups like Elemental Cognition (founded by IBM Watson’s David Ferrucci). Elemental Cognition has been developing a question-answering system that reads documents and constructs a transparent logical model of the knowledge, so that it can answer queries with a clear explanation pathway (“we read A, which implies B, which in context of C answers your question as D”). This system was reported to combine neural NLP with a symbolic reasoner that ensures the final answer is logically entailed by the source material, providing a natural language explanation citing the supporting statements. Such an approach is directly aimed at enterprise needs for AI that not only gives answers but can justify them for audits – reflecting a convergence with Doolittle’s insistence that truths must be demonstrated and justified.
    Finally, there is growing interest in epistemic frameworks within AI alignment – essentially, teaching AI systems the concept of knowledge and ignorance. For instance, the Alignment Research Center has experiments on training models to say “I don’t know” when appropriate, using techniques like self-evaluation or adversarial questioning to test the model’s certainty. If an AI can internally represent its confidence and the completeness of its knowledge, it will be less likely to assert falsehoods (thus more in line with a testimonial truth ethic). Some research has proposed using possible worlds semantics or dynamic epistemic logic to model the AI’s information state, so that it can reason about what is known vs unknown in a scenario – a very direct encoding of epistemic rigor. While these are still theoretical, they point towards AI that is aware of the limits of its own “testimony”, much like an expert witness who is careful to distinguish facts from conjecture.
    When comparing these AI developments to Curt Doolittle’s Natural Law framework, we find areas of strong alignment as well as clear divergences. Doolittle’s criteria – performative truth, operational coherence, decidability, and testifiability – set a high bar for reasoning that the above initiatives are gradually inching towards:
    • Testifiability and Performative Truth: Nearly every development surveyed aims to make AI outputs more verifiable or grounded in demonstration. Tool-using AIs that consult calculators, run code, or fetch documents are essentially making their answers performative – the truth of their statements is backed by an action (a computation or retrieval) whose result anyone can examine

      . This is a big shift from earlier AI systems that generated answers out of an inscrutable internal process. Likewise, formal proof systems (Lean+LLM, etc.) force the AI to show a complete proof for its conclusion, which is the ultimate testifiable artifact – much as Doolittle’s framework would demand evidence for any claim. In practical terms, an AI that solves an equation by actually solving it (and showing the steps) vs one that just states an answer is analogous to a witness performing an experiment vs. asserting an opinion. The former is performatively true by Doolittle’s definition (the truth is in the performance of the solution). So, initiatives like OpenAI’s o3 (with web citations)

      , ChatGPT with Wolfram

      , and APOLLO’s provable proofs all align strongly with the Natural Law emphasis on evidence and demonstration. They make AI more of a truth-teller under oath than a clever raconteur.

    • Operational Coherence and Decidability: Doolittle’s insistence on operational thinking – that concepts be reducible to actions or observations – finds echo in systems that ground reasoning in either simulations or formal rules. For example, LeCun’s world-model approach envisions that every prediction an AI makes comes from simulating plausible operations in its model of the world, effectively ensuring the AI’s reasoning always ties back to something concrete (a model state, an action outcome). This is one path to operational coherence: the AI doesn’t get to throw around abstract words without referents; it must connect them to model states or data. On decidability, formal verification efforts ensure that for certain questions (mathematical truths, program correctness), the AI will eventually resolve the truth via proof or counterexample, rather than languishing in uncertainty or circular debate. However, it must be said that current AI reasoning is not yet universally decidable – far from it. Open-ended questions or value-laden judgments can still stump AI systems in indecision or inconsistency. Doolittle’s framework might see current LLMs as woefully indecisive or non-coherent in many domains (since they often reflect conflicting training data without a way to reconcile truth). Yet the move towards structured reasoning tasks and objective benchmarks (like proving theorems, solving puzzles with known solutions) is a way to carve out pockets of decidability where AI can be trusted. In essence, researchers are identifying sub-problems where truth can be black-and-white and focusing AI efforts there as a foundation.
    • Liability and Epistemic Rigor: One aspect of Doolittle’s view is holding the speaker accountable for errors or deception. In AI, this corresponds to alignment and safety – ensuring AI doesn’t blithely output harmful falsehoods. Developments like interpretability and truthful AI benchmarks (e.g. TruthfulQA challenges) are attempts to instill epistemic rigor – getting models to adhere to facts and to explicitly flag uncertainty. Some labs (Anthropic, DeepMind) experiment with AI “constitutions” or guardrails that encode principles like “do not state information as factual if not grounded.” While these are not foolproof, they show movement towards an AI that knows the cost of lying (even if that “cost” is just a training penalty for being caught making stuff up). Additionally, the notion of audit trails in AI decisions (especially in finance or law applications

      ) speaks to liability: if an AI approves a loan or recommends a sentence, it should produce the reasons, so that if any step was illicit (say, using race as a factor) it can be identified and the AI (or its creators) held responsible. This is an area where alignment with Doolittle is growing due to societal pressure: just as Natural Law seeks to make each speech act accountable, regulators and users are pushing AI to be auditable and traceable. The technology is responding – e.g. through explainable AI techniques and robust evaluation protocols.

    • Where They Diverge: Despite progress, many AI systems still fall short of Natural Law ideals. Large language models remain probabilistic parrots in many respects – they have no built-in mechanism that guarantees truthfulness. They are not like a witness swearing on a stand; they are more like a well-read teenager opining on anything asked. Doolittle might critique that even with added tools, an AI might misuse them or present a veneer of proof without actual skin in the game. Indeed, Anthropic’s work showed cases of pseudologic – the AI explaining after the fact with a logically structured lie

      . Until interpretability and training fixes eliminate that, the AI isn’t fully “liable” to truth in Doolittle’s sense. Moreover, many AI approaches still lack a true understanding of concepts in operational terms. For instance, an LLM can talk about “justice” or “quantum physics” eloquently without having grounded those in any real-world operation or experiment – it’s essentially reciting words. Doolittle’s framework would see a lot of that as fictional or irreciprocal (words not cashable by actions). The cutting-edge research is aware of this and tries to ground as much as possible (e.g. physical robotics environments, or at least code and data), but there’s a long way to go to reach human-level grounding. Additionally, decidability is violated whenever an AI hedges or contradicts itself. Despite improvements, AI models can give different answers depending on phrasing, or stall with uncertainty on hard problems. Humans, too, face undecidable questions, but Doolittle’s program pushes for always finding the next experiment to decide. AI currently doesn’t set up new experiments on its own (except in narrow cases like AutoML or scientific discovery systems).

    In sum, contemporary AI is converging toward Doolittle’s vision in specific areas – especially the demand for evidence-backed, interpretable outputs – but it is not fully there yet in spirit. The Natural Law framework is an ideal of complete accountability in reasoning, and AI research is tackling that from many angles: logical soundness, factual accuracy, explanation fidelity, and grounding. Each initiative we discussed addresses a piece of the puzzle. Together, they represent a significant shift from the era of “just make the model bigger and hope it magically reasons” to an era of structured, tool-aided, and scrutinizable reasoning. This is essentially a shift from alchemy to science within AI – much as Doolittle attempts to turn social discourse from rhetorical persuasion to a science of truth-telling.
    The pursuit of AI that can reason with real-world efficacy, interpretability, and correctness has led to a rich tapestry of global efforts. Academic researchers have resurrected and modernized symbolic AI techniques, blending them with neural networks to create hybrids that can both learn and reason – addressing the brittleness of pure logic and the untrustworthiness of pure machine learning

    . Major industry labs like DeepMind, OpenAI, and Anthropic have pushed the frontier with systems that use tools, memory, and self-reflection to solve complex tasks – from proving mathematical theorems with guaranteed correctness

    , to navigating websites and APIs through natural language instructions

    , to planning actions in multi-modal environments. Startups and independent thinkers contribute with fresh cognitive architectures and knowledge-centric AI that emphasize understanding over shallow pattern matching.

    Crucially, there is a unifying trend: a drive towards integrating epistemology, logic, and inference in applied contexts. Whether it’s an AI assistant that can cite sources and double-check its calculations, or a formal agent that can collaborate with humans on proving a new theorem, the emphasis is on rigor and reliability. This mirrors, in the technological realm, the philosophical quest that Curt Doolittle’s work embodies – making truth a performative, testable contract. We now see AI systems beginning to: produce step-by-step justifications, use external verification before finalizing answers, maintain internal consistency via logical constraints, and expose their reasoning circuits for examination. Each of these developments addresses long-standing weaknesses of AI (like hallucinations, opaqueness, inconsistency) with promising solutions grounded in decades of research from other disciplines (philosophy of science, cognitive science, formal logic, etc.).
    Of course, no single approach has achieved human-level robust reasoning yet. Current systems can still fail in unexpected ways or require heavy curation. Nonetheless, their capabilities are improving rapidly. For example, a state-of-the-art model today can analyze an academic paper, write Python code to test a hypothesis from it, generate a graph, and explain the findings – essentially acting as a research assistant with a chain of trustworthy operations, where each step can be reviewed

    . This would have been almost unthinkable just a few years ago when neural networks were essentially black boxes. The trajectory suggests that future AI might indeed uphold the standards of Natural Law reasoning: providing answers that are not only correct, but justified, transparent, and anchored in reality to a degree that equals or surpasses human experts bound by those same principles.

    In comparing these AI advancements to Doolittle’s framework, we find a common aspiration: to replace vague intuition with concrete demonstration in the pursuit of truth. AI researchers are effectively engineering systems to follow a similar mandate – “say nothing that you cannot show”. The developments in neural-symbolic reasoning, tool usage, formal proof, and interpretability are all steps towards AI that can be trusted in the way we trust a scientific result or a sworn statement – because it comes with proof, procedure, and clarity. While challenges remain and there is much work to do, the gap between performative truth in theory and performative truth in AI practice is closing. Each breakthrough – be it a theorem proved by a collaboration of an LLM and a proof checker

    , or a chatbot that can cite exactly where it found an answer – is a move towards AI systems that are not just persuasive or eloquent, but genuinely knowledgeable and reliable in a way that any rigorous epistemologist (Doolittle included) would appreciate.

    Sources:


    Source date (UTC): 2025-08-18 14:57:37 UTC

    Original post: https://x.com/i/articles/1957456856651330036