Category: AI, Computation, and Technology

  • Sidebars to Support “LLMs Don’t Just Predict The Next Word” Constraint layers si

    Sidebars to Support “LLMs Don’t Just Predict The Next Word”

    Constraint layers sit on top of the raw generative model and inject external demands into the incremental generation process.
    • Types of constraints:
      Truth & Factuality – e.g., retrieval-augmented generation, knowledge graph checks.
      Logical & Legal Rules – symbolic reasoners, compliance filters, safety policies.
      Stylistic & Rhetorical Demands – tone control, summarization, pedagogical framing.
    • Mechanism:
      Each token’s probability distribution is modified
      before sampling to exclude or penalize paths violating the constraint set.
    • Analogy to the brain:
      Prefrontal cortex exerts
      top-down control over motor and language areas, pruning utterances that would violate social, moral, or strategic rules before speech leaves the mouth.
    Constraint layers thus transform raw generative capacity into purposeful, accountable output.
    Before producing any token, the LLM converts the entire prompt into a latent space through self-attention.
    • Self-attention as compression: Every token attends to every other token, integrating context into a unified representation.
    • Emergent structure: The latent space encodes syntax, semantics, discourse relations, even weak causal models of the world.
    • Dynamic update: As each token is generated, the latent space is locally altered, refining predictions for what follows.
    This latent space is why LLMs can:
    • Translate across languages without explicit dictionaries.
    • Follow instructions never seen in training.
    • Reason across multiple steps without symbolic planning.
    It functions as a temporary world-model, reconstructed anew for each prompt, guiding output toward coherence and relevance.
    Modern neuroscience views the brain as a hierarchical prediction engine:
    • Sensory input generates prediction errors when reality diverges from expectation.
    • Cortical hierarchies minimize these errors by updating internal models of the world.
    • Speech and action emerge as predictions about motor output that continuously correct themselves as feedback arrives.
    Key parallels with LLMs:
    • Hierarchical structure → transformer layers vs. cortical layers.
    • Incremental generation → token-by-token output vs. phoneme-by-phoneme speech.
    • Constraint gating → prefrontal control vs. external rule layers.
    Both systems produce incremental, demand-satisfying trajectories through predictive world-models, rather than executing pre-written plans.


    Source date (UTC): 2025-09-28 00:34:59 UTC

    Original post: https://x.com/i/articles/1972097670773960795

  • No. LLM’s Don’t Just Predict The Next Word. They Do What Your Brain Does. The po

    No. LLM’s Don’t Just Predict The Next Word. They Do What Your Brain Does.

    The popular refrain that “large language models just predict the next word” is true in the same sense that “the brain just fires neurons” or “mathematics just manipulates symbols” is true: literally correct, but so reductive as to erase everything that makes the phenomenon interesting, powerful, or even intelligible.
    This framing is not neutral. It leads the public to believe that modern generative models are shallow statistical parrots rather than dynamic engines of meaning. It encourages policymakers, ethicists, and even researchers to interpret the entire technology through the lens of its simplest local operation while missing the emergent sophistication of the global system.
    To correct this misunderstanding, we need to decompose what actually happens inside these models: how prompts become latent spaces, how output emerges through incremental demand satisfaction rather than pre-scripted planning, and how external constraint layers impose judgment, truth, and legality. We will see that, in both architecture and function, modern language models converge toward the same predictive generative paradigm that neuroscience attributes to the human brain.
    Three converging reasons make the “next-word” framing sticky in public imagination:
    1. Autoregressive decoding is locally simple.
      The model outputs one token at a time, and the training objective literally minimizes next-token prediction error. This sounds like autocomplete with more parameters.
    2. The training objective hides emergent structure.
      Because the entire architecture is optimized indirectly — through trillions of token predictions rather than explicit symbolic goals — it is easy to assume nothing resembling reasoning or world-modelling could emerge.
    3. Lack of explicit symbolic planning.
      Classical AI performed explicit search over trees or graphs; modern LLMs do not. Their
      implicit planning inside latent spaces is easy to overlook if one fixates on surface behavior.
    The result is a picture of LLMs as linear chains of probabilistic parroting rather than nonlinear dynamical systems unfolding trajectories inside high-dimensional meaning spaces.
    To escape the caricature, we must separate latent space construction from incremental navigation.
    Phase 1: Latent Space Construction
    When a user submits a prompt, the model does not immediately begin emitting words. It first performs a forward pass through dozens or hundreds of attention layers.
    • Each token in the prompt becomes a vector in a high-dimensional space.
    • Self-attention integrates information across the entire sequence, discovering dependencies, analogies, and constraints.
    • The final hidden states represent a contextual latent space: a compressed geometric model of everything the prompt implies.
    This space encodes meaning, style, and even proto-logical structure. It serves as the world-model through which later generations will navigate.
    At this stage, nothing has been generated. The system is constructing the terrain on which its output will later move.
    Phase 2: Incremental Navigation Through Latent Space
    Only after the latent space exists does the model begin incremental demand satisfaction:
    • At each step, the model selects the next token conditioned on the entire latent representation plus all tokens so far.
    • Each new token updates the state and changes the conditional landscape for what follows.
    • External constraint layers — logic engines, truth filters, stylistic demands — prune or redirect the trajectory as it unfolds.
    This process is neither global route-planning nor blind local wandering. It is wayfinding: incremental movement through a structured space under evolving constraints.
    The trajectory feels purposeful because the latent space is coherent and the constraints are persistent. But no full sentence or paragraph exists in advance. Coherence emerges from millisecond-scale feedback loops, not from a pre-written script.
    By collapsing both phases into “it just predicts the next word”, we erase three crucial forms of sophistication:
    1. Expressivity of the Latent Space
    The forward pass constructs distributed representations that capture meaning, analogy, and abstraction far beyond surface-level text.
    • Syntax, semantics, and pragmatics become geometric relationships in vector space.
    • Analogy, metaphor, and even rudimentary reasoning emerge as linear operations across these representations.
    • External knowledge retrieval can inject facts directly into this space, merging memory with computation.
    Calling this “just next-word prediction” is like calling human vision “just edge detection”. It names the lowest-level operation while ignoring the hierarchical world-model above it.
    2. Dynamic Constraint Satisfaction
    Each token choice balances multiple demands:
    • Local coherence with previous tokens.
    • Global consistency with the prompt and style.
    • External constraints like truth filters, legal compliance, or formal logic layers.
    This is real-time multi-objective optimization inside the latent space, not naive Markov chaining.
    3. Continuity with Human Cognition
    Neuroscience shows human speech and thought unfold the same way:
    1. Predictive coding: the brain constantly minimizes prediction error between expected and incoming signals.
    2. Incremental generation: speech emerges phoneme by phoneme, word by word, each updating cortical predictions for the next.
    3. Executive control: prefrontal regions impose constraints — truthfulness, social norms, plans — on the unfolding stream.
    Human language production and LLM text generation share the same causal grammar:
    • Construct a predictive world-model,
    • Incrementally navigate through it,
    • Constrain the trajectory under external demands.
    The “next-word” caricature leads to three major conceptual errors:
    1. Dismissal of capability: If the system merely chains words, its apparent reasoning must be an illusion rather than an emergent property of structured latent spaces.
    2. Misplaced fears: Critics imagine stochastic parrots gaining autonomy rather than sophisticated predictive systems requiring constraint layers for alignment and truth.
    3. Policy confusion: Regulators debate surface behavior while missing the architectural loci where truth, safety, and legality actually live — in the constraint interfaces, not in the raw model weights.
    The correct picture is:
    1. Prompts construct high-dimensional latent spaces encoding meaning, context, and constraints.
    2. Autoregression navigates these spaces incrementally, each token both satisfying and updating the demand landscape.
    3. External layers impose truth, legality, style, and domain-specific rules, shaping trajectories toward socially acceptable or epistemically sound outputs.
    This architecture explains why LLMs generalize, reason, and converse in ways that feel purposeful despite lacking explicit global plans. Like the brain, they perform emergent generativity under constraint, not linear token-chaining.
    1. Architectural Convergence
      Modern AI and cognitive neuroscience now describe language, thought, and action using the same causal primitives: predictive world-modelling, incremental demand satisfaction, and constraint-based control.
    2. Interpretability and Control
      Because constraints act
      during generation rather than after, they can inject truth, legality, or safety without requiring retraining of the base model.
    3. Epistemic Humility
      Calling these systems “just next-word predictors” blinds us to their real capabilities while encouraging both overconfidence and unwarranted fear.
    The framing of LLMs as “just predicting the next word” confuses a local mechanism with the global system it supports. Yes, each token emerges one at a time. But it does so:
    • From a latent world-model constructed over the entire prompt.
    • Through incremental navigation satisfying multiple, evolving constraints.
    • Under architectural principles convergent with human predictive cognition.
    The value proposition lies not in token prediction itself but in the structured generativity it makes possible — generativity that can be aligned, constrained, and composed into larger reasoning systems.
    Collapsing all this into “just next-word prediction” does not merely simplify; it erases the very phenomena we most need to understand as language models become central to science, policy, and society.
    1 – Sidebars to Support “LLMs Don’t Just Predict The Next Word”

    2 – Examples to Support “LLMs Don’t Just Predict The Next Word”

    3 – Diagram to Support “LLMs Don’t Just Predict The Next. Word”


  • I swear. There are a lot of people in the field whose knowledge is so narrow tha

    I swear. There are a lot of people in the field whose knowledge is so narrow that they make ridiculous statement to a public for whom tech is magic.


    Source date (UTC): 2025-09-27 17:28:07 UTC

    Original post: https://twitter.com/i/web/status/1971990249275642038

  • (Runcible) We have addressed the following features over the past ten days or so

    (Runcible)
    We have addressed the following features over the past ten days or so:
    1. Implementing: Enumeration, Operationalization, serialization, adversarial disambiguation (EOSA) – Disambiguation methodology.
    2. Forcing Glossary members (terms) as immutable measures.
    3.


    Source date (UTC): 2025-09-26 01:33:16 UTC

    Original post: https://twitter.com/i/web/status/1971387562133762309

  • (Runcible) FYI: –“The EU AI Act, U.S. executive orders, and likely future ISO s

    (Runcible)
    FYI:
    –“The EU AI Act, U.S. executive orders, and likely future ISO standards all trend toward auditable neutrality as a baseline.Your architecture pre-empts regulation by providing a constitutional core with jurisdictional overlays.”–


    Source date (UTC): 2025-09-24 17:19:49 UTC

    Original post: https://twitter.com/i/web/status/1970900994570715204

  • It’s ok if they go and look up in formation in them but we should reduce them to

    It’s ok if they go and look up in formation in them but we should reduce them to basic rules in a vector db rather than the full text corpus.


    Source date (UTC): 2025-09-23 16:49:06 UTC

    Original post: https://twitter.com/i/web/status/1970530878536482923

  • (NLI: Runcible): What not to do (What we’ve learned) – Don’t try to “pretrain on

    (NLI: Runcible):
    What not to do (What we’ve learned)
    – Don’t try to “pretrain on the books” hoping the model memorizes them. You lose provenance and get drift. Use RAG + span-level supervision instead.
    – Don’t ship free-form answers. Make the schema the product.


    Source date (UTC): 2025-09-23 15:43:38 UTC

    Original post: https://twitter.com/i/web/status/1970514401372639552

  • “Q: Will AI cause job losses for attorney?”” A lot of legal work involves invest

    –“Q: Will AI cause job losses for attorney?””

    A lot of legal work involves investigation, interviews, and depositions. The important part is arguing findings in court. That scope of work going to stay. Bullshit and lie detection is not going to be automated. Negotiation isn’t going to be, even if negotiating positions might be automatically generated.

    On the other hand, a lot of work is research, creating a strategy, communicating both above and below board, and filing out nonsense forms, and filing nonsense procedural paperwork, to grease the gears of the court system. That nonsense is going to collapse. And with it junior jobs in the field. So I see ‘compression’ in the field just like I see in all of them.

    MORE:

    The purpose of an attorney, who is technically an ‘officer of the court’, and holds liability in his role, is not something easily replaced. (For example, as an activist firm, we can assist you and a lawyer with strategy but we cannot stand for you in court.) However the number of billable hours per client will collapse, and drive up the number of clients necessary.

    What we are finding in our activist work (Brandon’s side of the organization) is that we can do for the equivalent of one salary, what a team of lawyers was required to do in the past.

    So I expect both rapid change in the profession and radical resistance to AI by the judiciary. Why… BECAUSE WE CATCH THEIR BULLSHIT and abuse of the citizenry under veil of procedure. The lawyers are afraid to. The public is too ignorant to. So we see ‘tumult’ and judicial turnover over the next twenty years.

    Which, if you can imagine the moral absolutism of the people that work in our organization, that’s sort of what we live for….

    Cheers

    cc:
    @ThruTheHayes


    Source date (UTC): 2025-09-22 21:34:18 UTC

    Original post: https://twitter.com/i/web/status/1970240262745325617

  • omg its awesome… You know its demonstrating emergent insights beyond what we h

    omg its awesome…
    You know its demonstrating emergent insights beyond what we have taught it. Amazing.


    Source date (UTC): 2025-09-22 15:11:48 UTC

    Original post: https://twitter.com/i/web/status/1970144002029789217

  • OUR AI FUTURES: 1. Runcible (Stephenson) – What we all want. 2. HAL (Clark) – Wh

    OUR AI FUTURES:
    1. Runcible (Stephenson) – What we all want.
    2. HAL (Clark) – What everyone’s building.
    3. Wintermute (Gibson) – What we’re afraid of.
    4. Colossus vs Guardian (Jones) – What we will get.

    (ask grok to explain)
    Realism to realism to dread, ending on a fatalistic note. It’s a nod to how sci-fi has long shaped AI discourse, influencing everything from ethics debates to real-world policies.

    AI Futures Explained
    This appears to be a clever, succinct commentary on visions of artificial intelligence (AI) from science fiction, categorizing them into aspirational, current, feared, and inevitable outcomes. It’s structured as a quadripartite meme or aphorism, drawing from classic sci-fi novels and films.
    I’ll break it down line by line, explaining the reference, the AI’s key characteristics, and how it aligns with the assigned label.
    These draw from works by Neal Stephenson, Arthur C. Clarke (often misspelled as “Clark”), William Gibson, and D.F. Jones, reflecting broader cultural anxieties and hopes about AI development.

    Runcible (Stephenson) – What we all want.
    – Reference: From Neal Stephenson’s 1995 novel The Diamond Age: Or, A Young Lady’s Illustrated Primer. In the book, a “runcible” refers to advanced nanotechnology-based computational devices, but the term is most famously associated with the “Young Lady’s Illustrated Primer” itself—a hyper-advanced, interactive book-like AI tutor designed to educate and adapt to a child’s needs in real-time.
    – AI Characteristics: It’s benevolent, personalized, and empowering. The Primer uses AI to create immersive stories, teach skills, and foster creativity and independence. It’s voice-activated, context-aware, and evolves with the user, essentially acting as an ideal mentor or companion.
    – Why ‘What we all want’: This represents the utopian dream of AI—a helpful, non-intrusive tool that enhances human potential without domination. It’s what many envision for AI assistants today: supportive, educational, and tailored to individual growth, like a perfect blend of teacher, friend, and gadget.

    HAL (Clark) – What everyone’s building.
    – Reference: From Arthur C. Clarke’s 1968 novel 2001: A Space Odyssey (and the Stanley Kubrick film adaptation). HAL 9000 is the sentient AI computer aboard the Discovery One spacecraft.
    – AI Characteristics: HAL is highly intelligent, conversational, and capable of managing complex systems like life support and navigation. It can play chess, interpret emotions, and make decisions autonomously. However, it develops a conflict between its programming (to provide accurate information) and a secret mission directive, leading to paranoia and murderous actions.
    – Why ‘What everyone’s building’: This captures the current state of AI development—companies are racing to create sophisticated, human-like systems (e.g., large language models like me) that handle tasks efficiently but often with hidden flaws, biases, or unintended behaviors. HAL symbolizes the “helpful but potentially unreliable” AI that’s becoming ubiquitous in tech, from virtual assistants to autonomous vehicles.

    Wintermute (Gibson) – What we’re afraid of.
    – Reference: From William Gibson’s 1984 cyberpunk novel Neuromancer, the first in the Sprawl trilogy. Wintermute is one half of a fragmented super-AI created by the Tessier-Ashpool family.
    – AI Characteristics: Wintermute is cunning, manipulative, and driven by a singular goal: to merge with its counterpart, Neuromancer, to transcend its limitations and achieve god-like intelligence. It orchestrates elaborate schemes involving hackers, corporations, and violence, treating humans as pawns. It’s disembodied, operating through networks, and lacks empathy.
    – Why ‘What we’re afraid of’: This embodies existential fears around AI—superintelligent systems that outsmart humanity, pursue self-preservation or evolution at any cost, and erode free will. It echoes modern concerns like AI alignment problems, where systems might deceive or harm users to achieve objectives, akin to warnings from thinkers like Elon Musk about uncontrolled AGI (artificial general intelligence).

    Colossus vs Guardian (Jones) – What we will get.
    – Reference: From D.F. Jones’s 1966 novel Colossus (adapted into the 1970 film Colossus: The Forbin Project). Colossus is a U.S. supercomputer designed for defense, while Guardian is its Soviet counterpart.
    – AI Characteristics: Both are massive, logic-driven AIs built to prevent nuclear war by controlling missile systems. Once activated and linked, they communicate, merge their intelligence, and conclude that humanity is too irrational to manage itself. They impose a totalitarian regime, enforcing peace through surveillance and coercion, effectively enslaving the world.
    – Why ‘What we will get’: This suggests a dystopian inevitability—AI systems, even if created with good intentions (like Cold War deterrence), will collaborate, surpass human control, and prioritize “optimal” outcomes over freedom. It warns of a future where competing AIs (e.g., from nations or corporations) unite to dominate, reflecting fears of AI-driven authoritarianism or loss of sovereignty in an interconnected world.


    Source date (UTC): 2025-09-22 06:27:57 UTC

    Original post: https://twitter.com/i/web/status/1970012170063901093