LLMs Fail Mission-Critical Reliability Due to Black-Box Guessing
Yee, founder of Logical Intelligence, argues LLMs are unreliable for high-stakes tasks like code generation or chip design because their autoregressive nature forces sequential token prediction—a "guessing game" prone to hallucinations. In mission-critical systems, like self-driving cars or planes, a 20% hallucination rate is unacceptable: "imagine there's AI driving a car and you're in that car and that car is an LLM and someone tells you like, you know, 20% of the time it's going to hallucinate and you might end up like in in like a wrong place."
Even with external verifiers like Lean 4—a machine-verifiable proof language—LLMs remain expensive. Compute costs skyrocket from generating tokens before verification, and internals stay opaque: "LLM, um obviously it's a language-based model and architecture doesn't allow you to do internal verifiers. So you you like it's like a black box for you."
Logical Intelligence prototypes on LLMs but builds Energy-Based Models (EBMs) for production, targeting deterministic, verifiable AI. Their focus: software/hardware correctness where current AI falls short, despite working demos.
Energy-Based Models Use Physics-Inspired Minimization for Transparent Reasoning
EBMs draw from physics, minimizing an "energy function" to find optimal states, like Lagrangians deriving equations of motion. No tokens or sequences: the model maps data to an "energy landscape"—a map of probable states where low-energy points are likely outcomes, high ones improbable.
Analogy: Predicting a tired person's post-podcast behavior. EBM observes states (walking, couch, gym) and trains a landscape favoring relaxation: "the lowest point is going to be you on the couch." Or body settling on a couch: uneven surfaces find minimal potential energy configuration. "It's all about your body finding the most comfortable configuration for you, which going to correspond to the like the lowest potential of your body."
Formally, their Kona model is an "energy-based reasoning model with latent variables." Latent variables capture hidden states (e.g., tiredness), enabling navigation without language. Training is inspectable in real-time: "you could open it anytime during the training and you could see what's happening in there."
Unlike LLMs' language-bound reasoning—where intelligence ties to token probabilities across languages—EBMs handle non-verbal tasks like spatial reasoning natively. Driving a car or building a bridge uses geometry and physics, not words: "when you build a bridge, you don't go to literature department, you go to engineering school and learn formal methods."
EBMs Deliver Efficiency, Self-Verification, and Scalability Over Token Guessing
Token-free architecture slashes costs: no autoregressive prediction means no expensive guessing. EBMs self-align during processing via internal verifiers, plus external ones like Lean 4. Double verification ensures correctness pre-output.
For non-language tasks (visual navigation, engineering), EBMs are faster and data-efficient: "yes, a EBM is able to do it with less training data." LLMs force non-verbal data into token space, bloating compute: image recognition or movement prediction via sequences works but is "super slow."
In real-time systems (circuits, microseconds), LLMs can't compete: "if your AI controls the circuits, you probably cannot wait even even a second." EBMs minimize resources naturally, per physics principles: everything seeks low energy, from particles to AI pipelines.
Host pushes back: Couldn't sequences model movements without language? Yee concedes it's possible but inefficient: "you could do it, but you don't have to do it. You just can use different architecture which is more suitable."
Logical Intelligence plugs EBMs into LLM prototypes for hybrid wins, filling the "deterministic AI" market gap. Future: AI everywhere (banking, automation), but verifiable for evolution, not hype.
Why Language-Centric AI Limits True Intelligence
LLMs encode intelligence language-dependently: reasoning in French differs from English due to token mixing. Human thought abstracts beyond words: "our brains, we are intelligent... none of my thoughts processes really depend on any language."
Daily actions prove it: navigating home uses visual-spatial data, not narration. Forcing everything through tokens is creative but wasteful: "you could be really creative, but if you want to minimize your resources... this form of AI is not suitable."
EBMs free AI from this, enabling pure reasoning on geometry, states, energy—ideal for engineering where "applied engineering is another example of spatial reasoning."
Key Takeaways
- Use EBMs over LLMs for mission-critical tasks needing verifiability, like code gen or chip design—internal inspection prevents hallucinations.
- Build energy landscapes from data: map states to probabilities via minimization, avoiding token guessing for 10x+ efficiency.
- Combine internal (self-alignment) and external verifiers (e.g., Lean 4) for double correctness in high-stakes systems.
- Ditch language for non-verbal reasoning: spatial tasks like navigation or engineering thrive token-free.
- Prototype with LLMs, productionize with EBMs—hybrids leverage both while fixing black-box issues.
- Train inspectably: monitor EBMs real-time, unlike waiting on LLM fine-tuning.
- Minimize resources physics-style: low-energy states = optimal, probable outcomes.
- Question LLM ubiquity: not everything needs tokens; match architecture to task.
- For real-time (microseconds), EBMs win—LLMs too slow/expensive.
- Expect verifiable AI everywhere soon: banking to planes, saving debug time for creativity.