Signals
Back to feed
4/10 Model Release 12 May 2026, 02:01 UTC

Thinking Machines launches interaction models, BioLLM uses human neurons, and Teutonic trains 80B decentralized LLM.

The simultaneous emergence of continuous multimodal interaction, biocomputing-based LLMs, and record-scale decentralized training signals a hard pivot away from traditional static architectures. BioLLM's use of 800k wetware neurons introduces unprecedented hardware paradigms, while Teutonic's 80B run proves decentralized compute can now challenge centralized training clusters.

The AI landscape is fracturing into highly specialized, non-traditional compute and architectural paradigms, highlighted by three major developments across multimodal interaction, biocomputing, and decentralized training.

What Happened & Technical Details First, Thinking Machines introduced "interaction models," shifting away from discrete prompt-response paradigms to continuous, real-time multimodal processing. These models natively integrate listening, watching, and thinking streams to enable human-like collaboration without traditional latency bottlenecks.

Second, and perhaps most radically, BioLLM emerged as the first "living language model." Operating on wetware rather than silicon, it utilizes 800,000 real human neurons grown on a chip. Input and output are encoded via standard LLM tokens, bridging biological neural networks with digital tokenization. This coincides with an upcoming non-invasive brain-computer interface (BCI) launch.

Finally, in the decentralized compute space, Teutonic has initiated training on an 80B parameter LLM, dethroning Bittensor's Covenant-72B. This marks the largest decentralized training run to date, utilizing a competitive miner network optimized for reducing perplexity across distributed nodes.

Why It Matters From an engineering perspective, these releases represent a divergence from standard transformer scaling on centralized GPU clusters. BioLLM is a proof-of-concept for neuromorphic wetware, potentially offering massive leaps in energy efficiency compared to traditional silicon, albeit with immense stability and scaling challenges. Teutonic’s 80B run proves that algorithmic improvements in distributed consensus and gradient synchronization are making decentralized training viable for frontier-class parameter counts. Meanwhile, Thinking Machines' continuous streaming architecture demands a fundamental rethink of context window management and state persistence in production environments.

What to Watch Next Keep a close eye on BioLLM's upcoming BCI launch to see if token-to-neuron latency and fidelity can support practical applications. For Teutonic, monitor the convergence rate and final perplexity scores of the 80B model to validate the efficiency of their decentralized training architecture. Finally, look for developer benchmarks on Thinking Machines' interaction models, specifically regarding streaming latency and memory overhead during prolonged continuous sessions.

biocomputing decentralized-ai multimodal bittensor wetware