What If You Never Needed an API Key Again? Building a Mesh LLM From Spare Compute
The current AI stack has a dependency most of us don't talk about: a handful of closed models from a handful of providers, and an API call standing between every agent and every action. Mic Neale — who helped build Goose, Block's open-source agentic coding system — thinks that's a problem worth solving at the infrastructure level. This talk introduces a working prototype of a decentralised mesh LLM: a system where individuals and small teams pool their spare compute to collectively run open models that none of them could run alone. When your GPU is idle, it contributes to the mesh. When you need capacity, the mesh is there. The result is access to capable open models without a cloud bill or an API dependency.
Mic will walk through how the mesh works technically — coordination, model sharding, latency management, and what happens when nodes drop out — and where it's headed: a model where contributors earn from their idle capacity, and where the economics of running frontier-class open models shift from "data centre required" to "your neighbourhood has enough."
Mic Neale
Mic Neale is a Principal Engineer at Block, where he works on Goose, the company’s open source AI coding agent. A co-founder of CloudBees and former open source architect at Red Hat, Mic built the early versions of the Drools inference engine and has spent over two decades at the intersection of developer tooling, distributed systems, and AI. With a background in electrical engineering from UNSW, he brings a pragmatic, builder-first perspective to how engineering teams adopt and deploy AI infrastructure at scale.