Predictions: Local LLMs on Apple Chips and the New LAN Parties
AI predictions: Apple releases chips for local LLMs, most tasks go offline, LAN parties are reborn with shared LLM compute instead of RAM.
Screenshot this:
Apple Will Ship Local LLM Chips
In N months Apple releases a new chip enabling local LLMs smart enough for basic tasks with sufficient tokens/second speed. Chinese companies keep pace.
Most AI Tasks Go Offline
Most tasks (browser control, file management, search) now happen locally. For smarter stuff — possibly with an Apple Plus subscription providing anonymity, extra security, and access to frontier models.
LAN Parties Are Reborn
LAN parties are reborn, but in a new form: you used to bring your RAM to a friend's house to run games with better graphics and FPS. Now — you bring your LLM compute to play a new kind of games. Together. 🥤
Original post in Telegram: @danokhlopkov
Dan Okhlopkov — AI agent practitioner. Building tools for TON Blockchain analysis and Telegram automation.