Tiiny AI Pocket Lab: World's Smallest AI Supercomputer Fits 120B-Parameter Model
World's Smallest AI PC Runs 120B-Parameter Model Locally

In a significant breakthrough for the field of artificial intelligence, a US-based deep-tech startup has unveiled a pocket-sized device capable of running some of the world's most advanced AI models entirely offline. Tiiny AI Inc. has launched the Tiiny AI Pocket Lab, officially verified by Guinness World Records as 'The Smallest MiniPC (100 LLM Locally)'. This innovation challenges the prevailing notion that powerful AI requires massive cloud infrastructure and data centres.

What is the Tiiny AI Pocket Lab?

The core achievement of the Tiiny AI Pocket Lab is its ability to run a large language model (LLM) with up to 120 billion parameters directly on the device, without any need for cloud connectivity, external servers, or expensive high-end GPUs. The compact device measures approximately 14.2 x 8 x 2.53 cm, weighs around 300 grams, and is designed to be a complete, energy-efficient AI inference system.

According to the company's vision, this technology aims to make powerful AI accessible and personal, moving "large-scale intelligence to the edge." The device operates within a 65W power envelope, offering high-level performance at a fraction of the energy consumption of traditional GPU-backed systems. This addresses growing concerns about the soaring energy costs and sustainability issues linked to cloud-based AI.

Power and Performance: Specs and Capabilities

Under the hood, the Pocket Lab is powered by an ARMv9.2 12-core CPU coupled with a dedicated neural processing unit (NPU), delivering about 190 TOPS of AI compute. It is equipped with 80GB of LPDDR5X memory and 1TB of storage. Tiiny AI states that the device operates in the 'golden zone of personal AI (10B-100B parameters),' which is suitable for over 80% of real-world applications.

The company claims its performance is comparable to advanced models like GPT-4o, enabling PhD-level reasoning, multi-step analysis, and deep contextual understanding. This is made possible by two core technologies: TurboSparse, a neuron-level sparse activation technique for efficient inference, and PowerInfer, an open-source engine that dynamically shares computation between the CPU and NPU.

Implications for Users and the AI Ecosystem

The launch of the Tiiny AI Pocket Lab represents a potential paradigm shift. It reduces dependence on the cloud, which the company identifies as the real bottleneck in today's AI ecosystem. For users, this translates to lower operational costs, reduced latency, enhanced privacy, and greater accessibility, especially for those in environments with limited or unreliable internet resources.

The device also features one-click installation for leading open-source models such as Llama, Mistral, Qwen, DeepSeek, and OpenAI GPT-OSS. This allows for easy deployment of various AI agents. The founding team, formed in 2024, includes engineers from prestigious institutions like MIT, Stanford, Intel, and Meta. The full suite of features is scheduled for release at the Consumer Electronics Show (CES) in January 2026.

This innovation arrives amidst a global debate on whether ever-larger AI models are cost-effective. Tiiny AI's approach demonstrates that advanced, private, and portable intelligence is not only possible but is now a tangible reality, potentially democratizing access to cutting-edge AI technology for individual users and developers worldwide.