What Is On-Device AI and Why Does It Protect Your Privacy?
On-device AI runs the entire model on your local hardware. No data leaves your device, no server processes your text, and no network connection is required. The model is trained in the cloud once, then shipped to your machine, where all inference happens locally. For writing tools, this means corrections and predictions fire in milliseconds with zero privacy exposure, regardless of your internet connection.
How does on-device AI work differently from cloud AI?
Every AI system involves two distinct phases: training and inference. Training is the process of building the model by exposing it to vast amounts of data, adjusting its parameters until it can make accurate predictions. This requires enormous compute and happens in data centres. Inference is what happens when you actually use the model: you provide an input and the model generates an output.
Cloud AI keeps inference on the server. When you type into Grammarly, your text travels over the internet to a Grammarly server, the model runs, and the correction comes back. This round trip adds 200-500ms of latency even on a fast connection, and it means your text has physically left your device.
On-device AI moves inference to your local hardware. The model is downloaded once, stored on your Mac, and runs entirely on your CPU, GPU, or dedicated Neural Engine. The response time drops to 10-50ms because there is no network in the loop. More importantly, your text never leaves your machine. There is nothing to intercept in transit, no server log to breach, and no third-party data retention policy to worry about.
What makes on-device AI possible now when it was not before?
Ten years ago, running a useful language model locally was not practical. The models capable of accurate grammar correction required billions of parameters and gigabytes of VRAM to run. Consumer hardware lacked the compute and memory bandwidth to execute them at interactive speeds.
Two advances changed this. First, model compression techniques improved dramatically. Quantisation reduces the precision of model weights from 32-bit floating point to 8-bit or even 4-bit integers, shrinking model size by up to 8x with minimal accuracy loss. Knowledge distillation trains smaller "student" models to replicate the behaviour of much larger "teacher" models, producing compact models that punch well above their size.
Second, Apple Silicon transformed the capabilities of consumer hardware. The Apple M1 chip (released 2020) introduced a 16-core Neural Engine rated at 11 TOPS (trillion operations per second). Its successor chips scaled this further: the Apple M2 Neural Engine processes 15.8 trillion operations per second. This dedicated hardware is purpose-built for the matrix multiplications that language models rely on, running them orders of magnitude more efficiently than a general-purpose CPU.
The combination of smaller, more efficient models and dramatically more capable local hardware opened the door for real-time, on-device language AI in consumer applications. Writing tools that required cloud servers in 2015 can now run entirely on a MacBook Air.
What are the tradeoffs of on-device vs cloud AI?
On-device AI has genuine advantages and genuine limitations. Understanding both helps you choose the right tool for the task.
On-device advantages:
- Zero latency from network round trips: responses in 10-50ms
- Works fully offline, including on planes and restricted networks
- No privacy exposure, no data retention, no server to breach
- No account required, no subscription, no service outage risk
Cloud AI advantages:
- Access to much larger models with broader capabilities
- Training data can be updated continuously for current knowledge
- Consistent experience across devices without local storage or compute
- Handles complex, open-ended tasks that require large context windows
For writing assistance, specifically spelling correction, grammar checking, and word prediction, the task is narrow enough that compact on-device models perform at equivalent quality to large cloud models. The difference matters more for open-ended tasks like essay generation or research, which require broader knowledge and longer context.
Which writing tools use on-device AI?
The landscape is more limited than marketing suggests. Most tools claiming "AI" writing assistance still route processing through cloud servers.
Charm uses fully on-device inference for all three of its core features. Spells (spelling correction, cyan glow), Polish (grammar correction, blue glow), and Oracle (word prediction, purple glow) all run on your Mac without any network activity. No account is required. The optional OpenAI API integration for enhanced grammar is explicitly opt-in, transparent, and disabled by default.
Apple Intelligence (macOS 15, Apple Silicon only) takes a hybrid approach. Writing tools for basic corrections run locally. More complex requests, such as drafting full emails or summarising documents, are routed to Apple's Private Cloud Compute infrastructure. Apple has published extensive technical claims about PCC privacy, but some data still leaves the device for complex tasks. Intel Mac users and anyone on macOS 14 cannot use Apple Intelligence at all.
macOS built-in autocorrect (NSSpellChecker) is dictionary-based and fully local. It processes nothing remotely, but it is also not AI: it uses static dictionary lookup rather than a language model, which limits its accuracy. It catches roughly 40% of typos and cannot handle Electron apps like Slack or VS Code.
Grammarly and LanguageTool (cloud default) process all text remotely. Their accuracy benefits from large cloud models, but every keystroke travels to their servers.
Frequently asked questions
What is on-device AI?
On-device AI runs a language model entirely on your local hardware, with no network connection required. The model was trained in the cloud, but inference - generating a prediction or correction - happens on your CPU, GPU, or Neural Engine. No text is transmitted, no server is contacted, and response time is measured in milliseconds.
Is on-device AI as good as cloud AI?
For focused tasks like spelling correction, grammar checking, and word prediction, on-device models are highly competitive. Cloud AI has an advantage for tasks requiring very large context windows or up-to-date world knowledge. For real-time writing assistance, on-device accuracy is more than sufficient, and the speed and privacy advantages are substantial.
Does on-device AI work offline?
Yes, fully. Because everything runs locally, on-device AI works with no internet connection. Charm works on a plane, in a cafe with no Wi-Fi, or on a corporate network that blocks outbound connections. Cloud tools like Grammarly require an active internet connection and fail silently if it drops.
How does Charm protect my privacy?
Charm runs its spelling, grammar, and word prediction models entirely on your Mac. No text is ever transmitted to any server. No account is required, so there is no user profile to associate your writing with. The app requires no network access for its core features, meaning there is nothing to intercept, breach, or subpoena.
What is the difference between on-device and cloud processing?
Cloud processing sends your input to a remote server, which runs the model and returns a result. This takes 200-500ms over a network and requires an internet connection. On-device processing runs the model on your local hardware in 10-50ms with no network involved. For writing assistance, compact on-device models perform at equivalent quality to large cloud models.
On-device AI for every app on your Mac.
Charm corrects spelling, polishes grammar, and predicts words locally. No cloud, no account, no data leaving your device. $9.99, yours forever.