On-device LLM verification for airgapped, mobile, and embedded environments.
Aare Verify Edge brings the same mathematical guarantees of our cloud API to environments where network connectivity isn't available or data cannot leave the device. Purpose-built for constrained environments with limited memory, compute, and power.
Instead of sending LLM output to a server, Edge runs a complete verification pipeline on-device: a Domain-Specific Language Model (DSLM) for entity extraction, and Z3 Lite for constraint solving. Same inputs, same outputs, different implementation.
A DSLM is a small model fine-tuned for one task: extracting the entities and patterns that matter for your specific validation domain. Unlike general-purpose LLMs, DSLMs are focused, quantized, and optimized for on-device inference.
Full Z3 is a powerful SMT solver that can handle arbitrary satisfiability problems: quantifiers, arrays, bit vectors, floating point. For policy verification, you don't need that machinery.
Z3 Lite is a purpose-built constraint solver for bounded policy verification:
HIPAA compliance involves two distinct challenges: detecting Protected Health Information (PHI) and verifying that disclosure rules are followed. Aare Edge handles both on-device.
The HIPAA Safe Harbor rule defines 18 categories of Protected Health Information that must be identified and protected. The HIPAA DSLM extracts these identifiers from unstructured text - handling variations in formatting, OCR errors, and natural language phrasing that regex cannot.
Detection alone isn't enough. HIPAA's Privacy Rule contains 76+ constraints governing when and how PHI can be used or disclosed. Z3 Lite evaluates these constraints against the extracted entities:
This entire pipeline runs on-device in under 100ms. No PHI leaves the device. Mathematical proof of compliance or violation.
| Component | Metric | Value |
|---|---|---|
| DSLM | Base model | Phi-3-mini with LoRA fine-tuning |
| Quantization | 4-bit | |
| Model size | ~500MB | |
| Inference time | 20-50ms | |
| Z3 Lite | Binary size | <1MB |
| Evaluation time | <10ms | |
| Total SDK | Cold start | ~500ms |
| Warm validation | <100ms | |
| Runtime memory | ~200MB |
Performance varies by domain. A DSLM for simple entity extraction will be smaller than one handling complex semantic boundaries.
iOS and Android native SDKs. Validate LLM responses in-app without network round-trips.
Edge gateways, industrial controllers, and embedded systems. Validate before transmitting to cloud.
TEEs, air-gapped servers, HSMs. Environments where external dependencies aren't allowed.
In-vehicle systems with limited connectivity. Real-time validation for voice assistants and displays.
FDA-regulated environments requiring on-device validation without cloud connectivity.
As LLMs move to watches and glasses, validation follows. Efficient enough to run on a watch.
They're not competing approaches. They're the same formal verification thesis,
implemented for different deployment constraints.
Native constraint solver for iOS (Swift) and Android (Kotlin). MIT licensed.
CoreML integration for iOS, ONNX for Android. Efficient on-device inference.
HIPAA Safe Harbor DSLM (18 PHI categories). Reference implementation with ontology and training pipeline.
DSLM training toolkit for your specific domain. Build validators for any compliance requirement.
SDK available under MIT license. Domain-specific DSLM weights under commercial license.
GitHub: https://github.com/aare-ai