Local-first
Inference, storage and computation default to the device. The cloud is opt-in, never required.
OmniFlux AI builds a family of mobile applications powered by a unified on-device inference engine. Zero servers. Zero tracking. Zero compromises.
Each app is purpose-built for a specific human moment, sharing the same uncompromising privacy guarantee.
On-device voice transcription, speaker diarization & smart summaries.
Multimodal visual understanding — entirely on your device.
Speech-to-text, speaker diarization & smart summaries — entirely on your phone.
All three stages happen on-device — your audio never leaves the phone.
A high-performance on-device inference runtime, optimized across CPU / GPU / NPU. One engine, every product.
Quantized models run on CPU, GPU and NPU with adaptive scheduling, delivering desktop-class speed on mobile silicon.
No analytics, no crash beacons, no account system. The app physically cannot phone home.
GGUF, MLC, MNN — pick from leading open-source models or bring your own.
Speech, Vision, NLP modules share a unified pipeline — features ship faster across products.
Compatible models
We believe the next generation of AI should empower individuals without surveilling them. We're building it, one focused app at a time.
Inference, storage and computation default to the device. The cloud is opt-in, never required.
We architect away the temptation to collect. No accounts. No telemetry. No backdoors.
We embrace open weights, open formats and transparent claims. Audit us, anytime.
Engineering notes, product stories and ideas from the team.
Cloud AI is a transitional architecture. Here's why we believe the next decade of useful AI lives in your pocket — and what it means for the products we are building.
A peek at the architecture choices behind our shared on-device inference runtime — and the constraints that shaped them.
Our flagship enters internal beta. A walkthrough of the streaming pipeline and what we learned shipping ASR on real phones.
Join our launch list for product releases, technical deep-dives and early access.