Deploy high-performance AI directly in your app.
For Free
From voice assistants to image recognition — all running on-device.
Summarize long text instantly
Process images with local models
Turn voice into actions or text
Sure! I will schedule a call with your colleague Max at 2PM tomorrow!
Build your own use case
Conversational AI, running on-device
Hi! I am great, and you? It seems you want to start a conversation. I'll...
Tag text by topic, intent, or sentiment
Topic: account management
Intent: cancellation request
Sentiment: angry
Summarize long text instantly
Process images with local models
Turn voice into actions or text
Sure! I will schedule a call with your colleague Max at 2PM tomorrow!
Build your own use case
Conversational AI, running on-device
Hi! I am great, and you? It seems you want to start a conversation. I'll...
Tag text by topic, intent, or sentiment
Topic: account management
Intent: cancellation request
Sentiment: angry
Build your own use case
Conversational AI, running on-device
Hi! I am great, and you? It seems you want to start a conversation. I'll...
Tag text by topic, intent, or sentiment
Topic: account management
Intent: cancellation request
Sentiment: angry
Summarize long text instantly
Process images with local models
Turn voice into actions or text
Sure! I will schedule a call with your colleague Max at 2PM tomorrow!
Build your own use case
Conversational AI, running on-device
Hi! I am great, and you? It seems you want to start a conversation. I'll...
Tag text by topic, intent, or sentiment
Topic: account management
Intent: cancellation request
Sentiment: angry
Summarize long text instantly
Process images with local models
Turn voice into actions or text
Sure! I will schedule a call with your colleague Max at 2PM tomorrow!
BUILD ONCE, SHIP ANYWHERE
Write your AI pipeline once and deploy it natively across mobile, desktop, and game engines — with hardware acceleration on every target.
macOS Easy Integration
Run a single model or orchestrate full AI agents — in just a few lines of code.
name: text-generation
stages:
- model: smollm2-360mfinal model = await Xybrid.model(modelId: 'smollm2-360m').load();
final result = await model.run(
envelope: Envelope.text(text: 'Explain quantum computing'),
);
print(result.text);Simple by design
From prototype to production in three steps.
Write a YAML config to chain models together — or call a single model directly with 3 lines of code.
Ship to iOS, Android, macOS, or Unity. One SDK, native acceleration on every device.
Push new model versions without app store updates. Track performance across your entire fleet.
From on-device inference to fleet-wide analytics — everything you need to ship AI at scale.
Run models directly on user devices. Zero network latency, full offline support, complete data privacy.
Chain ASR, LLM, and TTS into intelligent agents with a simple YAML config.
Push new model versions to devices without app store updates. Rollback instantly if needed.
Route inference to edge or cloud based on device capability, battery, connectivity, or custom rules.
CoreML, Metal, QNN — automatically use the fastest execution provider on every device.
Monitor inference latency, model usage, and device capabilities across your entire fleet.