Build realtime voice and chat agents for demanding applications. Integrated metrics and experiments to optimize for user outcomes. Deploy to hosted API endpoints or integrate via SDKs.
How is Agent Runtime different from other AI backends?
The Inworld Agent Runtime uniquely combines
lightning-fast C++ core for realtime multimodal conversational interactions.
built-in telemetry for deep user insights (traces & logs).
live A/B testing to accelerate improvements to the end user experience.
Is Agent Runtime free?
Yes, Agent Runtime is free. Consumption of models is the only thing you pay for. Agent Runtime itself incurs no cost or license fee. Learn more about model pricing here.
How do I get started?
Follow our quick start guide to deploy a realtime conversational AI endpoint in 3 minutes - Then integrate into your app.
Who is Inworld Agent Runtime designed for?
Inworld Agent Runtime is specifically designed for developers building realtime conversational AI and voice agents that scale to millions of concurrent users.
Use cases include language tutors, social media, AI companions, game characters, fitness coaches, social media, shopping agents, and more.
Can Inworld Agent Runtime work with my existing framework and stack?
Yes, you can use the Inworld CLI to deploy a hosted endpoint that can be easily called by any part of your existing stack.
What production-ready building blocks does Inworld Agent Runtime provide?
Developers get a full suite of pre-optimized nodes to construct any real-time AI pipeline that can scale to millions of users, including nodes for
model I/O (STT, LLM, TTS)
data engineering (prompt building, chunking)
flow logic (keyword matching, safety)
external tool calls (MCP integrations)
and more
What LLM, TTS, STT, and embedding models do you support?
Use Inworld’s state-of-the-art TTS models alongside your preferred LLM.
We support all major providers—OpenAI, Anthropic, Google, Mistral—and low-latency platforms like Fireworks, Groq, and Tenstorrent. Read supported models here.