Today, we’re launching the Inworld CLI, a unified toolkit for building, deploying, and optimizing realtime conversational AI and voice agents.
Watch the
Demo and view
Docs to get started.
Challenge
Until now, building realtime conversational AI meant facing:
- Performance Bottlenecks: Unpredictable latency from third-party APIs creates a jarring user experience. This is compounded by core language limitations, like Python's GIL, that block parallel execution and stall critical operations.
- High Development Overhead: Engineering resources are drained by maintenance. Teams spend more time debugging provider failures and integrating a complex patchwork of models than building new features, causing product velocity to stagnate.
- Slow Iteration Speed: Scattered conditional logic for different models and clients makes the entire system fragile. This fragility makes every change high-risk, paralyzing rapid A/B testing and stalling product improvements.
Inworld faced these very challenges as our customer base grew and expanded beyond games and into mobile apps, voice agents, ai companions, and more. We hence built
Inworld Runtime to solve them.
Inworld Runtime
Inworld Runtime is the AI backend for realtime conversational AI. You build your conversational AI with Inworld Runtime SDK, launch a hosted endpoint using Inworld CLI, and observe and optimize your conversational AI by running A/B experiment in the Inworld Portal.
Today, building with Inworld Runtime just became easier with the launch of Inworld CLI.
Inworld CLI
With Inworld CLI, developers can now build realtime conversational AI that are fast, easy to debug, and easy to optimize via A/B experiments.
- Build realtime experiences
- npm install -g @Inworld.ai/cli to install the Inworld CLI
- inworld login to login and generate api keys automatically
- inworld init to initialize conversational AI pipelines such as LLM -> TTS - preoptimized for latency and flexibility
- inworld run to test locally with instant feedback
- inworld deploy to create persistent, production-ready endpoints
- Monitor with clarity
- Integrated telemetry: Each request is automatically logged in dashboards, traces, and logs in Inworld Portal.
- Optimize continuously
- inworld graph variant register to run live A/B tests without client changes
Proven technology
Since launching Inworld Runtime earlier this year, we've seen developers build incredible realtime conversational AI experiences.
- Wishroll went from prototype to 1M users in 19 days with 20x cost reduction.
- Streamlabs built a real-time multimodal streaming assistant with under 500ms latency.
- Bible Chat scaled their AI-native voice features to millions. Inworld CLI builds on Runtime to help developers build agents more efficiently and reliably.
Get started with Inworld Runtime
Inworld Runtime is the best way to build and optimize realtime conversational AI and voice agents
Get started now with Inworld
CLI:
- Build a prod-ready conversational AI or voice agent
- Deploy it to Inworld Cloud as an endpoint so you can easily integrate into your app
- Monitor dashboards, traces, and logs in the Portal
- Improve user experience by Run live A/B experiments to identify the best model and prompt settings for your users