Intro to Inworld CLI - Build Realtime Conversational AI
Challenge
Until now, building realtime conversational AI meant facing:
Performance Bottlenecks: Unpredictable latency from third-party APIs creates a jarring user experience. This is compounded by core language limitations, like Python's GIL, that block parallel execution and stall critical operations.
High Development Overhead: Engineering resources are drained by maintenance. Teams spend more time debugging provider failures and integrating a complex patchwork of models than building new features, causing product velocity to stagnate.
Slow Iteration Speed: Scattered conditional logic for different models and clients makes the entire system fragile. This fragility makes every change high-risk, paralyzing rapid A/B testing and stalling product improvements.
Inworld faced these very challenges as our customer base grew and expanded beyond games and into mobile apps, voice agents, ai companions, and more. We hence built Inworld Runtime to solve them.
Inworld Runtime
Inworld Runtime is the AI backend for realtime conversational AI. You build your conversational AI with Inworld Runtime SDK, launch a hosted endpoint using Inworld CLI, and observe and optimize your conversational AI by running A/B experiment in the Inworld Portal.
Today, building with Inworld Runtime just became easier with the launch of Inworld CLI.
Inworld CLI
With Inworld CLI, developers can now build realtime conversational AI that are fast, easy to debug, and easy to optimize via A/B experiments.
Build realtime experiences
npm install -g @Inworld.ai/clito install the Inworld CLI
inworld loginto login and generate api keys automatically
inworld init to initialize conversational AI pipelines such as LLM -> TTS - preoptimized for latency and flexibility
inworld run to test locally with instant feedback
inworld deploy to create persistent, production-ready endpoints
Monitor with clarity
Integrated telemetry: Each request is automatically logged in dashboards, traces, and logs in Inworld Portal.
Optimize continuously
inworld graph variant register to run live A/B tests without client changes
Proven technology
Since launching Inworld Runtime earlier this year, we've seen developers build incredible realtime conversational AI experiences.
Wishroll went from prototype to 1M users in 19 days with 20x cost reduction.
Streamlabs built a real-time multimodal streaming assistant with under 500ms latency.
Bible Chat scaled their AI-native voice features to millions. Inworld CLI builds on Runtime to help developers build agents more efficiently and reliably.
Get started with Inworld Runtime
Inworld Runtime is the best way to build and optimize realtime conversational AI and voice agents