Activity
Mon
Wed
Fri
Sun
Jun
Jul
Aug
Sep
Oct
Nov
Dec
Jan
Feb
Mar
Apr
What is this?
Less
More

Memberships

Open Source Voice AI Community

909 members • Free

Voice AI HQ

436 members • Free

13 contributions to Open Source Voice AI Community
Voice AI in Robotics
This is more on the fun side of Voice AI. As we do more and more with voice and video AI and autonomous robotics my colleague Bin put together this cool autonomous driving interface and then crashed his Ferrari. "STOP RIGHT NOW!" then boom :) https://x.com/pham_blnh/status/2035427622046531874 https://huggingface.co/spaces/lerobot/visualize_dataset?path=%2Fbinhpham%2Fdriving-dataset%2Fepisode_0
Solving unwanted interruptions with Adaptive Interruption Handling
I assume most of you saw this already, but in case you missed it, you may find it helpful for solving the unwanted interruption problem. I know I've spoken to countless folks about this specific issue in the past and how frustrating it can be for users and developers of the system. https://www.youtube.com/watch?v=DSXCE7D4Kvs Full blog post here: https://livekit.com/blog/adaptive-interruption-handling
3
0
From SIP to AI: A Real Call Finally Worked
Sharing a small but exciting milestone from my learning journey 🚀 Over the last few weeks, I’ve been deep into voice infrastructure and SIP, and I finally have a full working setup: 📞 Local Phone → SIP (from Signalwire) → FreeSWITCH → Voice Agent → Response back to the caller The FreeSWITCH server is running on a Debian server on DigitalOcean, and everything is now talking to each other smoothly — SIP, RTP, real-time audio, and AI responses. I’m currently working with a client, and initially we’re aiming to scale this setup to handle ~100 concurrent calls, which is pushing me to really understand: - SIP call flows - Audio streaming - Server performance & scaling - Latency trade-offs vs managed platforms Honestly, this stuff is challenging but insanely exciting. Every time a real phone call hits the server and the agent responds correctly, it feels like magic — but the earned kind 😄 Just wanted to share this win and keep building. If anyone here is working on voice agents, SIP, or FreeSWITCH — would love to connect and exchange notes 🤝
1 like • Feb 8
Using a framework can reduce the burden of getting a voice agent runnings. It can let you focus on your business logic rather than all the plumbing. A good framework should make it straight forward to terminate your SIP trunk (FreeSwitch), distributed and load-balance to the calls to the correct agents (Dispatch), detect errors and provide mechanism to recover from the errors like if an LLM, STT, or TTS provider goes down should fail over to the next one. It should also provide great tools to analyze the telemetry data for each and every interaction and give you a solid foundation for how to do reproducible testing. How did you make your agent? Did you look at Pipecat or LiveKit? If you are serious about making voice agents and intend to make a production grade solution I think you may find this quick-start guide very helpful: https://docs.livekit.io/agents/start/voice-ai-quickstart/
Quick question
Hey everyone! How are you currently testing your voice agents?
1 like • Jan 8
I guess I am slow to the party here, but this thread was brought to my attention, so I will try to answer based on what I know about the topic. You can get good results using the built-in Agent Evals workflow when the agent is built on LiveKit Agents. Instead of trying to manually “play calls” over and over, write small behavioral tests (pytest + pytest-asyncio) that exercise a turn or two of conversation and assert what the agent should do — tone, tool usage, error handling, grounding, etc. A few things that have worked well: - Text-only evaluation first (fast + cheap), then only do audio tests when needed - judge() for qualitative checks, e.g., “did it politely greet and offer help?” - Tool-call assertions to confirm the right tool was called with the right args - Mocked tools to force errors and edge cases without touching real systems - Multi-turn tests so we can catch regressions in memory and workflows - Run the tests in CI with LLM keys as secrets, so every PR gets evaluated It’s been great for catching subtle regressions, especially when we tweak prompts or add new capabilities, without breaking older flows. I know many folks I have spoken to also find value in 3rd party tools like: https://getbluejay.ai/ https://hamming.ai/ If you are in the San Francisco area on January 27th and your focus for 2026 is agent reliability, we are having a meetup with industry experts on voice agent reliability techniques and testing. I will share the Luma here in a few days, in case you can join. Another great tool is to use Agent tasks (I've seen some call this templates) and Task groups, particularly if you need to use a specific flow across many customers. You can refine the task and reuse it. You can learn more here https://docs.livekit.io/agents/logic/tasks/ When diagnosing issues, it can be super helpful to use Langfuse or a feature like LiveKit Agents' observability. Here is a nice overview of how and why you would use something like this https://docs.livekit.io/deploy/observability/insights/ (YouTube demo)
1 like • Jan 9
For those that can make it and and are interested in voice AI reliability here is our meetup "Voice Mode - The Year of Reliability" on Jan 27th in San Francisco at Frontier Tower https://luma.com/voice-ai-meetup-the-year-of-reliability
Game Changer - New Potential Client - Need Assistance!!!
I have a meeting today with a potential client. He's the Director, PMO for a private detention and correctional conglomerate. They have educational re-entry programs, transportation operational division, real estate, etc. I want to implement Voice AI tools to their operations. I just want to start out doing a small project to collaborate with him to prove what I can do. What would be a good introduction statement? What kind of demo can I do? (Examples) Ultimately, what price do I charge? Your thoughts are much appreciated.
1 like • Dec '25
Yes there are lots of good ways to get started. LiveKit has an Agent builder where you can design in your browser initially then export the code when you want to start doing much more complex use cases. See here for getting started quickly if you want to go that path https://livekit.io/events/dev-day I am not super clear on your use case but I think a good start could be to connected your agent up to the content of the education program so the users can have an easy way to perform Q&A. I can imagine there will be a strong need for guard rails for a use case like this. Modality will also need to be thought about. Do you plan over telephone, in the browser, voice, text, video, etc. Answers to those will help steer the path to initial implementation. There are a lot of examples from AI enablement platforms. I think a good path forward for you is to find some example project that is a little like what you want to do and use that as your starting point. My personal feel on this is the space is moving very quickly and new and better models are coming around all the time. I would lean towards platforms that give you more flexibility to easily swap out a given model incase they go out of business or something substantially better is released from someone else. In this space there are a few primary paths for getting started. There are more "turn key" solutions which make it pretty quick to get started but may make more complex use cases harder to accomplish in the future. Or there are more flexible frameworks that will require you do to a little more lifting initially but makes it much easer to address future use cases. I am sort of proud of how LiveKit has bridged the gap by making it very easy to get up an running for the new comer as well as being the go to tool for the most advanced realtime AI use cases. Good luck. If you have specific questions I am sure many of us here would be glad to help try and answer.
1-10 of 13
Yepher Yepher
3
38points to level up
@yepher-yepher-2541
Voice AI Builder

Active 32d ago
Joined Nov 12, 2025