Activity
Mon
Wed
Fri
Sun
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
Jan
What is this?
Less
More

Memberships

6 contributions to Open Source Voice AI Community
4 likes • 3d
Give this one a go if you want to try Gemini Live tuned for being a snappy demo: https://askjohngeorge.com/demo Built with Pipecat and hosted on Pipecat Cloud.
0 likes • 1d
@Nour aka Sanava thank you 🙏 Fully vibe coded it chatting to Claude Code over the xmas holidays 😁
Why Livekit Cloud, when I can just use Livekit Open source with Speech to Speech models and wont be restricted by concurrency
Hi everyone, why do i need livekit cloud, when I can just use speech to speech models like Open-AI-Realtime-mini and livekit open source. This would save like crazy because then I am not limited by concurrency limits. Best
0 likes • Dec '25
What's your autoscaling plan for the workers?
3 likes • Dec '25
I'm more referring to the process management side of the speech-to-speech pipeline. Even without local GPU inference, every active call requires a persistent Python worker to bridge the WebRTC stream to OpenAI's WebSocket. If you just use standard EC2 autoscaling, how are you handling scale-in events? When AWS spins down an instance to save costs, it will kill the process, which means dropping active calls mid-sentence. You'd need custom lifecycle hooks to implement connection draining so the server waits for the call to finish before terminating. LiveKit Cloud (or a custom K8s operator) handles that stateful orchestration. Raw EC2 autoscaling groups don't afaik.
Voice agent observability with tracing
Are you using tracing in your voice agent? I thought about this today, because The team at LangChain built voice AI support into their agent debugging and monitoring tool, LangSmith. LangSmith is built around the concept of "tracing." If you've used OpenTelemetery for application logging, you're already familiar with tracing. If you haven't, think about it like this: a trace is a record of an operation that an application performs. Today's production voice agents are complex, multi-model, multi-modal, multi-turn systems! Tracing gives you leverage to understand what your agents are doing. This saves time during development. And it's critical in production. You can dig into what happened during each turn of any session. What did the user say and how was that processed by each model you're using in your voice agent? What was the latency for each inference operation? What audio and text was actually sent back to the user? You can also run analytics using tracing as your observability data. And you can use traces to build evals. Tanushree is an engineer at LangChain. Her video below shows using a local (on-device) model for transcription, then switching to using the OpenAI speech-to-text model running in the cloud. You can see the difference in accuracy. (Using Pipecat, switching between different models is a single-line code change.) Also, the video is fun! It's a French tutor. Which is a voice agent I definitely need. How to debug voice agents with LangSmith (video): https://youtu.be/0FmbIgzKAkQ LangSmith Pipecat integration docs page: https://docs.langchain.com/langsmith/trace-with-pipecat I always like to read the code for nifty Pipecat services like the LangSmith tracing processor. It's here, though I think this nice work will likely make its way into Pipecat core soon: https://github.com/langchain-ai/voice-agents-tracing/blob/main/pipecat/langsmith_processor.py
1 like • Dec '25
Thank you for sharing!
How I build production voice AI systems (workflow adopted by $4B unicorn engineering team)
Made a video walking through the AI-assisted development workflow I use for production voice AI projects. Quick context: I used this on a Vapi replacement project at a $4 billion unicorn. Within a couple of weeks, their engineering team had adopted the workflow without being asked — and they wanted me to present it division-wide. The video covers: - How I coordinate specialised AI agents for Pipecat, Vapi, and Twilio research - Building features that would normally take hours of documentation reading - Live implementation of JSON-based assistant configuration (swap providers via config file, no code changes) https://youtu.be/AG68VC_mOGY Happy to answer questions.
1 like • Nov '25
Thanks for watching Nour! 🫡
1 like • Dec '25
@Pablo Arce Mari thank you. I agree I need to work on making the videos more accessible and maybe prep some slides in the future. I'm working on it. LMK if you have any more suggestions 🫡
Pipecat VS Livekit
I'm just curious in what platforms are you building and the pros and cons of each one.
1 like • Nov '25
@Nir Simionovich thanks, really appreciate that! I'm slammed next week but should be free from Wednesday the 19th onwards. Happy to join if the timing works out.
1 like • Nov '25
@Arek Wu nice! I recognise the format from NotebookLM. I like to feed it codebase analysis from Claude and have it make me podcasts to listen to about stuff I'm figuring out. 😁
1-6 of 6
John George
3
35points to level up
@askjohngeorge
Hacking Voice AI 👨‍💻

Active 44m ago
Joined Nov 8, 2025
United Kingdom