Activity
Mon
Wed
Fri
Sun
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
Jan
Feb
Mar
Apr
What is this?
Less
More

Memberships

KVK Automates AI

596 members • Free

AI Marketing Hub

3.2k members • Free

SW Automation

8.3k members • Free

AI Workshop Lite

20.5k members • Free

Voice AI Alliance

3.4k members • Free

Open Source Voice AI Community

908 members • Free

AI Automation Agency Hub

312.5k members • Free

Flow State

1.4k members • Free

AI Automation Society

342k members • Free

2 contributions to Open Source Voice AI Community
Ultravox for realtime conversations
I'm fairly new to this but when I paired a custom LLM with Pipecat the latency wasn't what I was expecting. After some research I came across Ultravox and latency is top notch. But using it for a project in EdTech may end up making the business model less feasible... Does anyone else have experience with Ultravox or a custom build that is handling latency better?
1 like • 16d
@Sumeyye Bozkus - For realtime STS voice, I started with this config and recently have been testing for latency improvements. [ use case ex - restaurant order works well ] This may sound a like a bit of over-think but it works. Here is the gist of the flow. It is a modular option I can use/swap for multiple uses. Environment Stack: OS: Ubuntu 24.04 LTS (VPS / 8GB RAM). Orchestration: Docker Compose (Isolated Services). Primary Logic: n8n (v2.12.1 at the time) via HTTP Request nodes. A separate setup for RAG voice using a Telegram bot Website-to-LiveKit setup first with n8n logic to handle a "Voice-to-Voice" interaction Docker Stack: n8n, Redis, Postgres, ChromaDB (Core 0) AI Engine: llama.cpp + Llama-3.2-1B (Core 1) Voice Agent: Python LiveKit Worker + Faster-Whisper + Piper RAG Logic: n8n Webhook + Markdown Ingestion. Monitoring: Telegram Expert-in-the-loop Audio Networking: LiveKit Cloud (Free Tier) handles the WebRTC switchboard, The Worker (VPS): Python-based agent connects to LiveKit Cloud. It performs STT (Faster-Whisper Tiny), LLM (Llama-3.2-1B via llama-cpp), and TTS (Piper). The RAG Brain (n8n): The Agent Worker sends a text query to an n8n Webhook. n8n searches the Markdown files (Vector Store - Chromadb) and returns the context The n8n RAG Workflow - How it works: Created a workflow in n8n with these nodes: Webhook Node: Listens for POST requests from my Agent Worker VPS. Vector Store Node: Use a local engine (e.g., chromaDb) to index Markdown files. Code Node: Formats the retrieved context into a prompt. Response Node: Sends the context back to the Webhook. *** There is a python script call on the VPS for the LiveKit Agent Worker (Python). This script runs as a service on my VPS. It bridges the voice stream to my n8n workflow. This script connects to my LiveKit Cloud account for the audio stream and calls my n8n Webhook to retrieve Markdown context (RAG). The script uses Faster-Whisper for transcription and Piper for speech to decrease latency.
1 like • 15d
@Sumeyye Bozkus You can use slack, whatsapp, gmail, .. swap out and place your communication app of choice ~cheers Rc
Opensource Voice AI
Do we have any opensource orchestration platforms built on Livekit that work similar to vapi or Retell?
1 like • 16d
you can use n8n and MCP with LiveKit ~cheers Rc
1-2 of 2
Rc Hall
1
2points to level up
@rc-hall-4827
Own a Mobile Marketing business - Specialized in Real Estate, Restaurants, Home Services. Marketing includes PPC, SEO, Ai chatbots & Support Chatbots.

Active 2d ago
Joined Mar 16, 2026