Gemilive
Plug-and-play Gemini Multimodal Live
for your custom stack.
Abstracts away the brutal boilerplate of bridging native browser WebAudio, gapless PCM streaming, and live video frames to a secure Python backend proxy. Build conversational intelligence in six lines of code.
The "Proxy Problem"
While Google provides excellent core SDKs for the Gemini Multimodal Live API, integrating it securely into a production app usually kills a weekend. You can't put your API keys directly into a browser frontend, so you are forced to build a custom backend proxy. Suddenly, you're hand-wiring WebSockets to bridge raw 16kHz microphone streams from a JS frontend into a Python backend just to forward them to Gemini.
Gemilive effortlessly bridges that gap permanently.
Browser Frontend
gemilive-jsFastAPI Proxy
gemiliveGoogle Gemini
AI Cloud🐍 FastAPI Backend
main.pyfrom fastapi import FastAPI
from gemilive import mount_gemilive
app = FastAPI()
# Instantly proxies secure WebSockets to Gemini Live
# Handles upstream 16kHz PCM & downstream 24kHz audio
mount_gemilive(
app,
system_prompt="Be incredibly witty."
)
🌐 Browser Frontend
app.jsimport { GemiliveClient } from 'gemilive-js';
const client = new GemiliveClient("wss://api.yourdomain.com/ws/live");
// Manages mic extraction, downsampling, and canvas video snapping
await client.start();
client.onMessage = (msg) => console.log(msg);
// Audio automatically plays back purely gapless over WebAudio Timeline