Use gemini-2.0-flash for VAPI custom LLM
Vapi’s custom LLM example uses OpenAI’s ChatCompletionChunk for StreamingResponse, but Gemini streams using google.genai.types.GenerateContentResponse.
To work around this, I hacked the OpenAI proto to repurpose delta.content with Gemini’s output, but that feels hacky and wrong.
Does anyone have a clean example or guidance on how to properly stream gemini-2.0-flash as a StreamingResponse in a custom LLM using Vapi?