Explicit Context Caching
Support for Explicit Context Caching in the Gemini API. When applications frequently sends long, static system instructions and large document sets (100k+ tokens) in every prompt. Implementing an explicit cache would significantly reduce latency for users and lower the token overhead by allowing us to persist this context across multiple turns without re-sending the entire prefix.
1
vote
Training models is not part of the product direction of Firebase AI Logic. Other platforms like Vertex AI are more suitable for that.