September 30: Support for ingestion performance, search stability, and token accounting accuracy
New Features
💡 This release includes broad performance optimizations across ingestion, search, and summarization pipelines. Developers should see faster indexing for large document batches and improved throughput when embedding concurrent content streams.
We’ve also refined token accounting and reporting within the billing subsystem, ensuring real-time accuracy for both input and output token usage across all connected model providers.
Bugs Fixed
GPLA-4590: Resolved high-concurrency queue deadlocks affecting ingestion of large feed batches
GPLA-4594: Improved summarization retry logic for timeouts on long transcripts
GPLA-4601: Corrected mis-reported embedding-cost metrics for Gemini models
Last updated