Future Work
Supporting Other Models
Section titled “Supporting Other Models”Kubrick currently works with the Twelve Labs embedding model, but future updates could add compatibility for other commercial video multimodal embedding models as they emerge. We also see potential in exploring open-source models like LanguageBind, which would make on-premises and cost-effective deployments possible. This flexibility could give teams the choice between high-accuracy, fully managed APIs and privacy-focused, self-hosted solutions.
Query Rewriting for Improved Search
Section titled “Query Rewriting for Improved Search”Through our work with semantic video search, we’ve found that more descriptive queries tend to produce better results. In the future, Kubrick could include an AI-powered query rewriter that transforms short or vague prompts (e.g., “dog”) into richer descriptions (e.g., “scenes where a dog is in the main frame or visible in the background”). This would make searching more intuitive for all users and improve accuracy without requiring them to carefully craft their queries.
Re-ranking Search Results
Section titled “Re-ranking Search Results”Another possible enhancement is adding a re-ranking step to further improve search accuracy. After the initial search identifies the top matches, a large multimodal model could review and re-score the best 50 segments. This would help surface the most relevant clips first, particularly for complex or ambiguous searches, at the cost of adding latency.
VLM Integration
Section titled “VLM Integration”Looking ahead, Kubrick could support conversational video exploration by integrating Video-Language Models (VLMs) such as Twelve Labs Pegasus, Google Gemini Pro, or LLaVA. After retrieving relevant video segments, these models could answer natural-language questions like, “Summarize the events in the last two minutes,” providing context-aware responses that combine visual, scene, and audio understanding.
Cache Support for Ingestion Embeddings
Section titled “Cache Support for Ingestion Embeddings”Currently, only embedding requests for search queries utilise the cache layer. The ingestion pipeline generates a new embedding on every valid S3 Create event. This means that frequent S3 rename or move operations, or repeated uploads of identical videos under different names, will result in redundant calls to the external embedding API. Integrating the caching layer with the ingestion pipeline would eliminate the monetary and latency costs of calling the embedding API by recognizing and reusing embeddings for duplicate content.