December 10: Support for OpenAI GPT-4 Turbo, Llama 2 and Mistral models; query by example, bug fixes
New Features
💡 Graphlit now supports the OpenAI GPT-4 Turbo 128k model, both in Azure OpenAI and native OpenAI services. Added new model enum
GPT4_TURBO_VISION_128K.💡 Graphlit now supports Llama 2 7b, 13b, 70b models and Mistral 7b model, via Replicate. Developers can use their own Replicate API key, or be charged as credits for Graphlit usage.
💡 Graphlit now supports the Anthropic Claude 2.1 model. Added new model enum
CLAUDE_2_1.💡 Graphlit now supports the OpenAI GPT-4 Vision model for image descriptions and text extraction. Added new model enum
GPT4_TURBO_VISION_128K. See usage example in "Multimodal RAG" blog post.Added query by example to
contentsquery. Developers can specify one or more example contents, and query will use vector embeddings to return similar contents.Added query by example to
conversationsquery. Developers can specify one or more example conversations, and query will use vector embeddings to return similar conversations.Added vector search support for
conversationsqueries. Developers can provide search text which will use vector embeddings to return similar conversations.Added
promptSpecificationsmutation for directly prompting multiple models. This can be used to evaluate prompts against multiple models or compare different specification parameters in parallel.Added
promptStrategyfield to Specification, which supports multiple strategy types for preprocessing the prompt before being sent to the LLM model. For example,REWRITEprompt strategy will ask LLM to rewrite the incoming user prompt based on the previous conversation messages.Added
suggestConversationmutation, which returns a list of suggested followup questions based on the specified conversation and related contents. This can be used to auto-suggest questions for chatbot users.Added new summarization types:
CHAPTERS,QUESTIONSandPOSTS. See usage examples in the "LLMs for Podcasters" blog post.Added versioned model enums such as
GPT4_0613andGPT35_TURBO_16K_1106. Without version specified, such asGPT35_TURBO_16K, Graphlit will use the latest production model version, as defined by the LLM vendor.Added
lookupContentsquery to get multiple contents by id in one query.⚡ In Content type,
headlinefield was renamed toheadlinesand now returns an array of strings.⚡ Entity names are now limited to 1024 characters. Names will be truncated if they exceed the maximum length.
⚡ In SummarizationTypes enum,
BULLET_POINTSwas renamed toBULLETS.⚡ In ProjectStorage type,
originalTotalSizewas renamed tototalSize, andtotalRenditionSizefield was added.totalSizeis the sum of the ingested source file sizes, andtotalRenditionSizeis the sum of the source file sizes and any derived rendition sizes.⚡ In ConversationStrategy type,
strategyTypewas renamed totypefor consistency with rest of data model.⚡ In Specification type,
optimizeSearchConversationwas removed, and now is handled byOPTIMIZE_SEARCHprompt strategy.
Bugs Fixed
GPLA-1725: Should ignore RSS.xml from web feed sitemap
GPLA-1726: GPT-3.5 Turbo 16k LLM is adding "Citation #" to response
GPLA-1698: Workflow not applied to link-crawled content
GPLA-1692: Mismatched project storage total size, when some content has errored
GPLA-1237: Add relevance threshold for semantic search
Last updated