December 10: Support for OpenAI GPT-4 Turbo, Llama 2 and Mistral models; query by example, bug fixes
New Features
💡 Graphlit now supports the OpenAI GPT-4 Turbo 128k model, both in Azure OpenAI and native OpenAI services. Added new model enum
GPT4_TURBO_VISION_128K
.💡 Graphlit now supports Llama 2 7b, 13b, 70b models and Mistral 7b model, via Replicate. Developers can use their own Replicate API key, or be charged as credits for Graphlit usage.
💡 Graphlit now supports the Anthropic Claude 2.1 model. Added new model enum
CLAUDE_2_1
.💡 Graphlit now supports the OpenAI GPT-4 Vision model for image descriptions and text extraction. Added new model enum
GPT4_TURBO_VISION_128K
. See usage example in "Multimodal RAG" blog post.Added query by example to
contents
query. Developers can specify one or more example contents, and query will use vector embeddings to return similar contents.Added query by example to
conversations
query. Developers can specify one or more example conversations, and query will use vector embeddings to return similar conversations.Added vector search support for
conversations
queries. Developers can provide search text which will use vector embeddings to return similar conversations.Added
promptSpecifications
mutation for directly prompting multiple models. This can be used to evaluate prompts against multiple models or compare different specification parameters in parallel.Added
promptStrategy
field to Specification, which supports multiple strategy types for preprocessing the prompt before being sent to the LLM model. For example,REWRITE
prompt strategy will ask LLM to rewrite the incoming user prompt based on the previous conversation messages.Added
suggestConversation
mutation, which returns a list of suggested followup questions based on the specified conversation and related contents. This can be used to auto-suggest questions for chatbot users.Added new summarization types:
CHAPTERS
,QUESTIONS
andPOSTS
. See usage examples in the "LLMs for Podcasters" blog post.Added versioned model enums such as
GPT4_0613
andGPT35_TURBO_16K_1106
. Without version specified, such asGPT35_TURBO_16K
, Graphlit will use the latest production model version, as defined by the LLM vendor.Added
lookupContents
query to get multiple contents by id in one query.⚡ In Content type,
headline
field was renamed toheadlines
and now returns an array of strings.⚡ Entity names are now limited to 1024 characters. Names will be truncated if they exceed the maximum length.
⚡ In SummarizationTypes enum,
BULLET_POINTS
was renamed toBULLETS
.⚡ In ProjectStorage type,
originalTotalSize
was renamed tototalSize
, andtotalRenditionSize
field was added.totalSize
is the sum of the ingested source file sizes, andtotalRenditionSize
is the sum of the source file sizes and any derived rendition sizes.⚡ In ConversationStrategy type,
strategyType
was renamed totype
for consistency with rest of data model.⚡ In Specification type,
optimizeSearchConversation
was removed, and now is handled byOPTIMIZE_SEARCH
prompt strategy.
Bugs Fixed
GPLA-1725: Should ignore RSS.xml from web feed sitemap
GPLA-1726: GPT-3.5 Turbo 16k LLM is adding "Citation #" to response
GPLA-1698: Workflow not applied to link-crawled content
GPLA-1692: Mismatched project storage total size, when some content has errored
GPLA-1237: Add relevance threshold for semantic search
Last updated