Releases
v0.12.37
Release Notes
llama-index-core
[0.12.37]
Ensure Memory
returns at least one message (#18763 )
Separate text blocks with newlines when accessing message.content
(#18763 )
reset next_agent
in multi agent workflows (#18782 )
support sqlalchemy v1 in chat store (#18780 )
fix: broken hotpotqa dataset URL (#18764 )
Use get_tqdm_iterable
in SimpleDirectoryReader (#18722 )
Pass agent workflow kwargs into start event (#18747 )
fix(chunking): Ensure correct handling of multi-byte characters during AST node chunking (#18702 )
llama-index-llms-anthropic
[0.6.14]
Fixed DocumentBlock handling in OpenAI and Anthropic (#18769 )
llama-index-llms-bedrock-converse
[0.5.4]
Fix tool call parsing for bedrock converse (#18781 )
feat: add missing client params for bedrock (#18768 )
fix merging multiple tool calls in bedrock converse (#18761 )
llama-index-llms-openai
[0.3.42]
Fixed DocumentBlock handling in OpenAI and Anthropic (#18769 )
Remove tool-length check in openai (#18784 )
Add check for empty tool call delta, bump version (#18745 )
llama-index-llms-openai-like
[0.3.5]
Remove tool-length check in openai (#18784 )
llama-index-retrievers-vectorize
[0.1.0]
Add Vectorize retriever (#18685 )
llama-index-tools-desearch
[0.1.0]
Feature/desearch integration (#18738 )
You can’t perform that action at this time.