Releases: svilupp/PromptingTools.jl
Releases · svilupp/PromptingTools.jl
v0.73.0
PromptingTools v0.73.0
Breaking changes
Added
- Added a new
enable_retry!
function to enable the retry layer for HTTP requests. Main use case is to prevent rate limiting errors from external AI services, but you can fully customize the retry behavior for any HTTP status codes, etc. See?enable_retry!
for more information.
Updated
- Updated the
StreamCallbacks
dependency to0.6
to improve parsing of streaming responses with thinking from Anthropic.
Commits
Merged pull requests:
v0.72.3
PromptingTools v0.72.3
Breaking Changes
Added
- Added OpenAI's GPT 4.5 research preview (alias gpt45) to the model registry.
Commits
Merged pull requests:
v0.72.2
PromptingTools v0.72.2
Breaking Changes
Added
- Added Claude 3.7 Sonnet model to the model registry (alias
claude37
and updated defaultclaude
alias to point to it).
Commits
Merged pull requests:
v0.72.1
PromptingTools v0.72.1
Added
- Added Qwen models from OpenRouter API: Qwen Turbo, Plus and Max models with competitive pricing and 128K context window (aliases
orqwenturbo
,orqwenplus
andorqwenmax
, respectively).
Commits
Merged pull requests:
v0.72.0
PromptingTools v0.72.0
Breaking Changes
Added
- Added new Gemini 2.0 Pro and Flash models to the model registry (aliases
gem20p
andgem20f
, respectively). Added Gemini 2.0 Flash Lite Preview model (aliasgem20fl
) and Gemini 2.0 Flash Thinking Experimental model (aliasgem20ft
). - Added BM25 normalization kwarg to
RAGTools.jl
to enable 0-1 query-specific normalization of BM25 scores for easier filtering and comparison. See?RT.bm25
and?RT.max_bm25_score
for more information.
Commits
Merged pull requests:
v0.71.0
PromptingTools v0.71.0
Breaking changes
Added
- Added the new OpenAI's O3 Mini reasoning model to the model registry (alias
o3m
). - Added the new DeepSeek R1 Distill Llama 70b model hosted on GroqCloud, which beats GPT 4o in many benchmarks while being upto 10x cheaper (alias
glmr
- stands for Groq Llama Medium(70b) Reasoning). - Added experimental support for "thinking tokens" that can be found with DeepSeek API and the reasoning model R1 (alias
dsr
). If the thought chain is provided, we save it in themsg.extras[:reasoning_content]
field for advanced users.
Commits
Merged pull requests:
v0.70.0
PromptingTools v0.70.0
Breaking Changes
Added
- Added new DeepSeek V3 model via Deepseek (alias
dschat
or simplyds
, because they are in a category of their own), Fireworks.ai and Together.ai (fds
andtds
for hosted Deepseek V3, respectively). Added Qwen 2.5 Coder 32B (aliasfqwen25c
ortqwen25c
for Fireworks.ai and Together.ai, respectively). - Added the reasoning Qwen QwQ 32B hosted on Together.ai.
- Added the new OpenAI's O1 model to the model registry (alias
o1
). - Added DeepSeek Reasoner model alias (
dsreason
or simplydsr
) from DeepSeek - Added MiniMax model support and alias (
minimax
) for MiniMax-Text-01 model
Commits
Merged pull requests:
v0.69.1
PromptingTools v0.69.1
Fixed
- Added assertion in
response_to_message
for missing:tool_calls
key in the response message. It's model failure but it wasn't obvious from the original error. - Fixed error for usage information in CamelCase from OpenAI servers (Gemini proxy now sends it in CamelCase).
Commits
Merged pull requests:
Closed issues:
- Update Gemini Experimental model to 1206 (#253)
v0.69.0
PromptingTools v0.69.0
Added
- Added a new Gemini 2.0 Flash Experimental model (
gemini-2.0-flash-exp
) and updated the aliasgem20f
with it.
Commits
Merged pull requests:
v0.68.0
PromptingTools v0.68.0
Added
- Added a new
cache=:all_but_last
cache strategy for Anthropic models to enable caching of the entire conversation except for the last user message (useful for longer conversations that you want to re-use, but not continue). See the docstrings for more information on which cache strategy to use. Thank you to @Sixzero for this!
Commits
Merged pull requests: