ruby_llm 1.10.0 → 1.11.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/README.md +2 -2
- data/lib/ruby_llm/aliases.json +4 -4
- data/lib/ruby_llm/configuration.rb +1 -0
- data/lib/ruby_llm/models.json +920 -1005
- data/lib/ruby_llm/provider.rb +5 -1
- data/lib/ruby_llm/providers/anthropic/media.rb +2 -2
- data/lib/ruby_llm/providers/bedrock/chat.rb +10 -1
- data/lib/ruby_llm/providers/openai/media.rb +1 -1
- data/lib/ruby_llm/providers/xai/chat.rb +15 -0
- data/lib/ruby_llm/providers/xai/models.rb +75 -0
- data/lib/ruby_llm/providers/xai.rb +28 -0
- data/lib/ruby_llm/version.rb +1 -1
- data/lib/ruby_llm.rb +10 -8
- data/lib/tasks/models.rake +1 -0
- metadata +4 -1
checksums.yaml
CHANGED
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
---
|
|
2
2
|
SHA256:
|
|
3
|
-
metadata.gz:
|
|
4
|
-
data.tar.gz:
|
|
3
|
+
metadata.gz: c6cabf287dc7cd62616a9061131291f404c391be59f1468ccdb3b7f7cb9d2cb5
|
|
4
|
+
data.tar.gz: bf1123d01a7ddadcbfde46a2ffb6e378928bcbcc0b1cdb0b927b7905f0534c12
|
|
5
5
|
SHA512:
|
|
6
|
-
metadata.gz:
|
|
7
|
-
data.tar.gz:
|
|
6
|
+
metadata.gz: 12c1f267367279716c8e5e7fe6ef44327e4921b4e24516313f99954775a4c51d7e874d42184f466f8eb71ebe4f814b14a93b63b043c46096e7be4d804a15fc76
|
|
7
|
+
data.tar.gz: dd2d797e49e3c1741108b63f191ccc3252d1fec1868851802ceb33b65afec4609498cbdb8ee516800e48f04220fdd4179b43abb645177e8ed14ea25c1036916d
|
data/README.md
CHANGED
|
@@ -22,7 +22,7 @@ Battle tested at [<picture><source media="(prefers-color-scheme: dark)" srcset="
|
|
|
22
22
|
|
|
23
23
|
---
|
|
24
24
|
|
|
25
|
-
Build chatbots, AI agents, RAG applications. Works with OpenAI, Anthropic, Google, AWS, local models, and any OpenAI-compatible API.
|
|
25
|
+
Build chatbots, AI agents, RAG applications. Works with OpenAI, xAI, Anthropic, Google, AWS, local models, and any OpenAI-compatible API.
|
|
26
26
|
|
|
27
27
|
## Why RubyLLM?
|
|
28
28
|
|
|
@@ -124,7 +124,7 @@ response = chat.with_schema(ProductSchema).ask "Analyze this product", with: "pr
|
|
|
124
124
|
* **Async:** Fiber-based concurrency
|
|
125
125
|
* **Model registry:** 800+ models with capability detection and pricing
|
|
126
126
|
* **Extended thinking:** Control, view, and persist model deliberation
|
|
127
|
-
* **Providers:** OpenAI, Anthropic, Gemini, VertexAI, Bedrock, DeepSeek, Mistral, Ollama, OpenRouter, Perplexity, GPUStack, and any OpenAI-compatible API
|
|
127
|
+
* **Providers:** OpenAI, xAI, Anthropic, Gemini, VertexAI, Bedrock, DeepSeek, Mistral, Ollama, OpenRouter, Perplexity, GPUStack, and any OpenAI-compatible API
|
|
128
128
|
|
|
129
129
|
## Installation
|
|
130
130
|
|
data/lib/ruby_llm/aliases.json
CHANGED
|
@@ -120,10 +120,6 @@
|
|
|
120
120
|
"gemini": "gemini-2.5-flash-image",
|
|
121
121
|
"openrouter": "google/gemini-2.5-flash-image"
|
|
122
122
|
},
|
|
123
|
-
"gemini-2.5-flash-image-preview": {
|
|
124
|
-
"gemini": "gemini-2.5-flash-image-preview",
|
|
125
|
-
"openrouter": "google/gemini-2.5-flash-image-preview"
|
|
126
|
-
},
|
|
127
123
|
"gemini-2.5-flash-lite": {
|
|
128
124
|
"gemini": "gemini-2.5-flash-lite",
|
|
129
125
|
"openrouter": "google/gemini-2.5-flash-lite",
|
|
@@ -332,6 +328,10 @@
|
|
|
332
328
|
"openai": "gpt-5.2-chat-latest",
|
|
333
329
|
"openrouter": "openai/gpt-5.2-chat-latest"
|
|
334
330
|
},
|
|
331
|
+
"gpt-5.2-codex": {
|
|
332
|
+
"openai": "gpt-5.2-codex",
|
|
333
|
+
"openrouter": "openai/gpt-5.2-codex"
|
|
334
|
+
},
|
|
335
335
|
"gpt-5.2-pro": {
|
|
336
336
|
"openai": "gpt-5.2-pro",
|
|
337
337
|
"openrouter": "openai/gpt-5.2-pro"
|