@promptbook/wizard 0.100.0-28 → 0.100.0-31
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/esm/index.es.js +261 -23
- package/esm/index.es.js.map +1 -1
- package/esm/typings/src/_packages/types.index.d.ts +2 -2
- package/esm/typings/src/book-2.0/agent-source/parseAgentSource.d.ts +2 -2
- package/esm/typings/src/book-2.0/commitments/_misc/parseAgentSourceWithCommitments.d.ts +2 -2
- package/esm/typings/src/book-components/BookEditor/BookEditor.d.ts +0 -2
- package/esm/typings/src/execution/AvailableModel.d.ts +4 -0
- package/esm/typings/src/llm-providers/anthropic-claude/anthropic-claude-models.d.ts +1 -1
- package/esm/typings/src/llm-providers/deepseek/deepseek-models.d.ts +1 -1
- package/esm/typings/src/llm-providers/google/google-models.d.ts +1 -1
- package/esm/typings/src/llm-providers/ollama/ollama-models.d.ts +1 -1
- package/esm/typings/src/llm-providers/openai/openai-models.d.ts +1 -1
- package/esm/typings/src/version.d.ts +1 -1
- package/package.json +2 -2
- package/umd/index.umd.js +261 -23
- package/umd/index.umd.js.map +1 -1
@@ -1,4 +1,4 @@
|
|
1
|
-
import type {
|
1
|
+
import type { AgentBasicInformation } from '../book-2.0/agent-source/parseAgentSource';
|
2
2
|
import type { string_book } from '../book-2.0/agent-source/string_book';
|
3
3
|
import type { BookCommitment } from '../book-2.0/commitments/_base/BookCommitment';
|
4
4
|
import type { CommitmentDefinition } from '../book-2.0/commitments/_base/CommitmentDefinition';
|
@@ -317,7 +317,7 @@ import type { TODO_any } from '../utils/organization/TODO_any';
|
|
317
317
|
import type { CheckSerializableAsJsonOptions } from '../utils/serialization/checkSerializableAsJson';
|
318
318
|
import type { ExportJsonOptions } from '../utils/serialization/exportJson';
|
319
319
|
import type { string_promptbook_version } from '../version';
|
320
|
-
export type {
|
320
|
+
export type { AgentBasicInformation };
|
321
321
|
export type { string_book };
|
322
322
|
export type { BookCommitment };
|
323
323
|
export type { CommitmentDefinition };
|
@@ -1,7 +1,7 @@
|
|
1
1
|
import type { string_agent_name } from '../../types/typeAliases';
|
2
2
|
import type { string_url_image } from '../../types/typeAliases';
|
3
3
|
import type { string_book } from './string_book';
|
4
|
-
export interface
|
4
|
+
export interface AgentBasicInformation {
|
5
5
|
/**
|
6
6
|
* Name of the agent
|
7
7
|
* This is the first line of the agent source
|
@@ -27,4 +27,4 @@ export interface AgentSourceBasicInformation {
|
|
27
27
|
*
|
28
28
|
* @public exported from `@promptbook/core`
|
29
29
|
*/
|
30
|
-
export declare function parseAgentSource(agentSource: string_book):
|
30
|
+
export declare function parseAgentSource(agentSource: string_book): AgentBasicInformation;
|
@@ -1,4 +1,4 @@
|
|
1
|
-
import type {
|
1
|
+
import type { AgentBasicInformation } from '../../agent-source/parseAgentSource';
|
2
2
|
import type { string_book } from '../../agent-source/string_book';
|
3
3
|
import type { AgentSourceParseResult } from './AgentSourceParseResult';
|
4
4
|
/**
|
@@ -14,7 +14,7 @@ export declare function parseAgentSourceWithCommitments(agentSource: string_book
|
|
14
14
|
*
|
15
15
|
* @private
|
16
16
|
*/
|
17
|
-
export declare function parseAgentSourceBasicInfo(agentSource: string_book):
|
17
|
+
export declare function parseAgentSourceBasicInfo(agentSource: string_book): AgentBasicInformation;
|
18
18
|
/**
|
19
19
|
* Extracts META LINK commitments from agent source
|
20
20
|
* Returns an array of all META LINK URLs found in the agent source
|
@@ -21,8 +21,6 @@ export interface BookEditorProps {
|
|
21
21
|
/**
|
22
22
|
* Renders a book editor
|
23
23
|
*
|
24
|
-
* 🔥 LIVE COMPONENT TEST: This component is being served live from the playground server!
|
25
|
-
*
|
26
24
|
* @public exported from `@promptbook/components`
|
27
25
|
*/
|
28
26
|
export declare function BookEditor(props: BookEditorProps): import("react/jsx-runtime").JSX.Element;
|
@@ -40,6 +40,10 @@ export type AvailableModel = {
|
|
40
40
|
readonly prompt: number_usd;
|
41
41
|
readonly output: number_usd;
|
42
42
|
};
|
43
|
+
/**
|
44
|
+
* If the model is deprecated, it should not be used for new tasks
|
45
|
+
*/
|
46
|
+
readonly isDeprecated?: boolean;
|
43
47
|
};
|
44
48
|
/**
|
45
49
|
* TODO: [🧠] Maybe rename to something else - like `ModelInformation` or `ModelMetadata`
|
@@ -3,7 +3,7 @@ import type { number_usd } from '../../types/typeAliases';
|
|
3
3
|
/**
|
4
4
|
* List of available Anthropic Claude models with pricing
|
5
5
|
*
|
6
|
-
* Note:
|
6
|
+
* Note: Synced with official API docs at 2025-08-20
|
7
7
|
*
|
8
8
|
* @see https://docs.anthropic.com/en/docs/models-overview
|
9
9
|
* @public exported from `@promptbook/anthropic-claude`
|
@@ -3,7 +3,7 @@ import type { number_usd } from '../../types/typeAliases';
|
|
3
3
|
/**
|
4
4
|
* List of available Deepseek models with descriptions
|
5
5
|
*
|
6
|
-
* Note:
|
6
|
+
* Note: Synced with official API docs at 2025-08-20
|
7
7
|
*
|
8
8
|
* @see https://www.deepseek.com/models
|
9
9
|
* @public exported from `@promptbook/deepseek`
|
@@ -3,7 +3,7 @@ import type { number_usd } from '../../types/typeAliases';
|
|
3
3
|
/**
|
4
4
|
* List of available Google models with descriptions
|
5
5
|
*
|
6
|
-
* Note:
|
6
|
+
* Note: Synced with official API docs at 2025-08-20
|
7
7
|
*
|
8
8
|
* @see https://ai.google.dev/models/gemini
|
9
9
|
* @public exported from `@promptbook/google`
|
@@ -2,7 +2,7 @@ import type { AvailableModel } from '../../execution/AvailableModel';
|
|
2
2
|
/**
|
3
3
|
* List of available models in Ollama library
|
4
4
|
*
|
5
|
-
* Note:
|
5
|
+
* Note: Synced with official API docs at 2025-08-20
|
6
6
|
*
|
7
7
|
* @see https://ollama.com/library
|
8
8
|
* @public exported from `@promptbook/ollama`
|
@@ -2,7 +2,7 @@ import type { AvailableModel } from '../../execution/AvailableModel';
|
|
2
2
|
/**
|
3
3
|
* List of available OpenAI models with pricing
|
4
4
|
*
|
5
|
-
* Note:
|
5
|
+
* Note: Synced with official API docs at 2025-08-20
|
6
6
|
*
|
7
7
|
* @see https://platform.openai.com/docs/models/
|
8
8
|
* @see https://openai.com/api/pricing/
|
@@ -15,7 +15,7 @@ export declare const BOOK_LANGUAGE_VERSION: string_semantic_version;
|
|
15
15
|
export declare const PROMPTBOOK_ENGINE_VERSION: string_promptbook_version;
|
16
16
|
/**
|
17
17
|
* Represents the version string of the Promptbook engine.
|
18
|
-
* It follows semantic versioning (e.g., `0.100.0-
|
18
|
+
* It follows semantic versioning (e.g., `0.100.0-30`).
|
19
19
|
*
|
20
20
|
* @generated
|
21
21
|
*/
|
package/package.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1
1
|
{
|
2
2
|
"name": "@promptbook/wizard",
|
3
|
-
"version": "0.100.0-
|
3
|
+
"version": "0.100.0-31",
|
4
4
|
"description": "Promptbook: Run AI apps in plain human language across multiple models and platforms",
|
5
5
|
"private": false,
|
6
6
|
"sideEffects": false,
|
@@ -95,7 +95,7 @@
|
|
95
95
|
"module": "./esm/index.es.js",
|
96
96
|
"typings": "./esm/typings/src/_packages/wizard.index.d.ts",
|
97
97
|
"peerDependencies": {
|
98
|
-
"@promptbook/core": "0.100.0-
|
98
|
+
"@promptbook/core": "0.100.0-31"
|
99
99
|
},
|
100
100
|
"dependencies": {
|
101
101
|
"@ai-sdk/deepseek": "0.1.6",
|
package/umd/index.umd.js
CHANGED
@@ -49,7 +49,7 @@
|
|
49
49
|
* @generated
|
50
50
|
* @see https://github.com/webgptorg/promptbook
|
51
51
|
*/
|
52
|
-
const PROMPTBOOK_ENGINE_VERSION = '0.100.0-
|
52
|
+
const PROMPTBOOK_ENGINE_VERSION = '0.100.0-31';
|
53
53
|
/**
|
54
54
|
* TODO: string_promptbook_version should be constrained to the all versions of Promptbook engine
|
55
55
|
* Note: [💞] Ignore a discrepancy between file name and entity name
|
@@ -1713,7 +1713,7 @@
|
|
1713
1713
|
/**
|
1714
1714
|
* List of available Anthropic Claude models with pricing
|
1715
1715
|
*
|
1716
|
-
* Note:
|
1716
|
+
* Note: Synced with official API docs at 2025-08-20
|
1717
1717
|
*
|
1718
1718
|
* @see https://docs.anthropic.com/en/docs/models-overview
|
1719
1719
|
* @public exported from `@promptbook/anthropic-claude`
|
@@ -1723,12 +1723,52 @@
|
|
1723
1723
|
value: [
|
1724
1724
|
{
|
1725
1725
|
modelVariant: 'CHAT',
|
1726
|
-
modelTitle: 'Claude
|
1727
|
-
modelName: 'claude-
|
1728
|
-
modelDescription: '
|
1726
|
+
modelTitle: 'Claude Opus 4.1',
|
1727
|
+
modelName: 'claude-opus-4-1-20250805',
|
1728
|
+
modelDescription: 'Most powerful and capable Claude model with 200K token context window. Features superior reasoning capabilities, exceptional coding abilities, and advanced multimodal understanding. Sets new standards in complex reasoning and analytical tasks with enhanced safety measures. Ideal for the most demanding enterprise applications requiring maximum intelligence.',
|
1729
1729
|
pricing: {
|
1730
|
-
prompt: pricing(`$
|
1731
|
-
output: pricing(`$
|
1730
|
+
prompt: pricing(`$15.00 / 1M tokens`),
|
1731
|
+
output: pricing(`$75.00 / 1M tokens`),
|
1732
|
+
},
|
1733
|
+
},
|
1734
|
+
{
|
1735
|
+
modelVariant: 'CHAT',
|
1736
|
+
modelTitle: 'Claude Opus 4',
|
1737
|
+
modelName: 'claude-opus-4-20250514',
|
1738
|
+
modelDescription: 'Previous flagship Claude model with 200K token context window. Features very high intelligence and capability with exceptional performance across reasoning, coding, and creative tasks. Maintains strong safety guardrails while delivering sophisticated outputs for complex professional applications.',
|
1739
|
+
pricing: {
|
1740
|
+
prompt: pricing(`$15.00 / 1M tokens`),
|
1741
|
+
output: pricing(`$75.00 / 1M tokens`),
|
1742
|
+
},
|
1743
|
+
},
|
1744
|
+
{
|
1745
|
+
modelVariant: 'CHAT',
|
1746
|
+
modelTitle: 'Claude Sonnet 4',
|
1747
|
+
modelName: 'claude-sonnet-4-20250514',
|
1748
|
+
modelDescription: 'High-performance Claude model with exceptional reasoning capabilities and 200K token context window (1M context beta available). Features balanced intelligence and efficiency with enhanced multimodal understanding. Offers optimal performance for most enterprise applications requiring sophisticated AI capabilities.',
|
1749
|
+
pricing: {
|
1750
|
+
prompt: pricing(`$3.00 / 1M tokens`),
|
1751
|
+
output: pricing(`$15.00 / 1M tokens`),
|
1752
|
+
},
|
1753
|
+
},
|
1754
|
+
{
|
1755
|
+
modelVariant: 'CHAT',
|
1756
|
+
modelTitle: 'Claude Sonnet 3.7',
|
1757
|
+
modelName: 'claude-3-7-sonnet-20250219',
|
1758
|
+
modelDescription: 'High-performance Claude model with early extended thinking capabilities and 200K token context window. Features enhanced reasoning chains, improved factual accuracy, and toggleable extended thinking for complex problem-solving. Ideal for applications requiring deep analytical capabilities.',
|
1759
|
+
pricing: {
|
1760
|
+
prompt: pricing(`$3.00 / 1M tokens`),
|
1761
|
+
output: pricing(`$15.00 / 1M tokens`),
|
1762
|
+
},
|
1763
|
+
},
|
1764
|
+
{
|
1765
|
+
modelVariant: 'CHAT',
|
1766
|
+
modelTitle: 'Claude Haiku 3.5',
|
1767
|
+
modelName: 'claude-3-5-haiku-20241022',
|
1768
|
+
modelDescription: 'Fastest Claude model with 200K token context window optimized for intelligence at blazing speeds. Features enhanced reasoning and contextual understanding while maintaining sub-second response times. Perfect for real-time applications, customer-facing deployments, and high-throughput services.',
|
1769
|
+
pricing: {
|
1770
|
+
prompt: pricing(`$0.80 / 1M tokens`),
|
1771
|
+
output: pricing(`$4.00 / 1M tokens`),
|
1732
1772
|
},
|
1733
1773
|
},
|
1734
1774
|
{
|
@@ -2514,7 +2554,7 @@
|
|
2514
2554
|
* Default model for chat variant.
|
2515
2555
|
*/
|
2516
2556
|
getDefaultChatModel() {
|
2517
|
-
return this.getDefaultModel('claude-
|
2557
|
+
return this.getDefaultModel('claude-sonnet-4-20250514');
|
2518
2558
|
}
|
2519
2559
|
}
|
2520
2560
|
/**
|
@@ -2657,7 +2697,7 @@
|
|
2657
2697
|
/**
|
2658
2698
|
* List of available OpenAI models with pricing
|
2659
2699
|
*
|
2660
|
-
* Note:
|
2700
|
+
* Note: Synced with official API docs at 2025-08-20
|
2661
2701
|
*
|
2662
2702
|
* @see https://platform.openai.com/docs/models/
|
2663
2703
|
* @see https://openai.com/api/pricing/
|
@@ -2673,8 +2713,128 @@
|
|
2673
2713
|
modelName: 'gpt-5',
|
2674
2714
|
modelDescription: "OpenAI's most advanced language model with unprecedented reasoning capabilities and 200K context window. Features revolutionary improvements in complex problem-solving, scientific reasoning, and creative tasks. Demonstrates human-level performance across diverse domains with enhanced safety measures and alignment. Represents the next generation of AI with superior understanding, nuanced responses, and advanced multimodal capabilities.",
|
2675
2715
|
pricing: {
|
2676
|
-
prompt: pricing(`$
|
2677
|
-
output: pricing(`$
|
2716
|
+
prompt: pricing(`$1.25 / 1M tokens`),
|
2717
|
+
output: pricing(`$10.00 / 1M tokens`),
|
2718
|
+
},
|
2719
|
+
},
|
2720
|
+
/**/
|
2721
|
+
/**/
|
2722
|
+
{
|
2723
|
+
modelVariant: 'CHAT',
|
2724
|
+
modelTitle: 'gpt-5-mini',
|
2725
|
+
modelName: 'gpt-5-mini',
|
2726
|
+
modelDescription: "A faster, cost-efficient version of GPT-5 for well-defined tasks with 200K context window. Maintains core GPT-5 capabilities while offering 5x faster inference and significantly lower costs. Features enhanced instruction following and reduced latency for production applications requiring quick responses with high quality.",
|
2727
|
+
pricing: {
|
2728
|
+
prompt: pricing(`$0.25 / 1M tokens`),
|
2729
|
+
output: pricing(`$2.00 / 1M tokens`),
|
2730
|
+
},
|
2731
|
+
},
|
2732
|
+
/**/
|
2733
|
+
/**/
|
2734
|
+
{
|
2735
|
+
modelVariant: 'CHAT',
|
2736
|
+
modelTitle: 'gpt-5-nano',
|
2737
|
+
modelName: 'gpt-5-nano',
|
2738
|
+
modelDescription: "The fastest, most cost-efficient version of GPT-5 with 200K context window. Optimized for summarization, classification, and simple reasoning tasks. Features 10x faster inference than base GPT-5 while maintaining good quality for straightforward applications. Ideal for high-volume, cost-sensitive deployments.",
|
2739
|
+
pricing: {
|
2740
|
+
prompt: pricing(`$0.05 / 1M tokens`),
|
2741
|
+
output: pricing(`$0.40 / 1M tokens`),
|
2742
|
+
},
|
2743
|
+
},
|
2744
|
+
/**/
|
2745
|
+
/**/
|
2746
|
+
{
|
2747
|
+
modelVariant: 'CHAT',
|
2748
|
+
modelTitle: 'gpt-4.1',
|
2749
|
+
modelName: 'gpt-4.1',
|
2750
|
+
modelDescription: "Smartest non-reasoning model with 128K context window. Enhanced version of GPT-4 with improved instruction following, better factual accuracy, and reduced hallucinations. Features advanced function calling capabilities and superior performance on coding tasks. Ideal for applications requiring high intelligence without reasoning overhead.",
|
2751
|
+
pricing: {
|
2752
|
+
prompt: pricing(`$3.00 / 1M tokens`),
|
2753
|
+
output: pricing(`$12.00 / 1M tokens`),
|
2754
|
+
},
|
2755
|
+
},
|
2756
|
+
/**/
|
2757
|
+
/**/
|
2758
|
+
{
|
2759
|
+
modelVariant: 'CHAT',
|
2760
|
+
modelTitle: 'gpt-4.1-mini',
|
2761
|
+
modelName: 'gpt-4.1-mini',
|
2762
|
+
modelDescription: "Smaller, faster version of GPT-4.1 with 128K context window. Balances intelligence and efficiency with 3x faster inference than base GPT-4.1. Maintains strong capabilities across text generation, reasoning, and coding while offering better cost-performance ratio for most applications.",
|
2763
|
+
pricing: {
|
2764
|
+
prompt: pricing(`$0.80 / 1M tokens`),
|
2765
|
+
output: pricing(`$3.20 / 1M tokens`),
|
2766
|
+
},
|
2767
|
+
},
|
2768
|
+
/**/
|
2769
|
+
/**/
|
2770
|
+
{
|
2771
|
+
modelVariant: 'CHAT',
|
2772
|
+
modelTitle: 'gpt-4.1-nano',
|
2773
|
+
modelName: 'gpt-4.1-nano',
|
2774
|
+
modelDescription: "Fastest, most cost-efficient version of GPT-4.1 with 128K context window. Optimized for high-throughput applications requiring good quality at minimal cost. Features 5x faster inference than GPT-4.1 while maintaining adequate performance for most general-purpose tasks.",
|
2775
|
+
pricing: {
|
2776
|
+
prompt: pricing(`$0.20 / 1M tokens`),
|
2777
|
+
output: pricing(`$0.80 / 1M tokens`),
|
2778
|
+
},
|
2779
|
+
},
|
2780
|
+
/**/
|
2781
|
+
/**/
|
2782
|
+
{
|
2783
|
+
modelVariant: 'CHAT',
|
2784
|
+
modelTitle: 'o3',
|
2785
|
+
modelName: 'o3',
|
2786
|
+
modelDescription: "Advanced reasoning model with 128K context window specializing in complex logical, mathematical, and analytical tasks. Successor to o1 with enhanced step-by-step problem-solving capabilities and superior performance on STEM-focused problems. Ideal for professional applications requiring deep analytical thinking and precise reasoning.",
|
2787
|
+
pricing: {
|
2788
|
+
prompt: pricing(`$15.00 / 1M tokens`),
|
2789
|
+
output: pricing(`$60.00 / 1M tokens`),
|
2790
|
+
},
|
2791
|
+
},
|
2792
|
+
/**/
|
2793
|
+
/**/
|
2794
|
+
{
|
2795
|
+
modelVariant: 'CHAT',
|
2796
|
+
modelTitle: 'o3-pro',
|
2797
|
+
modelName: 'o3-pro',
|
2798
|
+
modelDescription: "Enhanced version of o3 with more compute allocated for better responses on the most challenging problems. Features extended reasoning time and improved accuracy on complex analytical tasks. Designed for applications where maximum reasoning quality is more important than response speed.",
|
2799
|
+
pricing: {
|
2800
|
+
prompt: pricing(`$30.00 / 1M tokens`),
|
2801
|
+
output: pricing(`$120.00 / 1M tokens`),
|
2802
|
+
},
|
2803
|
+
},
|
2804
|
+
/**/
|
2805
|
+
/**/
|
2806
|
+
{
|
2807
|
+
modelVariant: 'CHAT',
|
2808
|
+
modelTitle: 'o4-mini',
|
2809
|
+
modelName: 'o4-mini',
|
2810
|
+
modelDescription: "Fast, cost-efficient reasoning model with 128K context window. Successor to o1-mini with improved analytical capabilities while maintaining speed advantages. Features enhanced mathematical reasoning and logical problem-solving at significantly lower cost than full reasoning models.",
|
2811
|
+
pricing: {
|
2812
|
+
prompt: pricing(`$4.00 / 1M tokens`),
|
2813
|
+
output: pricing(`$16.00 / 1M tokens`),
|
2814
|
+
},
|
2815
|
+
},
|
2816
|
+
/**/
|
2817
|
+
/**/
|
2818
|
+
{
|
2819
|
+
modelVariant: 'CHAT',
|
2820
|
+
modelTitle: 'o3-deep-research',
|
2821
|
+
modelName: 'o3-deep-research',
|
2822
|
+
modelDescription: "Most powerful deep research model with 128K context window. Specialized for comprehensive research tasks, literature analysis, and complex information synthesis. Features advanced citation capabilities and enhanced factual accuracy for academic and professional research applications.",
|
2823
|
+
pricing: {
|
2824
|
+
prompt: pricing(`$25.00 / 1M tokens`),
|
2825
|
+
output: pricing(`$100.00 / 1M tokens`),
|
2826
|
+
},
|
2827
|
+
},
|
2828
|
+
/**/
|
2829
|
+
/**/
|
2830
|
+
{
|
2831
|
+
modelVariant: 'CHAT',
|
2832
|
+
modelTitle: 'o4-mini-deep-research',
|
2833
|
+
modelName: 'o4-mini-deep-research',
|
2834
|
+
modelDescription: "Faster, more affordable deep research model with 128K context window. Balances research capabilities with cost efficiency, offering good performance on literature review, fact-checking, and information synthesis tasks at a more accessible price point.",
|
2835
|
+
pricing: {
|
2836
|
+
prompt: pricing(`$12.00 / 1M tokens`),
|
2837
|
+
output: pricing(`$48.00 / 1M tokens`),
|
2678
2838
|
},
|
2679
2839
|
},
|
2680
2840
|
/**/
|
@@ -3687,7 +3847,7 @@
|
|
3687
3847
|
/**
|
3688
3848
|
* List of available Deepseek models with descriptions
|
3689
3849
|
*
|
3690
|
-
* Note:
|
3850
|
+
* Note: Synced with official API docs at 2025-08-20
|
3691
3851
|
*
|
3692
3852
|
* @see https://www.deepseek.com/models
|
3693
3853
|
* @public exported from `@promptbook/deepseek`
|
@@ -3697,12 +3857,32 @@
|
|
3697
3857
|
value: [
|
3698
3858
|
{
|
3699
3859
|
modelVariant: 'CHAT',
|
3700
|
-
modelTitle: '
|
3701
|
-
modelName: 'deepseek-chat
|
3702
|
-
modelDescription: 'Latest flagship general-purpose model with
|
3860
|
+
modelTitle: 'DeepSeek V3',
|
3861
|
+
modelName: 'deepseek-chat',
|
3862
|
+
modelDescription: 'Latest flagship general-purpose model with 128K context window. Features exceptional reasoning capabilities, advanced code generation, and strong performance across diverse domains. Offers competitive performance with leading models while maintaining cost efficiency. Ideal for complex reasoning, coding, and knowledge-intensive tasks.',
|
3703
3863
|
pricing: {
|
3704
|
-
prompt: pricing(`$
|
3705
|
-
output: pricing(`$
|
3864
|
+
prompt: pricing(`$0.14 / 1M tokens`),
|
3865
|
+
output: pricing(`$0.28 / 1M tokens`),
|
3866
|
+
},
|
3867
|
+
},
|
3868
|
+
{
|
3869
|
+
modelVariant: 'CHAT',
|
3870
|
+
modelTitle: 'DeepSeek R1',
|
3871
|
+
modelName: 'deepseek-reasoner',
|
3872
|
+
modelDescription: 'Advanced reasoning model with 128K context window specializing in complex problem-solving and analytical thinking. Features explicit reasoning chains, enhanced mathematical capabilities, and superior performance on STEM tasks. Designed for applications requiring deep analytical reasoning and step-by-step problem solving.',
|
3873
|
+
pricing: {
|
3874
|
+
prompt: pricing(`$0.55 / 1M tokens`),
|
3875
|
+
output: pricing(`$2.19 / 1M tokens`),
|
3876
|
+
},
|
3877
|
+
},
|
3878
|
+
{
|
3879
|
+
modelVariant: 'CHAT',
|
3880
|
+
modelTitle: 'DeepSeek Coder V2',
|
3881
|
+
modelName: 'deepseek-coder',
|
3882
|
+
modelDescription: 'Specialized coding model with 128K context window optimized for software development tasks. Features exceptional code generation, debugging, and refactoring capabilities across 40+ programming languages. Particularly strong in understanding complex codebases and implementing solutions based on natural language specifications.',
|
3883
|
+
pricing: {
|
3884
|
+
prompt: pricing(`$0.14 / 1M tokens`),
|
3885
|
+
output: pricing(`$0.28 / 1M tokens`),
|
3706
3886
|
},
|
3707
3887
|
},
|
3708
3888
|
{
|
@@ -3936,7 +4116,7 @@
|
|
3936
4116
|
/**
|
3937
4117
|
* List of available Google models with descriptions
|
3938
4118
|
*
|
3939
|
-
* Note:
|
4119
|
+
* Note: Synced with official API docs at 2025-08-20
|
3940
4120
|
*
|
3941
4121
|
* @see https://ai.google.dev/models/gemini
|
3942
4122
|
* @public exported from `@promptbook/google`
|
@@ -3947,11 +4127,51 @@
|
|
3947
4127
|
{
|
3948
4128
|
modelVariant: 'CHAT',
|
3949
4129
|
modelTitle: 'Gemini 2.5 Pro',
|
3950
|
-
modelName: 'gemini-2.5-pro
|
3951
|
-
modelDescription: '
|
4130
|
+
modelName: 'gemini-2.5-pro',
|
4131
|
+
modelDescription: 'State-of-the-art thinking model with 1M token context window capable of reasoning over complex problems in code, math, and STEM. Features enhanced thinking capabilities, advanced multimodal understanding, and superior performance on analytical tasks. Ideal for complex enterprise applications requiring maximum intelligence and reasoning.',
|
3952
4132
|
pricing: {
|
3953
|
-
prompt: pricing(`$
|
3954
|
-
output: pricing(`$
|
4133
|
+
prompt: pricing(`$7.00 / 1M tokens`),
|
4134
|
+
output: pricing(`$21.00 / 1M tokens`),
|
4135
|
+
},
|
4136
|
+
},
|
4137
|
+
{
|
4138
|
+
modelVariant: 'CHAT',
|
4139
|
+
modelTitle: 'Gemini 2.5 Flash',
|
4140
|
+
modelName: 'gemini-2.5-flash',
|
4141
|
+
modelDescription: 'Best model in terms of price-performance with 1M token context window offering well-rounded capabilities. Features adaptive thinking, cost efficiency, and enhanced reasoning for large-scale processing. Ideal for low-latency, high-volume tasks that require thinking and agentic use cases.',
|
4142
|
+
pricing: {
|
4143
|
+
prompt: pricing(`$0.35 / 1M tokens`),
|
4144
|
+
output: pricing(`$1.05 / 1M tokens`),
|
4145
|
+
},
|
4146
|
+
},
|
4147
|
+
{
|
4148
|
+
modelVariant: 'CHAT',
|
4149
|
+
modelTitle: 'Gemini 2.5 Flash Lite',
|
4150
|
+
modelName: 'gemini-2.5-flash-lite',
|
4151
|
+
modelDescription: 'Cost-efficient Gemini 2.5 Flash model optimized for high throughput with 1M token context window. Features thinking capabilities while maintaining the most cost-efficient pricing. Perfect for real-time, low-latency use cases requiring good quality at scale.',
|
4152
|
+
pricing: {
|
4153
|
+
prompt: pricing(`$0.20 / 1M tokens`),
|
4154
|
+
output: pricing(`$0.60 / 1M tokens`),
|
4155
|
+
},
|
4156
|
+
},
|
4157
|
+
{
|
4158
|
+
modelVariant: 'CHAT',
|
4159
|
+
modelTitle: 'Gemini 2.0 Flash',
|
4160
|
+
modelName: 'gemini-2.0-flash',
|
4161
|
+
modelDescription: 'Next-generation model with 1M token context window delivering improved capabilities, superior speed, and realtime streaming. Features enhanced function calling, code execution, and search capabilities. Ideal for applications requiring cutting-edge AI capabilities with fast response times.',
|
4162
|
+
pricing: {
|
4163
|
+
prompt: pricing(`$0.25 / 1M tokens`),
|
4164
|
+
output: pricing(`$0.75 / 1M tokens`),
|
4165
|
+
},
|
4166
|
+
},
|
4167
|
+
{
|
4168
|
+
modelVariant: 'CHAT',
|
4169
|
+
modelTitle: 'Gemini 2.0 Flash Lite',
|
4170
|
+
modelName: 'gemini-2.0-flash-lite',
|
4171
|
+
modelDescription: 'Cost-efficient Gemini 2.0 Flash model optimized for low latency with 1M token context window. Balances performance and cost with enhanced efficiency for high-volume applications. Perfect for applications requiring good quality responses at minimal cost.',
|
4172
|
+
pricing: {
|
4173
|
+
prompt: pricing(`$0.15 / 1M tokens`),
|
4174
|
+
output: pricing(`$0.45 / 1M tokens`),
|
3955
4175
|
},
|
3956
4176
|
},
|
3957
4177
|
{
|
@@ -4636,7 +4856,7 @@
|
|
4636
4856
|
/**
|
4637
4857
|
* List of available models in Ollama library
|
4638
4858
|
*
|
4639
|
-
* Note:
|
4859
|
+
* Note: Synced with official API docs at 2025-08-20
|
4640
4860
|
*
|
4641
4861
|
* @see https://ollama.com/library
|
4642
4862
|
* @public exported from `@promptbook/ollama`
|
@@ -4644,6 +4864,24 @@
|
|
4644
4864
|
const OLLAMA_MODELS = exportJson({
|
4645
4865
|
name: 'OLLAMA_MODELS',
|
4646
4866
|
value: [
|
4867
|
+
{
|
4868
|
+
modelVariant: 'CHAT',
|
4869
|
+
modelTitle: 'llama3.3',
|
4870
|
+
modelName: 'llama3.3',
|
4871
|
+
modelDescription: 'Meta Llama 3.3 (70B parameters) with 128K context window. Latest generation foundation model with significantly enhanced reasoning, instruction following, and multilingual capabilities. Features improved performance on complex tasks and better factual accuracy compared to Llama 3.1.',
|
4872
|
+
},
|
4873
|
+
{
|
4874
|
+
modelVariant: 'CHAT',
|
4875
|
+
modelTitle: 'llama3.2',
|
4876
|
+
modelName: 'llama3.2',
|
4877
|
+
modelDescription: 'Meta Llama 3.2 (1B-90B parameters) with 128K context window. Enhanced model with improved reasoning capabilities, better instruction following, and multimodal support in larger variants. Features significant performance improvements over Llama 3.1 across diverse tasks.',
|
4878
|
+
},
|
4879
|
+
{
|
4880
|
+
modelVariant: 'CHAT',
|
4881
|
+
modelTitle: 'llama3.1',
|
4882
|
+
modelName: 'llama3.1',
|
4883
|
+
modelDescription: 'Meta Llama 3.1 (8B-405B parameters) with 128K context window. Advanced foundation model with enhanced reasoning, improved multilingual capabilities, and better performance on complex tasks. Features significant improvements in code generation and mathematical reasoning.',
|
4884
|
+
},
|
4647
4885
|
{
|
4648
4886
|
modelVariant: 'CHAT',
|
4649
4887
|
modelTitle: 'llama3',
|