@promptbook/vercel 0.94.0-0 → 0.94.0-1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/README.md CHANGED
@@ -120,8 +120,6 @@ Rest of the documentation is common for **entire promptbook ecosystem**:
120
120
 
121
121
  During the computer revolution, we have seen [multiple generations of computer languages](https://github.com/webgptorg/promptbook/discussions/180), from the physical rewiring of the vacuum tubes through low-level machine code to the high-level languages like Python or JavaScript. And now, we're on the edge of the **next revolution**!
122
122
 
123
-
124
-
125
123
  It's a revolution of writing software in **plain human language** that is understandable and executable by both humans and machines – and it's going to change everything!
126
124
 
127
125
  The incredible growth in power of microprocessors and the Moore's Law have been the driving force behind the ever-more powerful languages, and it's been an amazing journey! Similarly, the large language models (like GPT or Claude) are the next big thing in language technology, and they're set to transform the way we interact with computers.
@@ -307,8 +305,6 @@ Personas can have access to different knowledge, tools and actions. They can als
307
305
 
308
306
  - [PERSONA](https://github.com/webgptorg/promptbook/blob/main/documents/commands/PERSONA.md)
309
307
 
310
-
311
-
312
308
  ### **3. How:** Knowledge, Instruments and Actions
313
309
 
314
310
  The resources used by the personas are used to do the work.
@@ -369,6 +365,7 @@ Or you can install them separately:
369
365
  - **[@promptbook/vercel](https://www.npmjs.com/package/@promptbook/vercel)** - Adapter for Vercel functionalities
370
366
  - **[@promptbook/google](https://www.npmjs.com/package/@promptbook/google)** - Integration with Google's Gemini API
371
367
  - **[@promptbook/deepseek](https://www.npmjs.com/package/@promptbook/deepseek)** - Integration with [DeepSeek API](https://www.deepseek.com/)
368
+ - **[@promptbook/ollama](https://www.npmjs.com/package/@promptbook/ollama)** - Integration with [Ollama](https://ollama.com/) API
372
369
  - **[@promptbook/azure-openai](https://www.npmjs.com/package/@promptbook/azure-openai)** - Execution tools for Azure OpenAI API
373
370
 
374
371
  - **[@promptbook/fake-llm](https://www.npmjs.com/package/@promptbook/fake-llm)** - Mocked execution tools for testing the library and saving the tokens
@@ -407,8 +404,6 @@ The following glossary is used to clarify certain concepts:
407
404
 
408
405
  _Note: This section is not complete dictionary, more list of general AI / LLM terms that has connection with Promptbook_
409
406
 
410
-
411
-
412
407
  ### 💯 Core concepts
413
408
 
414
409
  - [📚 Collection of pipelines](https://github.com/webgptorg/promptbook/discussions/65)
package/esm/index.es.js CHANGED
@@ -16,7 +16,7 @@ const BOOK_LANGUAGE_VERSION = '1.0.0';
16
16
  * @generated
17
17
  * @see https://github.com/webgptorg/promptbook
18
18
  */
19
- const PROMPTBOOK_ENGINE_VERSION = '0.94.0-0';
19
+ const PROMPTBOOK_ENGINE_VERSION = '0.94.0-1';
20
20
  /**
21
21
  * TODO: string_promptbook_version should be constrained to the all versions of Promptbook engine
22
22
  * Note: [💞] Ignore a discrepancy between file name and entity name
@@ -8,6 +8,8 @@ import { _DeepseekMetadataRegistration } from '../llm-providers/deepseek/registe
8
8
  import { _DeepseekRegistration } from '../llm-providers/deepseek/register-constructor';
9
9
  import { _GoogleMetadataRegistration } from '../llm-providers/google/register-configuration';
10
10
  import { _GoogleRegistration } from '../llm-providers/google/register-constructor';
11
+ import { _OllamaMetadataRegistration } from '../llm-providers/ollama/register-configuration';
12
+ import { _OllamaRegistration } from '../llm-providers/ollama/register-constructor';
11
13
  import { _OpenAiMetadataRegistration } from '../llm-providers/openai/register-configuration';
12
14
  import { _OpenAiAssistantMetadataRegistration } from '../llm-providers/openai/register-configuration';
13
15
  import { _OpenAiRegistration } from '../llm-providers/openai/register-constructor';
@@ -36,6 +38,8 @@ export { _DeepseekMetadataRegistration };
36
38
  export { _DeepseekRegistration };
37
39
  export { _GoogleMetadataRegistration };
38
40
  export { _GoogleRegistration };
41
+ export { _OllamaMetadataRegistration };
42
+ export { _OllamaRegistration };
39
43
  export { _OpenAiMetadataRegistration };
40
44
  export { _OpenAiAssistantMetadataRegistration };
41
45
  export { _OpenAiRegistration };
@@ -108,6 +108,7 @@ import { _AnthropicClaudeMetadataRegistration } from '../llm-providers/anthropic
108
108
  import { _AzureOpenAiMetadataRegistration } from '../llm-providers/azure-openai/register-configuration';
109
109
  import { _DeepseekMetadataRegistration } from '../llm-providers/deepseek/register-configuration';
110
110
  import { _GoogleMetadataRegistration } from '../llm-providers/google/register-configuration';
111
+ import { _OllamaMetadataRegistration } from '../llm-providers/ollama/register-configuration';
111
112
  import { _OpenAiMetadataRegistration } from '../llm-providers/openai/register-configuration';
112
113
  import { _OpenAiAssistantMetadataRegistration } from '../llm-providers/openai/register-configuration';
113
114
  import { migratePipeline } from '../migrations/migratePipeline';
@@ -255,6 +256,7 @@ export { _AnthropicClaudeMetadataRegistration };
255
256
  export { _AzureOpenAiMetadataRegistration };
256
257
  export { _DeepseekMetadataRegistration };
257
258
  export { _GoogleMetadataRegistration };
259
+ export { _OllamaMetadataRegistration };
258
260
  export { _OpenAiMetadataRegistration };
259
261
  export { _OpenAiAssistantMetadataRegistration };
260
262
  export { migratePipeline };
@@ -0,0 +1,8 @@
1
+ import { BOOK_LANGUAGE_VERSION, PROMPTBOOK_ENGINE_VERSION } from '../version';
2
+ import { createOllamaExecutionTools } from '../llm-providers/ollama/createOllamaExecutionTools';
3
+ import { OllamaExecutionTools } from '../llm-providers/ollama/OllamaExecutionTools';
4
+ import { _OllamaRegistration } from '../llm-providers/ollama/register-constructor';
5
+ export { BOOK_LANGUAGE_VERSION, PROMPTBOOK_ENGINE_VERSION };
6
+ export { createOllamaExecutionTools };
7
+ export { OllamaExecutionTools };
8
+ export { _OllamaRegistration };
@@ -79,6 +79,7 @@ import type { AnthropicClaudeExecutionToolsProxiedOptions } from '../llm-provide
79
79
  import type { AzureOpenAiExecutionToolsOptions } from '../llm-providers/azure-openai/AzureOpenAiExecutionToolsOptions';
80
80
  import type { DeepseekExecutionToolsOptions } from '../llm-providers/deepseek/DeepseekExecutionToolsOptions';
81
81
  import type { GoogleExecutionToolsOptions } from '../llm-providers/google/GoogleExecutionToolsOptions';
82
+ import type { OllamaExecutionToolsOptions } from '../llm-providers/ollama/OllamaExecutionToolsOptions';
82
83
  import type { OpenAiAssistantExecutionToolsOptions } from '../llm-providers/openai/OpenAiAssistantExecutionToolsOptions';
83
84
  import type { OpenAiExecutionToolsOptions } from '../llm-providers/openai/OpenAiExecutionToolsOptions';
84
85
  import type { VercelExecutionToolsOptions } from '../llm-providers/vercel/VercelExecutionToolsOptions';
@@ -377,6 +378,7 @@ export type { AnthropicClaudeExecutionToolsProxiedOptions };
377
378
  export type { AzureOpenAiExecutionToolsOptions };
378
379
  export type { DeepseekExecutionToolsOptions };
379
380
  export type { GoogleExecutionToolsOptions };
381
+ export type { OllamaExecutionToolsOptions };
380
382
  export type { OpenAiAssistantExecutionToolsOptions };
381
383
  export type { OpenAiExecutionToolsOptions };
382
384
  export type { VercelExecutionToolsOptions };
@@ -7,6 +7,8 @@ import { _DeepseekMetadataRegistration } from '../llm-providers/deepseek/registe
7
7
  import { _DeepseekRegistration } from '../llm-providers/deepseek/register-constructor';
8
8
  import { _GoogleMetadataRegistration } from '../llm-providers/google/register-configuration';
9
9
  import { _GoogleRegistration } from '../llm-providers/google/register-constructor';
10
+ import { _OllamaMetadataRegistration } from '../llm-providers/ollama/register-configuration';
11
+ import { _OllamaRegistration } from '../llm-providers/ollama/register-constructor';
10
12
  import { _OpenAiMetadataRegistration } from '../llm-providers/openai/register-configuration';
11
13
  import { _OpenAiAssistantMetadataRegistration } from '../llm-providers/openai/register-configuration';
12
14
  import { _OpenAiRegistration } from '../llm-providers/openai/register-constructor';
@@ -35,6 +37,8 @@ export { _DeepseekMetadataRegistration };
35
37
  export { _DeepseekRegistration };
36
38
  export { _GoogleMetadataRegistration };
37
39
  export { _GoogleRegistration };
40
+ export { _OllamaMetadataRegistration };
41
+ export { _OllamaRegistration };
38
42
  export { _OpenAiMetadataRegistration };
39
43
  export { _OpenAiAssistantMetadataRegistration };
40
44
  export { _OpenAiRegistration };
@@ -72,6 +72,8 @@ export type AbstractTask<TTaskResult extends AbstractTaskResult> = {
72
72
  readonly taskId: task_id;
73
73
  /**
74
74
  * Human-readable title of the task - used for displaying in the UI
75
+ *
76
+ * For example name of the book which is being executed
75
77
  */
76
78
  readonly title: string;
77
79
  /**
@@ -99,7 +101,7 @@ export type AbstractTask<TTaskResult extends AbstractTaskResult> = {
99
101
  /**
100
102
  * Gets just the current value which is mutated during the task processing
101
103
  */
102
- currentValue: PartialDeep<TTaskResult>;
104
+ readonly currentValue: PartialDeep<TTaskResult>;
103
105
  /**
104
106
  * List of errors that occurred during the task processing
105
107
  */
@@ -0,0 +1,19 @@
1
+ import type { AvailableModel } from '../../execution/AvailableModel';
2
+ import type { LlmExecutionTools } from '../../execution/LlmExecutionTools';
3
+ import type { ChatPromptResult } from '../../execution/PromptResult';
4
+ import type { OllamaExecutionToolsOptions } from './OllamaExecutionToolsOptions';
5
+ /**
6
+ * Execution Tools for calling a local Ollama model via HTTP API
7
+ *
8
+ * @public exported from `@promptbook/ollama`
9
+ */
10
+ export declare class OllamaExecutionTools implements LlmExecutionTools {
11
+ protected readonly options: OllamaExecutionToolsOptions;
12
+ private limiter;
13
+ constructor(options: OllamaExecutionToolsOptions);
14
+ get title(): string;
15
+ get description(): string;
16
+ checkConfiguration(): Promise<void>;
17
+ listModels(): Promise<ReadonlyArray<AvailableModel>>;
18
+ callChatModel(prompt: Pick<import('../../types/Prompt').Prompt, 'content' | 'parameters' | 'modelRequirements'>): Promise<ChatPromptResult>;
19
+ }
@@ -0,0 +1,12 @@
1
+ export interface OllamaExecutionToolsOptions {
2
+ /** Base URL of Ollama API, e.g., http://localhost:11434 */
3
+ baseUrl: string;
4
+ /** Model name to use for requests */
5
+ model: string;
6
+ /** Optional rate limit: max requests per minute */
7
+ maxRequestsPerMinute?: number;
8
+ /** Verbose logging */
9
+ isVerbose?: boolean;
10
+ /** Optional user identifier */
11
+ userId?: string;
12
+ }
@@ -0,0 +1,11 @@
1
+ import { OllamaExecutionTools } from "./OllamaExecutionTools";
2
+ import { OllamaExecutionToolsOptions } from "./OllamaExecutionToolsOptions";
3
+ /**
4
+ * Execution Tools for calling Ollama API
5
+ *
6
+ * @public exported from `@promptbook/ollama`
7
+ */
8
+ export declare const createOllamaExecutionTools: ((options: OllamaExecutionToolsOptions) => OllamaExecutionTools) & {
9
+ packageName: string;
10
+ className: string;
11
+ };
@@ -0,0 +1,6 @@
1
+ #!/usr/bin/env ts-node
2
+ export {};
3
+ /**
4
+ * TODO: [main] !!3 Test here that `systemMessage`, `temperature` and `seed` are working correctly
5
+ * Note: [⚫] Code in this file should never be published in any package
6
+ */
@@ -0,0 +1,14 @@
1
+ import type { Registration } from '../../utils/$Register';
2
+ /**
3
+ * Registration of LLM provider metadata
4
+ *
5
+ * Warning: This is not useful for the end user, it is just a side effect of the mechanism that handles all available LLM tools
6
+ *
7
+ * @public exported from `@promptbook/core`
8
+ * @public exported from `@promptbook/wizzard`
9
+ * @public exported from `@promptbook/cli`
10
+ */
11
+ export declare const _OllamaMetadataRegistration: Registration;
12
+ /**
13
+ * Note: [💞] Ignore a discrepancy between file name and entity name
14
+ */
@@ -0,0 +1,15 @@
1
+ import type { Registration } from '../../utils/$Register';
2
+ /**
3
+ * Registration of LLM provider
4
+ *
5
+ * Warning: This is not useful for the end user, it is just a side effect of the mechanism that handles all available LLM tools
6
+ *
7
+ * @public exported from `@promptbook/ollama`
8
+ * @public exported from `@promptbook/wizzard`
9
+ * @public exported from `@promptbook/cli`
10
+ */
11
+ export declare const _OllamaRegistration: Registration;
12
+ /**
13
+ * TODO: [🎶] Naming "constructor" vs "creator" vs "factory"
14
+ * Note: [💞] Ignore a discrepancy between file name and entity name
15
+ */
@@ -15,7 +15,7 @@ export declare const BOOK_LANGUAGE_VERSION: string_semantic_version;
15
15
  export declare const PROMPTBOOK_ENGINE_VERSION: string_promptbook_version;
16
16
  /**
17
17
  * Represents the version string of the Promptbook engine.
18
- * It follows semantic versioning (e.g., `0.93.0`).
18
+ * It follows semantic versioning (e.g., `0.94.0-0`).
19
19
  *
20
20
  * @generated
21
21
  */
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@promptbook/vercel",
3
- "version": "0.94.0-0",
3
+ "version": "0.94.0-1",
4
4
  "description": "Promptbook: Run AI apps in plain human language across multiple models and platforms",
5
5
  "private": false,
6
6
  "sideEffects": false,
@@ -63,7 +63,7 @@
63
63
  "module": "./esm/index.es.js",
64
64
  "typings": "./esm/typings/src/_packages/vercel.index.d.ts",
65
65
  "peerDependencies": {
66
- "@promptbook/core": "0.94.0-0"
66
+ "@promptbook/core": "0.94.0-1"
67
67
  },
68
68
  "dependencies": {
69
69
  "colors": "1.4.0",
package/umd/index.umd.js CHANGED
@@ -23,7 +23,7 @@
23
23
  * @generated
24
24
  * @see https://github.com/webgptorg/promptbook
25
25
  */
26
- const PROMPTBOOK_ENGINE_VERSION = '0.94.0-0';
26
+ const PROMPTBOOK_ENGINE_VERSION = '0.94.0-1';
27
27
  /**
28
28
  * TODO: string_promptbook_version should be constrained to the all versions of Promptbook engine
29
29
  * Note: [💞] Ignore a discrepancy between file name and entity name