@promptbook/openai 0.94.0-0 → 0.94.0-1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/README.md CHANGED
@@ -263,8 +263,6 @@ Rest of the documentation is common for **entire promptbook ecosystem**:
263
263
 
264
264
  During the computer revolution, we have seen [multiple generations of computer languages](https://github.com/webgptorg/promptbook/discussions/180), from the physical rewiring of the vacuum tubes through low-level machine code to the high-level languages like Python or JavaScript. And now, we're on the edge of the **next revolution**!
265
265
 
266
-
267
-
268
266
  It's a revolution of writing software in **plain human language** that is understandable and executable by both humans and machines – and it's going to change everything!
269
267
 
270
268
  The incredible growth in power of microprocessors and the Moore's Law have been the driving force behind the ever-more powerful languages, and it's been an amazing journey! Similarly, the large language models (like GPT or Claude) are the next big thing in language technology, and they're set to transform the way we interact with computers.
@@ -450,8 +448,6 @@ Personas can have access to different knowledge, tools and actions. They can als
450
448
 
451
449
  - [PERSONA](https://github.com/webgptorg/promptbook/blob/main/documents/commands/PERSONA.md)
452
450
 
453
-
454
-
455
451
  ### **3. How:** Knowledge, Instruments and Actions
456
452
 
457
453
  The resources used by the personas are used to do the work.
@@ -512,6 +508,7 @@ Or you can install them separately:
512
508
  - **[@promptbook/vercel](https://www.npmjs.com/package/@promptbook/vercel)** - Adapter for Vercel functionalities
513
509
  - **[@promptbook/google](https://www.npmjs.com/package/@promptbook/google)** - Integration with Google's Gemini API
514
510
  - **[@promptbook/deepseek](https://www.npmjs.com/package/@promptbook/deepseek)** - Integration with [DeepSeek API](https://www.deepseek.com/)
511
+ - **[@promptbook/ollama](https://www.npmjs.com/package/@promptbook/ollama)** - Integration with [Ollama](https://ollama.com/) API
515
512
  - **[@promptbook/azure-openai](https://www.npmjs.com/package/@promptbook/azure-openai)** - Execution tools for Azure OpenAI API
516
513
 
517
514
  - **[@promptbook/fake-llm](https://www.npmjs.com/package/@promptbook/fake-llm)** - Mocked execution tools for testing the library and saving the tokens
@@ -550,8 +547,6 @@ The following glossary is used to clarify certain concepts:
550
547
 
551
548
  _Note: This section is not complete dictionary, more list of general AI / LLM terms that has connection with Promptbook_
552
549
 
553
-
554
-
555
550
  ### 💯 Core concepts
556
551
 
557
552
  - [📚 Collection of pipelines](https://github.com/webgptorg/promptbook/discussions/65)
package/esm/index.es.js CHANGED
@@ -18,7 +18,7 @@ const BOOK_LANGUAGE_VERSION = '1.0.0';
18
18
  * @generated
19
19
  * @see https://github.com/webgptorg/promptbook
20
20
  */
21
- const PROMPTBOOK_ENGINE_VERSION = '0.94.0-0';
21
+ const PROMPTBOOK_ENGINE_VERSION = '0.94.0-1';
22
22
  /**
23
23
  * TODO: string_promptbook_version should be constrained to the all versions of Promptbook engine
24
24
  * Note: [💞] Ignore a discrepancy between file name and entity name
@@ -8,6 +8,8 @@ import { _DeepseekMetadataRegistration } from '../llm-providers/deepseek/registe
8
8
  import { _DeepseekRegistration } from '../llm-providers/deepseek/register-constructor';
9
9
  import { _GoogleMetadataRegistration } from '../llm-providers/google/register-configuration';
10
10
  import { _GoogleRegistration } from '../llm-providers/google/register-constructor';
11
+ import { _OllamaMetadataRegistration } from '../llm-providers/ollama/register-configuration';
12
+ import { _OllamaRegistration } from '../llm-providers/ollama/register-constructor';
11
13
  import { _OpenAiMetadataRegistration } from '../llm-providers/openai/register-configuration';
12
14
  import { _OpenAiAssistantMetadataRegistration } from '../llm-providers/openai/register-configuration';
13
15
  import { _OpenAiRegistration } from '../llm-providers/openai/register-constructor';
@@ -36,6 +38,8 @@ export { _DeepseekMetadataRegistration };
36
38
  export { _DeepseekRegistration };
37
39
  export { _GoogleMetadataRegistration };
38
40
  export { _GoogleRegistration };
41
+ export { _OllamaMetadataRegistration };
42
+ export { _OllamaRegistration };
39
43
  export { _OpenAiMetadataRegistration };
40
44
  export { _OpenAiAssistantMetadataRegistration };
41
45
  export { _OpenAiRegistration };
@@ -108,6 +108,7 @@ import { _AnthropicClaudeMetadataRegistration } from '../llm-providers/anthropic
108
108
  import { _AzureOpenAiMetadataRegistration } from '../llm-providers/azure-openai/register-configuration';
109
109
  import { _DeepseekMetadataRegistration } from '../llm-providers/deepseek/register-configuration';
110
110
  import { _GoogleMetadataRegistration } from '../llm-providers/google/register-configuration';
111
+ import { _OllamaMetadataRegistration } from '../llm-providers/ollama/register-configuration';
111
112
  import { _OpenAiMetadataRegistration } from '../llm-providers/openai/register-configuration';
112
113
  import { _OpenAiAssistantMetadataRegistration } from '../llm-providers/openai/register-configuration';
113
114
  import { migratePipeline } from '../migrations/migratePipeline';
@@ -255,6 +256,7 @@ export { _AnthropicClaudeMetadataRegistration };
255
256
  export { _AzureOpenAiMetadataRegistration };
256
257
  export { _DeepseekMetadataRegistration };
257
258
  export { _GoogleMetadataRegistration };
259
+ export { _OllamaMetadataRegistration };
258
260
  export { _OpenAiMetadataRegistration };
259
261
  export { _OpenAiAssistantMetadataRegistration };
260
262
  export { migratePipeline };
@@ -0,0 +1,8 @@
1
+ import { BOOK_LANGUAGE_VERSION, PROMPTBOOK_ENGINE_VERSION } from '../version';
2
+ import { createOllamaExecutionTools } from '../llm-providers/ollama/createOllamaExecutionTools';
3
+ import { OllamaExecutionTools } from '../llm-providers/ollama/OllamaExecutionTools';
4
+ import { _OllamaRegistration } from '../llm-providers/ollama/register-constructor';
5
+ export { BOOK_LANGUAGE_VERSION, PROMPTBOOK_ENGINE_VERSION };
6
+ export { createOllamaExecutionTools };
7
+ export { OllamaExecutionTools };
8
+ export { _OllamaRegistration };
@@ -79,6 +79,7 @@ import type { AnthropicClaudeExecutionToolsProxiedOptions } from '../llm-provide
79
79
  import type { AzureOpenAiExecutionToolsOptions } from '../llm-providers/azure-openai/AzureOpenAiExecutionToolsOptions';
80
80
  import type { DeepseekExecutionToolsOptions } from '../llm-providers/deepseek/DeepseekExecutionToolsOptions';
81
81
  import type { GoogleExecutionToolsOptions } from '../llm-providers/google/GoogleExecutionToolsOptions';
82
+ import type { OllamaExecutionToolsOptions } from '../llm-providers/ollama/OllamaExecutionToolsOptions';
82
83
  import type { OpenAiAssistantExecutionToolsOptions } from '../llm-providers/openai/OpenAiAssistantExecutionToolsOptions';
83
84
  import type { OpenAiExecutionToolsOptions } from '../llm-providers/openai/OpenAiExecutionToolsOptions';
84
85
  import type { VercelExecutionToolsOptions } from '../llm-providers/vercel/VercelExecutionToolsOptions';
@@ -377,6 +378,7 @@ export type { AnthropicClaudeExecutionToolsProxiedOptions };
377
378
  export type { AzureOpenAiExecutionToolsOptions };
378
379
  export type { DeepseekExecutionToolsOptions };
379
380
  export type { GoogleExecutionToolsOptions };
381
+ export type { OllamaExecutionToolsOptions };
380
382
  export type { OpenAiAssistantExecutionToolsOptions };
381
383
  export type { OpenAiExecutionToolsOptions };
382
384
  export type { VercelExecutionToolsOptions };
@@ -7,6 +7,8 @@ import { _DeepseekMetadataRegistration } from '../llm-providers/deepseek/registe
7
7
  import { _DeepseekRegistration } from '../llm-providers/deepseek/register-constructor';
8
8
  import { _GoogleMetadataRegistration } from '../llm-providers/google/register-configuration';
9
9
  import { _GoogleRegistration } from '../llm-providers/google/register-constructor';
10
+ import { _OllamaMetadataRegistration } from '../llm-providers/ollama/register-configuration';
11
+ import { _OllamaRegistration } from '../llm-providers/ollama/register-constructor';
10
12
  import { _OpenAiMetadataRegistration } from '../llm-providers/openai/register-configuration';
11
13
  import { _OpenAiAssistantMetadataRegistration } from '../llm-providers/openai/register-configuration';
12
14
  import { _OpenAiRegistration } from '../llm-providers/openai/register-constructor';
@@ -35,6 +37,8 @@ export { _DeepseekMetadataRegistration };
35
37
  export { _DeepseekRegistration };
36
38
  export { _GoogleMetadataRegistration };
37
39
  export { _GoogleRegistration };
40
+ export { _OllamaMetadataRegistration };
41
+ export { _OllamaRegistration };
38
42
  export { _OpenAiMetadataRegistration };
39
43
  export { _OpenAiAssistantMetadataRegistration };
40
44
  export { _OpenAiRegistration };
@@ -72,6 +72,8 @@ export type AbstractTask<TTaskResult extends AbstractTaskResult> = {
72
72
  readonly taskId: task_id;
73
73
  /**
74
74
  * Human-readable title of the task - used for displaying in the UI
75
+ *
76
+ * For example name of the book which is being executed
75
77
  */
76
78
  readonly title: string;
77
79
  /**
@@ -99,7 +101,7 @@ export type AbstractTask<TTaskResult extends AbstractTaskResult> = {
99
101
  /**
100
102
  * Gets just the current value which is mutated during the task processing
101
103
  */
102
- currentValue: PartialDeep<TTaskResult>;
104
+ readonly currentValue: PartialDeep<TTaskResult>;
103
105
  /**
104
106
  * List of errors that occurred during the task processing
105
107
  */
@@ -0,0 +1,19 @@
1
+ import type { AvailableModel } from '../../execution/AvailableModel';
2
+ import type { LlmExecutionTools } from '../../execution/LlmExecutionTools';
3
+ import type { ChatPromptResult } from '../../execution/PromptResult';
4
+ import type { OllamaExecutionToolsOptions } from './OllamaExecutionToolsOptions';
5
+ /**
6
+ * Execution Tools for calling a local Ollama model via HTTP API
7
+ *
8
+ * @public exported from `@promptbook/ollama`
9
+ */
10
+ export declare class OllamaExecutionTools implements LlmExecutionTools {
11
+ protected readonly options: OllamaExecutionToolsOptions;
12
+ private limiter;
13
+ constructor(options: OllamaExecutionToolsOptions);
14
+ get title(): string;
15
+ get description(): string;
16
+ checkConfiguration(): Promise<void>;
17
+ listModels(): Promise<ReadonlyArray<AvailableModel>>;
18
+ callChatModel(prompt: Pick<import('../../types/Prompt').Prompt, 'content' | 'parameters' | 'modelRequirements'>): Promise<ChatPromptResult>;
19
+ }
@@ -0,0 +1,12 @@
1
+ export interface OllamaExecutionToolsOptions {
2
+ /** Base URL of Ollama API, e.g., http://localhost:11434 */
3
+ baseUrl: string;
4
+ /** Model name to use for requests */
5
+ model: string;
6
+ /** Optional rate limit: max requests per minute */
7
+ maxRequestsPerMinute?: number;
8
+ /** Verbose logging */
9
+ isVerbose?: boolean;
10
+ /** Optional user identifier */
11
+ userId?: string;
12
+ }
@@ -0,0 +1,11 @@
1
+ import { OllamaExecutionTools } from "./OllamaExecutionTools";
2
+ import { OllamaExecutionToolsOptions } from "./OllamaExecutionToolsOptions";
3
+ /**
4
+ * Execution Tools for calling Ollama API
5
+ *
6
+ * @public exported from `@promptbook/ollama`
7
+ */
8
+ export declare const createOllamaExecutionTools: ((options: OllamaExecutionToolsOptions) => OllamaExecutionTools) & {
9
+ packageName: string;
10
+ className: string;
11
+ };
@@ -0,0 +1,6 @@
1
+ #!/usr/bin/env ts-node
2
+ export {};
3
+ /**
4
+ * TODO: [main] !!3 Test here that `systemMessage`, `temperature` and `seed` are working correctly
5
+ * Note: [⚫] Code in this file should never be published in any package
6
+ */
@@ -0,0 +1,14 @@
1
+ import type { Registration } from '../../utils/$Register';
2
+ /**
3
+ * Registration of LLM provider metadata
4
+ *
5
+ * Warning: This is not useful for the end user, it is just a side effect of the mechanism that handles all available LLM tools
6
+ *
7
+ * @public exported from `@promptbook/core`
8
+ * @public exported from `@promptbook/wizzard`
9
+ * @public exported from `@promptbook/cli`
10
+ */
11
+ export declare const _OllamaMetadataRegistration: Registration;
12
+ /**
13
+ * Note: [💞] Ignore a discrepancy between file name and entity name
14
+ */
@@ -0,0 +1,15 @@
1
+ import type { Registration } from '../../utils/$Register';
2
+ /**
3
+ * Registration of LLM provider
4
+ *
5
+ * Warning: This is not useful for the end user, it is just a side effect of the mechanism that handles all available LLM tools
6
+ *
7
+ * @public exported from `@promptbook/ollama`
8
+ * @public exported from `@promptbook/wizzard`
9
+ * @public exported from `@promptbook/cli`
10
+ */
11
+ export declare const _OllamaRegistration: Registration;
12
+ /**
13
+ * TODO: [🎶] Naming "constructor" vs "creator" vs "factory"
14
+ * Note: [💞] Ignore a discrepancy between file name and entity name
15
+ */
@@ -15,7 +15,7 @@ export declare const BOOK_LANGUAGE_VERSION: string_semantic_version;
15
15
  export declare const PROMPTBOOK_ENGINE_VERSION: string_promptbook_version;
16
16
  /**
17
17
  * Represents the version string of the Promptbook engine.
18
- * It follows semantic versioning (e.g., `0.93.0`).
18
+ * It follows semantic versioning (e.g., `0.94.0-0`).
19
19
  *
20
20
  * @generated
21
21
  */
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@promptbook/openai",
3
- "version": "0.94.0-0",
3
+ "version": "0.94.0-1",
4
4
  "description": "Promptbook: Run AI apps in plain human language across multiple models and platforms",
5
5
  "private": false,
6
6
  "sideEffects": false,
@@ -63,7 +63,7 @@
63
63
  "module": "./esm/index.es.js",
64
64
  "typings": "./esm/typings/src/_packages/openai.index.d.ts",
65
65
  "peerDependencies": {
66
- "@promptbook/core": "0.94.0-0"
66
+ "@promptbook/core": "0.94.0-1"
67
67
  },
68
68
  "dependencies": {
69
69
  "bottleneck": "^2.19.5",
package/umd/index.umd.js CHANGED
@@ -25,7 +25,7 @@
25
25
  * @generated
26
26
  * @see https://github.com/webgptorg/promptbook
27
27
  */
28
- const PROMPTBOOK_ENGINE_VERSION = '0.94.0-0';
28
+ const PROMPTBOOK_ENGINE_VERSION = '0.94.0-1';
29
29
  /**
30
30
  * TODO: string_promptbook_version should be constrained to the all versions of Promptbook engine
31
31
  * Note: [💞] Ignore a discrepancy between file name and entity name