@fre4x/openai 1.1.0-beta.3 → 1.1.0-beta.4
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +9 -8
- package/dist/index.js +13968 -17780
- package/package.json +3 -3
package/README.md
CHANGED
|
@@ -4,17 +4,17 @@
|
|
|
4
4
|
|
|
5
5
|
Part of **[FRE4X-B1TE](../)** — a monorepo of MCP servers built for autonomous agents.
|
|
6
6
|
|
|
7
|
-
This B1TE bridges your agent directly to OpenAI's advanced models — enabling it to see with GPT-4o, imagine with
|
|
7
|
+
This B1TE bridges your agent directly to OpenAI's advanced models — enabling it to see with GPT-4o, imagine with GPT Image 2, and hear with Whisper.
|
|
8
8
|
|
|
9
9
|
## Tools
|
|
10
10
|
|
|
11
11
|
| Tool | Capability |
|
|
12
12
|
|------|-----------|
|
|
13
|
-
| `
|
|
14
|
-
| `
|
|
15
|
-
| `
|
|
16
|
-
| `
|
|
17
|
-
| `
|
|
13
|
+
| `analyze_media` | Analyze an image from a `file://` or `http(s)://` URL. Prompt is optional. |
|
|
14
|
+
| `list_models` | List available OpenAI models with pagination. |
|
|
15
|
+
| `generate_image` | Image synthesis via GPT Image 2. Embeds base64 image data and optionally saves to `output_dir`. |
|
|
16
|
+
| `transcribe_audio` | Transcribe local audio files using Whisper (`whisper-1`) and auto-save a `.txt` transcript. |
|
|
17
|
+
| `generate_speech` | Generate speech audio from text and optionally save it locally. |
|
|
18
18
|
|
|
19
19
|
## Setup
|
|
20
20
|
|
|
@@ -42,11 +42,12 @@ This B1TE bridges your agent directly to OpenAI's advanced models — enabling i
|
|
|
42
42
|
## Features
|
|
43
43
|
|
|
44
44
|
- **Protocol Compliance**: Implements the split `content` (Markdown) and `structuredContent` (Raw Data) model for high-fidelity agent interaction.
|
|
45
|
-
- **Agent-Sufficient Output**: `content.text` carries enough detail for autonomous follow-up, including model IDs, generated image
|
|
45
|
+
- **Agent-Sufficient Output**: `content.text` carries enough detail for autonomous follow-up, including model IDs, generated image metadata, revised prompts, saved paths, transcript metadata, and speech metadata.
|
|
46
46
|
- **Strict Validation**: All tools utilize `outputSchema` for reliable data parsing and reduced hallucinations.
|
|
47
47
|
- **Privacy & Security**: Aggressively masks implementation details (HTTP codes, library names) in error messages.
|
|
48
48
|
- **Token Efficiency**: Mandatory pagination for model listing.
|
|
49
|
-
- **Vision Input Contract**: `
|
|
49
|
+
- **Vision Input Contract**: `analyze_media` accepts only `file://` and `http(s)://` URLs. `data:` URLs are rejected at validation time.
|
|
50
|
+
- **Transcript Saving**: `transcribe_audio` always writes `openai_transcription_{number}.txt` into the current working directory unless `output_dir` overrides the destination.
|
|
50
51
|
|
|
51
52
|
## Development
|
|
52
53
|
|