@blank-utils/llm 0.5.4 → 0.5.5
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +17 -30
- package/package.json +1 -1
package/README.md
CHANGED
|
@@ -326,18 +326,18 @@ Conditional rendering components:
|
|
|
326
326
|
|
|
327
327
|
### Default Chat Interface explained
|
|
328
328
|
|
|
329
|
-
The `<Chat>` and `<ChatApp>` components provide a production-ready
|
|
329
|
+
The `<Chat>` and `<ChatApp>` components provide a production-ready, minimalist, borderless "Terminal" interface inspired by brutalist and sacred computer aesthetics.
|
|
330
330
|
|
|
331
331
|
**Key Features & Usage Points:**
|
|
332
332
|
|
|
333
|
-
- **✨ Zero Config**: Just drop it in. No CSS files to import, no state to manage.
|
|
333
|
+
- **✨ Zero Config**: Just drop it in. No CSS files to import, no state to manage. All spacing, monospace typography, and layout are handled out-of-the-box.
|
|
334
334
|
- **🎨 Rich Text Rendering**:
|
|
335
335
|
- **Global Markdown**: Bold, cursives, lists, tables.
|
|
336
|
-
- **Code Blocks**: Syntax highlighting for 20+ languages.
|
|
336
|
+
- **Code Blocks**: Syntax highlighting for 20+ languages cleanly integrated without borders.
|
|
337
337
|
- **Diagrams**: Renders `mermaid` diagrams automatically.
|
|
338
338
|
- **Math**: Supports LateX expressions.
|
|
339
339
|
- **⚡ Eager Interaction**: Users can type and send messages _while_ the model is still initializing. The chat controls the queue.
|
|
340
|
-
- **🌗
|
|
340
|
+
- **🌗 Multimodal & Files**: Deeply integrated multimodal chat input. Simply drag and drop PDFs for automatic local text extraction, or drag-and-drop/paste images for seamless processing against vision models (e.g., `phi-3.5-vision`). WebLLM crash mitigations for dynamic image cropping are automatically handled.
|
|
341
341
|
- **🔄 Model Switching**:
|
|
342
342
|
- If using `<ChatApp />`, a model selector dropdown is included automatically.
|
|
343
343
|
- If using `<Chat />`, pass `onModelChange` to enable the dropdown.
|
|
@@ -497,32 +497,19 @@ You can use either the **alias** (short name) or the **full model ID** when spec
|
|
|
497
497
|
|
|
498
498
|
### WebLLM Backend (WebGPU)
|
|
499
499
|
|
|
500
|
-
|
|
501
|
-
|
|
502
|
-
|
|
503
|
-
|
|
504
|
-
|
|
505
|
-
|
|
506
|
-
|
|
507
|
-
|
|
508
|
-
|
|
509
|
-
|
|
510
|
-
|
|
511
|
-
|
|
512
|
-
|
|
513
|
-
| `qwen-2.5-coder-0.5b` | Qwen 2.5 Coder 0.5B | Code-focused |
|
|
514
|
-
| `qwen-2.5-coder-1.5b` | Qwen 2.5 Coder 1.5B | Code-focused |
|
|
515
|
-
| `gemma-2-2b` | Gemma 2 2B | Google, efficient |
|
|
516
|
-
| `gemma-2-2b-1k` | Gemma 2 2B (1K ctx) | Lower memory |
|
|
517
|
-
| `gemma-2-9b` | Gemma 2 9B | Large |
|
|
518
|
-
| `smollm2-135m` | SmolLM2 135M | Ultra lightweight |
|
|
519
|
-
| `smollm2-360m` | SmolLM2 360M | Lightweight |
|
|
520
|
-
| `smollm2-1.7b` | SmolLM2 1.7B | Small |
|
|
521
|
-
| `mistral-7b` | Mistral 7B v0.3 | General purpose |
|
|
522
|
-
| `deepseek-r1-qwen-7b` | DeepSeek R1 Distill Qwen 7B | Reasoning |
|
|
523
|
-
| `deepseek-r1-llama-8b` | DeepSeek R1 Distill Llama 8B | Reasoning |
|
|
524
|
-
| `hermes-3-llama-3.2-3b` | Hermes 3 Llama 3.2 3B | Function calling |
|
|
525
|
-
| `hermes-3-llama-3.1-8b` | Hermes 3 Llama 3.1 8B | Function calling |
|
|
500
|
+
The WebLLM backend supports **over 80+ WebGPU-accelerated models**, organized into major ecosystems. All models are strongly typed and natively supported:
|
|
501
|
+
|
|
502
|
+
- **Microsoft Phi**: `phi-3.5-mini`, `phi-3.5-vision` (Multimodal support!), `phi-3-mini-4k`, `phi-2`, `phi-1.5`
|
|
503
|
+
- **Meta Llama**: `llama-3.2-1b`/`3b`, `llama-3.1-8b`/`70b`, `llama-3-8b`/`70b`, `llama-2-7b`/`13b`
|
|
504
|
+
- **Alibaba Qwen**: `qwen3` (`0.6b` to `8b`), `qwen-2.5` (`0.5b` to `7b`), `qwen-2.5-coder` (`0.5b` to `7b`), `qwen-2.5-math` (`1.5b`)
|
|
505
|
+
- **Google Gemma**: `gemma-2-2b`, `gemma-2-9b`, `gemma-2-2b-jpn`, `gemma-2b`
|
|
506
|
+
- **HuggingFace SmolLM2**: `smollm2-135m` (~360MB VRAM!), `smollm2-360m`, `smollm2-1.7b`
|
|
507
|
+
- **Mistral & Ministral**: `mistral-7b` (v0.2/v0.3), `ministral-3-3b-base`/`reasoning`/`instruct`
|
|
508
|
+
- **DeepSeek Reasoning**: `deepseek-r1-qwen-7b`, `deepseek-r1-llama-8b`
|
|
509
|
+
- **NousResearch Hermes**: `hermes-3-llama-3.2-3b`/`3.1-8b`, `hermes-2-theta/pro-llama`, `hermes-pro/openhermes-mistral`
|
|
510
|
+
- **Other**: `tinyllama-1.1b`, `redpajama-3b`, `stablelm-2-zephyr-1.6b`, `wizardmath-7b`
|
|
511
|
+
|
|
512
|
+
_Note: Many models have `-1k` variants (e.g. `llama-3.1-8b-1k`) with smaller context windows intentionally configured to require significantly less VRAM on constrained devices._
|
|
526
513
|
|
|
527
514
|
### Transformers.js Backend (CPU / WASM)
|
|
528
515
|
|
package/package.json
CHANGED