@gmickel/gno 0.22.3 → 0.22.5

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (2) hide show
  1. package/README.md +49 -0
  2. package/package.json +1 -1
package/README.md CHANGED
@@ -27,11 +27,42 @@ GNO is a local knowledge engine that turns your documents into a searchable, con
27
27
  - [How It Works](#how-it-works)
28
28
  - [Features](#features)
29
29
  - [Local Models](#local-models)
30
+ - [Fine-Tuned Models](#fine-tuned-models)
30
31
  - [Architecture](#architecture)
31
32
  - [Development](#development)
32
33
 
33
34
  ---
34
35
 
36
+ ## What's New in v0.22
37
+
38
+ - **Promoted Slim Retrieval Model**: published `slim-retrieval-v1` on Hugging Face for direct `hf:` installation in GNO
39
+ - **Fine-Tuning Workflow**: local MLX LoRA training, portable GGUF export, automatic checkpoint selection, promotion bundles, and repeatable benchmark comparisons
40
+ - **Autonomous Search Harness**: bounded candidate search with early-stop guards, repeated incumbent confirmation, and promotion targets
41
+ - **Public Docs & Site**: fine-tuned model docs and feature pages now point at the published HF model and the `slim-tuned` preset
42
+
43
+ ### Fine-Tuned Model Quick Use
44
+
45
+ ```yaml
46
+ models:
47
+ activePreset: slim-tuned
48
+ presets:
49
+ - id: slim-tuned
50
+ name: GNO Slim Retrieval v1
51
+ embed: hf:gpustack/bge-m3-GGUF/bge-m3-Q4_K_M.gguf
52
+ rerank: hf:ggml-org/Qwen3-Reranker-0.6B-Q8_0-GGUF/qwen3-reranker-0.6b-q8_0.gguf
53
+ gen: hf:guiltylemon/gno-expansion-slim-retrieval-v1/gno-expansion-auto-entity-lock-default-mix-lr95-f16.gguf
54
+ ```
55
+
56
+ Then:
57
+
58
+ ```bash
59
+ gno models use slim-tuned
60
+ gno models pull --gen
61
+ gno query "ECONNREFUSED 127.0.0.1:5432" --thorough
62
+ ```
63
+
64
+ > Full guide: [Fine-Tuned Models](https://gno.sh/docs/FINE-TUNED-MODELS/) · [Feature page](https://gno.sh/features/fine-tuned-models/)
65
+
35
66
  ## What's New in v0.21
36
67
 
37
68
  - **Ask CLI Query Modes**: `gno ask` now accepts repeatable `--query-mode term|intent|hyde` entries, matching the existing Ask API and Web controls
@@ -447,6 +478,24 @@ gno models use slim
447
478
  gno models pull --all # Optional: pre-download models (auto-downloads on first use)
448
479
  ```
449
480
 
481
+ ## Fine-Tuned Models
482
+
483
+ GNO now has a published promoted retrieval model for the default slim path:
484
+
485
+ - model repo: `guiltylemon/gno-expansion-slim-retrieval-v1`
486
+ - recommended preset id: `slim-tuned`
487
+ - runtime URI:
488
+ - `hf:guiltylemon/gno-expansion-slim-retrieval-v1/gno-expansion-auto-entity-lock-default-mix-lr95-f16.gguf`
489
+
490
+ Use it when you want the tuned retrieval expansion path immediately, without running local fine-tuning yourself.
491
+
492
+ For private/internal products, use the same workflow but keep the final GGUF private and point `gen:` at a `file:` URI instead of publishing to Hugging Face.
493
+
494
+ See:
495
+
496
+ - [Fine-Tuned Models docs](https://gno.sh/docs/FINE-TUNED-MODELS/)
497
+ - [Fine-Tuned Models feature page](https://gno.sh/features/fine-tuned-models/)
498
+
450
499
  ### HTTP Backends (Remote GPU)
451
500
 
452
501
  Offload inference to a GPU server on your network:
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@gmickel/gno",
3
- "version": "0.22.3",
3
+ "version": "0.22.5",
4
4
  "description": "Local semantic search for your documents. Index Markdown, PDF, and Office files with hybrid BM25 + vector search.",
5
5
  "keywords": [
6
6
  "embeddings",