@ruvector/edge-net 0.5.0 → 0.5.3

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -0,0 +1,214 @@
1
+ {
2
+ "version": "1.0.0",
3
+ "updated": "2026-01-03T00:00:00.000Z",
4
+ "gcs_bucket": "ruvector-models",
5
+ "ipfs_gateway": "https://ipfs.io/ipfs",
6
+ "models": {
7
+ "minilm-l6": {
8
+ "name": "MiniLM-L6-v2",
9
+ "type": "embedding",
10
+ "huggingface": "Xenova/all-MiniLM-L6-v2",
11
+ "dimensions": 384,
12
+ "size": "22MB",
13
+ "tier": 1,
14
+ "quantized": ["int8", "fp16"],
15
+ "description": "Fast, good quality embeddings for edge deployment",
16
+ "recommended_for": ["edge-minimal", "low-memory"],
17
+ "artifacts": {}
18
+ },
19
+ "e5-small": {
20
+ "name": "E5-Small-v2",
21
+ "type": "embedding",
22
+ "huggingface": "Xenova/e5-small-v2",
23
+ "dimensions": 384,
24
+ "size": "28MB",
25
+ "tier": 1,
26
+ "quantized": ["int8", "fp16"],
27
+ "description": "Microsoft E5 - excellent for retrieval tasks",
28
+ "recommended_for": ["retrieval", "semantic-search"],
29
+ "artifacts": {}
30
+ },
31
+ "bge-small": {
32
+ "name": "BGE-Small-EN-v1.5",
33
+ "type": "embedding",
34
+ "huggingface": "Xenova/bge-small-en-v1.5",
35
+ "dimensions": 384,
36
+ "size": "33MB",
37
+ "tier": 2,
38
+ "quantized": ["int8", "fp16"],
39
+ "description": "BAAI BGE - best for retrieval and ranking",
40
+ "recommended_for": ["retrieval", "reranking"],
41
+ "artifacts": {}
42
+ },
43
+ "gte-small": {
44
+ "name": "GTE-Small",
45
+ "type": "embedding",
46
+ "huggingface": "Xenova/gte-small",
47
+ "dimensions": 384,
48
+ "size": "67MB",
49
+ "tier": 2,
50
+ "quantized": ["int8", "fp16"],
51
+ "description": "General Text Embeddings - high quality",
52
+ "recommended_for": ["general", "quality"],
53
+ "artifacts": {}
54
+ },
55
+ "gte-base": {
56
+ "name": "GTE-Base",
57
+ "type": "embedding",
58
+ "huggingface": "Xenova/gte-base",
59
+ "dimensions": 768,
60
+ "size": "100MB",
61
+ "tier": 3,
62
+ "quantized": ["int8", "fp16"],
63
+ "description": "GTE Base - 768 dimensions for higher quality",
64
+ "recommended_for": ["cloud", "high-quality"],
65
+ "artifacts": {}
66
+ },
67
+ "multilingual-e5": {
68
+ "name": "Multilingual-E5-Small",
69
+ "type": "embedding",
70
+ "huggingface": "Xenova/multilingual-e5-small",
71
+ "dimensions": 384,
72
+ "size": "118MB",
73
+ "tier": 3,
74
+ "quantized": ["int8", "fp16"],
75
+ "description": "Supports 100+ languages",
76
+ "recommended_for": ["multilingual", "international"],
77
+ "artifacts": {}
78
+ },
79
+ "distilgpt2": {
80
+ "name": "DistilGPT2",
81
+ "type": "generation",
82
+ "huggingface": "Xenova/distilgpt2",
83
+ "size": "82MB",
84
+ "tier": 1,
85
+ "quantized": ["int8", "int4", "fp16"],
86
+ "capabilities": ["general", "completion"],
87
+ "description": "Fast distilled GPT-2 for text generation",
88
+ "recommended_for": ["edge", "fast-inference"],
89
+ "artifacts": {}
90
+ },
91
+ "tinystories": {
92
+ "name": "TinyStories-33M",
93
+ "type": "generation",
94
+ "huggingface": "Xenova/TinyStories-33M",
95
+ "size": "65MB",
96
+ "tier": 1,
97
+ "quantized": ["int8", "int4"],
98
+ "capabilities": ["stories", "creative"],
99
+ "description": "Ultra-small model trained on children's stories",
100
+ "recommended_for": ["creative", "stories", "minimal"],
101
+ "artifacts": {}
102
+ },
103
+ "starcoder-tiny": {
104
+ "name": "TinyStarCoder-Py",
105
+ "type": "generation",
106
+ "huggingface": "Xenova/tiny_starcoder_py",
107
+ "size": "40MB",
108
+ "tier": 1,
109
+ "quantized": ["int8", "int4"],
110
+ "capabilities": ["code", "python"],
111
+ "description": "Ultra-small Python code generation",
112
+ "recommended_for": ["code", "python", "edge"],
113
+ "artifacts": {}
114
+ },
115
+ "phi-1.5": {
116
+ "name": "Phi-1.5",
117
+ "type": "generation",
118
+ "huggingface": "Xenova/phi-1_5",
119
+ "size": "280MB",
120
+ "tier": 2,
121
+ "quantized": ["int8", "int4", "fp16"],
122
+ "capabilities": ["code", "reasoning", "math"],
123
+ "description": "Microsoft Phi-1.5 - excellent code and reasoning",
124
+ "recommended_for": ["code", "reasoning", "balanced"],
125
+ "artifacts": {}
126
+ },
127
+ "codegen-350m": {
128
+ "name": "CodeGen-350M-Mono",
129
+ "type": "generation",
130
+ "huggingface": "Xenova/codegen-350M-mono",
131
+ "size": "320MB",
132
+ "tier": 2,
133
+ "quantized": ["int8", "int4", "fp16"],
134
+ "capabilities": ["code", "python"],
135
+ "description": "Salesforce CodeGen - Python specialist",
136
+ "recommended_for": ["code", "python"],
137
+ "artifacts": {}
138
+ },
139
+ "qwen-0.5b": {
140
+ "name": "Qwen-1.5-0.5B",
141
+ "type": "generation",
142
+ "huggingface": "Xenova/Qwen1.5-0.5B",
143
+ "size": "430MB",
144
+ "tier": 3,
145
+ "quantized": ["int8", "int4", "fp16"],
146
+ "capabilities": ["multilingual", "general", "code"],
147
+ "description": "Alibaba Qwen 0.5B - multilingual capabilities",
148
+ "recommended_for": ["multilingual", "general"],
149
+ "artifacts": {}
150
+ },
151
+ "phi-2": {
152
+ "name": "Phi-2",
153
+ "type": "generation",
154
+ "huggingface": "Xenova/phi-2",
155
+ "size": "550MB",
156
+ "tier": 3,
157
+ "quantized": ["int8", "int4", "fp16"],
158
+ "capabilities": ["code", "reasoning", "math", "general"],
159
+ "description": "Microsoft Phi-2 - advanced reasoning model",
160
+ "recommended_for": ["reasoning", "code", "quality"],
161
+ "artifacts": {}
162
+ },
163
+ "gemma-2b": {
164
+ "name": "Gemma-2B-IT",
165
+ "type": "generation",
166
+ "huggingface": "Xenova/gemma-2b-it",
167
+ "size": "1.1GB",
168
+ "tier": 4,
169
+ "quantized": ["int8", "int4", "fp16"],
170
+ "capabilities": ["instruction", "general", "code", "reasoning"],
171
+ "description": "Google Gemma 2B instruction-tuned",
172
+ "recommended_for": ["cloud", "high-quality", "instruction"],
173
+ "artifacts": {}
174
+ }
175
+ },
176
+ "profiles": {
177
+ "edge-minimal": {
178
+ "description": "Minimal footprint for constrained edge devices",
179
+ "embedding": "minilm-l6",
180
+ "generation": "tinystories",
181
+ "total_size": "~87MB",
182
+ "quantization": "int4"
183
+ },
184
+ "edge-balanced": {
185
+ "description": "Best quality/size ratio for edge deployment",
186
+ "embedding": "e5-small",
187
+ "generation": "phi-1.5",
188
+ "total_size": "~308MB",
189
+ "quantization": "int8"
190
+ },
191
+ "edge-code": {
192
+ "description": "Optimized for code generation tasks",
193
+ "embedding": "bge-small",
194
+ "generation": "starcoder-tiny",
195
+ "total_size": "~73MB",
196
+ "quantization": "int8"
197
+ },
198
+ "edge-full": {
199
+ "description": "Maximum quality on edge devices",
200
+ "embedding": "gte-base",
201
+ "generation": "phi-2",
202
+ "total_size": "~650MB",
203
+ "quantization": "int8"
204
+ },
205
+ "cloud-optimal": {
206
+ "description": "Best quality for cloud/server deployment",
207
+ "embedding": "gte-base",
208
+ "generation": "gemma-2b",
209
+ "total_size": "~1.2GB",
210
+ "quantization": "fp16"
211
+ }
212
+ },
213
+ "adapters": {}
214
+ }