lemonade-sdk 8.0.5__py3-none-any.whl → 8.1.0__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of lemonade-sdk might be problematic. Click here for more details.

@@ -1,4 +1,4 @@
1
- from typing import Optional, Union, List, Any
1
+ from typing import Optional, Union, List
2
2
 
3
3
  from pydantic import BaseModel
4
4
 
@@ -18,9 +18,6 @@ class LoadConfig(BaseModel):
18
18
  model_name: str
19
19
  checkpoint: Optional[str] = None
20
20
  recipe: Optional[str] = None
21
- # Indicates the maximum prompt length allowed for that specific
22
- # checkpoint + recipe combination
23
- max_prompt_length: Optional[int] = None
24
21
  # Indicates whether the model is a reasoning model, like DeepSeek
25
22
  reasoning: Optional[bool] = False
26
23
  # Indicates which Multimodal Projector (mmproj) file to use
@@ -39,19 +39,16 @@
39
39
  "Llama-3.2-1B-Instruct-Hybrid": {
40
40
  "checkpoint": "amd/Llama-3.2-1B-Instruct-awq-g128-int4-asym-fp16-onnx-hybrid",
41
41
  "recipe": "oga-hybrid",
42
- "max_prompt_length": 3000,
43
42
  "suggested": true
44
43
  },
45
44
  "Llama-3.2-3B-Instruct-Hybrid": {
46
45
  "checkpoint": "amd/Llama-3.2-3B-Instruct-awq-g128-int4-asym-fp16-onnx-hybrid",
47
46
  "recipe": "oga-hybrid",
48
- "max_prompt_length": 2000,
49
47
  "suggested": true
50
48
  },
51
49
  "Phi-3-Mini-Instruct-Hybrid": {
52
50
  "checkpoint": "amd/Phi-3-mini-4k-instruct-awq-g128-int4-asym-fp16-onnx-hybrid",
53
51
  "recipe": "oga-hybrid",
54
- "max_prompt_length": 2000,
55
52
  "suggested": true
56
53
  },
57
54
  "Phi-3.5-Mini-Instruct-Hybrid": {
@@ -62,13 +59,26 @@
62
59
  "Qwen-1.5-7B-Chat-Hybrid": {
63
60
  "checkpoint": "amd/Qwen1.5-7B-Chat-awq-g128-int4-asym-fp16-onnx-hybrid",
64
61
  "recipe": "oga-hybrid",
65
- "max_prompt_length": 3000,
62
+ "suggested": true
63
+ },
64
+ "Qwen-2.5-7B-Instruct-Hybrid": {
65
+ "checkpoint": "amd/Qwen2.5-7B-Instruct-awq-uint4-asym-g128-lmhead-g32-fp16-onnx-hybrid",
66
+ "recipe": "oga-hybrid",
67
+ "suggested": true
68
+ },
69
+ "Qwen-2.5-3B-Instruct-Hybrid": {
70
+ "checkpoint": "amd/Qwen2.5-3B-Instruct-awq-uint4-asym-g128-lmhead-g32-fp16-onnx-hybrid",
71
+ "recipe": "oga-hybrid",
72
+ "suggested": true
73
+ },
74
+ "Qwen-2.5-1.5B-Instruct-Hybrid": {
75
+ "checkpoint": "amd/Qwen2.5-1.5B-Instruct-awq-uint4-asym-g128-lmhead-g32-fp16-onnx-hybrid",
76
+ "recipe": "oga-hybrid",
66
77
  "suggested": true
67
78
  },
68
79
  "DeepSeek-R1-Distill-Llama-8B-Hybrid": {
69
80
  "checkpoint": "amd/DeepSeek-R1-Distill-Llama-8B-awq-asym-uint4-g128-lmhead-onnx-hybrid",
70
81
  "recipe": "oga-hybrid",
71
- "max_prompt_length": 2000,
72
82
  "suggested": true,
73
83
  "labels": ["reasoning"]
74
84
  },
@@ -76,25 +86,32 @@
76
86
  "checkpoint": "amd/DeepSeek-R1-Distill-Qwen-7B-awq-asym-uint4-g128-lmhead-onnx-hybrid",
77
87
  "recipe": "oga-hybrid",
78
88
  "max_prompt_length": 2000,
79
- "suggested": true,
89
+ "suggested": false,
80
90
  "labels": ["reasoning"]
81
91
  },
82
92
  "Mistral-7B-v0.3-Instruct-Hybrid": {
83
93
  "checkpoint": "amd/Mistral-7B-Instruct-v0.3-awq-g128-int4-asym-fp16-onnx-hybrid",
84
94
  "recipe": "oga-hybrid",
85
- "max_prompt_length": 2000,
86
95
  "suggested": true
87
96
  },
88
97
  "Llama-3.1-8B-Instruct-Hybrid": {
89
98
  "checkpoint": "amd/Llama-3.1-8B-Instruct-awq-asym-uint4-g128-lmhead-onnx-hybrid",
90
99
  "recipe": "oga-hybrid",
91
- "max_prompt_length": 2000,
92
100
  "suggested": true
93
101
  },
94
102
  "Llama-xLAM-2-8b-fc-r-Hybrid": {
95
103
  "checkpoint": "amd/Llama-xLAM-2-8b-fc-r-awq-g128-int4-asym-bfp16-onnx-hybrid",
96
104
  "recipe": "oga-hybrid",
97
- "max_prompt_length": 2000,
105
+ "suggested": true
106
+ },
107
+ "Qwen-2.5-7B-Instruct-NPU": {
108
+ "checkpoint": "amd/Qwen2.5-7B-Instruct-awq-g128-int4-asym-bf16-onnx-ryzen-strix",
109
+ "recipe": "oga-npu",
110
+ "suggested": true
111
+ },
112
+ "Qwen-2.5-1.5B-Instruct-NPU": {
113
+ "checkpoint": "amd/Qwen2.5-1.5B-Instruct-awq-g128-int4-asym-bf16-onnx-ryzen-strix",
114
+ "recipe": "oga-npu",
98
115
  "suggested": true
99
116
  },
100
117
  "Llama-3.2-1B-Instruct-DirectML": {
@@ -169,6 +186,12 @@
169
186
  "suggested": true,
170
187
  "labels": ["reasoning"]
171
188
  },
189
+ "Qwen3-30B-A3B-Instruct-2507-GGUF": {
190
+ "checkpoint": "unsloth/Qwen3-30B-A3B-Instruct-2507-GGUF:Qwen3-30B-A3B-Instruct-2507-Q4_0.gguf",
191
+ "recipe": "llamacpp",
192
+ "suggested": true,
193
+ "labels": ["coding"]
194
+ },
172
195
  "Gemma-3-4b-it-GGUF": {
173
196
  "checkpoint": "ggml-org/gemma-3-4b-it-GGUF:Q4_K_M",
174
197
  "mmproj": "mmproj-model-f16.gguf",
@@ -213,5 +236,17 @@
213
236
  "recipe": "llamacpp",
214
237
  "suggested": false,
215
238
  "labels": ["reranking"]
239
+ },
240
+ "Devstral-Small-2507-GGUF":{
241
+ "checkpoint": "mistralai/Devstral-Small-2507_gguf:Q4_K_M",
242
+ "recipe": "llamacpp",
243
+ "suggested": true,
244
+ "labels": ["coding"]
245
+ },
246
+ "Qwen2.5-Coder-32B-Instruct-GGUF": {
247
+ "checkpoint": "Qwen/Qwen2.5-Coder-32B-Instruct-GGUF:Q4_K_M",
248
+ "recipe": "llamacpp",
249
+ "suggested": true,
250
+ "labels": ["reasoning", "coding"]
216
251
  }
217
252
  }