opengradient 0.5.1__py3-none-any.whl → 0.5.2__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- opengradient/client.py +12 -4
- opengradient/types.py +29 -29
- {opengradient-0.5.1.dist-info → opengradient-0.5.2.dist-info}/METADATA +1 -1
- {opengradient-0.5.1.dist-info → opengradient-0.5.2.dist-info}/RECORD +8 -8
- {opengradient-0.5.1.dist-info → opengradient-0.5.2.dist-info}/WHEEL +0 -0
- {opengradient-0.5.1.dist-info → opengradient-0.5.2.dist-info}/entry_points.txt +0 -0
- {opengradient-0.5.1.dist-info → opengradient-0.5.2.dist-info}/licenses/LICENSE +0 -0
- {opengradient-0.5.1.dist-info → opengradient-0.5.2.dist-info}/top_level.txt +0 -0
opengradient/client.py
CHANGED
|
@@ -465,7 +465,7 @@ class Client:
|
|
|
465
465
|
return OpenGradientError("That model CID is not supported yet for TEE inference")
|
|
466
466
|
|
|
467
467
|
return self._external_llm_completion(
|
|
468
|
-
model=model_cid,
|
|
468
|
+
model=model_cid.split('/')[1],
|
|
469
469
|
prompt=prompt,
|
|
470
470
|
max_tokens=max_tokens,
|
|
471
471
|
stop_sequence=stop_sequence,
|
|
@@ -479,12 +479,16 @@ class Client:
|
|
|
479
479
|
|
|
480
480
|
if model_cid not in [llm.value for llm in LLM]:
|
|
481
481
|
raise OpenGradientError("That model CID is not yet supported for inference")
|
|
482
|
+
|
|
483
|
+
model_name = model_cid
|
|
484
|
+
if model_cid in [llm.value for llm in TEE_LLM]:
|
|
485
|
+
model_name = model_cid.split('/')[1]
|
|
482
486
|
|
|
483
487
|
contract = self._blockchain.eth.contract(address=self._inference_hub_contract_address, abi=self._inference_abi)
|
|
484
488
|
|
|
485
489
|
llm_request = {
|
|
486
490
|
"mode": inference_mode.value,
|
|
487
|
-
"modelCID":
|
|
491
|
+
"modelCID": model_name,
|
|
488
492
|
"prompt": prompt,
|
|
489
493
|
"max_tokens": max_tokens,
|
|
490
494
|
"stop_sequence": stop_sequence or [],
|
|
@@ -670,7 +674,7 @@ class Client:
|
|
|
670
674
|
return OpenGradientError("That model CID is not supported yet for TEE inference")
|
|
671
675
|
|
|
672
676
|
return self._external_llm_chat(
|
|
673
|
-
model=model_cid,
|
|
677
|
+
model=model_cid.split('/')[1],
|
|
674
678
|
messages=messages,
|
|
675
679
|
max_tokens=max_tokens,
|
|
676
680
|
stop_sequence=stop_sequence,
|
|
@@ -686,6 +690,10 @@ class Client:
|
|
|
686
690
|
|
|
687
691
|
if model_cid not in [llm.value for llm in LLM]:
|
|
688
692
|
raise OpenGradientError("That model CID is not yet supported for inference")
|
|
693
|
+
|
|
694
|
+
model_name = model_cid
|
|
695
|
+
if model_cid in [llm.value for llm in TEE_LLM]:
|
|
696
|
+
model_name = model_cid.split('/')[1]
|
|
689
697
|
|
|
690
698
|
contract = self._blockchain.eth.contract(address=self._inference_hub_contract_address, abi=self._inference_abi)
|
|
691
699
|
|
|
@@ -713,7 +721,7 @@ class Client:
|
|
|
713
721
|
|
|
714
722
|
llm_request = {
|
|
715
723
|
"mode": inference_mode.value,
|
|
716
|
-
"modelCID":
|
|
724
|
+
"modelCID": model_name,
|
|
717
725
|
"messages": messages,
|
|
718
726
|
"max_tokens": max_tokens,
|
|
719
727
|
"stop_sequence": stop_sequence or [],
|
opengradient/types.py
CHANGED
|
@@ -202,27 +202,27 @@ class LLM(str, Enum):
|
|
|
202
202
|
DOBBY_LEASHED_3_1_8B = "SentientAGI/Dobby-Mini-Leashed-Llama-3.1-8B"
|
|
203
203
|
|
|
204
204
|
# OpenAI models via TEE
|
|
205
|
-
GPT_4_1_2025_04_14 = "gpt-4.1-2025-04-14"
|
|
206
|
-
GPT_4O = "gpt-4o"
|
|
207
|
-
O4_MINI = "o4-mini"
|
|
205
|
+
GPT_4_1_2025_04_14 = "OpenAI/gpt-4.1-2025-04-14"
|
|
206
|
+
GPT_4O = "OpenAI/gpt-4o"
|
|
207
|
+
O4_MINI = "OpenAI/o4-mini"
|
|
208
208
|
|
|
209
209
|
# Anthropic models via TEE
|
|
210
|
-
CLAUDE_3_7_SONNET = "claude-3.7-sonnet"
|
|
211
|
-
CLAUDE_3_5_HAIKU = "claude-3.5-haiku"
|
|
212
|
-
CLAUDE_4_0_SONNET = "claude-4.0-sonnet"
|
|
210
|
+
CLAUDE_3_7_SONNET = "Anthropic/claude-3.7-sonnet"
|
|
211
|
+
CLAUDE_3_5_HAIKU = "Anthropic/claude-3.5-haiku"
|
|
212
|
+
CLAUDE_4_0_SONNET = "Anthropic/claude-4.0-sonnet"
|
|
213
213
|
|
|
214
214
|
# Google models via TEE
|
|
215
|
-
GEMINI_2_5_FLASH = "gemini-2.5-flash"
|
|
216
|
-
GEMINI_2_5_PRO = "gemini-2.5-pro"
|
|
217
|
-
GEMINI_2_0_FLASH = "gemini-2.0-flash"
|
|
215
|
+
GEMINI_2_5_FLASH = "Google/gemini-2.5-flash"
|
|
216
|
+
GEMINI_2_5_PRO = "Google/gemini-2.5-pro"
|
|
217
|
+
GEMINI_2_0_FLASH = "Google/gemini-2.0-flash"
|
|
218
218
|
|
|
219
219
|
# xAI Grok models via TEE
|
|
220
|
-
GROK_3_MINI_BETA = "grok-3-mini-beta"
|
|
221
|
-
GROK_3_BETA = "grok-3-beta"
|
|
220
|
+
GROK_3_MINI_BETA = "xAI/grok-3-mini-beta"
|
|
221
|
+
GROK_3_BETA = "xAI/grok-3-beta"
|
|
222
222
|
GROK_2_1212 = "grok-2-1212"
|
|
223
|
-
GROK_2_VISION_LATEST = "grok-2-vision-latest"
|
|
224
|
-
GROK_4_1_FAST = "grok-4.1-fast"
|
|
225
|
-
GROK_4_1_FAST_NON_REASONING = "grok-4-1-fast-non-reasoning"
|
|
223
|
+
GROK_2_VISION_LATEST = "xAI/grok-2-vision-latest"
|
|
224
|
+
GROK_4_1_FAST = "xAI/grok-4.1-fast"
|
|
225
|
+
GROK_4_1_FAST_NON_REASONING = "xAI/grok-4-1-fast-non-reasoning"
|
|
226
226
|
|
|
227
227
|
class TEE_LLM(str, Enum):
|
|
228
228
|
"""Enum for LLM models available for TEE execution"""
|
|
@@ -231,27 +231,27 @@ class TEE_LLM(str, Enum):
|
|
|
231
231
|
# META_LLAMA_3_1_70B_INSTRUCT = "meta-llama/Llama-3.1-70B-Instruct"
|
|
232
232
|
|
|
233
233
|
# OpenAI models via TEE
|
|
234
|
-
GPT_4_1_2025_04_14 = "gpt-4.1-2025-04-14"
|
|
235
|
-
GPT_4O = "gpt-4o"
|
|
236
|
-
O4_MINI = "o4-mini"
|
|
234
|
+
GPT_4_1_2025_04_14 = "OpenAI/gpt-4.1-2025-04-14"
|
|
235
|
+
GPT_4O = "OpenAI/gpt-4o"
|
|
236
|
+
O4_MINI = "OpenAI/o4-mini"
|
|
237
237
|
|
|
238
238
|
# Anthropic models via TEE
|
|
239
|
-
CLAUDE_3_7_SONNET = "claude-3.7-sonnet"
|
|
240
|
-
CLAUDE_3_5_HAIKU = "claude-3.5-haiku"
|
|
241
|
-
CLAUDE_4_0_SONNET = "claude-4.0-sonnet"
|
|
239
|
+
CLAUDE_3_7_SONNET = "Anthropic/claude-3.7-sonnet"
|
|
240
|
+
CLAUDE_3_5_HAIKU = "Anthropic/claude-3.5-haiku"
|
|
241
|
+
CLAUDE_4_0_SONNET = "Anthropic/claude-4.0-sonnet"
|
|
242
242
|
|
|
243
243
|
# Google models via TEE
|
|
244
|
-
GEMINI_2_5_FLASH = "gemini-2.5-flash"
|
|
245
|
-
GEMINI_2_5_PRO = "gemini-2.5-pro"
|
|
246
|
-
GEMINI_2_0_FLASH = "gemini-2.0-flash"
|
|
244
|
+
GEMINI_2_5_FLASH = "Google/gemini-2.5-flash"
|
|
245
|
+
GEMINI_2_5_PRO = "Google/gemini-2.5-pro"
|
|
246
|
+
GEMINI_2_0_FLASH = "Google/gemini-2.0-flash"
|
|
247
247
|
|
|
248
248
|
# xAI Grok models via TEE
|
|
249
|
-
GROK_3_MINI_BETA = "grok-3-mini-beta"
|
|
250
|
-
GROK_3_BETA = "grok-3-beta"
|
|
251
|
-
GROK_2_1212 = "grok-2-1212"
|
|
252
|
-
GROK_2_VISION_LATEST = "grok-2-vision-latest"
|
|
253
|
-
GROK_4_1_FAST = "grok-4.1-fast"
|
|
254
|
-
GROK_4_1_FAST_NON_REASONING = "grok-4-1-fast-non-reasoning"
|
|
249
|
+
GROK_3_MINI_BETA = "xAI/grok-3-mini-beta"
|
|
250
|
+
GROK_3_BETA = "xAI/grok-3-beta"
|
|
251
|
+
GROK_2_1212 = "xAI/grok-2-1212"
|
|
252
|
+
GROK_2_VISION_LATEST = "xAI/grok-2-vision-latest"
|
|
253
|
+
GROK_4_1_FAST = "xAI/grok-4.1-fast"
|
|
254
|
+
GROK_4_1_FAST_NON_REASONING = "xAI/grok-4-1-fast-non-reasoning"
|
|
255
255
|
|
|
256
256
|
|
|
257
257
|
@dataclass
|
|
@@ -1,10 +1,10 @@
|
|
|
1
1
|
opengradient/__init__.py,sha256=wVg0KTFNBl7RnZF9huR5-m_q1E7tO-YyQwY7AD9JFoc,12635
|
|
2
2
|
opengradient/account.py,sha256=5wrYpws_1lozjOFjLCTHtxgoxK-LmObDAaVy9eDcJY4,1145
|
|
3
3
|
opengradient/cli.py,sha256=QzjH_KS6TF8gm_L1otFWA-oHkJ5SSfizFoRn0xR0b70,29162
|
|
4
|
-
opengradient/client.py,sha256=
|
|
4
|
+
opengradient/client.py,sha256=Seid3raffwAcrc3qVIINu259iO-BdaGXGT5O2VaT6ac,62356
|
|
5
5
|
opengradient/defaults.py,sha256=w8-dr5ciF2TGnqbm_ib0Yz4U0YL5ikpNqkcPVpmXzP8,673
|
|
6
6
|
opengradient/exceptions.py,sha256=88tfegboGtlehQcwhxsl6ZzhLJWZWlkf_bkHTiCtXpo,3391
|
|
7
|
-
opengradient/types.py,sha256=
|
|
7
|
+
opengradient/types.py,sha256=ygnQXoGJPv9i3daS0oduUsmUNoPGx6Oozkt-Yy7Nn6s,7548
|
|
8
8
|
opengradient/utils.py,sha256=ZUq4OBIml2vsC0tRqus4Zwb_e3g4woo00apByrafuVw,8058
|
|
9
9
|
opengradient/abi/InferencePrecompile.abi,sha256=reepTHg6Q01UrFP0Gexc-JayplsvOLPfG7jrEZ-cV28,10197
|
|
10
10
|
opengradient/abi/PriceHistoryInference.abi,sha256=ZB3fZdx1kaFlp2wt1vTbTZZG1k8HPvmNtkG5Q8Bnajw,5098
|
|
@@ -27,9 +27,9 @@ opengradient/workflow_models/constants.py,sha256=viIkb_LGcfVprqQNaA80gBTj6cfYam0
|
|
|
27
27
|
opengradient/workflow_models/types.py,sha256=Z22hF6c8Y4D2GlzVEIBODGwsqSjSrQvUcpZ7R-mIJdI,409
|
|
28
28
|
opengradient/workflow_models/utils.py,sha256=ySfpuiOBqLTlfto6ZxZf2vc7K6RGIja0l4eaVm5AOzY,1503
|
|
29
29
|
opengradient/workflow_models/workflow_models.py,sha256=d4C_gs39DAfy4cdY9Ee6GMXpPfzwvKFpmxzK1A7LNgU,3900
|
|
30
|
-
opengradient-0.5.
|
|
31
|
-
opengradient-0.5.
|
|
32
|
-
opengradient-0.5.
|
|
33
|
-
opengradient-0.5.
|
|
34
|
-
opengradient-0.5.
|
|
35
|
-
opengradient-0.5.
|
|
30
|
+
opengradient-0.5.2.dist-info/licenses/LICENSE,sha256=xEcvQ3AxZOtDkrqkys2Mm6Y9diEnaSeQRKvxi-JGnNA,1069
|
|
31
|
+
opengradient-0.5.2.dist-info/METADATA,sha256=vKpBpoNs4RhhXcz275SIoYrIpL550Z7ykk_9zS0-CdA,3992
|
|
32
|
+
opengradient-0.5.2.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
|
|
33
|
+
opengradient-0.5.2.dist-info/entry_points.txt,sha256=yUKTaJx8RXnybkob0J62wVBiCp_1agVbgw9uzsmaeJc,54
|
|
34
|
+
opengradient-0.5.2.dist-info/top_level.txt,sha256=oC1zimVLa2Yi1LQz8c7x-0IQm92milb5ax8gHBHwDqU,13
|
|
35
|
+
opengradient-0.5.2.dist-info/RECORD,,
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|