nvidia-nat-llama-index 1.3a20250819__py3-none-any.whl → 1.3.0.dev2__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -12,18 +12,35 @@
12
12
  # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13
13
  # See the License for the specific language governing permissions and
14
14
  # limitations under the License.
15
- # pylint: disable=unused-argument
16
15
 
17
16
  from nat.builder.builder import Builder
18
17
  from nat.builder.framework_enum import LLMFrameworkEnum
19
18
  from nat.cli.register_workflow import register_embedder_client
20
19
  from nat.data_models.retry_mixin import RetryMixin
20
+ from nat.embedder.azure_openai_embedder import AzureOpenAIEmbedderModelConfig
21
21
  from nat.embedder.nim_embedder import NIMEmbedderModelConfig
22
+ from nat.embedder.openai_embedder import OpenAIEmbedderModelConfig
22
23
  from nat.utils.exception_handlers.automatic_retries import patch_with_retry
23
24
 
24
25
 
26
+ @register_embedder_client(config_type=AzureOpenAIEmbedderModelConfig, wrapper_type=LLMFrameworkEnum.LLAMA_INDEX)
27
+ async def azure_openai_llama_index(embedder_config: AzureOpenAIEmbedderModelConfig, _builder: Builder):
28
+
29
+ from llama_index.embeddings.azure_openai import AzureOpenAIEmbedding
30
+
31
+ client = AzureOpenAIEmbedding(**embedder_config.model_dump(exclude={"type"}, by_alias=True))
32
+
33
+ if isinstance(embedder_config, RetryMixin):
34
+ client = patch_with_retry(client,
35
+ retries=embedder_config.num_retries,
36
+ retry_codes=embedder_config.retry_on_status_codes,
37
+ retry_on_messages=embedder_config.retry_on_errors)
38
+
39
+ yield client
40
+
41
+
25
42
  @register_embedder_client(config_type=NIMEmbedderModelConfig, wrapper_type=LLMFrameworkEnum.LLAMA_INDEX)
26
- async def nim_llamaindex(embedder_config: NIMEmbedderModelConfig, builder: Builder):
43
+ async def nim_llamaindex(embedder_config: NIMEmbedderModelConfig, _builder: Builder):
27
44
 
28
45
  from llama_index.embeddings.nvidia import NVIDIAEmbedding # pylint: disable=no-name-in-module
29
46
 
@@ -33,7 +50,15 @@ async def nim_llamaindex(embedder_config: NIMEmbedderModelConfig, builder: Build
33
50
  embedder_config.model_name,
34
51
  }
35
52
 
36
- client = NVIDIAEmbedding(**config_obj)
53
+ yield NVIDIAEmbedding(**config_obj)
54
+
55
+
56
+ @register_embedder_client(config_type=OpenAIEmbedderModelConfig, wrapper_type=LLMFrameworkEnum.LLAMA_INDEX)
57
+ async def openai_llama_index(embedder_config: OpenAIEmbedderModelConfig, _builder: Builder):
58
+
59
+ from llama_index.embeddings.openai import OpenAIEmbedding
60
+
61
+ client = OpenAIEmbedding(**embedder_config.model_dump(exclude={"type"}, by_alias=True))
37
62
 
38
63
  if isinstance(embedder_config, RetryMixin):
39
64
  client = patch_with_retry(client,
@@ -18,22 +18,38 @@ from nat.builder.framework_enum import LLMFrameworkEnum
18
18
  from nat.cli.register_workflow import register_llm_client
19
19
  from nat.data_models.retry_mixin import RetryMixin
20
20
  from nat.llm.aws_bedrock_llm import AWSBedrockModelConfig
21
+ from nat.llm.azure_openai_llm import AzureOpenAIModelConfig
21
22
  from nat.llm.nim_llm import NIMModelConfig
22
23
  from nat.llm.openai_llm import OpenAIModelConfig
23
24
  from nat.utils.exception_handlers.automatic_retries import patch_with_retry
24
25
 
25
26
 
26
- @register_llm_client(config_type=NIMModelConfig, wrapper_type=LLMFrameworkEnum.LLAMA_INDEX)
27
- async def nim_llama_index(llm_config: NIMModelConfig, builder: Builder):
27
+ @register_llm_client(config_type=AWSBedrockModelConfig, wrapper_type=LLMFrameworkEnum.LLAMA_INDEX)
28
+ async def aws_bedrock_llama_index(llm_config: AWSBedrockModelConfig, _builder: Builder):
28
29
 
29
- from llama_index.llms.nvidia import NVIDIA
30
+ from llama_index.llms.bedrock import Bedrock
30
31
 
31
- kwargs = llm_config.model_dump(exclude={"type"}, by_alias=True)
32
+ kwargs = llm_config.model_dump(exclude={"type", "max_tokens"}, by_alias=True)
32
33
 
33
- if ("base_url" in kwargs and kwargs["base_url"] is None):
34
- del kwargs["base_url"]
34
+ llm = Bedrock(**kwargs)
35
35
 
36
- llm = NVIDIA(**kwargs)
36
+ if isinstance(llm_config, RetryMixin):
37
+ llm = patch_with_retry(llm,
38
+ retries=llm_config.num_retries,
39
+ retry_codes=llm_config.retry_on_status_codes,
40
+ retry_on_messages=llm_config.retry_on_errors)
41
+
42
+ yield llm
43
+
44
+
45
+ @register_llm_client(config_type=AzureOpenAIModelConfig, wrapper_type=LLMFrameworkEnum.LLAMA_INDEX)
46
+ async def azure_openai_llama_index(llm_config: AzureOpenAIModelConfig, _builder: Builder):
47
+
48
+ from llama_index.llms.azure_openai import AzureOpenAI
49
+
50
+ kwargs = llm_config.model_dump(exclude={"type"}, by_alias=True)
51
+
52
+ llm = AzureOpenAI(**kwargs)
37
53
 
38
54
  if isinstance(llm_config, RetryMixin):
39
55
  llm = patch_with_retry(llm,
@@ -44,17 +60,17 @@ async def nim_llama_index(llm_config: NIMModelConfig, builder: Builder):
44
60
  yield llm
45
61
 
46
62
 
47
- @register_llm_client(config_type=OpenAIModelConfig, wrapper_type=LLMFrameworkEnum.LLAMA_INDEX)
48
- async def openai_llama_index(llm_config: OpenAIModelConfig, builder: Builder):
63
+ @register_llm_client(config_type=NIMModelConfig, wrapper_type=LLMFrameworkEnum.LLAMA_INDEX)
64
+ async def nim_llama_index(llm_config: NIMModelConfig, _builder: Builder):
49
65
 
50
- from llama_index.llms.openai import OpenAI
66
+ from llama_index.llms.nvidia import NVIDIA
51
67
 
52
68
  kwargs = llm_config.model_dump(exclude={"type"}, by_alias=True)
53
69
 
54
70
  if ("base_url" in kwargs and kwargs["base_url"] is None):
55
71
  del kwargs["base_url"]
56
72
 
57
- llm = OpenAI(**kwargs)
73
+ llm = NVIDIA(**kwargs)
58
74
 
59
75
  if isinstance(llm_config, RetryMixin):
60
76
  llm = patch_with_retry(llm,
@@ -65,14 +81,17 @@ async def openai_llama_index(llm_config: OpenAIModelConfig, builder: Builder):
65
81
  yield llm
66
82
 
67
83
 
68
- @register_llm_client(config_type=AWSBedrockModelConfig, wrapper_type=LLMFrameworkEnum.LLAMA_INDEX)
69
- async def aws_bedrock_llama_index(llm_config: AWSBedrockModelConfig, builder: Builder):
84
+ @register_llm_client(config_type=OpenAIModelConfig, wrapper_type=LLMFrameworkEnum.LLAMA_INDEX)
85
+ async def openai_llama_index(llm_config: OpenAIModelConfig, _builder: Builder):
70
86
 
71
- from llama_index.llms.bedrock import Bedrock
87
+ from llama_index.llms.openai import OpenAI
72
88
 
73
- kwargs = llm_config.model_dump(exclude={"type", "max_tokens"}, by_alias=True)
89
+ kwargs = llm_config.model_dump(exclude={"type"}, by_alias=True)
74
90
 
75
- llm = Bedrock(**kwargs)
91
+ if ("base_url" in kwargs and kwargs["base_url"] is None):
92
+ del kwargs["base_url"]
93
+
94
+ llm = OpenAI(**kwargs)
76
95
 
77
96
  if isinstance(llm_config, RetryMixin):
78
97
  llm = patch_with_retry(llm,
@@ -19,6 +19,6 @@
19
19
 
20
20
  # Import any providers which need to be automatically registered here
21
21
 
22
+ from . import embedder
22
23
  from . import llm
23
24
  from . import tool_wrapper
24
- from . import embedder
@@ -1,18 +1,22 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: nvidia-nat-llama-index
3
- Version: 1.3a20250819
3
+ Version: 1.3.0.dev2
4
4
  Summary: Subpackage for Llama-Index integration in NeMo Agent toolkit
5
5
  Keywords: ai,rag,agents
6
6
  Classifier: Programming Language :: Python
7
7
  Requires-Python: <3.13,>=3.11
8
8
  Description-Content-Type: text/markdown
9
- Requires-Dist: nvidia-nat==v1.3a20250819
10
- Requires-Dist: llama-index-core==0.12.21
11
- Requires-Dist: llama-index-embeddings-nvidia==0.3.1
12
- Requires-Dist: llama-index-llms-bedrock==0.3.8
13
- Requires-Dist: llama-index-llms-nvidia==0.3.1
14
- Requires-Dist: llama-index-readers-file==0.4.4
15
- Requires-Dist: llama-index==0.12.21
9
+ Requires-Dist: nvidia-nat==v1.3.0-dev2
10
+ Requires-Dist: llama-index-core~=0.12.21
11
+ Requires-Dist: llama-index-embeddings-azure-openai~=0.3.1
12
+ Requires-Dist: llama-index-embeddings-nvidia~=0.3.1
13
+ Requires-Dist: llama-index-embeddings-openai~=0.3.1
14
+ Requires-Dist: llama-index-llms-azure-openai~=0.3.1
15
+ Requires-Dist: llama-index-llms-bedrock~=0.3.8
16
+ Requires-Dist: llama-index-llms-nvidia~=0.3.1
17
+ Requires-Dist: llama-index-llms-openai~=0.3.38
18
+ Requires-Dist: llama-index-readers-file~=0.4.4
19
+ Requires-Dist: llama-index~=0.12.21
16
20
 
17
21
  <!--
18
22
  SPDX-FileCopyrightText: Copyright (c) 2025, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
@@ -0,0 +1,11 @@
1
+ nat/meta/pypi.md,sha256=s9C3pgWB0HLIXTx5QPryNOWN0O2fIRIap0p9_zCHlTs,1112
2
+ nat/plugins/llama_index/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
3
+ nat/plugins/llama_index/embedder.py,sha256=ROGlJJxwIZXV3Sg93Z9RIelOOhxuJcrw0B1id43tn_U,3130
4
+ nat/plugins/llama_index/llm.py,sha256=PtXibAiZWCv1pjEeP5A77Tyi-rVUGCHc81jEoIdy3p8,3949
5
+ nat/plugins/llama_index/register.py,sha256=ICmUCqrNvWTDFe6Zjm6B6KUkgqsyqfFPTHMGU74KsA4,891
6
+ nat/plugins/llama_index/tool_wrapper.py,sha256=VFKMIIeLdWqHwW2Ax11E2w-_9w3ow6Iuhra1Hk78RYM,1387
7
+ nvidia_nat_llama_index-1.3.0.dev2.dist-info/METADATA,sha256=uKukv46PucXHXNFo51uQa_FkYeR6JHtk91E5W8cIzxY,1908
8
+ nvidia_nat_llama_index-1.3.0.dev2.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
9
+ nvidia_nat_llama_index-1.3.0.dev2.dist-info/entry_points.txt,sha256=2LqRRju5448P2v8B3y6TSPnk-nOd5T3AmV5JibCnoQc,68
10
+ nvidia_nat_llama_index-1.3.0.dev2.dist-info/top_level.txt,sha256=8-CJ2cP6-f0ZReXe5Hzqp-5pvzzHz-5Ds5H2bGqh1-U,4
11
+ nvidia_nat_llama_index-1.3.0.dev2.dist-info/RECORD,,
@@ -1,11 +0,0 @@
1
- nat/meta/pypi.md,sha256=s9C3pgWB0HLIXTx5QPryNOWN0O2fIRIap0p9_zCHlTs,1112
2
- nat/plugins/llama_index/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
3
- nat/plugins/llama_index/embedder.py,sha256=L7Wx9QJH2Sh0y6kmgisjtTZfDC9YcuuusKl0Tn_7JgA,1921
4
- nat/plugins/llama_index/llm.py,sha256=VyuIqks2G9Zcs9a0X7jJyASzSUdtL0Pns0v4_rFEsJE,3220
5
- nat/plugins/llama_index/register.py,sha256=DelI6uth8dO8h_7czPW4Akfr_ajXDBwhEa1dJzRf5u4,891
6
- nat/plugins/llama_index/tool_wrapper.py,sha256=VFKMIIeLdWqHwW2Ax11E2w-_9w3ow6Iuhra1Hk78RYM,1387
7
- nvidia_nat_llama_index-1.3a20250819.dist-info/METADATA,sha256=D_tmaDrZz1Asy1TrV034A8iVKbyafsUmcVe4Q1xwvOA,1703
8
- nvidia_nat_llama_index-1.3a20250819.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
9
- nvidia_nat_llama_index-1.3a20250819.dist-info/entry_points.txt,sha256=2LqRRju5448P2v8B3y6TSPnk-nOd5T3AmV5JibCnoQc,68
10
- nvidia_nat_llama_index-1.3a20250819.dist-info/top_level.txt,sha256=8-CJ2cP6-f0ZReXe5Hzqp-5pvzzHz-5Ds5H2bGqh1-U,4
11
- nvidia_nat_llama_index-1.3a20250819.dist-info/RECORD,,