nvidia-nat-llama-index 1.3.dev0__py3-none-any.whl → 1.3.0.dev2__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- nat/plugins/llama_index/embedder.py +28 -3
- nat/plugins/llama_index/llm.py +35 -16
- nat/plugins/llama_index/register.py +1 -1
- {nvidia_nat_llama_index-1.3.dev0.dist-info → nvidia_nat_llama_index-1.3.0.dev2.dist-info}/METADATA +12 -8
- nvidia_nat_llama_index-1.3.0.dev2.dist-info/RECORD +11 -0
- nvidia_nat_llama_index-1.3.dev0.dist-info/RECORD +0 -11
- {nvidia_nat_llama_index-1.3.dev0.dist-info → nvidia_nat_llama_index-1.3.0.dev2.dist-info}/WHEEL +0 -0
- {nvidia_nat_llama_index-1.3.dev0.dist-info → nvidia_nat_llama_index-1.3.0.dev2.dist-info}/entry_points.txt +0 -0
- {nvidia_nat_llama_index-1.3.dev0.dist-info → nvidia_nat_llama_index-1.3.0.dev2.dist-info}/top_level.txt +0 -0
@@ -12,18 +12,35 @@
|
|
12
12
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
13
13
|
# See the License for the specific language governing permissions and
|
14
14
|
# limitations under the License.
|
15
|
-
# pylint: disable=unused-argument
|
16
15
|
|
17
16
|
from nat.builder.builder import Builder
|
18
17
|
from nat.builder.framework_enum import LLMFrameworkEnum
|
19
18
|
from nat.cli.register_workflow import register_embedder_client
|
20
19
|
from nat.data_models.retry_mixin import RetryMixin
|
20
|
+
from nat.embedder.azure_openai_embedder import AzureOpenAIEmbedderModelConfig
|
21
21
|
from nat.embedder.nim_embedder import NIMEmbedderModelConfig
|
22
|
+
from nat.embedder.openai_embedder import OpenAIEmbedderModelConfig
|
22
23
|
from nat.utils.exception_handlers.automatic_retries import patch_with_retry
|
23
24
|
|
24
25
|
|
26
|
+
@register_embedder_client(config_type=AzureOpenAIEmbedderModelConfig, wrapper_type=LLMFrameworkEnum.LLAMA_INDEX)
|
27
|
+
async def azure_openai_llama_index(embedder_config: AzureOpenAIEmbedderModelConfig, _builder: Builder):
|
28
|
+
|
29
|
+
from llama_index.embeddings.azure_openai import AzureOpenAIEmbedding
|
30
|
+
|
31
|
+
client = AzureOpenAIEmbedding(**embedder_config.model_dump(exclude={"type"}, by_alias=True))
|
32
|
+
|
33
|
+
if isinstance(embedder_config, RetryMixin):
|
34
|
+
client = patch_with_retry(client,
|
35
|
+
retries=embedder_config.num_retries,
|
36
|
+
retry_codes=embedder_config.retry_on_status_codes,
|
37
|
+
retry_on_messages=embedder_config.retry_on_errors)
|
38
|
+
|
39
|
+
yield client
|
40
|
+
|
41
|
+
|
25
42
|
@register_embedder_client(config_type=NIMEmbedderModelConfig, wrapper_type=LLMFrameworkEnum.LLAMA_INDEX)
|
26
|
-
async def nim_llamaindex(embedder_config: NIMEmbedderModelConfig,
|
43
|
+
async def nim_llamaindex(embedder_config: NIMEmbedderModelConfig, _builder: Builder):
|
27
44
|
|
28
45
|
from llama_index.embeddings.nvidia import NVIDIAEmbedding # pylint: disable=no-name-in-module
|
29
46
|
|
@@ -33,7 +50,15 @@ async def nim_llamaindex(embedder_config: NIMEmbedderModelConfig, builder: Build
|
|
33
50
|
embedder_config.model_name,
|
34
51
|
}
|
35
52
|
|
36
|
-
|
53
|
+
yield NVIDIAEmbedding(**config_obj)
|
54
|
+
|
55
|
+
|
56
|
+
@register_embedder_client(config_type=OpenAIEmbedderModelConfig, wrapper_type=LLMFrameworkEnum.LLAMA_INDEX)
|
57
|
+
async def openai_llama_index(embedder_config: OpenAIEmbedderModelConfig, _builder: Builder):
|
58
|
+
|
59
|
+
from llama_index.embeddings.openai import OpenAIEmbedding
|
60
|
+
|
61
|
+
client = OpenAIEmbedding(**embedder_config.model_dump(exclude={"type"}, by_alias=True))
|
37
62
|
|
38
63
|
if isinstance(embedder_config, RetryMixin):
|
39
64
|
client = patch_with_retry(client,
|
nat/plugins/llama_index/llm.py
CHANGED
@@ -18,22 +18,38 @@ from nat.builder.framework_enum import LLMFrameworkEnum
|
|
18
18
|
from nat.cli.register_workflow import register_llm_client
|
19
19
|
from nat.data_models.retry_mixin import RetryMixin
|
20
20
|
from nat.llm.aws_bedrock_llm import AWSBedrockModelConfig
|
21
|
+
from nat.llm.azure_openai_llm import AzureOpenAIModelConfig
|
21
22
|
from nat.llm.nim_llm import NIMModelConfig
|
22
23
|
from nat.llm.openai_llm import OpenAIModelConfig
|
23
24
|
from nat.utils.exception_handlers.automatic_retries import patch_with_retry
|
24
25
|
|
25
26
|
|
26
|
-
@register_llm_client(config_type=
|
27
|
-
async def
|
27
|
+
@register_llm_client(config_type=AWSBedrockModelConfig, wrapper_type=LLMFrameworkEnum.LLAMA_INDEX)
|
28
|
+
async def aws_bedrock_llama_index(llm_config: AWSBedrockModelConfig, _builder: Builder):
|
28
29
|
|
29
|
-
from llama_index.llms.
|
30
|
+
from llama_index.llms.bedrock import Bedrock
|
30
31
|
|
31
|
-
kwargs = llm_config.model_dump(exclude={"type"}, by_alias=True)
|
32
|
+
kwargs = llm_config.model_dump(exclude={"type", "max_tokens"}, by_alias=True)
|
32
33
|
|
33
|
-
|
34
|
-
del kwargs["base_url"]
|
34
|
+
llm = Bedrock(**kwargs)
|
35
35
|
|
36
|
-
|
36
|
+
if isinstance(llm_config, RetryMixin):
|
37
|
+
llm = patch_with_retry(llm,
|
38
|
+
retries=llm_config.num_retries,
|
39
|
+
retry_codes=llm_config.retry_on_status_codes,
|
40
|
+
retry_on_messages=llm_config.retry_on_errors)
|
41
|
+
|
42
|
+
yield llm
|
43
|
+
|
44
|
+
|
45
|
+
@register_llm_client(config_type=AzureOpenAIModelConfig, wrapper_type=LLMFrameworkEnum.LLAMA_INDEX)
|
46
|
+
async def azure_openai_llama_index(llm_config: AzureOpenAIModelConfig, _builder: Builder):
|
47
|
+
|
48
|
+
from llama_index.llms.azure_openai import AzureOpenAI
|
49
|
+
|
50
|
+
kwargs = llm_config.model_dump(exclude={"type"}, by_alias=True)
|
51
|
+
|
52
|
+
llm = AzureOpenAI(**kwargs)
|
37
53
|
|
38
54
|
if isinstance(llm_config, RetryMixin):
|
39
55
|
llm = patch_with_retry(llm,
|
@@ -44,17 +60,17 @@ async def nim_llama_index(llm_config: NIMModelConfig, builder: Builder):
|
|
44
60
|
yield llm
|
45
61
|
|
46
62
|
|
47
|
-
@register_llm_client(config_type=
|
48
|
-
async def
|
63
|
+
@register_llm_client(config_type=NIMModelConfig, wrapper_type=LLMFrameworkEnum.LLAMA_INDEX)
|
64
|
+
async def nim_llama_index(llm_config: NIMModelConfig, _builder: Builder):
|
49
65
|
|
50
|
-
from llama_index.llms.
|
66
|
+
from llama_index.llms.nvidia import NVIDIA
|
51
67
|
|
52
68
|
kwargs = llm_config.model_dump(exclude={"type"}, by_alias=True)
|
53
69
|
|
54
70
|
if ("base_url" in kwargs and kwargs["base_url"] is None):
|
55
71
|
del kwargs["base_url"]
|
56
72
|
|
57
|
-
llm =
|
73
|
+
llm = NVIDIA(**kwargs)
|
58
74
|
|
59
75
|
if isinstance(llm_config, RetryMixin):
|
60
76
|
llm = patch_with_retry(llm,
|
@@ -65,14 +81,17 @@ async def openai_llama_index(llm_config: OpenAIModelConfig, builder: Builder):
|
|
65
81
|
yield llm
|
66
82
|
|
67
83
|
|
68
|
-
@register_llm_client(config_type=
|
69
|
-
async def
|
84
|
+
@register_llm_client(config_type=OpenAIModelConfig, wrapper_type=LLMFrameworkEnum.LLAMA_INDEX)
|
85
|
+
async def openai_llama_index(llm_config: OpenAIModelConfig, _builder: Builder):
|
70
86
|
|
71
|
-
from llama_index.llms.
|
87
|
+
from llama_index.llms.openai import OpenAI
|
72
88
|
|
73
|
-
kwargs = llm_config.model_dump(exclude={"type"
|
89
|
+
kwargs = llm_config.model_dump(exclude={"type"}, by_alias=True)
|
74
90
|
|
75
|
-
|
91
|
+
if ("base_url" in kwargs and kwargs["base_url"] is None):
|
92
|
+
del kwargs["base_url"]
|
93
|
+
|
94
|
+
llm = OpenAI(**kwargs)
|
76
95
|
|
77
96
|
if isinstance(llm_config, RetryMixin):
|
78
97
|
llm = patch_with_retry(llm,
|
{nvidia_nat_llama_index-1.3.dev0.dist-info → nvidia_nat_llama_index-1.3.0.dev2.dist-info}/METADATA
RENAMED
@@ -1,18 +1,22 @@
|
|
1
1
|
Metadata-Version: 2.4
|
2
2
|
Name: nvidia-nat-llama-index
|
3
|
-
Version: 1.3.
|
3
|
+
Version: 1.3.0.dev2
|
4
4
|
Summary: Subpackage for Llama-Index integration in NeMo Agent toolkit
|
5
5
|
Keywords: ai,rag,agents
|
6
6
|
Classifier: Programming Language :: Python
|
7
7
|
Requires-Python: <3.13,>=3.11
|
8
8
|
Description-Content-Type: text/markdown
|
9
|
-
Requires-Dist: nvidia-nat==v1.3-
|
10
|
-
Requires-Dist: llama-index-core
|
11
|
-
Requires-Dist: llama-index-embeddings-
|
12
|
-
Requires-Dist: llama-index-
|
13
|
-
Requires-Dist: llama-index-
|
14
|
-
Requires-Dist: llama-index-
|
15
|
-
Requires-Dist: llama-index
|
9
|
+
Requires-Dist: nvidia-nat==v1.3.0-dev2
|
10
|
+
Requires-Dist: llama-index-core~=0.12.21
|
11
|
+
Requires-Dist: llama-index-embeddings-azure-openai~=0.3.1
|
12
|
+
Requires-Dist: llama-index-embeddings-nvidia~=0.3.1
|
13
|
+
Requires-Dist: llama-index-embeddings-openai~=0.3.1
|
14
|
+
Requires-Dist: llama-index-llms-azure-openai~=0.3.1
|
15
|
+
Requires-Dist: llama-index-llms-bedrock~=0.3.8
|
16
|
+
Requires-Dist: llama-index-llms-nvidia~=0.3.1
|
17
|
+
Requires-Dist: llama-index-llms-openai~=0.3.38
|
18
|
+
Requires-Dist: llama-index-readers-file~=0.4.4
|
19
|
+
Requires-Dist: llama-index~=0.12.21
|
16
20
|
|
17
21
|
<!--
|
18
22
|
SPDX-FileCopyrightText: Copyright (c) 2025, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
|
@@ -0,0 +1,11 @@
|
|
1
|
+
nat/meta/pypi.md,sha256=s9C3pgWB0HLIXTx5QPryNOWN0O2fIRIap0p9_zCHlTs,1112
|
2
|
+
nat/plugins/llama_index/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
3
|
+
nat/plugins/llama_index/embedder.py,sha256=ROGlJJxwIZXV3Sg93Z9RIelOOhxuJcrw0B1id43tn_U,3130
|
4
|
+
nat/plugins/llama_index/llm.py,sha256=PtXibAiZWCv1pjEeP5A77Tyi-rVUGCHc81jEoIdy3p8,3949
|
5
|
+
nat/plugins/llama_index/register.py,sha256=ICmUCqrNvWTDFe6Zjm6B6KUkgqsyqfFPTHMGU74KsA4,891
|
6
|
+
nat/plugins/llama_index/tool_wrapper.py,sha256=VFKMIIeLdWqHwW2Ax11E2w-_9w3ow6Iuhra1Hk78RYM,1387
|
7
|
+
nvidia_nat_llama_index-1.3.0.dev2.dist-info/METADATA,sha256=uKukv46PucXHXNFo51uQa_FkYeR6JHtk91E5W8cIzxY,1908
|
8
|
+
nvidia_nat_llama_index-1.3.0.dev2.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
|
9
|
+
nvidia_nat_llama_index-1.3.0.dev2.dist-info/entry_points.txt,sha256=2LqRRju5448P2v8B3y6TSPnk-nOd5T3AmV5JibCnoQc,68
|
10
|
+
nvidia_nat_llama_index-1.3.0.dev2.dist-info/top_level.txt,sha256=8-CJ2cP6-f0ZReXe5Hzqp-5pvzzHz-5Ds5H2bGqh1-U,4
|
11
|
+
nvidia_nat_llama_index-1.3.0.dev2.dist-info/RECORD,,
|
@@ -1,11 +0,0 @@
|
|
1
|
-
nat/meta/pypi.md,sha256=s9C3pgWB0HLIXTx5QPryNOWN0O2fIRIap0p9_zCHlTs,1112
|
2
|
-
nat/plugins/llama_index/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
3
|
-
nat/plugins/llama_index/embedder.py,sha256=L7Wx9QJH2Sh0y6kmgisjtTZfDC9YcuuusKl0Tn_7JgA,1921
|
4
|
-
nat/plugins/llama_index/llm.py,sha256=VyuIqks2G9Zcs9a0X7jJyASzSUdtL0Pns0v4_rFEsJE,3220
|
5
|
-
nat/plugins/llama_index/register.py,sha256=DelI6uth8dO8h_7czPW4Akfr_ajXDBwhEa1dJzRf5u4,891
|
6
|
-
nat/plugins/llama_index/tool_wrapper.py,sha256=VFKMIIeLdWqHwW2Ax11E2w-_9w3ow6Iuhra1Hk78RYM,1387
|
7
|
-
nvidia_nat_llama_index-1.3.dev0.dist-info/METADATA,sha256=EP1nCqjUnjLBuvlh_JKpFIG_rebxwmKzo0MjDejM0MY,1694
|
8
|
-
nvidia_nat_llama_index-1.3.dev0.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
|
9
|
-
nvidia_nat_llama_index-1.3.dev0.dist-info/entry_points.txt,sha256=2LqRRju5448P2v8B3y6TSPnk-nOd5T3AmV5JibCnoQc,68
|
10
|
-
nvidia_nat_llama_index-1.3.dev0.dist-info/top_level.txt,sha256=8-CJ2cP6-f0ZReXe5Hzqp-5pvzzHz-5Ds5H2bGqh1-U,4
|
11
|
-
nvidia_nat_llama_index-1.3.dev0.dist-info/RECORD,,
|
{nvidia_nat_llama_index-1.3.dev0.dist-info → nvidia_nat_llama_index-1.3.0.dev2.dist-info}/WHEEL
RENAMED
File without changes
|
File without changes
|
File without changes
|