nvidia-nat 1.4.0a20251102__py3-none-any.whl → 1.4.0a20251112__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- nat/cli/commands/workflow/workflow_commands.py +3 -2
- nat/eval/dataset_handler/dataset_filter.py +34 -2
- nat/eval/evaluate.py +1 -1
- nat/eval/utils/weave_eval.py +17 -3
- nat/front_ends/fastapi/fastapi_front_end_config.py +7 -0
- nat/front_ends/fastapi/fastapi_front_end_plugin.py +13 -7
- nat/front_ends/fastapi/fastapi_front_end_plugin_worker.py +20 -14
- nat/llm/aws_bedrock_llm.py +11 -9
- nat/llm/azure_openai_llm.py +12 -4
- nat/llm/litellm_llm.py +11 -4
- nat/llm/nim_llm.py +11 -9
- nat/llm/openai_llm.py +12 -9
- nat/tool/code_execution/code_sandbox.py +3 -6
- nat/tool/code_execution/local_sandbox/Dockerfile.sandbox +19 -32
- nat/tool/code_execution/local_sandbox/local_sandbox_server.py +5 -0
- nat/tool/code_execution/local_sandbox/sandbox.requirements.txt +2 -0
- nat/tool/code_execution/local_sandbox/start_local_sandbox.sh +10 -4
- nat/tool/server_tools.py +15 -2
- nat/utils/__init__.py +8 -4
- nat/utils/io/yaml_tools.py +73 -3
- {nvidia_nat-1.4.0a20251102.dist-info → nvidia_nat-1.4.0a20251112.dist-info}/METADATA +3 -1
- {nvidia_nat-1.4.0a20251102.dist-info → nvidia_nat-1.4.0a20251112.dist-info}/RECORD +27 -30
- nat/data_models/temperature_mixin.py +0 -44
- nat/data_models/top_p_mixin.py +0 -44
- nat/tool/code_execution/test_code_execution_sandbox.py +0 -414
- {nvidia_nat-1.4.0a20251102.dist-info → nvidia_nat-1.4.0a20251112.dist-info}/WHEEL +0 -0
- {nvidia_nat-1.4.0a20251102.dist-info → nvidia_nat-1.4.0a20251112.dist-info}/entry_points.txt +0 -0
- {nvidia_nat-1.4.0a20251102.dist-info → nvidia_nat-1.4.0a20251112.dist-info}/licenses/LICENSE-3rd-party.txt +0 -0
- {nvidia_nat-1.4.0a20251102.dist-info → nvidia_nat-1.4.0a20251112.dist-info}/licenses/LICENSE.md +0 -0
- {nvidia_nat-1.4.0a20251102.dist-info → nvidia_nat-1.4.0a20251112.dist-info}/top_level.txt +0 -0
|
@@ -1,414 +0,0 @@
|
|
|
1
|
-
# SPDX-FileCopyrightText: Copyright (c) 2025, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
|
|
2
|
-
# SPDX-License-Identifier: Apache-2.0
|
|
3
|
-
#
|
|
4
|
-
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
5
|
-
# you may not use this file except in compliance with the License.
|
|
6
|
-
# You may obtain a copy of the License at
|
|
7
|
-
#
|
|
8
|
-
# http://www.apache.org/licenses/LICENSE-2.0
|
|
9
|
-
#
|
|
10
|
-
# Unless required by applicable law or agreed to in writing, software
|
|
11
|
-
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
12
|
-
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
13
|
-
# See the License for the specific language governing permissions and
|
|
14
|
-
# limitations under the License.
|
|
15
|
-
"""
|
|
16
|
-
Test suite for Code Execution Sandbox using pytest.
|
|
17
|
-
|
|
18
|
-
This module provides comprehensive testing for the code execution sandbox service,
|
|
19
|
-
replacing the original bash script with a more maintainable Python implementation.
|
|
20
|
-
"""
|
|
21
|
-
|
|
22
|
-
import os
|
|
23
|
-
from typing import Any
|
|
24
|
-
|
|
25
|
-
import pytest
|
|
26
|
-
import requests
|
|
27
|
-
from requests.exceptions import ConnectionError
|
|
28
|
-
from requests.exceptions import RequestException
|
|
29
|
-
from requests.exceptions import Timeout
|
|
30
|
-
|
|
31
|
-
|
|
32
|
-
class TestCodeExecutionSandbox:
|
|
33
|
-
"""Test suite for the Code Execution Sandbox service."""
|
|
34
|
-
|
|
35
|
-
@pytest.fixture(scope="class")
|
|
36
|
-
def sandbox_config(self):
|
|
37
|
-
"""Configuration for sandbox testing."""
|
|
38
|
-
return {
|
|
39
|
-
"url": os.environ.get("SANDBOX_URL", "http://127.0.0.1:6000/execute"),
|
|
40
|
-
"timeout": int(os.environ.get("SANDBOX_TIMEOUT", "30")),
|
|
41
|
-
"connection_timeout": 5
|
|
42
|
-
}
|
|
43
|
-
|
|
44
|
-
@pytest.fixture(scope="class", autouse=True)
|
|
45
|
-
def check_sandbox_running(self, sandbox_config):
|
|
46
|
-
"""Check if sandbox server is running before running tests."""
|
|
47
|
-
try:
|
|
48
|
-
_ = requests.get(sandbox_config["url"], timeout=sandbox_config["connection_timeout"])
|
|
49
|
-
print(f"✓ Sandbox server is running at {sandbox_config['url']}")
|
|
50
|
-
except (ConnectionError, Timeout, RequestException):
|
|
51
|
-
pytest.skip(
|
|
52
|
-
f"Sandbox server is not running at {sandbox_config['url']}. "
|
|
53
|
-
"Please start it with: cd src/nat/tool/code_execution/local_sandbox && ./start_local_sandbox.sh")
|
|
54
|
-
|
|
55
|
-
def execute_code(self, sandbox_config: dict[str, Any], code: str, language: str = "python") -> dict[str, Any]:
|
|
56
|
-
"""
|
|
57
|
-
Execute code in the sandbox and return the response.
|
|
58
|
-
|
|
59
|
-
Args:
|
|
60
|
-
sandbox_config: Configuration dictionary
|
|
61
|
-
code: Code to execute
|
|
62
|
-
language: Programming language (default: python)
|
|
63
|
-
|
|
64
|
-
Returns:
|
|
65
|
-
dictionary containing the response from the sandbox
|
|
66
|
-
"""
|
|
67
|
-
payload = {"generated_code": code, "timeout": sandbox_config["timeout"], "language": language}
|
|
68
|
-
|
|
69
|
-
response = requests.post(
|
|
70
|
-
sandbox_config["url"],
|
|
71
|
-
json=payload,
|
|
72
|
-
timeout=sandbox_config["timeout"] + 5 # Add buffer to request timeout
|
|
73
|
-
)
|
|
74
|
-
|
|
75
|
-
# Ensure we got a response
|
|
76
|
-
response.raise_for_status()
|
|
77
|
-
return response.json()
|
|
78
|
-
|
|
79
|
-
def test_simple_print(self, sandbox_config):
|
|
80
|
-
"""Test simple print statement execution."""
|
|
81
|
-
code = "print('Hello, World!')"
|
|
82
|
-
result = self.execute_code(sandbox_config, code)
|
|
83
|
-
|
|
84
|
-
assert result["process_status"] == "completed"
|
|
85
|
-
assert "Hello, World!" in result["stdout"]
|
|
86
|
-
assert result["stderr"] == ""
|
|
87
|
-
|
|
88
|
-
def test_basic_arithmetic(self, sandbox_config):
|
|
89
|
-
"""Test basic arithmetic operations."""
|
|
90
|
-
code = """
|
|
91
|
-
result = 2 + 3
|
|
92
|
-
print(f'Result: {result}')
|
|
93
|
-
"""
|
|
94
|
-
result = self.execute_code(sandbox_config, code)
|
|
95
|
-
|
|
96
|
-
assert result["process_status"] == "completed"
|
|
97
|
-
assert "Result: 5" in result["stdout"]
|
|
98
|
-
assert result["stderr"] == ""
|
|
99
|
-
|
|
100
|
-
def test_numpy_operations(self, sandbox_config):
|
|
101
|
-
"""Test numpy dependency availability and operations."""
|
|
102
|
-
code = """
|
|
103
|
-
import numpy as np
|
|
104
|
-
arr = np.array([1, 2, 3, 4, 5])
|
|
105
|
-
print(f'Array: {arr}')
|
|
106
|
-
print(f'Mean: {np.mean(arr)}')
|
|
107
|
-
"""
|
|
108
|
-
result = self.execute_code(sandbox_config, code)
|
|
109
|
-
|
|
110
|
-
assert result["process_status"] == "completed"
|
|
111
|
-
assert "Array: [1 2 3 4 5]" in result["stdout"]
|
|
112
|
-
assert "Mean: 3.0" in result["stdout"]
|
|
113
|
-
assert result["stderr"] == ""
|
|
114
|
-
|
|
115
|
-
def test_pandas_operations(self, sandbox_config):
|
|
116
|
-
"""Test pandas dependency availability and operations."""
|
|
117
|
-
code = """
|
|
118
|
-
import pandas as pd
|
|
119
|
-
df = pd.DataFrame({'A': [1, 2, 3], 'B': [4, 5, 6]})
|
|
120
|
-
print(df)
|
|
121
|
-
print(f'Sum of column A: {df["A"].sum()}')
|
|
122
|
-
"""
|
|
123
|
-
result = self.execute_code(sandbox_config, code)
|
|
124
|
-
|
|
125
|
-
assert result["process_status"] == "completed"
|
|
126
|
-
assert "Sum of column A: 6" in result["stdout"]
|
|
127
|
-
assert result["stderr"] == ""
|
|
128
|
-
|
|
129
|
-
def test_plotly_import(self, sandbox_config):
|
|
130
|
-
"""Test plotly dependency availability."""
|
|
131
|
-
code = """
|
|
132
|
-
import plotly.graph_objects as go
|
|
133
|
-
print('Plotly imported successfully')
|
|
134
|
-
fig = go.Figure()
|
|
135
|
-
fig.add_trace(go.Scatter(x=[1, 2, 3], y=[4, 5, 6]))
|
|
136
|
-
print('Plot created successfully')
|
|
137
|
-
"""
|
|
138
|
-
result = self.execute_code(sandbox_config, code)
|
|
139
|
-
|
|
140
|
-
assert result["process_status"] == "completed"
|
|
141
|
-
assert "Plotly imported successfully" in result["stdout"]
|
|
142
|
-
assert "Plot created successfully" in result["stdout"]
|
|
143
|
-
assert result["stderr"] == ""
|
|
144
|
-
|
|
145
|
-
def test_syntax_error_handling(self, sandbox_config):
|
|
146
|
-
"""Test handling of syntax errors."""
|
|
147
|
-
code = """
|
|
148
|
-
print('Hello World'
|
|
149
|
-
# Missing closing parenthesis
|
|
150
|
-
"""
|
|
151
|
-
result = self.execute_code(sandbox_config, code)
|
|
152
|
-
|
|
153
|
-
assert result["process_status"] == "error"
|
|
154
|
-
assert "SyntaxError" in result["stderr"] or "SyntaxError" in result["stdout"]
|
|
155
|
-
|
|
156
|
-
def test_runtime_error_handling(self, sandbox_config):
|
|
157
|
-
"""Test handling of runtime errors."""
|
|
158
|
-
code = """
|
|
159
|
-
x = 1 / 0
|
|
160
|
-
print('This should not print')
|
|
161
|
-
"""
|
|
162
|
-
result = self.execute_code(sandbox_config, code)
|
|
163
|
-
|
|
164
|
-
assert result["process_status"] == "error"
|
|
165
|
-
assert "ZeroDivisionError" in result["stderr"] or "ZeroDivisionError" in result["stdout"]
|
|
166
|
-
|
|
167
|
-
def test_import_error_handling(self, sandbox_config):
|
|
168
|
-
"""Test handling of import errors."""
|
|
169
|
-
code = """
|
|
170
|
-
import nonexistent_module
|
|
171
|
-
print('This should not print')
|
|
172
|
-
"""
|
|
173
|
-
result = self.execute_code(sandbox_config, code)
|
|
174
|
-
|
|
175
|
-
assert result["process_status"] == "error"
|
|
176
|
-
assert "ModuleNotFoundError" in result["stderr"] or "ImportError" in result["stderr"]
|
|
177
|
-
|
|
178
|
-
def test_mixed_output(self, sandbox_config):
|
|
179
|
-
"""Test code that produces both stdout and stderr output."""
|
|
180
|
-
code = """
|
|
181
|
-
import sys
|
|
182
|
-
print('This goes to stdout')
|
|
183
|
-
print('This goes to stderr', file=sys.stderr)
|
|
184
|
-
print('Back to stdout')
|
|
185
|
-
"""
|
|
186
|
-
result = self.execute_code(sandbox_config, code)
|
|
187
|
-
|
|
188
|
-
assert result["process_status"] == "completed"
|
|
189
|
-
assert "This goes to stdout" in result["stdout"]
|
|
190
|
-
assert "Back to stdout" in result["stdout"]
|
|
191
|
-
assert "This goes to stderr" in result["stderr"]
|
|
192
|
-
|
|
193
|
-
def test_long_running_code(self, sandbox_config):
|
|
194
|
-
"""Test code that takes some time to execute but completes within timeout."""
|
|
195
|
-
code = """
|
|
196
|
-
import time
|
|
197
|
-
for i in range(3):
|
|
198
|
-
print(f'Iteration {i}')
|
|
199
|
-
time.sleep(0.5)
|
|
200
|
-
print('Completed')
|
|
201
|
-
"""
|
|
202
|
-
result = self.execute_code(sandbox_config, code)
|
|
203
|
-
|
|
204
|
-
assert result["process_status"] == "completed"
|
|
205
|
-
assert "Iteration 0" in result["stdout"]
|
|
206
|
-
assert "Iteration 1" in result["stdout"]
|
|
207
|
-
assert "Iteration 2" in result["stdout"]
|
|
208
|
-
assert "Completed" in result["stdout"]
|
|
209
|
-
assert result["stderr"] == ""
|
|
210
|
-
|
|
211
|
-
def test_file_operations(self, sandbox_config):
|
|
212
|
-
"""Test basic file operations in the sandbox."""
|
|
213
|
-
code = """
|
|
214
|
-
import os
|
|
215
|
-
print(f'Current directory: {os.getcwd()}')
|
|
216
|
-
with open('test_file.txt', 'w') as f:
|
|
217
|
-
f.write('Hello, World!')
|
|
218
|
-
with open('test_file.txt', 'r') as f:
|
|
219
|
-
content = f.read()
|
|
220
|
-
print(f'File content: {content}')
|
|
221
|
-
os.remove('test_file.txt')
|
|
222
|
-
print('File operations completed')
|
|
223
|
-
"""
|
|
224
|
-
result = self.execute_code(sandbox_config, code)
|
|
225
|
-
|
|
226
|
-
assert result["process_status"] == "completed"
|
|
227
|
-
assert "File content: Hello, World!" in result["stdout"]
|
|
228
|
-
assert "File operations completed" in result["stdout"]
|
|
229
|
-
assert result["stderr"] == ""
|
|
230
|
-
|
|
231
|
-
def test_file_persistence_create(self, sandbox_config):
|
|
232
|
-
"""Test file persistence - create various file types."""
|
|
233
|
-
code = """
|
|
234
|
-
import os
|
|
235
|
-
import pandas as pd
|
|
236
|
-
import numpy as np
|
|
237
|
-
print('Current directory:', os.getcwd())
|
|
238
|
-
print('Directory contents:', os.listdir('.'))
|
|
239
|
-
|
|
240
|
-
# Create a test file
|
|
241
|
-
with open('persistence_test.txt', 'w') as f:
|
|
242
|
-
f.write('Hello from sandbox persistence test!')
|
|
243
|
-
|
|
244
|
-
# Create a CSV file
|
|
245
|
-
df = pd.DataFrame({'A': [1, 2, 3], 'B': [4, 5, 6]})
|
|
246
|
-
df.to_csv('persistence_test.csv', index=False)
|
|
247
|
-
|
|
248
|
-
# Create a numpy array file
|
|
249
|
-
arr = np.array([1, 2, 3, 4, 5])
|
|
250
|
-
np.save('persistence_test.npy', arr)
|
|
251
|
-
|
|
252
|
-
print('Files created:')
|
|
253
|
-
for file in os.listdir('.'):
|
|
254
|
-
if 'persistence_test' in file:
|
|
255
|
-
print(' -', file)
|
|
256
|
-
"""
|
|
257
|
-
result = self.execute_code(sandbox_config, code)
|
|
258
|
-
|
|
259
|
-
assert result["process_status"] == "completed"
|
|
260
|
-
assert "persistence_test.txt" in result["stdout"]
|
|
261
|
-
assert "persistence_test.csv" in result["stdout"]
|
|
262
|
-
assert "persistence_test.npy" in result["stdout"]
|
|
263
|
-
assert result["stderr"] == ""
|
|
264
|
-
|
|
265
|
-
def test_file_persistence_read(self, sandbox_config):
|
|
266
|
-
"""Test file persistence - read back created files."""
|
|
267
|
-
code = """
|
|
268
|
-
import pandas as pd
|
|
269
|
-
import numpy as np
|
|
270
|
-
|
|
271
|
-
# Read back the files we created
|
|
272
|
-
print('=== Reading persistence_test.txt ===')
|
|
273
|
-
with open('persistence_test.txt', 'r') as f:
|
|
274
|
-
content = f.read()
|
|
275
|
-
print(f'Content: {content}')
|
|
276
|
-
|
|
277
|
-
print('\\n=== Reading persistence_test.csv ===')
|
|
278
|
-
df = pd.read_csv('persistence_test.csv')
|
|
279
|
-
print(df)
|
|
280
|
-
print(f'DataFrame shape: {df.shape}')
|
|
281
|
-
|
|
282
|
-
print('\\n=== Reading persistence_test.npy ===')
|
|
283
|
-
arr = np.load('persistence_test.npy')
|
|
284
|
-
print(f'Array: {arr}')
|
|
285
|
-
print(f'Array sum: {np.sum(arr)}')
|
|
286
|
-
|
|
287
|
-
print('\\n=== File persistence test PASSED! ===')
|
|
288
|
-
"""
|
|
289
|
-
result = self.execute_code(sandbox_config, code)
|
|
290
|
-
|
|
291
|
-
assert result["process_status"] == "completed"
|
|
292
|
-
assert "Content: Hello from sandbox persistence test!" in result["stdout"]
|
|
293
|
-
assert "DataFrame shape: (3, 2)" in result["stdout"]
|
|
294
|
-
assert "Array: [1 2 3 4 5]" in result["stdout"]
|
|
295
|
-
assert "Array sum: 15" in result["stdout"]
|
|
296
|
-
assert "File persistence test PASSED!" in result["stdout"]
|
|
297
|
-
assert result["stderr"] == ""
|
|
298
|
-
|
|
299
|
-
def test_json_operations(self, sandbox_config):
|
|
300
|
-
"""Test JSON file operations for persistence."""
|
|
301
|
-
code = """
|
|
302
|
-
import json
|
|
303
|
-
import os
|
|
304
|
-
|
|
305
|
-
# Create a complex JSON file
|
|
306
|
-
data = {
|
|
307
|
-
'test_name': 'sandbox_persistence',
|
|
308
|
-
'timestamp': '2024-07-03',
|
|
309
|
-
'results': {
|
|
310
|
-
'numpy_test': True,
|
|
311
|
-
'pandas_test': True,
|
|
312
|
-
'file_operations': True
|
|
313
|
-
},
|
|
314
|
-
'metrics': [1.5, 2.3, 3.7, 4.1],
|
|
315
|
-
'metadata': {
|
|
316
|
-
'working_dir': os.getcwd(),
|
|
317
|
-
'python_version': '3.x'
|
|
318
|
-
}
|
|
319
|
-
}
|
|
320
|
-
|
|
321
|
-
# Save JSON file
|
|
322
|
-
with open('persistence_test.json', 'w') as f:
|
|
323
|
-
json.dump(data, f, indent=2)
|
|
324
|
-
|
|
325
|
-
# Read it back
|
|
326
|
-
with open('persistence_test.json', 'r') as f:
|
|
327
|
-
loaded_data = json.load(f)
|
|
328
|
-
|
|
329
|
-
print('JSON file created and loaded successfully')
|
|
330
|
-
print(f'Test name: {loaded_data["test_name"]}')
|
|
331
|
-
print(f'Results count: {len(loaded_data["results"])}')
|
|
332
|
-
print(f'Metrics: {loaded_data["metrics"]}')
|
|
333
|
-
print('JSON persistence test completed!')
|
|
334
|
-
"""
|
|
335
|
-
result = self.execute_code(sandbox_config, code)
|
|
336
|
-
|
|
337
|
-
assert result["process_status"] == "completed"
|
|
338
|
-
assert "JSON file created and loaded successfully" in result["stdout"]
|
|
339
|
-
assert "Test name: sandbox_persistence" in result["stdout"]
|
|
340
|
-
assert "Results count: 3" in result["stdout"]
|
|
341
|
-
assert "JSON persistence test completed!" in result["stdout"]
|
|
342
|
-
assert result["stderr"] == ""
|
|
343
|
-
|
|
344
|
-
def test_missing_generated_code_field(self, sandbox_config):
|
|
345
|
-
"""Test request missing the generated_code field."""
|
|
346
|
-
payload = {"timeout": 10, "language": "python"}
|
|
347
|
-
|
|
348
|
-
response = requests.post(sandbox_config["url"], json=payload)
|
|
349
|
-
|
|
350
|
-
# Should return an error status code or error in response
|
|
351
|
-
assert response.status_code != 200 or "error" in response.json()
|
|
352
|
-
|
|
353
|
-
def test_missing_timeout_field(self, sandbox_config):
|
|
354
|
-
"""Test request missing the timeout field."""
|
|
355
|
-
payload = {"generated_code": "print('test')", "language": "python"}
|
|
356
|
-
|
|
357
|
-
response = requests.post(sandbox_config["url"], json=payload)
|
|
358
|
-
|
|
359
|
-
# Should return error for missing timeout field
|
|
360
|
-
result = response.json()
|
|
361
|
-
assert response.status_code == 400 and result["process_status"] == "error"
|
|
362
|
-
|
|
363
|
-
def test_invalid_json(self, sandbox_config):
|
|
364
|
-
"""Test request with invalid JSON."""
|
|
365
|
-
invalid_json = '{"generated_code": "print("test")", "timeout": 10}'
|
|
366
|
-
|
|
367
|
-
response = requests.post(sandbox_config["url"], data=invalid_json, headers={"Content-Type": "application/json"})
|
|
368
|
-
|
|
369
|
-
# Should return error for invalid JSON
|
|
370
|
-
assert response.status_code != 200
|
|
371
|
-
|
|
372
|
-
def test_non_json_request(self, sandbox_config):
|
|
373
|
-
"""Test request with non-JSON content."""
|
|
374
|
-
response = requests.post(sandbox_config["url"], data="This is not JSON", headers={"Content-Type": "text/plain"})
|
|
375
|
-
|
|
376
|
-
# Should return error for non-JSON content
|
|
377
|
-
assert response.status_code != 200
|
|
378
|
-
|
|
379
|
-
def test_timeout_too_low(self, sandbox_config):
|
|
380
|
-
"""Test request with timeout too low."""
|
|
381
|
-
code = """
|
|
382
|
-
import time
|
|
383
|
-
time.sleep(2.0)
|
|
384
|
-
"""
|
|
385
|
-
payload = {"generated_code": code, "timeout": 1, "language": "python"}
|
|
386
|
-
response = requests.post(sandbox_config["url"], json=payload)
|
|
387
|
-
assert response.json()["process_status"] == "timeout"
|
|
388
|
-
assert response.status_code == 200
|
|
389
|
-
|
|
390
|
-
|
|
391
|
-
# Pytest configuration and fixtures for command-line options
|
|
392
|
-
def pytest_addoption(parser):
|
|
393
|
-
"""Add custom command-line options for pytest."""
|
|
394
|
-
parser.addoption("--sandbox-url",
|
|
395
|
-
action="store",
|
|
396
|
-
default="http://127.0.0.1:6000/execute",
|
|
397
|
-
help="Sandbox URL for testing")
|
|
398
|
-
parser.addoption("--sandbox-timeout",
|
|
399
|
-
action="store",
|
|
400
|
-
type=int,
|
|
401
|
-
default=30,
|
|
402
|
-
help="Timeout in seconds for sandbox operations")
|
|
403
|
-
|
|
404
|
-
|
|
405
|
-
@pytest.fixture(scope="session", autouse=True)
|
|
406
|
-
def setup_environment(request):
|
|
407
|
-
"""Setup environment variables from command-line options."""
|
|
408
|
-
os.environ["SANDBOX_URL"] = request.config.getoption("--sandbox-url", "http://127.0.0.1:6000/execute")
|
|
409
|
-
os.environ["SANDBOX_TIMEOUT"] = str(request.config.getoption("--sandbox-timeout", 30))
|
|
410
|
-
|
|
411
|
-
|
|
412
|
-
if __name__ == "__main__":
|
|
413
|
-
# Allow running as a script
|
|
414
|
-
pytest.main([__file__, "-v"])
|
|
File without changes
|
{nvidia_nat-1.4.0a20251102.dist-info → nvidia_nat-1.4.0a20251112.dist-info}/entry_points.txt
RENAMED
|
File without changes
|
|
File without changes
|
{nvidia_nat-1.4.0a20251102.dist-info → nvidia_nat-1.4.0a20251112.dist-info}/licenses/LICENSE.md
RENAMED
|
File without changes
|
|
File without changes
|