llama-cloud 0.1.37__py3-none-any.whl → 0.1.38__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of llama-cloud might be problematic. Click here for more details.
- llama_cloud/__init__.py +8 -2
- llama_cloud/client.py +3 -0
- llama_cloud/resources/__init__.py +2 -0
- llama_cloud/resources/alpha/__init__.py +2 -0
- llama_cloud/resources/alpha/client.py +118 -0
- llama_cloud/resources/beta/client.py +126 -30
- llama_cloud/resources/chat_apps/client.py +32 -8
- llama_cloud/resources/classifier/client.py +139 -11
- llama_cloud/resources/data_sinks/client.py +32 -8
- llama_cloud/resources/data_sources/client.py +32 -8
- llama_cloud/resources/data_sources/types/data_source_update_component.py +2 -0
- llama_cloud/resources/embedding_model_configs/client.py +48 -12
- llama_cloud/resources/files/client.py +176 -42
- llama_cloud/resources/jobs/client.py +12 -6
- llama_cloud/resources/llama_extract/client.py +138 -32
- llama_cloud/resources/organizations/client.py +18 -4
- llama_cloud/resources/parsing/client.py +16 -4
- llama_cloud/resources/pipelines/client.py +32 -8
- llama_cloud/resources/projects/client.py +78 -18
- llama_cloud/resources/reports/client.py +126 -30
- llama_cloud/resources/retrievers/client.py +48 -12
- llama_cloud/types/__init__.py +6 -2
- llama_cloud/types/agent_deployment_summary.py +1 -0
- llama_cloud/types/classify_job.py +2 -0
- llama_cloud/types/cloud_jira_data_source_v_2.py +52 -0
- llama_cloud/types/cloud_jira_data_source_v_2_api_version.py +21 -0
- llama_cloud/types/configurable_data_source_names.py +4 -0
- llama_cloud/types/data_source_component.py +2 -0
- llama_cloud/types/data_source_create_component.py +2 -0
- llama_cloud/types/data_source_reader_version_metadata_reader_version.py +9 -1
- llama_cloud/types/{classify_job_with_status.py → paginated_response_classify_job.py} +5 -18
- llama_cloud/types/pipeline_data_source_component.py +2 -0
- llama_cloud/types/usage_response_active_alerts_item.py +4 -0
- {llama_cloud-0.1.37.dist-info → llama_cloud-0.1.38.dist-info}/METADATA +2 -3
- {llama_cloud-0.1.37.dist-info → llama_cloud-0.1.38.dist-info}/RECORD +37 -33
- {llama_cloud-0.1.37.dist-info → llama_cloud-0.1.38.dist-info}/WHEEL +1 -1
- {llama_cloud-0.1.37.dist-info → llama_cloud-0.1.38.dist-info}/LICENSE +0 -0
|
@@ -39,6 +39,7 @@ class RetrieversClient:
|
|
|
39
39
|
name: typing.Optional[str] = None,
|
|
40
40
|
project_id: typing.Optional[str] = None,
|
|
41
41
|
organization_id: typing.Optional[str] = None,
|
|
42
|
+
project_id: typing.Optional[str] = None,
|
|
42
43
|
) -> typing.List[Retriever]:
|
|
43
44
|
"""
|
|
44
45
|
List Retrievers for a project.
|
|
@@ -49,6 +50,8 @@ class RetrieversClient:
|
|
|
49
50
|
- project_id: typing.Optional[str].
|
|
50
51
|
|
|
51
52
|
- organization_id: typing.Optional[str].
|
|
53
|
+
|
|
54
|
+
- project_id: typing.Optional[str].
|
|
52
55
|
---
|
|
53
56
|
from llama_cloud.client import LlamaCloud
|
|
54
57
|
|
|
@@ -61,7 +64,7 @@ class RetrieversClient:
|
|
|
61
64
|
"GET",
|
|
62
65
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/retrievers"),
|
|
63
66
|
params=remove_none_from_dict({"name": name, "project_id": project_id, "organization_id": organization_id}),
|
|
64
|
-
headers=self._client_wrapper.get_headers(),
|
|
67
|
+
headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
|
|
65
68
|
timeout=60,
|
|
66
69
|
)
|
|
67
70
|
if 200 <= _response.status_code < 300:
|
|
@@ -80,6 +83,7 @@ class RetrieversClient:
|
|
|
80
83
|
project_id: typing.Optional[str] = None,
|
|
81
84
|
organization_id: typing.Optional[str] = None,
|
|
82
85
|
request: RetrieverCreate,
|
|
86
|
+
project_id: typing.Optional[str] = None,
|
|
83
87
|
) -> Retriever:
|
|
84
88
|
"""
|
|
85
89
|
Create a new Retriever.
|
|
@@ -90,6 +94,8 @@ class RetrieversClient:
|
|
|
90
94
|
- organization_id: typing.Optional[str].
|
|
91
95
|
|
|
92
96
|
- request: RetrieverCreate.
|
|
97
|
+
|
|
98
|
+
- project_id: typing.Optional[str].
|
|
93
99
|
---
|
|
94
100
|
from llama_cloud import RetrieverCreate
|
|
95
101
|
from llama_cloud.client import LlamaCloud
|
|
@@ -108,7 +114,7 @@ class RetrieversClient:
|
|
|
108
114
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/retrievers"),
|
|
109
115
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
110
116
|
json=jsonable_encoder(request),
|
|
111
|
-
headers=self._client_wrapper.get_headers(),
|
|
117
|
+
headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
|
|
112
118
|
timeout=60,
|
|
113
119
|
)
|
|
114
120
|
if 200 <= _response.status_code < 300:
|
|
@@ -127,6 +133,7 @@ class RetrieversClient:
|
|
|
127
133
|
project_id: typing.Optional[str] = None,
|
|
128
134
|
organization_id: typing.Optional[str] = None,
|
|
129
135
|
request: RetrieverCreate,
|
|
136
|
+
project_id: typing.Optional[str] = None,
|
|
130
137
|
) -> Retriever:
|
|
131
138
|
"""
|
|
132
139
|
Upsert a new Retriever.
|
|
@@ -137,6 +144,8 @@ class RetrieversClient:
|
|
|
137
144
|
- organization_id: typing.Optional[str].
|
|
138
145
|
|
|
139
146
|
- request: RetrieverCreate.
|
|
147
|
+
|
|
148
|
+
- project_id: typing.Optional[str].
|
|
140
149
|
---
|
|
141
150
|
from llama_cloud import RetrieverCreate
|
|
142
151
|
from llama_cloud.client import LlamaCloud
|
|
@@ -155,7 +164,7 @@ class RetrieversClient:
|
|
|
155
164
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/retrievers"),
|
|
156
165
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
157
166
|
json=jsonable_encoder(request),
|
|
158
|
-
headers=self._client_wrapper.get_headers(),
|
|
167
|
+
headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
|
|
159
168
|
timeout=60,
|
|
160
169
|
)
|
|
161
170
|
if 200 <= _response.status_code < 300:
|
|
@@ -174,6 +183,7 @@ class RetrieversClient:
|
|
|
174
183
|
*,
|
|
175
184
|
project_id: typing.Optional[str] = None,
|
|
176
185
|
organization_id: typing.Optional[str] = None,
|
|
186
|
+
project_id: typing.Optional[str] = None,
|
|
177
187
|
) -> Retriever:
|
|
178
188
|
"""
|
|
179
189
|
Get a Retriever by ID.
|
|
@@ -184,6 +194,8 @@ class RetrieversClient:
|
|
|
184
194
|
- project_id: typing.Optional[str].
|
|
185
195
|
|
|
186
196
|
- organization_id: typing.Optional[str].
|
|
197
|
+
|
|
198
|
+
- project_id: typing.Optional[str].
|
|
187
199
|
---
|
|
188
200
|
from llama_cloud.client import LlamaCloud
|
|
189
201
|
|
|
@@ -198,7 +210,7 @@ class RetrieversClient:
|
|
|
198
210
|
"GET",
|
|
199
211
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", f"api/v1/retrievers/{retriever_id}"),
|
|
200
212
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
201
|
-
headers=self._client_wrapper.get_headers(),
|
|
213
|
+
headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
|
|
202
214
|
timeout=60,
|
|
203
215
|
)
|
|
204
216
|
if 200 <= _response.status_code < 300:
|
|
@@ -301,6 +313,7 @@ class RetrieversClient:
|
|
|
301
313
|
rerank_top_n: typing.Optional[int] = OMIT,
|
|
302
314
|
rerank_config: typing.Optional[ReRankConfig] = OMIT,
|
|
303
315
|
query: str,
|
|
316
|
+
project_id: typing.Optional[str] = None,
|
|
304
317
|
) -> CompositeRetrievalResult:
|
|
305
318
|
"""
|
|
306
319
|
Retrieve data using a Retriever.
|
|
@@ -319,6 +332,8 @@ class RetrieversClient:
|
|
|
319
332
|
- rerank_config: typing.Optional[ReRankConfig]. The rerank configuration for composite retrieval.
|
|
320
333
|
|
|
321
334
|
- query: str. The query to retrieve against.
|
|
335
|
+
|
|
336
|
+
- project_id: typing.Optional[str].
|
|
322
337
|
---
|
|
323
338
|
from llama_cloud import CompositeRetrievalMode, ReRankConfig, ReRankerType
|
|
324
339
|
from llama_cloud.client import LlamaCloud
|
|
@@ -349,7 +364,7 @@ class RetrieversClient:
|
|
|
349
364
|
),
|
|
350
365
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
351
366
|
json=jsonable_encoder(_request),
|
|
352
|
-
headers=self._client_wrapper.get_headers(),
|
|
367
|
+
headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
|
|
353
368
|
timeout=60,
|
|
354
369
|
)
|
|
355
370
|
if 200 <= _response.status_code < 300:
|
|
@@ -372,6 +387,7 @@ class RetrieversClient:
|
|
|
372
387
|
rerank_config: typing.Optional[ReRankConfig] = OMIT,
|
|
373
388
|
query: str,
|
|
374
389
|
pipelines: typing.Optional[typing.List[RetrieverPipeline]] = OMIT,
|
|
390
|
+
project_id: typing.Optional[str] = None,
|
|
375
391
|
) -> CompositeRetrievalResult:
|
|
376
392
|
"""
|
|
377
393
|
Retrieve data using specified pipelines without creating a persistent retriever.
|
|
@@ -390,6 +406,8 @@ class RetrieversClient:
|
|
|
390
406
|
- query: str. The query to retrieve against.
|
|
391
407
|
|
|
392
408
|
- pipelines: typing.Optional[typing.List[RetrieverPipeline]]. The pipelines to use for retrieval.
|
|
409
|
+
|
|
410
|
+
- project_id: typing.Optional[str].
|
|
393
411
|
---
|
|
394
412
|
from llama_cloud import CompositeRetrievalMode, ReRankConfig, ReRankerType
|
|
395
413
|
from llama_cloud.client import LlamaCloud
|
|
@@ -419,7 +437,7 @@ class RetrieversClient:
|
|
|
419
437
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/retrievers/retrieve"),
|
|
420
438
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
421
439
|
json=jsonable_encoder(_request),
|
|
422
|
-
headers=self._client_wrapper.get_headers(),
|
|
440
|
+
headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
|
|
423
441
|
timeout=60,
|
|
424
442
|
)
|
|
425
443
|
if 200 <= _response.status_code < 300:
|
|
@@ -443,6 +461,7 @@ class AsyncRetrieversClient:
|
|
|
443
461
|
name: typing.Optional[str] = None,
|
|
444
462
|
project_id: typing.Optional[str] = None,
|
|
445
463
|
organization_id: typing.Optional[str] = None,
|
|
464
|
+
project_id: typing.Optional[str] = None,
|
|
446
465
|
) -> typing.List[Retriever]:
|
|
447
466
|
"""
|
|
448
467
|
List Retrievers for a project.
|
|
@@ -453,6 +472,8 @@ class AsyncRetrieversClient:
|
|
|
453
472
|
- project_id: typing.Optional[str].
|
|
454
473
|
|
|
455
474
|
- organization_id: typing.Optional[str].
|
|
475
|
+
|
|
476
|
+
- project_id: typing.Optional[str].
|
|
456
477
|
---
|
|
457
478
|
from llama_cloud.client import AsyncLlamaCloud
|
|
458
479
|
|
|
@@ -465,7 +486,7 @@ class AsyncRetrieversClient:
|
|
|
465
486
|
"GET",
|
|
466
487
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/retrievers"),
|
|
467
488
|
params=remove_none_from_dict({"name": name, "project_id": project_id, "organization_id": organization_id}),
|
|
468
|
-
headers=self._client_wrapper.get_headers(),
|
|
489
|
+
headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
|
|
469
490
|
timeout=60,
|
|
470
491
|
)
|
|
471
492
|
if 200 <= _response.status_code < 300:
|
|
@@ -484,6 +505,7 @@ class AsyncRetrieversClient:
|
|
|
484
505
|
project_id: typing.Optional[str] = None,
|
|
485
506
|
organization_id: typing.Optional[str] = None,
|
|
486
507
|
request: RetrieverCreate,
|
|
508
|
+
project_id: typing.Optional[str] = None,
|
|
487
509
|
) -> Retriever:
|
|
488
510
|
"""
|
|
489
511
|
Create a new Retriever.
|
|
@@ -494,6 +516,8 @@ class AsyncRetrieversClient:
|
|
|
494
516
|
- organization_id: typing.Optional[str].
|
|
495
517
|
|
|
496
518
|
- request: RetrieverCreate.
|
|
519
|
+
|
|
520
|
+
- project_id: typing.Optional[str].
|
|
497
521
|
---
|
|
498
522
|
from llama_cloud import RetrieverCreate
|
|
499
523
|
from llama_cloud.client import AsyncLlamaCloud
|
|
@@ -512,7 +536,7 @@ class AsyncRetrieversClient:
|
|
|
512
536
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/retrievers"),
|
|
513
537
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
514
538
|
json=jsonable_encoder(request),
|
|
515
|
-
headers=self._client_wrapper.get_headers(),
|
|
539
|
+
headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
|
|
516
540
|
timeout=60,
|
|
517
541
|
)
|
|
518
542
|
if 200 <= _response.status_code < 300:
|
|
@@ -531,6 +555,7 @@ class AsyncRetrieversClient:
|
|
|
531
555
|
project_id: typing.Optional[str] = None,
|
|
532
556
|
organization_id: typing.Optional[str] = None,
|
|
533
557
|
request: RetrieverCreate,
|
|
558
|
+
project_id: typing.Optional[str] = None,
|
|
534
559
|
) -> Retriever:
|
|
535
560
|
"""
|
|
536
561
|
Upsert a new Retriever.
|
|
@@ -541,6 +566,8 @@ class AsyncRetrieversClient:
|
|
|
541
566
|
- organization_id: typing.Optional[str].
|
|
542
567
|
|
|
543
568
|
- request: RetrieverCreate.
|
|
569
|
+
|
|
570
|
+
- project_id: typing.Optional[str].
|
|
544
571
|
---
|
|
545
572
|
from llama_cloud import RetrieverCreate
|
|
546
573
|
from llama_cloud.client import AsyncLlamaCloud
|
|
@@ -559,7 +586,7 @@ class AsyncRetrieversClient:
|
|
|
559
586
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/retrievers"),
|
|
560
587
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
561
588
|
json=jsonable_encoder(request),
|
|
562
|
-
headers=self._client_wrapper.get_headers(),
|
|
589
|
+
headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
|
|
563
590
|
timeout=60,
|
|
564
591
|
)
|
|
565
592
|
if 200 <= _response.status_code < 300:
|
|
@@ -578,6 +605,7 @@ class AsyncRetrieversClient:
|
|
|
578
605
|
*,
|
|
579
606
|
project_id: typing.Optional[str] = None,
|
|
580
607
|
organization_id: typing.Optional[str] = None,
|
|
608
|
+
project_id: typing.Optional[str] = None,
|
|
581
609
|
) -> Retriever:
|
|
582
610
|
"""
|
|
583
611
|
Get a Retriever by ID.
|
|
@@ -588,6 +616,8 @@ class AsyncRetrieversClient:
|
|
|
588
616
|
- project_id: typing.Optional[str].
|
|
589
617
|
|
|
590
618
|
- organization_id: typing.Optional[str].
|
|
619
|
+
|
|
620
|
+
- project_id: typing.Optional[str].
|
|
591
621
|
---
|
|
592
622
|
from llama_cloud.client import AsyncLlamaCloud
|
|
593
623
|
|
|
@@ -602,7 +632,7 @@ class AsyncRetrieversClient:
|
|
|
602
632
|
"GET",
|
|
603
633
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", f"api/v1/retrievers/{retriever_id}"),
|
|
604
634
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
605
|
-
headers=self._client_wrapper.get_headers(),
|
|
635
|
+
headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
|
|
606
636
|
timeout=60,
|
|
607
637
|
)
|
|
608
638
|
if 200 <= _response.status_code < 300:
|
|
@@ -705,6 +735,7 @@ class AsyncRetrieversClient:
|
|
|
705
735
|
rerank_top_n: typing.Optional[int] = OMIT,
|
|
706
736
|
rerank_config: typing.Optional[ReRankConfig] = OMIT,
|
|
707
737
|
query: str,
|
|
738
|
+
project_id: typing.Optional[str] = None,
|
|
708
739
|
) -> CompositeRetrievalResult:
|
|
709
740
|
"""
|
|
710
741
|
Retrieve data using a Retriever.
|
|
@@ -723,6 +754,8 @@ class AsyncRetrieversClient:
|
|
|
723
754
|
- rerank_config: typing.Optional[ReRankConfig]. The rerank configuration for composite retrieval.
|
|
724
755
|
|
|
725
756
|
- query: str. The query to retrieve against.
|
|
757
|
+
|
|
758
|
+
- project_id: typing.Optional[str].
|
|
726
759
|
---
|
|
727
760
|
from llama_cloud import CompositeRetrievalMode, ReRankConfig, ReRankerType
|
|
728
761
|
from llama_cloud.client import AsyncLlamaCloud
|
|
@@ -753,7 +786,7 @@ class AsyncRetrieversClient:
|
|
|
753
786
|
),
|
|
754
787
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
755
788
|
json=jsonable_encoder(_request),
|
|
756
|
-
headers=self._client_wrapper.get_headers(),
|
|
789
|
+
headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
|
|
757
790
|
timeout=60,
|
|
758
791
|
)
|
|
759
792
|
if 200 <= _response.status_code < 300:
|
|
@@ -776,6 +809,7 @@ class AsyncRetrieversClient:
|
|
|
776
809
|
rerank_config: typing.Optional[ReRankConfig] = OMIT,
|
|
777
810
|
query: str,
|
|
778
811
|
pipelines: typing.Optional[typing.List[RetrieverPipeline]] = OMIT,
|
|
812
|
+
project_id: typing.Optional[str] = None,
|
|
779
813
|
) -> CompositeRetrievalResult:
|
|
780
814
|
"""
|
|
781
815
|
Retrieve data using specified pipelines without creating a persistent retriever.
|
|
@@ -794,6 +828,8 @@ class AsyncRetrieversClient:
|
|
|
794
828
|
- query: str. The query to retrieve against.
|
|
795
829
|
|
|
796
830
|
- pipelines: typing.Optional[typing.List[RetrieverPipeline]]. The pipelines to use for retrieval.
|
|
831
|
+
|
|
832
|
+
- project_id: typing.Optional[str].
|
|
797
833
|
---
|
|
798
834
|
from llama_cloud import CompositeRetrievalMode, ReRankConfig, ReRankerType
|
|
799
835
|
from llama_cloud.client import AsyncLlamaCloud
|
|
@@ -823,7 +859,7 @@ class AsyncRetrieversClient:
|
|
|
823
859
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/retrievers/retrieve"),
|
|
824
860
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
825
861
|
json=jsonable_encoder(_request),
|
|
826
|
-
headers=self._client_wrapper.get_headers(),
|
|
862
|
+
headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
|
|
827
863
|
timeout=60,
|
|
828
864
|
)
|
|
829
865
|
if 200 <= _response.status_code < 300:
|
llama_cloud/types/__init__.py
CHANGED
|
@@ -44,7 +44,6 @@ from .classification_result import ClassificationResult
|
|
|
44
44
|
from .classifier_rule import ClassifierRule
|
|
45
45
|
from .classify_job import ClassifyJob
|
|
46
46
|
from .classify_job_results import ClassifyJobResults
|
|
47
|
-
from .classify_job_with_status import ClassifyJobWithStatus
|
|
48
47
|
from .classify_parsing_configuration import ClassifyParsingConfiguration
|
|
49
48
|
from .cloud_astra_db_vector_store import CloudAstraDbVectorStore
|
|
50
49
|
from .cloud_az_storage_blob_data_source import CloudAzStorageBlobDataSource
|
|
@@ -54,6 +53,8 @@ from .cloud_confluence_data_source import CloudConfluenceDataSource
|
|
|
54
53
|
from .cloud_document import CloudDocument
|
|
55
54
|
from .cloud_document_create import CloudDocumentCreate
|
|
56
55
|
from .cloud_jira_data_source import CloudJiraDataSource
|
|
56
|
+
from .cloud_jira_data_source_v_2 import CloudJiraDataSourceV2
|
|
57
|
+
from .cloud_jira_data_source_v_2_api_version import CloudJiraDataSourceV2ApiVersion
|
|
57
58
|
from .cloud_milvus_vector_store import CloudMilvusVectorStore
|
|
58
59
|
from .cloud_mongo_db_atlas_vector_search import CloudMongoDbAtlasVectorSearch
|
|
59
60
|
from .cloud_notion_page_data_source import CloudNotionPageDataSource
|
|
@@ -246,6 +247,7 @@ from .paginated_list_pipeline_files_response import PaginatedListPipelineFilesRe
|
|
|
246
247
|
from .paginated_report_response import PaginatedReportResponse
|
|
247
248
|
from .paginated_response_agent_data import PaginatedResponseAgentData
|
|
248
249
|
from .paginated_response_aggregate_group import PaginatedResponseAggregateGroup
|
|
250
|
+
from .paginated_response_classify_job import PaginatedResponseClassifyJob
|
|
249
251
|
from .paginated_response_quota_configuration import PaginatedResponseQuotaConfiguration
|
|
250
252
|
from .parse_job_config import ParseJobConfig
|
|
251
253
|
from .parse_job_config_priority import ParseJobConfigPriority
|
|
@@ -437,7 +439,6 @@ __all__ = [
|
|
|
437
439
|
"ClassifierRule",
|
|
438
440
|
"ClassifyJob",
|
|
439
441
|
"ClassifyJobResults",
|
|
440
|
-
"ClassifyJobWithStatus",
|
|
441
442
|
"ClassifyParsingConfiguration",
|
|
442
443
|
"CloudAstraDbVectorStore",
|
|
443
444
|
"CloudAzStorageBlobDataSource",
|
|
@@ -447,6 +448,8 @@ __all__ = [
|
|
|
447
448
|
"CloudDocument",
|
|
448
449
|
"CloudDocumentCreate",
|
|
449
450
|
"CloudJiraDataSource",
|
|
451
|
+
"CloudJiraDataSourceV2",
|
|
452
|
+
"CloudJiraDataSourceV2ApiVersion",
|
|
450
453
|
"CloudMilvusVectorStore",
|
|
451
454
|
"CloudMongoDbAtlasVectorSearch",
|
|
452
455
|
"CloudNotionPageDataSource",
|
|
@@ -631,6 +634,7 @@ __all__ = [
|
|
|
631
634
|
"PaginatedReportResponse",
|
|
632
635
|
"PaginatedResponseAgentData",
|
|
633
636
|
"PaginatedResponseAggregateGroup",
|
|
637
|
+
"PaginatedResponseClassifyJob",
|
|
634
638
|
"PaginatedResponseQuotaConfiguration",
|
|
635
639
|
"ParseJobConfig",
|
|
636
640
|
"ParseJobConfigPriority",
|
|
@@ -23,6 +23,7 @@ class AgentDeploymentSummary(pydantic.BaseModel):
|
|
|
23
23
|
display_name: str = pydantic.Field(description="Display name of the deployed app")
|
|
24
24
|
created_at: dt.datetime = pydantic.Field(description="Timestamp when the app deployment was created")
|
|
25
25
|
updated_at: dt.datetime = pydantic.Field(description="Timestamp when the app deployment was last updated")
|
|
26
|
+
api_key_id: typing.Optional[str]
|
|
26
27
|
|
|
27
28
|
def json(self, **kwargs: typing.Any) -> str:
|
|
28
29
|
kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
|
|
@@ -6,6 +6,7 @@ import typing
|
|
|
6
6
|
from ..core.datetime_utils import serialize_datetime
|
|
7
7
|
from .classifier_rule import ClassifierRule
|
|
8
8
|
from .classify_parsing_configuration import ClassifyParsingConfiguration
|
|
9
|
+
from .status_enum import StatusEnum
|
|
9
10
|
|
|
10
11
|
try:
|
|
11
12
|
import pydantic
|
|
@@ -27,6 +28,7 @@ class ClassifyJob(pydantic.BaseModel):
|
|
|
27
28
|
rules: typing.List[ClassifierRule] = pydantic.Field(description="The rules to classify the files")
|
|
28
29
|
user_id: str = pydantic.Field(description="The ID of the user")
|
|
29
30
|
project_id: str = pydantic.Field(description="The ID of the project")
|
|
31
|
+
status: StatusEnum = pydantic.Field(description="The status of the classify job")
|
|
30
32
|
parsing_configuration: typing.Optional[ClassifyParsingConfiguration] = pydantic.Field(
|
|
31
33
|
description="The configuration for the parsing job"
|
|
32
34
|
)
|
|
@@ -0,0 +1,52 @@
|
|
|
1
|
+
# This file was auto-generated by Fern from our API Definition.
|
|
2
|
+
|
|
3
|
+
import datetime as dt
|
|
4
|
+
import typing
|
|
5
|
+
|
|
6
|
+
from ..core.datetime_utils import serialize_datetime
|
|
7
|
+
from .cloud_jira_data_source_v_2_api_version import CloudJiraDataSourceV2ApiVersion
|
|
8
|
+
|
|
9
|
+
try:
|
|
10
|
+
import pydantic
|
|
11
|
+
if pydantic.__version__.startswith("1."):
|
|
12
|
+
raise ImportError
|
|
13
|
+
import pydantic.v1 as pydantic # type: ignore
|
|
14
|
+
except ImportError:
|
|
15
|
+
import pydantic # type: ignore
|
|
16
|
+
|
|
17
|
+
|
|
18
|
+
class CloudJiraDataSourceV2(pydantic.BaseModel):
|
|
19
|
+
"""
|
|
20
|
+
Cloud Jira Data Source integrating JiraReaderV2.
|
|
21
|
+
"""
|
|
22
|
+
|
|
23
|
+
supports_access_control: typing.Optional[bool]
|
|
24
|
+
email: typing.Optional[str]
|
|
25
|
+
api_token: typing.Optional[str]
|
|
26
|
+
server_url: str = pydantic.Field(description="The server url for Jira Cloud.")
|
|
27
|
+
cloud_id: typing.Optional[str]
|
|
28
|
+
authentication_mechanism: str = pydantic.Field(description="Type of Authentication for connecting to Jira APIs.")
|
|
29
|
+
api_version: typing.Optional[CloudJiraDataSourceV2ApiVersion] = pydantic.Field(
|
|
30
|
+
description="Jira REST API version to use (2 or 3). 3 supports Atlassian Document Format (ADF)."
|
|
31
|
+
)
|
|
32
|
+
query: str = pydantic.Field(description="JQL (Jira Query Language) query to search.")
|
|
33
|
+
fields: typing.Optional[typing.List[str]]
|
|
34
|
+
expand: typing.Optional[str]
|
|
35
|
+
requests_per_minute: typing.Optional[int]
|
|
36
|
+
get_permissions: typing.Optional[bool] = pydantic.Field(
|
|
37
|
+
description="Whether to fetch project role permissions and issue-level security"
|
|
38
|
+
)
|
|
39
|
+
class_name: typing.Optional[str]
|
|
40
|
+
|
|
41
|
+
def json(self, **kwargs: typing.Any) -> str:
|
|
42
|
+
kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
|
|
43
|
+
return super().json(**kwargs_with_defaults)
|
|
44
|
+
|
|
45
|
+
def dict(self, **kwargs: typing.Any) -> typing.Dict[str, typing.Any]:
|
|
46
|
+
kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
|
|
47
|
+
return super().dict(**kwargs_with_defaults)
|
|
48
|
+
|
|
49
|
+
class Config:
|
|
50
|
+
frozen = True
|
|
51
|
+
smart_union = True
|
|
52
|
+
json_encoders = {dt.datetime: serialize_datetime}
|
|
@@ -0,0 +1,21 @@
|
|
|
1
|
+
# This file was auto-generated by Fern from our API Definition.
|
|
2
|
+
|
|
3
|
+
import enum
|
|
4
|
+
import typing
|
|
5
|
+
|
|
6
|
+
T_Result = typing.TypeVar("T_Result")
|
|
7
|
+
|
|
8
|
+
|
|
9
|
+
class CloudJiraDataSourceV2ApiVersion(str, enum.Enum):
|
|
10
|
+
"""
|
|
11
|
+
Jira REST API version to use (2 or 3). 3 supports Atlassian Document Format (ADF).
|
|
12
|
+
"""
|
|
13
|
+
|
|
14
|
+
TWO = "2"
|
|
15
|
+
THREE = "3"
|
|
16
|
+
|
|
17
|
+
def visit(self, two: typing.Callable[[], T_Result], three: typing.Callable[[], T_Result]) -> T_Result:
|
|
18
|
+
if self is CloudJiraDataSourceV2ApiVersion.TWO:
|
|
19
|
+
return two()
|
|
20
|
+
if self is CloudJiraDataSourceV2ApiVersion.THREE:
|
|
21
|
+
return three()
|
|
@@ -16,6 +16,7 @@ class ConfigurableDataSourceNames(str, enum.Enum):
|
|
|
16
16
|
NOTION_PAGE = "NOTION_PAGE"
|
|
17
17
|
CONFLUENCE = "CONFLUENCE"
|
|
18
18
|
JIRA = "JIRA"
|
|
19
|
+
JIRA_V_2 = "JIRA_V2"
|
|
19
20
|
BOX = "BOX"
|
|
20
21
|
|
|
21
22
|
def visit(
|
|
@@ -29,6 +30,7 @@ class ConfigurableDataSourceNames(str, enum.Enum):
|
|
|
29
30
|
notion_page: typing.Callable[[], T_Result],
|
|
30
31
|
confluence: typing.Callable[[], T_Result],
|
|
31
32
|
jira: typing.Callable[[], T_Result],
|
|
33
|
+
jira_v_2: typing.Callable[[], T_Result],
|
|
32
34
|
box: typing.Callable[[], T_Result],
|
|
33
35
|
) -> T_Result:
|
|
34
36
|
if self is ConfigurableDataSourceNames.S_3:
|
|
@@ -49,5 +51,7 @@ class ConfigurableDataSourceNames(str, enum.Enum):
|
|
|
49
51
|
return confluence()
|
|
50
52
|
if self is ConfigurableDataSourceNames.JIRA:
|
|
51
53
|
return jira()
|
|
54
|
+
if self is ConfigurableDataSourceNames.JIRA_V_2:
|
|
55
|
+
return jira_v_2()
|
|
52
56
|
if self is ConfigurableDataSourceNames.BOX:
|
|
53
57
|
return box()
|
|
@@ -6,6 +6,7 @@ from .cloud_az_storage_blob_data_source import CloudAzStorageBlobDataSource
|
|
|
6
6
|
from .cloud_box_data_source import CloudBoxDataSource
|
|
7
7
|
from .cloud_confluence_data_source import CloudConfluenceDataSource
|
|
8
8
|
from .cloud_jira_data_source import CloudJiraDataSource
|
|
9
|
+
from .cloud_jira_data_source_v_2 import CloudJiraDataSourceV2
|
|
9
10
|
from .cloud_notion_page_data_source import CloudNotionPageDataSource
|
|
10
11
|
from .cloud_one_drive_data_source import CloudOneDriveDataSource
|
|
11
12
|
from .cloud_s_3_data_source import CloudS3DataSource
|
|
@@ -22,5 +23,6 @@ DataSourceComponent = typing.Union[
|
|
|
22
23
|
CloudNotionPageDataSource,
|
|
23
24
|
CloudConfluenceDataSource,
|
|
24
25
|
CloudJiraDataSource,
|
|
26
|
+
CloudJiraDataSourceV2,
|
|
25
27
|
CloudBoxDataSource,
|
|
26
28
|
]
|
|
@@ -6,6 +6,7 @@ from .cloud_az_storage_blob_data_source import CloudAzStorageBlobDataSource
|
|
|
6
6
|
from .cloud_box_data_source import CloudBoxDataSource
|
|
7
7
|
from .cloud_confluence_data_source import CloudConfluenceDataSource
|
|
8
8
|
from .cloud_jira_data_source import CloudJiraDataSource
|
|
9
|
+
from .cloud_jira_data_source_v_2 import CloudJiraDataSourceV2
|
|
9
10
|
from .cloud_notion_page_data_source import CloudNotionPageDataSource
|
|
10
11
|
from .cloud_one_drive_data_source import CloudOneDriveDataSource
|
|
11
12
|
from .cloud_s_3_data_source import CloudS3DataSource
|
|
@@ -22,5 +23,6 @@ DataSourceCreateComponent = typing.Union[
|
|
|
22
23
|
CloudNotionPageDataSource,
|
|
23
24
|
CloudConfluenceDataSource,
|
|
24
25
|
CloudJiraDataSource,
|
|
26
|
+
CloudJiraDataSourceV2,
|
|
25
27
|
CloudBoxDataSource,
|
|
26
28
|
]
|
|
@@ -9,9 +9,17 @@ T_Result = typing.TypeVar("T_Result")
|
|
|
9
9
|
class DataSourceReaderVersionMetadataReaderVersion(str, enum.Enum):
|
|
10
10
|
ONE_0 = "1.0"
|
|
11
11
|
TWO_0 = "2.0"
|
|
12
|
+
TWO_1 = "2.1"
|
|
12
13
|
|
|
13
|
-
def visit(
|
|
14
|
+
def visit(
|
|
15
|
+
self,
|
|
16
|
+
one_0: typing.Callable[[], T_Result],
|
|
17
|
+
two_0: typing.Callable[[], T_Result],
|
|
18
|
+
two_1: typing.Callable[[], T_Result],
|
|
19
|
+
) -> T_Result:
|
|
14
20
|
if self is DataSourceReaderVersionMetadataReaderVersion.ONE_0:
|
|
15
21
|
return one_0()
|
|
16
22
|
if self is DataSourceReaderVersionMetadataReaderVersion.TWO_0:
|
|
17
23
|
return two_0()
|
|
24
|
+
if self is DataSourceReaderVersionMetadataReaderVersion.TWO_1:
|
|
25
|
+
return two_1()
|
|
@@ -4,9 +4,7 @@ import datetime as dt
|
|
|
4
4
|
import typing
|
|
5
5
|
|
|
6
6
|
from ..core.datetime_utils import serialize_datetime
|
|
7
|
-
from .
|
|
8
|
-
from .classify_parsing_configuration import ClassifyParsingConfiguration
|
|
9
|
-
from .status_enum import StatusEnum
|
|
7
|
+
from .classify_job import ClassifyJob
|
|
10
8
|
|
|
11
9
|
try:
|
|
12
10
|
import pydantic
|
|
@@ -17,21 +15,10 @@ except ImportError:
|
|
|
17
15
|
import pydantic # type: ignore
|
|
18
16
|
|
|
19
17
|
|
|
20
|
-
class
|
|
21
|
-
""
|
|
22
|
-
|
|
23
|
-
|
|
24
|
-
|
|
25
|
-
id: str = pydantic.Field(description="Unique identifier")
|
|
26
|
-
created_at: typing.Optional[dt.datetime]
|
|
27
|
-
updated_at: typing.Optional[dt.datetime]
|
|
28
|
-
rules: typing.List[ClassifierRule] = pydantic.Field(description="The rules to classify the files")
|
|
29
|
-
user_id: str = pydantic.Field(description="The ID of the user")
|
|
30
|
-
project_id: str = pydantic.Field(description="The ID of the project")
|
|
31
|
-
parsing_configuration: typing.Optional[ClassifyParsingConfiguration] = pydantic.Field(
|
|
32
|
-
description="The configuration for the parsing job"
|
|
33
|
-
)
|
|
34
|
-
status: StatusEnum = pydantic.Field(description="The status of the classify job")
|
|
18
|
+
class PaginatedResponseClassifyJob(pydantic.BaseModel):
|
|
19
|
+
items: typing.List[ClassifyJob] = pydantic.Field(description="The list of items.")
|
|
20
|
+
next_page_token: typing.Optional[str]
|
|
21
|
+
total_size: typing.Optional[int]
|
|
35
22
|
|
|
36
23
|
def json(self, **kwargs: typing.Any) -> str:
|
|
37
24
|
kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
|
|
@@ -6,6 +6,7 @@ from .cloud_az_storage_blob_data_source import CloudAzStorageBlobDataSource
|
|
|
6
6
|
from .cloud_box_data_source import CloudBoxDataSource
|
|
7
7
|
from .cloud_confluence_data_source import CloudConfluenceDataSource
|
|
8
8
|
from .cloud_jira_data_source import CloudJiraDataSource
|
|
9
|
+
from .cloud_jira_data_source_v_2 import CloudJiraDataSourceV2
|
|
9
10
|
from .cloud_notion_page_data_source import CloudNotionPageDataSource
|
|
10
11
|
from .cloud_one_drive_data_source import CloudOneDriveDataSource
|
|
11
12
|
from .cloud_s_3_data_source import CloudS3DataSource
|
|
@@ -22,5 +23,6 @@ PipelineDataSourceComponent = typing.Union[
|
|
|
22
23
|
CloudNotionPageDataSource,
|
|
23
24
|
CloudConfluenceDataSource,
|
|
24
25
|
CloudJiraDataSource,
|
|
26
|
+
CloudJiraDataSourceV2,
|
|
25
27
|
CloudBoxDataSource,
|
|
26
28
|
]
|
|
@@ -12,6 +12,7 @@ class UsageResponseActiveAlertsItem(str, enum.Enum):
|
|
|
12
12
|
CONFIGURED_SPEND_LIMIT_EXCEEDED = "configured_spend_limit_exceeded"
|
|
13
13
|
FREE_CREDITS_EXHAUSTED = "free_credits_exhausted"
|
|
14
14
|
INTERNAL_SPENDING_ALERT = "internal_spending_alert"
|
|
15
|
+
HAS_SPENDING_ALERT = "has_spending_alert"
|
|
15
16
|
|
|
16
17
|
def visit(
|
|
17
18
|
self,
|
|
@@ -20,6 +21,7 @@ class UsageResponseActiveAlertsItem(str, enum.Enum):
|
|
|
20
21
|
configured_spend_limit_exceeded: typing.Callable[[], T_Result],
|
|
21
22
|
free_credits_exhausted: typing.Callable[[], T_Result],
|
|
22
23
|
internal_spending_alert: typing.Callable[[], T_Result],
|
|
24
|
+
has_spending_alert: typing.Callable[[], T_Result],
|
|
23
25
|
) -> T_Result:
|
|
24
26
|
if self is UsageResponseActiveAlertsItem.PLAN_SPEND_LIMIT_EXCEEDED:
|
|
25
27
|
return plan_spend_limit_exceeded()
|
|
@@ -31,3 +33,5 @@ class UsageResponseActiveAlertsItem(str, enum.Enum):
|
|
|
31
33
|
return free_credits_exhausted()
|
|
32
34
|
if self is UsageResponseActiveAlertsItem.INTERNAL_SPENDING_ALERT:
|
|
33
35
|
return internal_spending_alert()
|
|
36
|
+
if self is UsageResponseActiveAlertsItem.HAS_SPENDING_ALERT:
|
|
37
|
+
return has_spending_alert()
|
|
@@ -1,6 +1,6 @@
|
|
|
1
|
-
Metadata-Version: 2.
|
|
1
|
+
Metadata-Version: 2.1
|
|
2
2
|
Name: llama-cloud
|
|
3
|
-
Version: 0.1.
|
|
3
|
+
Version: 0.1.38
|
|
4
4
|
Summary:
|
|
5
5
|
License: MIT
|
|
6
6
|
Author: Logan Markewich
|
|
@@ -13,7 +13,6 @@ Classifier: Programming Language :: Python :: 3.9
|
|
|
13
13
|
Classifier: Programming Language :: Python :: 3.10
|
|
14
14
|
Classifier: Programming Language :: Python :: 3.11
|
|
15
15
|
Classifier: Programming Language :: Python :: 3.12
|
|
16
|
-
Classifier: Programming Language :: Python :: 3.13
|
|
17
16
|
Requires-Dist: certifi (>=2024.7.4)
|
|
18
17
|
Requires-Dist: httpx (>=0.20.0)
|
|
19
18
|
Requires-Dist: pydantic (>=1.10)
|