llama-cloud 0.1.38__py3-none-any.whl → 0.1.39__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of llama-cloud might be problematic. Click here for more details.
- llama_cloud/resources/alpha/client.py +2 -8
- llama_cloud/resources/beta/client.py +30 -126
- llama_cloud/resources/chat_apps/client.py +8 -32
- llama_cloud/resources/classifier/client.py +8 -32
- llama_cloud/resources/data_sinks/client.py +8 -32
- llama_cloud/resources/data_sources/client.py +8 -32
- llama_cloud/resources/embedding_model_configs/client.py +12 -48
- llama_cloud/resources/files/client.py +42 -176
- llama_cloud/resources/jobs/client.py +2 -8
- llama_cloud/resources/llama_extract/client.py +32 -138
- llama_cloud/resources/organizations/client.py +4 -18
- llama_cloud/resources/parsing/client.py +4 -16
- llama_cloud/resources/pipelines/client.py +8 -32
- llama_cloud/resources/projects/client.py +18 -78
- llama_cloud/resources/reports/client.py +30 -126
- llama_cloud/resources/retrievers/client.py +12 -48
- llama_cloud/types/file.py +1 -1
- llama_cloud/types/pipeline_file.py +4 -4
- {llama_cloud-0.1.38.dist-info → llama_cloud-0.1.39.dist-info}/METADATA +1 -1
- {llama_cloud-0.1.38.dist-info → llama_cloud-0.1.39.dist-info}/RECORD +22 -22
- {llama_cloud-0.1.38.dist-info → llama_cloud-0.1.39.dist-info}/LICENSE +0 -0
- {llama_cloud-0.1.38.dist-info → llama_cloud-0.1.39.dist-info}/WHEEL +0 -0
|
@@ -39,7 +39,6 @@ class RetrieversClient:
|
|
|
39
39
|
name: typing.Optional[str] = None,
|
|
40
40
|
project_id: typing.Optional[str] = None,
|
|
41
41
|
organization_id: typing.Optional[str] = None,
|
|
42
|
-
project_id: typing.Optional[str] = None,
|
|
43
42
|
) -> typing.List[Retriever]:
|
|
44
43
|
"""
|
|
45
44
|
List Retrievers for a project.
|
|
@@ -50,8 +49,6 @@ class RetrieversClient:
|
|
|
50
49
|
- project_id: typing.Optional[str].
|
|
51
50
|
|
|
52
51
|
- organization_id: typing.Optional[str].
|
|
53
|
-
|
|
54
|
-
- project_id: typing.Optional[str].
|
|
55
52
|
---
|
|
56
53
|
from llama_cloud.client import LlamaCloud
|
|
57
54
|
|
|
@@ -64,7 +61,7 @@ class RetrieversClient:
|
|
|
64
61
|
"GET",
|
|
65
62
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/retrievers"),
|
|
66
63
|
params=remove_none_from_dict({"name": name, "project_id": project_id, "organization_id": organization_id}),
|
|
67
|
-
headers=
|
|
64
|
+
headers=self._client_wrapper.get_headers(),
|
|
68
65
|
timeout=60,
|
|
69
66
|
)
|
|
70
67
|
if 200 <= _response.status_code < 300:
|
|
@@ -83,7 +80,6 @@ class RetrieversClient:
|
|
|
83
80
|
project_id: typing.Optional[str] = None,
|
|
84
81
|
organization_id: typing.Optional[str] = None,
|
|
85
82
|
request: RetrieverCreate,
|
|
86
|
-
project_id: typing.Optional[str] = None,
|
|
87
83
|
) -> Retriever:
|
|
88
84
|
"""
|
|
89
85
|
Create a new Retriever.
|
|
@@ -94,8 +90,6 @@ class RetrieversClient:
|
|
|
94
90
|
- organization_id: typing.Optional[str].
|
|
95
91
|
|
|
96
92
|
- request: RetrieverCreate.
|
|
97
|
-
|
|
98
|
-
- project_id: typing.Optional[str].
|
|
99
93
|
---
|
|
100
94
|
from llama_cloud import RetrieverCreate
|
|
101
95
|
from llama_cloud.client import LlamaCloud
|
|
@@ -114,7 +108,7 @@ class RetrieversClient:
|
|
|
114
108
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/retrievers"),
|
|
115
109
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
116
110
|
json=jsonable_encoder(request),
|
|
117
|
-
headers=
|
|
111
|
+
headers=self._client_wrapper.get_headers(),
|
|
118
112
|
timeout=60,
|
|
119
113
|
)
|
|
120
114
|
if 200 <= _response.status_code < 300:
|
|
@@ -133,7 +127,6 @@ class RetrieversClient:
|
|
|
133
127
|
project_id: typing.Optional[str] = None,
|
|
134
128
|
organization_id: typing.Optional[str] = None,
|
|
135
129
|
request: RetrieverCreate,
|
|
136
|
-
project_id: typing.Optional[str] = None,
|
|
137
130
|
) -> Retriever:
|
|
138
131
|
"""
|
|
139
132
|
Upsert a new Retriever.
|
|
@@ -144,8 +137,6 @@ class RetrieversClient:
|
|
|
144
137
|
- organization_id: typing.Optional[str].
|
|
145
138
|
|
|
146
139
|
- request: RetrieverCreate.
|
|
147
|
-
|
|
148
|
-
- project_id: typing.Optional[str].
|
|
149
140
|
---
|
|
150
141
|
from llama_cloud import RetrieverCreate
|
|
151
142
|
from llama_cloud.client import LlamaCloud
|
|
@@ -164,7 +155,7 @@ class RetrieversClient:
|
|
|
164
155
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/retrievers"),
|
|
165
156
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
166
157
|
json=jsonable_encoder(request),
|
|
167
|
-
headers=
|
|
158
|
+
headers=self._client_wrapper.get_headers(),
|
|
168
159
|
timeout=60,
|
|
169
160
|
)
|
|
170
161
|
if 200 <= _response.status_code < 300:
|
|
@@ -183,7 +174,6 @@ class RetrieversClient:
|
|
|
183
174
|
*,
|
|
184
175
|
project_id: typing.Optional[str] = None,
|
|
185
176
|
organization_id: typing.Optional[str] = None,
|
|
186
|
-
project_id: typing.Optional[str] = None,
|
|
187
177
|
) -> Retriever:
|
|
188
178
|
"""
|
|
189
179
|
Get a Retriever by ID.
|
|
@@ -194,8 +184,6 @@ class RetrieversClient:
|
|
|
194
184
|
- project_id: typing.Optional[str].
|
|
195
185
|
|
|
196
186
|
- organization_id: typing.Optional[str].
|
|
197
|
-
|
|
198
|
-
- project_id: typing.Optional[str].
|
|
199
187
|
---
|
|
200
188
|
from llama_cloud.client import LlamaCloud
|
|
201
189
|
|
|
@@ -210,7 +198,7 @@ class RetrieversClient:
|
|
|
210
198
|
"GET",
|
|
211
199
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", f"api/v1/retrievers/{retriever_id}"),
|
|
212
200
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
213
|
-
headers=
|
|
201
|
+
headers=self._client_wrapper.get_headers(),
|
|
214
202
|
timeout=60,
|
|
215
203
|
)
|
|
216
204
|
if 200 <= _response.status_code < 300:
|
|
@@ -313,7 +301,6 @@ class RetrieversClient:
|
|
|
313
301
|
rerank_top_n: typing.Optional[int] = OMIT,
|
|
314
302
|
rerank_config: typing.Optional[ReRankConfig] = OMIT,
|
|
315
303
|
query: str,
|
|
316
|
-
project_id: typing.Optional[str] = None,
|
|
317
304
|
) -> CompositeRetrievalResult:
|
|
318
305
|
"""
|
|
319
306
|
Retrieve data using a Retriever.
|
|
@@ -332,8 +319,6 @@ class RetrieversClient:
|
|
|
332
319
|
- rerank_config: typing.Optional[ReRankConfig]. The rerank configuration for composite retrieval.
|
|
333
320
|
|
|
334
321
|
- query: str. The query to retrieve against.
|
|
335
|
-
|
|
336
|
-
- project_id: typing.Optional[str].
|
|
337
322
|
---
|
|
338
323
|
from llama_cloud import CompositeRetrievalMode, ReRankConfig, ReRankerType
|
|
339
324
|
from llama_cloud.client import LlamaCloud
|
|
@@ -364,7 +349,7 @@ class RetrieversClient:
|
|
|
364
349
|
),
|
|
365
350
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
366
351
|
json=jsonable_encoder(_request),
|
|
367
|
-
headers=
|
|
352
|
+
headers=self._client_wrapper.get_headers(),
|
|
368
353
|
timeout=60,
|
|
369
354
|
)
|
|
370
355
|
if 200 <= _response.status_code < 300:
|
|
@@ -387,7 +372,6 @@ class RetrieversClient:
|
|
|
387
372
|
rerank_config: typing.Optional[ReRankConfig] = OMIT,
|
|
388
373
|
query: str,
|
|
389
374
|
pipelines: typing.Optional[typing.List[RetrieverPipeline]] = OMIT,
|
|
390
|
-
project_id: typing.Optional[str] = None,
|
|
391
375
|
) -> CompositeRetrievalResult:
|
|
392
376
|
"""
|
|
393
377
|
Retrieve data using specified pipelines without creating a persistent retriever.
|
|
@@ -406,8 +390,6 @@ class RetrieversClient:
|
|
|
406
390
|
- query: str. The query to retrieve against.
|
|
407
391
|
|
|
408
392
|
- pipelines: typing.Optional[typing.List[RetrieverPipeline]]. The pipelines to use for retrieval.
|
|
409
|
-
|
|
410
|
-
- project_id: typing.Optional[str].
|
|
411
393
|
---
|
|
412
394
|
from llama_cloud import CompositeRetrievalMode, ReRankConfig, ReRankerType
|
|
413
395
|
from llama_cloud.client import LlamaCloud
|
|
@@ -437,7 +419,7 @@ class RetrieversClient:
|
|
|
437
419
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/retrievers/retrieve"),
|
|
438
420
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
439
421
|
json=jsonable_encoder(_request),
|
|
440
|
-
headers=
|
|
422
|
+
headers=self._client_wrapper.get_headers(),
|
|
441
423
|
timeout=60,
|
|
442
424
|
)
|
|
443
425
|
if 200 <= _response.status_code < 300:
|
|
@@ -461,7 +443,6 @@ class AsyncRetrieversClient:
|
|
|
461
443
|
name: typing.Optional[str] = None,
|
|
462
444
|
project_id: typing.Optional[str] = None,
|
|
463
445
|
organization_id: typing.Optional[str] = None,
|
|
464
|
-
project_id: typing.Optional[str] = None,
|
|
465
446
|
) -> typing.List[Retriever]:
|
|
466
447
|
"""
|
|
467
448
|
List Retrievers for a project.
|
|
@@ -472,8 +453,6 @@ class AsyncRetrieversClient:
|
|
|
472
453
|
- project_id: typing.Optional[str].
|
|
473
454
|
|
|
474
455
|
- organization_id: typing.Optional[str].
|
|
475
|
-
|
|
476
|
-
- project_id: typing.Optional[str].
|
|
477
456
|
---
|
|
478
457
|
from llama_cloud.client import AsyncLlamaCloud
|
|
479
458
|
|
|
@@ -486,7 +465,7 @@ class AsyncRetrieversClient:
|
|
|
486
465
|
"GET",
|
|
487
466
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/retrievers"),
|
|
488
467
|
params=remove_none_from_dict({"name": name, "project_id": project_id, "organization_id": organization_id}),
|
|
489
|
-
headers=
|
|
468
|
+
headers=self._client_wrapper.get_headers(),
|
|
490
469
|
timeout=60,
|
|
491
470
|
)
|
|
492
471
|
if 200 <= _response.status_code < 300:
|
|
@@ -505,7 +484,6 @@ class AsyncRetrieversClient:
|
|
|
505
484
|
project_id: typing.Optional[str] = None,
|
|
506
485
|
organization_id: typing.Optional[str] = None,
|
|
507
486
|
request: RetrieverCreate,
|
|
508
|
-
project_id: typing.Optional[str] = None,
|
|
509
487
|
) -> Retriever:
|
|
510
488
|
"""
|
|
511
489
|
Create a new Retriever.
|
|
@@ -516,8 +494,6 @@ class AsyncRetrieversClient:
|
|
|
516
494
|
- organization_id: typing.Optional[str].
|
|
517
495
|
|
|
518
496
|
- request: RetrieverCreate.
|
|
519
|
-
|
|
520
|
-
- project_id: typing.Optional[str].
|
|
521
497
|
---
|
|
522
498
|
from llama_cloud import RetrieverCreate
|
|
523
499
|
from llama_cloud.client import AsyncLlamaCloud
|
|
@@ -536,7 +512,7 @@ class AsyncRetrieversClient:
|
|
|
536
512
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/retrievers"),
|
|
537
513
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
538
514
|
json=jsonable_encoder(request),
|
|
539
|
-
headers=
|
|
515
|
+
headers=self._client_wrapper.get_headers(),
|
|
540
516
|
timeout=60,
|
|
541
517
|
)
|
|
542
518
|
if 200 <= _response.status_code < 300:
|
|
@@ -555,7 +531,6 @@ class AsyncRetrieversClient:
|
|
|
555
531
|
project_id: typing.Optional[str] = None,
|
|
556
532
|
organization_id: typing.Optional[str] = None,
|
|
557
533
|
request: RetrieverCreate,
|
|
558
|
-
project_id: typing.Optional[str] = None,
|
|
559
534
|
) -> Retriever:
|
|
560
535
|
"""
|
|
561
536
|
Upsert a new Retriever.
|
|
@@ -566,8 +541,6 @@ class AsyncRetrieversClient:
|
|
|
566
541
|
- organization_id: typing.Optional[str].
|
|
567
542
|
|
|
568
543
|
- request: RetrieverCreate.
|
|
569
|
-
|
|
570
|
-
- project_id: typing.Optional[str].
|
|
571
544
|
---
|
|
572
545
|
from llama_cloud import RetrieverCreate
|
|
573
546
|
from llama_cloud.client import AsyncLlamaCloud
|
|
@@ -586,7 +559,7 @@ class AsyncRetrieversClient:
|
|
|
586
559
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/retrievers"),
|
|
587
560
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
588
561
|
json=jsonable_encoder(request),
|
|
589
|
-
headers=
|
|
562
|
+
headers=self._client_wrapper.get_headers(),
|
|
590
563
|
timeout=60,
|
|
591
564
|
)
|
|
592
565
|
if 200 <= _response.status_code < 300:
|
|
@@ -605,7 +578,6 @@ class AsyncRetrieversClient:
|
|
|
605
578
|
*,
|
|
606
579
|
project_id: typing.Optional[str] = None,
|
|
607
580
|
organization_id: typing.Optional[str] = None,
|
|
608
|
-
project_id: typing.Optional[str] = None,
|
|
609
581
|
) -> Retriever:
|
|
610
582
|
"""
|
|
611
583
|
Get a Retriever by ID.
|
|
@@ -616,8 +588,6 @@ class AsyncRetrieversClient:
|
|
|
616
588
|
- project_id: typing.Optional[str].
|
|
617
589
|
|
|
618
590
|
- organization_id: typing.Optional[str].
|
|
619
|
-
|
|
620
|
-
- project_id: typing.Optional[str].
|
|
621
591
|
---
|
|
622
592
|
from llama_cloud.client import AsyncLlamaCloud
|
|
623
593
|
|
|
@@ -632,7 +602,7 @@ class AsyncRetrieversClient:
|
|
|
632
602
|
"GET",
|
|
633
603
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", f"api/v1/retrievers/{retriever_id}"),
|
|
634
604
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
635
|
-
headers=
|
|
605
|
+
headers=self._client_wrapper.get_headers(),
|
|
636
606
|
timeout=60,
|
|
637
607
|
)
|
|
638
608
|
if 200 <= _response.status_code < 300:
|
|
@@ -735,7 +705,6 @@ class AsyncRetrieversClient:
|
|
|
735
705
|
rerank_top_n: typing.Optional[int] = OMIT,
|
|
736
706
|
rerank_config: typing.Optional[ReRankConfig] = OMIT,
|
|
737
707
|
query: str,
|
|
738
|
-
project_id: typing.Optional[str] = None,
|
|
739
708
|
) -> CompositeRetrievalResult:
|
|
740
709
|
"""
|
|
741
710
|
Retrieve data using a Retriever.
|
|
@@ -754,8 +723,6 @@ class AsyncRetrieversClient:
|
|
|
754
723
|
- rerank_config: typing.Optional[ReRankConfig]. The rerank configuration for composite retrieval.
|
|
755
724
|
|
|
756
725
|
- query: str. The query to retrieve against.
|
|
757
|
-
|
|
758
|
-
- project_id: typing.Optional[str].
|
|
759
726
|
---
|
|
760
727
|
from llama_cloud import CompositeRetrievalMode, ReRankConfig, ReRankerType
|
|
761
728
|
from llama_cloud.client import AsyncLlamaCloud
|
|
@@ -786,7 +753,7 @@ class AsyncRetrieversClient:
|
|
|
786
753
|
),
|
|
787
754
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
788
755
|
json=jsonable_encoder(_request),
|
|
789
|
-
headers=
|
|
756
|
+
headers=self._client_wrapper.get_headers(),
|
|
790
757
|
timeout=60,
|
|
791
758
|
)
|
|
792
759
|
if 200 <= _response.status_code < 300:
|
|
@@ -809,7 +776,6 @@ class AsyncRetrieversClient:
|
|
|
809
776
|
rerank_config: typing.Optional[ReRankConfig] = OMIT,
|
|
810
777
|
query: str,
|
|
811
778
|
pipelines: typing.Optional[typing.List[RetrieverPipeline]] = OMIT,
|
|
812
|
-
project_id: typing.Optional[str] = None,
|
|
813
779
|
) -> CompositeRetrievalResult:
|
|
814
780
|
"""
|
|
815
781
|
Retrieve data using specified pipelines without creating a persistent retriever.
|
|
@@ -828,8 +794,6 @@ class AsyncRetrieversClient:
|
|
|
828
794
|
- query: str. The query to retrieve against.
|
|
829
795
|
|
|
830
796
|
- pipelines: typing.Optional[typing.List[RetrieverPipeline]]. The pipelines to use for retrieval.
|
|
831
|
-
|
|
832
|
-
- project_id: typing.Optional[str].
|
|
833
797
|
---
|
|
834
798
|
from llama_cloud import CompositeRetrievalMode, ReRankConfig, ReRankerType
|
|
835
799
|
from llama_cloud.client import AsyncLlamaCloud
|
|
@@ -859,7 +823,7 @@ class AsyncRetrieversClient:
|
|
|
859
823
|
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/retrievers/retrieve"),
|
|
860
824
|
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
861
825
|
json=jsonable_encoder(_request),
|
|
862
|
-
headers=
|
|
826
|
+
headers=self._client_wrapper.get_headers(),
|
|
863
827
|
timeout=60,
|
|
864
828
|
)
|
|
865
829
|
if 200 <= _response.status_code < 300:
|
llama_cloud/types/file.py
CHANGED
|
@@ -25,7 +25,7 @@ class File(pydantic.BaseModel):
|
|
|
25
25
|
created_at: typing.Optional[dt.datetime]
|
|
26
26
|
updated_at: typing.Optional[dt.datetime]
|
|
27
27
|
name: str
|
|
28
|
-
external_file_id: str
|
|
28
|
+
external_file_id: typing.Optional[str]
|
|
29
29
|
file_size: typing.Optional[int]
|
|
30
30
|
file_type: typing.Optional[str]
|
|
31
31
|
project_id: str = pydantic.Field(description="The ID of the project that the file belongs to")
|
|
@@ -31,14 +31,14 @@ class PipelineFile(pydantic.BaseModel):
|
|
|
31
31
|
external_file_id: typing.Optional[str]
|
|
32
32
|
file_size: typing.Optional[int]
|
|
33
33
|
file_type: typing.Optional[str]
|
|
34
|
-
project_id: str
|
|
34
|
+
project_id: typing.Optional[str]
|
|
35
35
|
last_modified_at: typing.Optional[dt.datetime]
|
|
36
|
-
resource_info: typing.Optional[typing.Dict[str, typing.Optional[PipelineFileResourceInfoValue]]]
|
|
37
|
-
permission_info: typing.Optional[typing.Dict[str, typing.Optional[PipelineFilePermissionInfoValue]]]
|
|
38
|
-
data_source_id: typing.Optional[str]
|
|
39
36
|
file_id: typing.Optional[str]
|
|
40
37
|
pipeline_id: str = pydantic.Field(description="The ID of the pipeline that the file is associated with")
|
|
38
|
+
resource_info: typing.Optional[typing.Dict[str, typing.Optional[PipelineFileResourceInfoValue]]]
|
|
39
|
+
permission_info: typing.Optional[typing.Dict[str, typing.Optional[PipelineFilePermissionInfoValue]]]
|
|
41
40
|
custom_metadata: typing.Optional[typing.Dict[str, typing.Optional[PipelineFileCustomMetadataValue]]]
|
|
41
|
+
data_source_id: typing.Optional[str]
|
|
42
42
|
config_hash: typing.Optional[typing.Dict[str, typing.Optional[PipelineFileConfigHashValue]]]
|
|
43
43
|
indexed_page_count: typing.Optional[int]
|
|
44
44
|
status: typing.Optional[PipelineFileStatus]
|
|
@@ -15,36 +15,36 @@ llama_cloud/resources/admin/client.py,sha256=YIYy9kU1_xaE0gkpmZZbCgLzZj6XSrAUplS
|
|
|
15
15
|
llama_cloud/resources/agent_deployments/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
16
16
|
llama_cloud/resources/agent_deployments/client.py,sha256=3EOzOjmRs4KISgJ566enq3FCuN3YtskjO0OHqQGtkQ0,6122
|
|
17
17
|
llama_cloud/resources/alpha/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
18
|
-
llama_cloud/resources/alpha/client.py,sha256=
|
|
18
|
+
llama_cloud/resources/alpha/client.py,sha256=d5cRIUykNpnVryuxWBPUpmo-2L1vMIDeZIF3DvTIx7E,4322
|
|
19
19
|
llama_cloud/resources/beta/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
20
|
-
llama_cloud/resources/beta/client.py,sha256=
|
|
20
|
+
llama_cloud/resources/beta/client.py,sha256=_GNkHQxyZxhZOkLIRzfCw6PexQx-E8r_7R-3Wd9Y0uE,63128
|
|
21
21
|
llama_cloud/resources/chat_apps/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
22
|
-
llama_cloud/resources/chat_apps/client.py,sha256=
|
|
22
|
+
llama_cloud/resources/chat_apps/client.py,sha256=orSI8rpQbUwVEToolEeiEi5Qe--suXFvfu6D9JDii5I,23595
|
|
23
23
|
llama_cloud/resources/classifier/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
24
|
-
llama_cloud/resources/classifier/client.py,sha256=
|
|
24
|
+
llama_cloud/resources/classifier/client.py,sha256=Q1vdRTSe-QnZyFucTlxHIRmdCGa-PFf7XUazik7s59E,17618
|
|
25
25
|
llama_cloud/resources/data_sinks/__init__.py,sha256=ZHUjn3HbKhq_7QS1q74r2m5RGKF5lxcvF2P6pGvpcis,147
|
|
26
|
-
llama_cloud/resources/data_sinks/client.py,sha256=
|
|
26
|
+
llama_cloud/resources/data_sinks/client.py,sha256=GpD6FhbGqkg2oUToyMG6J8hPxG_iG7W5ZJRo0qg3yzk,20639
|
|
27
27
|
llama_cloud/resources/data_sinks/types/__init__.py,sha256=M1aTcufJwiEZo9B0KmYj9PfkSd6I1ooFt9tpIRGwgg8,168
|
|
28
28
|
llama_cloud/resources/data_sinks/types/data_sink_update_component.py,sha256=ynPdEg844hZaD6EcAK0jrMY_vogtvmLTZ7FZSwWcor8,912
|
|
29
29
|
llama_cloud/resources/data_sources/__init__.py,sha256=McURkcNBGHXH1hmRDRmZI1dRzJrekCTHZsgv03r2oZI,227
|
|
30
|
-
llama_cloud/resources/data_sources/client.py,sha256=
|
|
30
|
+
llama_cloud/resources/data_sources/client.py,sha256=SZFm8bW5nkaXringdSnmxHqvVjKM7cNNOtqVXjgTKhc,21855
|
|
31
31
|
llama_cloud/resources/data_sources/types/__init__.py,sha256=Cd5xEECTzXqQSfJALfJPSjudlSLeb3RENeJVi8vwPbM,303
|
|
32
32
|
llama_cloud/resources/data_sources/types/data_source_update_component.py,sha256=_jQY6FhcvenWdzi27SK1bSY8muXKLRkXlVrTqEWgKKc,1159
|
|
33
33
|
llama_cloud/resources/data_sources/types/data_source_update_custom_metadata_value.py,sha256=3aFC-p8MSxjhOu2nFtqk0pixj6RqNqcFnbOYngUdZUk,215
|
|
34
34
|
llama_cloud/resources/embedding_model_configs/__init__.py,sha256=cXDtKKq-gj7yjFjdQ5GrGyPs-T5tRV_0JjUMGlAbdUs,1115
|
|
35
|
-
llama_cloud/resources/embedding_model_configs/client.py,sha256=
|
|
35
|
+
llama_cloud/resources/embedding_model_configs/client.py,sha256=2JDvZJtSger9QJ8luPct-2zvwjaJAR8VcKsTZ1wgYTE,17769
|
|
36
36
|
llama_cloud/resources/embedding_model_configs/types/__init__.py,sha256=6-rcDwJhw_0shz3CjrPvlYBYXJJ1bLn-PpplhOsQ79w,1156
|
|
37
37
|
llama_cloud/resources/embedding_model_configs/types/embedding_model_config_create_embedding_config.py,sha256=SQCHJk0AmBbKS5XKdcEJxhDhIMLQCmCI13IHC28v7vQ,3054
|
|
38
38
|
llama_cloud/resources/evals/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
39
39
|
llama_cloud/resources/evals/client.py,sha256=v2AyeQV0hVgC6xoP2gJNgneJMaeXALV1hIeirYGxlPw,3242
|
|
40
40
|
llama_cloud/resources/files/__init__.py,sha256=Ws53l-S3kyAGFinYPOb9WpN84DtbFn6gLYZtI2akBLQ,169
|
|
41
|
-
llama_cloud/resources/files/client.py,sha256=
|
|
41
|
+
llama_cloud/resources/files/client.py,sha256=Crd0IR0cV5fld4jUGAHE8VsIbw7vCYrOIyBTSwDyitA,56242
|
|
42
42
|
llama_cloud/resources/files/types/__init__.py,sha256=ZZuDQsYsxmQ9VwpfN7oqftzGRnFTR2EMYdCa7zARo4g,204
|
|
43
43
|
llama_cloud/resources/files/types/file_create_from_url_resource_info_value.py,sha256=Wc8wFgujOO5pZvbbh2TMMzpa37GKZd14GYNJ9bdq7BE,214
|
|
44
44
|
llama_cloud/resources/jobs/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
45
|
-
llama_cloud/resources/jobs/client.py,sha256=
|
|
45
|
+
llama_cloud/resources/jobs/client.py,sha256=b2R_Oj2OCtcv-IIJNz9aq42hDgrOk_huqTSJhTB9VaA,6202
|
|
46
46
|
llama_cloud/resources/llama_extract/__init__.py,sha256=V6VZ8hQXwAuvOOZyk43nnbINoDQqEr03AjKQPhYKluk,997
|
|
47
|
-
llama_cloud/resources/llama_extract/client.py,sha256=
|
|
47
|
+
llama_cloud/resources/llama_extract/client.py,sha256=B_qhVsk-Qs81qrFOVgWqcvelSB3TLWFJCibnn--3BjE,83096
|
|
48
48
|
llama_cloud/resources/llama_extract/types/__init__.py,sha256=2Iu4w5LXZY2Govr1RzahIfY0b84y658SQjMDtj7rH_0,1497
|
|
49
49
|
llama_cloud/resources/llama_extract/types/extract_agent_create_data_schema.py,sha256=zB31hJQ8hKaIsPkfTWiX5hqsPVFMyyeWEDZ_Aq237jo,305
|
|
50
50
|
llama_cloud/resources/llama_extract/types/extract_agent_create_data_schema_zero_value.py,sha256=xoyXH3f0Y5beMWBxmtXSz6QoB_df_-0QBsYdjBhZnGw,217
|
|
@@ -57,24 +57,24 @@ llama_cloud/resources/llama_extract/types/extract_schema_validate_request_data_s
|
|
|
57
57
|
llama_cloud/resources/llama_extract/types/extract_stateless_request_data_schema.py,sha256=lBblR9zgjJsbWL-2bDisCj7EQiX6aky6GQ4tuMr3LtU,325
|
|
58
58
|
llama_cloud/resources/llama_extract/types/extract_stateless_request_data_schema_zero_value.py,sha256=4-ONLmkrEP36ZH0qRXp3sbXCtLVNQQX4dLXFeF4u47g,222
|
|
59
59
|
llama_cloud/resources/organizations/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
60
|
-
llama_cloud/resources/organizations/client.py,sha256=
|
|
60
|
+
llama_cloud/resources/organizations/client.py,sha256=RoN-nkN7VeRZnrrElXhaPrgQFzGMHgNY41_XpbCXP0g,56623
|
|
61
61
|
llama_cloud/resources/parsing/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
62
|
-
llama_cloud/resources/parsing/client.py,sha256=
|
|
62
|
+
llama_cloud/resources/parsing/client.py,sha256=K9qYu5v99OqbMmeeeCioQDUnT085A5keCencSwES-gc,88995
|
|
63
63
|
llama_cloud/resources/pipelines/__init__.py,sha256=zyvVEOF_krvEZkCIj_kZoMKfhDqHo_R32a1mv9CriQc,1193
|
|
64
|
-
llama_cloud/resources/pipelines/client.py,sha256=
|
|
64
|
+
llama_cloud/resources/pipelines/client.py,sha256=VAqAm0oY_nXGkMPqXuzPEHS9kPtpuOE5sxfyqlzXuSI,134738
|
|
65
65
|
llama_cloud/resources/pipelines/types/__init__.py,sha256=C68NQ5QzA0dFXf9oePFFGmV1vn96jcAp-QAznSgoRYQ,1375
|
|
66
66
|
llama_cloud/resources/pipelines/types/pipeline_file_update_custom_metadata_value.py,sha256=trI48WLxPcAqV9207Q6-3cj1nl4EGlZpw7En56ZsPgg,217
|
|
67
67
|
llama_cloud/resources/pipelines/types/pipeline_update_embedding_config.py,sha256=c8FF64fDrBMX_2RX4uY3CjbNc0Ss_AUJ4Eqs-KeV4Wc,2874
|
|
68
68
|
llama_cloud/resources/pipelines/types/pipeline_update_transform_config.py,sha256=KbkyULMv-qeS3qRd31ia6pd5rOdypS0o2UL42NRcA7E,321
|
|
69
69
|
llama_cloud/resources/pipelines/types/retrieval_params_search_filters_inference_schema_value.py,sha256=hZWXYlTib0af85ECcerC4xD-bUQe8rG3Q6G1jFTMQcI,228
|
|
70
70
|
llama_cloud/resources/projects/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
71
|
-
llama_cloud/resources/projects/client.py,sha256=
|
|
71
|
+
llama_cloud/resources/projects/client.py,sha256=PF36iWtSa5amUt3q56YwLypOZjclIXSubCRv9NttpLs,25404
|
|
72
72
|
llama_cloud/resources/reports/__init__.py,sha256=cruYbQ1bIuJbRpkfaQY7ajUEslffjd7KzvzMzbtPH94,217
|
|
73
|
-
llama_cloud/resources/reports/client.py,sha256=
|
|
73
|
+
llama_cloud/resources/reports/client.py,sha256=kHjtXVVc1Xi3T1GyBvSW5K4mTdr6xQwZA3vw-liRKBg,46736
|
|
74
74
|
llama_cloud/resources/reports/types/__init__.py,sha256=LfwDYrI4RcQu-o42iAe7HkcwHww2YU90lOonBPTmZIk,291
|
|
75
75
|
llama_cloud/resources/reports/types/update_report_plan_api_v_1_reports_report_id_plan_patch_request_action.py,sha256=Qh-MSeRvDBfNb5hoLELivv1pLtrYVf52WVoP7G8V34A,807
|
|
76
76
|
llama_cloud/resources/retrievers/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
77
|
-
llama_cloud/resources/retrievers/client.py,sha256=
|
|
77
|
+
llama_cloud/resources/retrievers/client.py,sha256=z2LhmA-cZVFzr9P6loeCZYnJbvSIk0QitFeVFp-IyZk,32126
|
|
78
78
|
llama_cloud/resources/users/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
79
79
|
llama_cloud/resources/users/client.py,sha256=A2s8e2syQHkkSwPz-Lrt_Zxp1K-8nqJqj5EafE6NWYs,5545
|
|
80
80
|
llama_cloud/types/__init__.py,sha256=HSHS4fNorQ2l79VYOegTeh4KG7LnIzpSvUOhTELh3pQ,33920
|
|
@@ -194,7 +194,7 @@ llama_cloud/types/extract_state.py,sha256=TNeVAXXKZaiM2srlbQlzRSn4_TDpR4xyT_yQhJ
|
|
|
194
194
|
llama_cloud/types/extract_target.py,sha256=Gt-FNqblzcjdfq1hxsqEjWWu-HNLXdKy4w98nog52Ms,478
|
|
195
195
|
llama_cloud/types/fail_page_mode.py,sha256=n4fgPpiEB5siPoEg0Sux4COg7ElNybjshxDoUihZwRU,786
|
|
196
196
|
llama_cloud/types/failure_handling_config.py,sha256=EmAQW0qm7-JTSYFwhmIWxqkVNWym_AyAJIMEmeI9Cqc,1216
|
|
197
|
-
llama_cloud/types/file.py,sha256=
|
|
197
|
+
llama_cloud/types/file.py,sha256=sXdF-cdHL3k1-DPIxAjYpb-kNHzcOAV_earVoYITzUA,1765
|
|
198
198
|
llama_cloud/types/file_classification.py,sha256=jKzAc_3rg0Usyf3TNr-bI5HZn9zGIj9vYH90RKoDtiY,1418
|
|
199
199
|
llama_cloud/types/file_count_by_status_response.py,sha256=WuorbZvKjDs9Ql1hUiQu4gN5iCm8d6fr92KLyHpRvQU,1356
|
|
200
200
|
llama_cloud/types/file_create.py,sha256=eLUC50CzXOdAR_P2mBtX_R7kGteIVbP1V3LzuP1s0Xs,1629
|
|
@@ -308,7 +308,7 @@ llama_cloud/types/pipeline_data_source_custom_metadata_value.py,sha256=8n3r60sxM
|
|
|
308
308
|
llama_cloud/types/pipeline_data_source_status.py,sha256=BD4xoftwp9lWC8EjJTnf3boIG_AyzjLPuP4qJxGhmcc,1039
|
|
309
309
|
llama_cloud/types/pipeline_deployment.py,sha256=eVBrz032aPb2cqtIIVYT5MTHQvBNm89XazoNrRWVugo,1356
|
|
310
310
|
llama_cloud/types/pipeline_embedding_config.py,sha256=7NJzlabQLFUFsvj7fye-oKLPasaXCWJBm-XuLxy-xmQ,3112
|
|
311
|
-
llama_cloud/types/pipeline_file.py,sha256=
|
|
311
|
+
llama_cloud/types/pipeline_file.py,sha256=zoE1A4pdD7S4cgDtR_aVToQ08JDt_siUJTYsdayG8s4,2510
|
|
312
312
|
llama_cloud/types/pipeline_file_config_hash_value.py,sha256=4lvLnDpzNAHdiMkGJTTNDTu3p3H7Nxw5MR1Mzte7-_M,201
|
|
313
313
|
llama_cloud/types/pipeline_file_create.py,sha256=yoMIzWED0ktKerE48kgzInBa3d0aNGO5JjTtDTDAn4A,1310
|
|
314
314
|
llama_cloud/types/pipeline_file_create_custom_metadata_value.py,sha256=olVj5yhQFx1QqWO1Wv9d6AtL-YyYO9_OYtOfcD2ZeGY,217
|
|
@@ -398,7 +398,7 @@ llama_cloud/types/vertex_embedding_mode.py,sha256=yY23FjuWU_DkXjBb3JoKV4SCMqel2B
|
|
|
398
398
|
llama_cloud/types/vertex_text_embedding.py,sha256=-C4fNCYfFl36ATdBMGFVPpiHIKxjk0KB1ERA2Ec20aU,1932
|
|
399
399
|
llama_cloud/types/webhook_configuration.py,sha256=_Xm15whrWoKNBuCoO5y_NunA-ByhCAYK87LnC4W-Pzg,1350
|
|
400
400
|
llama_cloud/types/webhook_configuration_webhook_events_item.py,sha256=OL3moFO_6hsKZYSBQBsSHmWA0NgLcLJgBPZfABwT60c,2544
|
|
401
|
-
llama_cloud-0.1.
|
|
402
|
-
llama_cloud-0.1.
|
|
403
|
-
llama_cloud-0.1.
|
|
404
|
-
llama_cloud-0.1.
|
|
401
|
+
llama_cloud-0.1.39.dist-info/LICENSE,sha256=_iNqtPcw1Ue7dZKwOwgPtbegMUkWVy15hC7bffAdNmY,1067
|
|
402
|
+
llama_cloud-0.1.39.dist-info/METADATA,sha256=WNoZded1_WK8ATxqn2o6TaLGwsn2uijfjy07RnZd8bI,1143
|
|
403
|
+
llama_cloud-0.1.39.dist-info/WHEEL,sha256=sP946D7jFCHeNz5Iq4fL4Lu-PrWrFsgfLXbbkciIZwg,88
|
|
404
|
+
llama_cloud-0.1.39.dist-info/RECORD,,
|
|
File without changes
|
|
File without changes
|