llama-cloud 0.0.1__py3-none-any.whl → 0.0.2__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of llama-cloud might be problematic. Click here for more details.
- llama_cloud/client.py +2 -2
- llama_cloud/resources/api_keys/client.py +16 -16
- llama_cloud/resources/billing/client.py +12 -12
- llama_cloud/resources/component_definitions/client.py +12 -12
- llama_cloud/resources/data_sinks/client.py +24 -24
- llama_cloud/resources/data_sources/client.py +24 -24
- llama_cloud/resources/deprecated/client.py +48 -48
- llama_cloud/resources/evals/client.py +40 -40
- llama_cloud/resources/files/client.py +24 -24
- llama_cloud/resources/parsing/client.py +48 -48
- llama_cloud/resources/pipelines/client.py +223 -152
- llama_cloud/resources/projects/client.py +64 -64
- {llama_cloud-0.0.1.dist-info → llama_cloud-0.0.2.dist-info}/METADATA +1 -1
- {llama_cloud-0.0.1.dist-info → llama_cloud-0.0.2.dist-info}/RECORD +16 -16
- {llama_cloud-0.0.1.dist-info → llama_cloud-0.0.2.dist-info}/LICENSE +0 -0
- {llama_cloud-0.0.1.dist-info → llama_cloud-0.0.2.dist-info}/WHEEL +0 -0
|
@@ -38,9 +38,9 @@ class ProjectsClient:
|
|
|
38
38
|
Parameters:
|
|
39
39
|
- project_name: typing.Optional[str].
|
|
40
40
|
---
|
|
41
|
-
from platform.client import
|
|
41
|
+
from platform.client import LlamaCloud
|
|
42
42
|
|
|
43
|
-
client =
|
|
43
|
+
client = LlamaCloud(
|
|
44
44
|
token="YOUR_TOKEN",
|
|
45
45
|
base_url="https://yourhost.com/path/to/api",
|
|
46
46
|
)
|
|
@@ -71,9 +71,9 @@ class ProjectsClient:
|
|
|
71
71
|
- request: ProjectCreate.
|
|
72
72
|
---
|
|
73
73
|
from platform import ProjectCreate
|
|
74
|
-
from platform.client import
|
|
74
|
+
from platform.client import LlamaCloud
|
|
75
75
|
|
|
76
|
-
client =
|
|
76
|
+
client = LlamaCloud(
|
|
77
77
|
token="YOUR_TOKEN",
|
|
78
78
|
base_url="https://yourhost.com/path/to/api",
|
|
79
79
|
)
|
|
@@ -109,9 +109,9 @@ class ProjectsClient:
|
|
|
109
109
|
- request: ProjectCreate.
|
|
110
110
|
---
|
|
111
111
|
from platform import ProjectCreate
|
|
112
|
-
from platform.client import
|
|
112
|
+
from platform.client import LlamaCloud
|
|
113
113
|
|
|
114
|
-
client =
|
|
114
|
+
client = LlamaCloud(
|
|
115
115
|
token="YOUR_TOKEN",
|
|
116
116
|
base_url="https://yourhost.com/path/to/api",
|
|
117
117
|
)
|
|
@@ -145,9 +145,9 @@ class ProjectsClient:
|
|
|
145
145
|
Parameters:
|
|
146
146
|
- project_id: str.
|
|
147
147
|
---
|
|
148
|
-
from platform.client import
|
|
148
|
+
from platform.client import LlamaCloud
|
|
149
149
|
|
|
150
|
-
client =
|
|
150
|
+
client = LlamaCloud(
|
|
151
151
|
token="YOUR_TOKEN",
|
|
152
152
|
base_url="https://yourhost.com/path/to/api",
|
|
153
153
|
)
|
|
@@ -180,9 +180,9 @@ class ProjectsClient:
|
|
|
180
180
|
|
|
181
181
|
- name: str.
|
|
182
182
|
---
|
|
183
|
-
from platform.client import
|
|
183
|
+
from platform.client import LlamaCloud
|
|
184
184
|
|
|
185
|
-
client =
|
|
185
|
+
client = LlamaCloud(
|
|
186
186
|
token="YOUR_TOKEN",
|
|
187
187
|
base_url="https://yourhost.com/path/to/api",
|
|
188
188
|
)
|
|
@@ -215,9 +215,9 @@ class ProjectsClient:
|
|
|
215
215
|
Parameters:
|
|
216
216
|
- project_id: str.
|
|
217
217
|
---
|
|
218
|
-
from platform.client import
|
|
218
|
+
from platform.client import LlamaCloud
|
|
219
219
|
|
|
220
|
-
client =
|
|
220
|
+
client = LlamaCloud(
|
|
221
221
|
token="YOUR_TOKEN",
|
|
222
222
|
base_url="https://yourhost.com/path/to/api",
|
|
223
223
|
)
|
|
@@ -248,9 +248,9 @@ class ProjectsClient:
|
|
|
248
248
|
Parameters:
|
|
249
249
|
- project_id: str.
|
|
250
250
|
---
|
|
251
|
-
from platform.client import
|
|
251
|
+
from platform.client import LlamaCloud
|
|
252
252
|
|
|
253
|
-
client =
|
|
253
|
+
client = LlamaCloud(
|
|
254
254
|
token="YOUR_TOKEN",
|
|
255
255
|
base_url="https://yourhost.com/path/to/api",
|
|
256
256
|
)
|
|
@@ -285,9 +285,9 @@ class ProjectsClient:
|
|
|
285
285
|
|
|
286
286
|
- name: str. The name of the EvalDataset.
|
|
287
287
|
---
|
|
288
|
-
from platform.client import
|
|
288
|
+
from platform.client import LlamaCloud
|
|
289
289
|
|
|
290
|
-
client =
|
|
290
|
+
client = LlamaCloud(
|
|
291
291
|
token="YOUR_TOKEN",
|
|
292
292
|
base_url="https://yourhost.com/path/to/api",
|
|
293
293
|
)
|
|
@@ -328,9 +328,9 @@ class ProjectsClient:
|
|
|
328
328
|
|
|
329
329
|
- results: typing.Dict[str, typing.List[LocalEval]]. The eval results.
|
|
330
330
|
---
|
|
331
|
-
from platform.client import
|
|
331
|
+
from platform.client import LlamaCloud
|
|
332
332
|
|
|
333
|
-
client =
|
|
333
|
+
client = LlamaCloud(
|
|
334
334
|
token="YOUR_TOKEN",
|
|
335
335
|
base_url="https://yourhost.com/path/to/api",
|
|
336
336
|
)
|
|
@@ -366,9 +366,9 @@ class ProjectsClient:
|
|
|
366
366
|
Parameters:
|
|
367
367
|
- project_id: str.
|
|
368
368
|
---
|
|
369
|
-
from platform.client import
|
|
369
|
+
from platform.client import LlamaCloud
|
|
370
370
|
|
|
371
|
-
client =
|
|
371
|
+
client = LlamaCloud(
|
|
372
372
|
token="YOUR_TOKEN",
|
|
373
373
|
base_url="https://yourhost.com/path/to/api",
|
|
374
374
|
)
|
|
@@ -399,9 +399,9 @@ class ProjectsClient:
|
|
|
399
399
|
Parameters:
|
|
400
400
|
- project_id: str.
|
|
401
401
|
---
|
|
402
|
-
from platform.client import
|
|
402
|
+
from platform.client import LlamaCloud
|
|
403
403
|
|
|
404
|
-
client =
|
|
404
|
+
client = LlamaCloud(
|
|
405
405
|
token="YOUR_TOKEN",
|
|
406
406
|
base_url="https://yourhost.com/path/to/api",
|
|
407
407
|
)
|
|
@@ -436,9 +436,9 @@ class ProjectsClient:
|
|
|
436
436
|
|
|
437
437
|
- local_eval_set_id: str.
|
|
438
438
|
---
|
|
439
|
-
from platform.client import
|
|
439
|
+
from platform.client import LlamaCloud
|
|
440
440
|
|
|
441
|
-
client =
|
|
441
|
+
client = LlamaCloud(
|
|
442
442
|
token="YOUR_TOKEN",
|
|
443
443
|
base_url="https://yourhost.com/path/to/api",
|
|
444
444
|
)
|
|
@@ -473,9 +473,9 @@ class ProjectsClient:
|
|
|
473
473
|
Parameters:
|
|
474
474
|
- project_id: str.
|
|
475
475
|
---
|
|
476
|
-
from platform.client import
|
|
476
|
+
from platform.client import LlamaCloud
|
|
477
477
|
|
|
478
|
-
client =
|
|
478
|
+
client = LlamaCloud(
|
|
479
479
|
token="YOUR_TOKEN",
|
|
480
480
|
base_url="https://yourhost.com/path/to/api",
|
|
481
481
|
)
|
|
@@ -509,9 +509,9 @@ class ProjectsClient:
|
|
|
509
509
|
- request: PromptMixinPrompts.
|
|
510
510
|
---
|
|
511
511
|
from platform import PromptMixinPrompts
|
|
512
|
-
from platform.client import
|
|
512
|
+
from platform.client import LlamaCloud
|
|
513
513
|
|
|
514
|
-
client =
|
|
514
|
+
client = LlamaCloud(
|
|
515
515
|
token="YOUR_TOKEN",
|
|
516
516
|
base_url="https://yourhost.com/path/to/api",
|
|
517
517
|
)
|
|
@@ -555,9 +555,9 @@ class ProjectsClient:
|
|
|
555
555
|
- request: PromptMixinPrompts.
|
|
556
556
|
---
|
|
557
557
|
from platform import PromptMixinPrompts
|
|
558
|
-
from platform.client import
|
|
558
|
+
from platform.client import LlamaCloud
|
|
559
559
|
|
|
560
|
-
client =
|
|
560
|
+
client = LlamaCloud(
|
|
561
561
|
token="YOUR_TOKEN",
|
|
562
562
|
base_url="https://yourhost.com/path/to/api",
|
|
563
563
|
)
|
|
@@ -599,9 +599,9 @@ class ProjectsClient:
|
|
|
599
599
|
|
|
600
600
|
- prompt_set_id: str.
|
|
601
601
|
---
|
|
602
|
-
from platform.client import
|
|
602
|
+
from platform.client import LlamaCloud
|
|
603
603
|
|
|
604
|
-
client =
|
|
604
|
+
client = LlamaCloud(
|
|
605
605
|
token="YOUR_TOKEN",
|
|
606
606
|
base_url="https://yourhost.com/path/to/api",
|
|
607
607
|
)
|
|
@@ -640,9 +640,9 @@ class AsyncProjectsClient:
|
|
|
640
640
|
Parameters:
|
|
641
641
|
- project_name: typing.Optional[str].
|
|
642
642
|
---
|
|
643
|
-
from platform.client import
|
|
643
|
+
from platform.client import AsyncLlamaCloud
|
|
644
644
|
|
|
645
|
-
client =
|
|
645
|
+
client = AsyncLlamaCloud(
|
|
646
646
|
token="YOUR_TOKEN",
|
|
647
647
|
base_url="https://yourhost.com/path/to/api",
|
|
648
648
|
)
|
|
@@ -673,9 +673,9 @@ class AsyncProjectsClient:
|
|
|
673
673
|
- request: ProjectCreate.
|
|
674
674
|
---
|
|
675
675
|
from platform import ProjectCreate
|
|
676
|
-
from platform.client import
|
|
676
|
+
from platform.client import AsyncLlamaCloud
|
|
677
677
|
|
|
678
|
-
client =
|
|
678
|
+
client = AsyncLlamaCloud(
|
|
679
679
|
token="YOUR_TOKEN",
|
|
680
680
|
base_url="https://yourhost.com/path/to/api",
|
|
681
681
|
)
|
|
@@ -711,9 +711,9 @@ class AsyncProjectsClient:
|
|
|
711
711
|
- request: ProjectCreate.
|
|
712
712
|
---
|
|
713
713
|
from platform import ProjectCreate
|
|
714
|
-
from platform.client import
|
|
714
|
+
from platform.client import AsyncLlamaCloud
|
|
715
715
|
|
|
716
|
-
client =
|
|
716
|
+
client = AsyncLlamaCloud(
|
|
717
717
|
token="YOUR_TOKEN",
|
|
718
718
|
base_url="https://yourhost.com/path/to/api",
|
|
719
719
|
)
|
|
@@ -747,9 +747,9 @@ class AsyncProjectsClient:
|
|
|
747
747
|
Parameters:
|
|
748
748
|
- project_id: str.
|
|
749
749
|
---
|
|
750
|
-
from platform.client import
|
|
750
|
+
from platform.client import AsyncLlamaCloud
|
|
751
751
|
|
|
752
|
-
client =
|
|
752
|
+
client = AsyncLlamaCloud(
|
|
753
753
|
token="YOUR_TOKEN",
|
|
754
754
|
base_url="https://yourhost.com/path/to/api",
|
|
755
755
|
)
|
|
@@ -782,9 +782,9 @@ class AsyncProjectsClient:
|
|
|
782
782
|
|
|
783
783
|
- name: str.
|
|
784
784
|
---
|
|
785
|
-
from platform.client import
|
|
785
|
+
from platform.client import AsyncLlamaCloud
|
|
786
786
|
|
|
787
|
-
client =
|
|
787
|
+
client = AsyncLlamaCloud(
|
|
788
788
|
token="YOUR_TOKEN",
|
|
789
789
|
base_url="https://yourhost.com/path/to/api",
|
|
790
790
|
)
|
|
@@ -817,9 +817,9 @@ class AsyncProjectsClient:
|
|
|
817
817
|
Parameters:
|
|
818
818
|
- project_id: str.
|
|
819
819
|
---
|
|
820
|
-
from platform.client import
|
|
820
|
+
from platform.client import AsyncLlamaCloud
|
|
821
821
|
|
|
822
|
-
client =
|
|
822
|
+
client = AsyncLlamaCloud(
|
|
823
823
|
token="YOUR_TOKEN",
|
|
824
824
|
base_url="https://yourhost.com/path/to/api",
|
|
825
825
|
)
|
|
@@ -850,9 +850,9 @@ class AsyncProjectsClient:
|
|
|
850
850
|
Parameters:
|
|
851
851
|
- project_id: str.
|
|
852
852
|
---
|
|
853
|
-
from platform.client import
|
|
853
|
+
from platform.client import AsyncLlamaCloud
|
|
854
854
|
|
|
855
|
-
client =
|
|
855
|
+
client = AsyncLlamaCloud(
|
|
856
856
|
token="YOUR_TOKEN",
|
|
857
857
|
base_url="https://yourhost.com/path/to/api",
|
|
858
858
|
)
|
|
@@ -887,9 +887,9 @@ class AsyncProjectsClient:
|
|
|
887
887
|
|
|
888
888
|
- name: str. The name of the EvalDataset.
|
|
889
889
|
---
|
|
890
|
-
from platform.client import
|
|
890
|
+
from platform.client import AsyncLlamaCloud
|
|
891
891
|
|
|
892
|
-
client =
|
|
892
|
+
client = AsyncLlamaCloud(
|
|
893
893
|
token="YOUR_TOKEN",
|
|
894
894
|
base_url="https://yourhost.com/path/to/api",
|
|
895
895
|
)
|
|
@@ -930,9 +930,9 @@ class AsyncProjectsClient:
|
|
|
930
930
|
|
|
931
931
|
- results: typing.Dict[str, typing.List[LocalEval]]. The eval results.
|
|
932
932
|
---
|
|
933
|
-
from platform.client import
|
|
933
|
+
from platform.client import AsyncLlamaCloud
|
|
934
934
|
|
|
935
|
-
client =
|
|
935
|
+
client = AsyncLlamaCloud(
|
|
936
936
|
token="YOUR_TOKEN",
|
|
937
937
|
base_url="https://yourhost.com/path/to/api",
|
|
938
938
|
)
|
|
@@ -968,9 +968,9 @@ class AsyncProjectsClient:
|
|
|
968
968
|
Parameters:
|
|
969
969
|
- project_id: str.
|
|
970
970
|
---
|
|
971
|
-
from platform.client import
|
|
971
|
+
from platform.client import AsyncLlamaCloud
|
|
972
972
|
|
|
973
|
-
client =
|
|
973
|
+
client = AsyncLlamaCloud(
|
|
974
974
|
token="YOUR_TOKEN",
|
|
975
975
|
base_url="https://yourhost.com/path/to/api",
|
|
976
976
|
)
|
|
@@ -1001,9 +1001,9 @@ class AsyncProjectsClient:
|
|
|
1001
1001
|
Parameters:
|
|
1002
1002
|
- project_id: str.
|
|
1003
1003
|
---
|
|
1004
|
-
from platform.client import
|
|
1004
|
+
from platform.client import AsyncLlamaCloud
|
|
1005
1005
|
|
|
1006
|
-
client =
|
|
1006
|
+
client = AsyncLlamaCloud(
|
|
1007
1007
|
token="YOUR_TOKEN",
|
|
1008
1008
|
base_url="https://yourhost.com/path/to/api",
|
|
1009
1009
|
)
|
|
@@ -1038,9 +1038,9 @@ class AsyncProjectsClient:
|
|
|
1038
1038
|
|
|
1039
1039
|
- local_eval_set_id: str.
|
|
1040
1040
|
---
|
|
1041
|
-
from platform.client import
|
|
1041
|
+
from platform.client import AsyncLlamaCloud
|
|
1042
1042
|
|
|
1043
|
-
client =
|
|
1043
|
+
client = AsyncLlamaCloud(
|
|
1044
1044
|
token="YOUR_TOKEN",
|
|
1045
1045
|
base_url="https://yourhost.com/path/to/api",
|
|
1046
1046
|
)
|
|
@@ -1075,9 +1075,9 @@ class AsyncProjectsClient:
|
|
|
1075
1075
|
Parameters:
|
|
1076
1076
|
- project_id: str.
|
|
1077
1077
|
---
|
|
1078
|
-
from platform.client import
|
|
1078
|
+
from platform.client import AsyncLlamaCloud
|
|
1079
1079
|
|
|
1080
|
-
client =
|
|
1080
|
+
client = AsyncLlamaCloud(
|
|
1081
1081
|
token="YOUR_TOKEN",
|
|
1082
1082
|
base_url="https://yourhost.com/path/to/api",
|
|
1083
1083
|
)
|
|
@@ -1111,9 +1111,9 @@ class AsyncProjectsClient:
|
|
|
1111
1111
|
- request: PromptMixinPrompts.
|
|
1112
1112
|
---
|
|
1113
1113
|
from platform import PromptMixinPrompts
|
|
1114
|
-
from platform.client import
|
|
1114
|
+
from platform.client import AsyncLlamaCloud
|
|
1115
1115
|
|
|
1116
|
-
client =
|
|
1116
|
+
client = AsyncLlamaCloud(
|
|
1117
1117
|
token="YOUR_TOKEN",
|
|
1118
1118
|
base_url="https://yourhost.com/path/to/api",
|
|
1119
1119
|
)
|
|
@@ -1157,9 +1157,9 @@ class AsyncProjectsClient:
|
|
|
1157
1157
|
- request: PromptMixinPrompts.
|
|
1158
1158
|
---
|
|
1159
1159
|
from platform import PromptMixinPrompts
|
|
1160
|
-
from platform.client import
|
|
1160
|
+
from platform.client import AsyncLlamaCloud
|
|
1161
1161
|
|
|
1162
|
-
client =
|
|
1162
|
+
client = AsyncLlamaCloud(
|
|
1163
1163
|
token="YOUR_TOKEN",
|
|
1164
1164
|
base_url="https://yourhost.com/path/to/api",
|
|
1165
1165
|
)
|
|
@@ -1201,9 +1201,9 @@ class AsyncProjectsClient:
|
|
|
1201
1201
|
|
|
1202
1202
|
- prompt_set_id: str.
|
|
1203
1203
|
---
|
|
1204
|
-
from platform.client import
|
|
1204
|
+
from platform.client import AsyncLlamaCloud
|
|
1205
1205
|
|
|
1206
|
-
client =
|
|
1206
|
+
client = AsyncLlamaCloud(
|
|
1207
1207
|
token="YOUR_TOKEN",
|
|
1208
1208
|
base_url="https://yourhost.com/path/to/api",
|
|
1209
1209
|
)
|
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
llama_cloud/__init__.py,sha256=6HaLTVk7FR6IzYhV_u93-G3K6-XFGg40piOUmghj34k,7561
|
|
2
|
-
llama_cloud/client.py,sha256
|
|
2
|
+
llama_cloud/client.py,sha256=RCaROySFECX9t10dsg8lBbTLlds0QD_x2zU9x9HfKYA,3807
|
|
3
3
|
llama_cloud/core/__init__.py,sha256=QJS3CJ2TYP2E1Tge0CS6Z7r8LTNzJHQVX1hD3558eP0,519
|
|
4
4
|
llama_cloud/core/api_error.py,sha256=RE8LELok2QCjABadECTvtDp7qejA1VmINCh6TbqPwSE,426
|
|
5
5
|
llama_cloud/core/client_wrapper.py,sha256=xmj0jCdQ0ySzbSqHUWOkpRRy069y74I_HuXkWltcsVM,1507
|
|
@@ -10,38 +10,38 @@ llama_cloud/errors/__init__.py,sha256=pbbVUFtB9LCocA1RMWMMF_RKjsy5YkOKX5BAuE49w6
|
|
|
10
10
|
llama_cloud/errors/unprocessable_entity_error.py,sha256=FvR7XPlV3Xx5nu8HNlmLhBRdk4so_gCHjYT5PyZe6sM,313
|
|
11
11
|
llama_cloud/resources/__init__.py,sha256=oNfkOfXcF9JLlasW7zdX5eo3KNrfzgpfQlCpeyS8Gnw,1016
|
|
12
12
|
llama_cloud/resources/api_keys/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
13
|
-
llama_cloud/resources/api_keys/client.py,sha256=
|
|
13
|
+
llama_cloud/resources/api_keys/client.py,sha256=7i4XzUJTACS_NzJL3OcC_fo5Kq3SMpau0S8g4u9v3N0,11664
|
|
14
14
|
llama_cloud/resources/billing/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
15
|
-
llama_cloud/resources/billing/client.py,sha256=
|
|
15
|
+
llama_cloud/resources/billing/client.py,sha256=SVP1KgVWhyacXWhSdvGJ_4Bq4GqsEfBSXGngW7yaPJ8,9446
|
|
16
16
|
llama_cloud/resources/component_definitions/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
17
|
-
llama_cloud/resources/component_definitions/client.py,sha256=
|
|
17
|
+
llama_cloud/resources/component_definitions/client.py,sha256=eLpV6pOVx20tCUW90iWkDYEnJqMgEamgMWMw47tw4VU,7881
|
|
18
18
|
llama_cloud/resources/data_sinks/__init__.py,sha256=nsMEyxkVilxvQGSdJi0Z0yKZoTaTWewZIGJNoMwNDsw,205
|
|
19
|
-
llama_cloud/resources/data_sinks/client.py,sha256=
|
|
19
|
+
llama_cloud/resources/data_sinks/client.py,sha256=MRlyR6rGSZDuRENubstMDq79PQbi8s5svjdHPascikk,20243
|
|
20
20
|
llama_cloud/resources/data_sinks/types/__init__.py,sha256=M9AO57_TUUgjUcGOhxcROql5U7UbJDbEm7aQj3YqU2I,269
|
|
21
21
|
llama_cloud/resources/data_sinks/types/data_sink_update_component.py,sha256=TjBOpvPvUIyi-NT1Gv1vShMoe-jzDKc8UYaFfo7XOO8,249
|
|
22
22
|
llama_cloud/resources/data_sinks/types/data_sink_update_component_one.py,sha256=wNaRFihU7fW2nhLALvqcRZngS9-rudxn_dgmh_z6FRs,639
|
|
23
23
|
llama_cloud/resources/data_sources/__init__.py,sha256=CCs8ur4fvszPjy0GpTWmMjUAx0WykNgKDKFDNbkYLeM,289
|
|
24
|
-
llama_cloud/resources/data_sources/client.py,sha256=
|
|
24
|
+
llama_cloud/resources/data_sources/client.py,sha256=m6cu5eMJyoatzqNY7vjMa0SGj53sRZVY8pIKPyGevR0,21455
|
|
25
25
|
llama_cloud/resources/data_sources/types/__init__.py,sha256=iOdDXvAM6w80PR62JCscsTOwzDIXHHcG_Ypv18DEdic,410
|
|
26
26
|
llama_cloud/resources/data_sources/types/data_source_update_component.py,sha256=8MoJgdjYmN5WqntDpMXX34WJsf-Wsn0gYw_0t9SOTTA,257
|
|
27
27
|
llama_cloud/resources/data_sources/types/data_source_update_component_one.py,sha256=jfHjlwkUonW0Z73XhJ3w0BZpmptuXU205FWXS1Ucf44,742
|
|
28
28
|
llama_cloud/resources/data_sources/types/data_source_update_custom_metadata_value.py,sha256=3aFC-p8MSxjhOu2nFtqk0pixj6RqNqcFnbOYngUdZUk,215
|
|
29
29
|
llama_cloud/resources/deprecated/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
30
|
-
llama_cloud/resources/deprecated/client.py,sha256=
|
|
30
|
+
llama_cloud/resources/deprecated/client.py,sha256=hSGJara04Yg4RcGEKeZ2RjYggLz1OurrBTmZFXt6bfk,37288
|
|
31
31
|
llama_cloud/resources/evals/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
32
|
-
llama_cloud/resources/evals/client.py,sha256=
|
|
32
|
+
llama_cloud/resources/evals/client.py,sha256=uCxe3pPXHhNb3uOMzPEugE-SEoxc45SJ6LevaDaIFuk,28583
|
|
33
33
|
llama_cloud/resources/files/__init__.py,sha256=aZpyTj6KpZvA5XVwmuo1sIlRs7ba98btxVBZ6j5vIsI,155
|
|
34
|
-
llama_cloud/resources/files/client.py,sha256=
|
|
34
|
+
llama_cloud/resources/files/client.py,sha256=6b8eM0GfAFQXkKzufpDOfyq19alY9wfcHc8vUm4n7dQ,22746
|
|
35
35
|
llama_cloud/resources/files/types/__init__.py,sha256=ZWnnYWuDYZSfUJc7Jv3HyovzijdB--DTK4YB-uPcDsA,181
|
|
36
36
|
llama_cloud/resources/files/types/file_create_resource_info_value.py,sha256=R7Y-CJf7fnbvIqE3xOI5XOrmPwLbVJLC7zpxMu8Zopk,201
|
|
37
37
|
llama_cloud/resources/parsing/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
38
|
-
llama_cloud/resources/parsing/client.py,sha256=
|
|
38
|
+
llama_cloud/resources/parsing/client.py,sha256=2_9U7zElV58GDVeh9eCI-VAYRe5d0CKej_C60_4JEq0,37288
|
|
39
39
|
llama_cloud/resources/pipelines/__init__.py,sha256=H7yaFIN62vjuhU3TOKzzuf8qpxZRgw1xVa-eyig-2YU,175
|
|
40
|
-
llama_cloud/resources/pipelines/client.py,sha256=
|
|
40
|
+
llama_cloud/resources/pipelines/client.py,sha256=_ExQjfJ28fPZ5gQXOev8Hi0YhKL6xXuZs5IFVNhM7ZM,106484
|
|
41
41
|
llama_cloud/resources/pipelines/types/__init__.py,sha256=xuT4OBPLrRfEe-E3UVdJvRjl9jTp7tNBK_YzZBb6Kj8,212
|
|
42
42
|
llama_cloud/resources/pipelines/types/pipeline_file_update_custom_metadata_value.py,sha256=trI48WLxPcAqV9207Q6-3cj1nl4EGlZpw7En56ZsPgg,217
|
|
43
43
|
llama_cloud/resources/projects/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
44
|
-
llama_cloud/resources/projects/client.py,sha256=
|
|
44
|
+
llama_cloud/resources/projects/client.py,sha256=Ky2bq1O0cPBorT6r2jA_Z_YfbwDE4K0TRWkqVkWh_Hg,48354
|
|
45
45
|
llama_cloud/types/__init__.py,sha256=xKnYxctZU1sCSVv6_GNg4w6t_o9vJIBTcD9Lqgu32Ts,10347
|
|
46
46
|
llama_cloud/types/api_key.py,sha256=zEZG3fvKSlKeGkLOuRiMJRh_uv3fED_F3it_21rMHPo,1217
|
|
47
47
|
llama_cloud/types/azure_open_ai_embedding.py,sha256=zCk8TBWeDHM635nsSmRLSs-GFQWyHo8v7Ovv-8rtgGQ,3346
|
|
@@ -167,7 +167,7 @@ llama_cloud/types/token_text_splitter.py,sha256=DxBaqoExtsGeYvUdZPP_LYpEo0_-Oj31
|
|
|
167
167
|
llama_cloud/types/transformation_category_names.py,sha256=0xjYe-mDW9OKbTGqL5fSbTvqsfrG4LDu_stW_ubVLl4,582
|
|
168
168
|
llama_cloud/types/validation_error.py,sha256=2JhGNJouo8QpfrMBoT_JCwYSn1nFN2Nnq0p9uPLDH-U,992
|
|
169
169
|
llama_cloud/types/validation_error_loc_item.py,sha256=LAtjCHIllWRBFXvAZ5QZpp7CPXjdtN9EB7HrLVo6EP0,128
|
|
170
|
-
llama_cloud-0.0.
|
|
171
|
-
llama_cloud-0.0.
|
|
172
|
-
llama_cloud-0.0.
|
|
173
|
-
llama_cloud-0.0.
|
|
170
|
+
llama_cloud-0.0.2.dist-info/LICENSE,sha256=_iNqtPcw1Ue7dZKwOwgPtbegMUkWVy15hC7bffAdNmY,1067
|
|
171
|
+
llama_cloud-0.0.2.dist-info/METADATA,sha256=trCof17Q1p97KJ_fY_RoyqWbgo9fyZhOmtienxOsdno,750
|
|
172
|
+
llama_cloud-0.0.2.dist-info/WHEEL,sha256=FMvqSimYX_P7y0a7UY-_Mc83r5zkBZsCYPm7Lr0Bsq4,88
|
|
173
|
+
llama_cloud-0.0.2.dist-info/RECORD,,
|
|
File without changes
|
|
File without changes
|