llama-cloud 0.0.1__py3-none-any.whl → 0.0.2__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of llama-cloud might be problematic. Click here for more details.

@@ -38,9 +38,9 @@ class ProjectsClient:
38
38
  Parameters:
39
39
  - project_name: typing.Optional[str].
40
40
  ---
41
- from platform.client import PlatformApi
41
+ from platform.client import LlamaCloud
42
42
 
43
- client = PlatformApi(
43
+ client = LlamaCloud(
44
44
  token="YOUR_TOKEN",
45
45
  base_url="https://yourhost.com/path/to/api",
46
46
  )
@@ -71,9 +71,9 @@ class ProjectsClient:
71
71
  - request: ProjectCreate.
72
72
  ---
73
73
  from platform import ProjectCreate
74
- from platform.client import PlatformApi
74
+ from platform.client import LlamaCloud
75
75
 
76
- client = PlatformApi(
76
+ client = LlamaCloud(
77
77
  token="YOUR_TOKEN",
78
78
  base_url="https://yourhost.com/path/to/api",
79
79
  )
@@ -109,9 +109,9 @@ class ProjectsClient:
109
109
  - request: ProjectCreate.
110
110
  ---
111
111
  from platform import ProjectCreate
112
- from platform.client import PlatformApi
112
+ from platform.client import LlamaCloud
113
113
 
114
- client = PlatformApi(
114
+ client = LlamaCloud(
115
115
  token="YOUR_TOKEN",
116
116
  base_url="https://yourhost.com/path/to/api",
117
117
  )
@@ -145,9 +145,9 @@ class ProjectsClient:
145
145
  Parameters:
146
146
  - project_id: str.
147
147
  ---
148
- from platform.client import PlatformApi
148
+ from platform.client import LlamaCloud
149
149
 
150
- client = PlatformApi(
150
+ client = LlamaCloud(
151
151
  token="YOUR_TOKEN",
152
152
  base_url="https://yourhost.com/path/to/api",
153
153
  )
@@ -180,9 +180,9 @@ class ProjectsClient:
180
180
 
181
181
  - name: str.
182
182
  ---
183
- from platform.client import PlatformApi
183
+ from platform.client import LlamaCloud
184
184
 
185
- client = PlatformApi(
185
+ client = LlamaCloud(
186
186
  token="YOUR_TOKEN",
187
187
  base_url="https://yourhost.com/path/to/api",
188
188
  )
@@ -215,9 +215,9 @@ class ProjectsClient:
215
215
  Parameters:
216
216
  - project_id: str.
217
217
  ---
218
- from platform.client import PlatformApi
218
+ from platform.client import LlamaCloud
219
219
 
220
- client = PlatformApi(
220
+ client = LlamaCloud(
221
221
  token="YOUR_TOKEN",
222
222
  base_url="https://yourhost.com/path/to/api",
223
223
  )
@@ -248,9 +248,9 @@ class ProjectsClient:
248
248
  Parameters:
249
249
  - project_id: str.
250
250
  ---
251
- from platform.client import PlatformApi
251
+ from platform.client import LlamaCloud
252
252
 
253
- client = PlatformApi(
253
+ client = LlamaCloud(
254
254
  token="YOUR_TOKEN",
255
255
  base_url="https://yourhost.com/path/to/api",
256
256
  )
@@ -285,9 +285,9 @@ class ProjectsClient:
285
285
 
286
286
  - name: str. The name of the EvalDataset.
287
287
  ---
288
- from platform.client import PlatformApi
288
+ from platform.client import LlamaCloud
289
289
 
290
- client = PlatformApi(
290
+ client = LlamaCloud(
291
291
  token="YOUR_TOKEN",
292
292
  base_url="https://yourhost.com/path/to/api",
293
293
  )
@@ -328,9 +328,9 @@ class ProjectsClient:
328
328
 
329
329
  - results: typing.Dict[str, typing.List[LocalEval]]. The eval results.
330
330
  ---
331
- from platform.client import PlatformApi
331
+ from platform.client import LlamaCloud
332
332
 
333
- client = PlatformApi(
333
+ client = LlamaCloud(
334
334
  token="YOUR_TOKEN",
335
335
  base_url="https://yourhost.com/path/to/api",
336
336
  )
@@ -366,9 +366,9 @@ class ProjectsClient:
366
366
  Parameters:
367
367
  - project_id: str.
368
368
  ---
369
- from platform.client import PlatformApi
369
+ from platform.client import LlamaCloud
370
370
 
371
- client = PlatformApi(
371
+ client = LlamaCloud(
372
372
  token="YOUR_TOKEN",
373
373
  base_url="https://yourhost.com/path/to/api",
374
374
  )
@@ -399,9 +399,9 @@ class ProjectsClient:
399
399
  Parameters:
400
400
  - project_id: str.
401
401
  ---
402
- from platform.client import PlatformApi
402
+ from platform.client import LlamaCloud
403
403
 
404
- client = PlatformApi(
404
+ client = LlamaCloud(
405
405
  token="YOUR_TOKEN",
406
406
  base_url="https://yourhost.com/path/to/api",
407
407
  )
@@ -436,9 +436,9 @@ class ProjectsClient:
436
436
 
437
437
  - local_eval_set_id: str.
438
438
  ---
439
- from platform.client import PlatformApi
439
+ from platform.client import LlamaCloud
440
440
 
441
- client = PlatformApi(
441
+ client = LlamaCloud(
442
442
  token="YOUR_TOKEN",
443
443
  base_url="https://yourhost.com/path/to/api",
444
444
  )
@@ -473,9 +473,9 @@ class ProjectsClient:
473
473
  Parameters:
474
474
  - project_id: str.
475
475
  ---
476
- from platform.client import PlatformApi
476
+ from platform.client import LlamaCloud
477
477
 
478
- client = PlatformApi(
478
+ client = LlamaCloud(
479
479
  token="YOUR_TOKEN",
480
480
  base_url="https://yourhost.com/path/to/api",
481
481
  )
@@ -509,9 +509,9 @@ class ProjectsClient:
509
509
  - request: PromptMixinPrompts.
510
510
  ---
511
511
  from platform import PromptMixinPrompts
512
- from platform.client import PlatformApi
512
+ from platform.client import LlamaCloud
513
513
 
514
- client = PlatformApi(
514
+ client = LlamaCloud(
515
515
  token="YOUR_TOKEN",
516
516
  base_url="https://yourhost.com/path/to/api",
517
517
  )
@@ -555,9 +555,9 @@ class ProjectsClient:
555
555
  - request: PromptMixinPrompts.
556
556
  ---
557
557
  from platform import PromptMixinPrompts
558
- from platform.client import PlatformApi
558
+ from platform.client import LlamaCloud
559
559
 
560
- client = PlatformApi(
560
+ client = LlamaCloud(
561
561
  token="YOUR_TOKEN",
562
562
  base_url="https://yourhost.com/path/to/api",
563
563
  )
@@ -599,9 +599,9 @@ class ProjectsClient:
599
599
 
600
600
  - prompt_set_id: str.
601
601
  ---
602
- from platform.client import PlatformApi
602
+ from platform.client import LlamaCloud
603
603
 
604
- client = PlatformApi(
604
+ client = LlamaCloud(
605
605
  token="YOUR_TOKEN",
606
606
  base_url="https://yourhost.com/path/to/api",
607
607
  )
@@ -640,9 +640,9 @@ class AsyncProjectsClient:
640
640
  Parameters:
641
641
  - project_name: typing.Optional[str].
642
642
  ---
643
- from platform.client import AsyncPlatformApi
643
+ from platform.client import AsyncLlamaCloud
644
644
 
645
- client = AsyncPlatformApi(
645
+ client = AsyncLlamaCloud(
646
646
  token="YOUR_TOKEN",
647
647
  base_url="https://yourhost.com/path/to/api",
648
648
  )
@@ -673,9 +673,9 @@ class AsyncProjectsClient:
673
673
  - request: ProjectCreate.
674
674
  ---
675
675
  from platform import ProjectCreate
676
- from platform.client import AsyncPlatformApi
676
+ from platform.client import AsyncLlamaCloud
677
677
 
678
- client = AsyncPlatformApi(
678
+ client = AsyncLlamaCloud(
679
679
  token="YOUR_TOKEN",
680
680
  base_url="https://yourhost.com/path/to/api",
681
681
  )
@@ -711,9 +711,9 @@ class AsyncProjectsClient:
711
711
  - request: ProjectCreate.
712
712
  ---
713
713
  from platform import ProjectCreate
714
- from platform.client import AsyncPlatformApi
714
+ from platform.client import AsyncLlamaCloud
715
715
 
716
- client = AsyncPlatformApi(
716
+ client = AsyncLlamaCloud(
717
717
  token="YOUR_TOKEN",
718
718
  base_url="https://yourhost.com/path/to/api",
719
719
  )
@@ -747,9 +747,9 @@ class AsyncProjectsClient:
747
747
  Parameters:
748
748
  - project_id: str.
749
749
  ---
750
- from platform.client import AsyncPlatformApi
750
+ from platform.client import AsyncLlamaCloud
751
751
 
752
- client = AsyncPlatformApi(
752
+ client = AsyncLlamaCloud(
753
753
  token="YOUR_TOKEN",
754
754
  base_url="https://yourhost.com/path/to/api",
755
755
  )
@@ -782,9 +782,9 @@ class AsyncProjectsClient:
782
782
 
783
783
  - name: str.
784
784
  ---
785
- from platform.client import AsyncPlatformApi
785
+ from platform.client import AsyncLlamaCloud
786
786
 
787
- client = AsyncPlatformApi(
787
+ client = AsyncLlamaCloud(
788
788
  token="YOUR_TOKEN",
789
789
  base_url="https://yourhost.com/path/to/api",
790
790
  )
@@ -817,9 +817,9 @@ class AsyncProjectsClient:
817
817
  Parameters:
818
818
  - project_id: str.
819
819
  ---
820
- from platform.client import AsyncPlatformApi
820
+ from platform.client import AsyncLlamaCloud
821
821
 
822
- client = AsyncPlatformApi(
822
+ client = AsyncLlamaCloud(
823
823
  token="YOUR_TOKEN",
824
824
  base_url="https://yourhost.com/path/to/api",
825
825
  )
@@ -850,9 +850,9 @@ class AsyncProjectsClient:
850
850
  Parameters:
851
851
  - project_id: str.
852
852
  ---
853
- from platform.client import AsyncPlatformApi
853
+ from platform.client import AsyncLlamaCloud
854
854
 
855
- client = AsyncPlatformApi(
855
+ client = AsyncLlamaCloud(
856
856
  token="YOUR_TOKEN",
857
857
  base_url="https://yourhost.com/path/to/api",
858
858
  )
@@ -887,9 +887,9 @@ class AsyncProjectsClient:
887
887
 
888
888
  - name: str. The name of the EvalDataset.
889
889
  ---
890
- from platform.client import AsyncPlatformApi
890
+ from platform.client import AsyncLlamaCloud
891
891
 
892
- client = AsyncPlatformApi(
892
+ client = AsyncLlamaCloud(
893
893
  token="YOUR_TOKEN",
894
894
  base_url="https://yourhost.com/path/to/api",
895
895
  )
@@ -930,9 +930,9 @@ class AsyncProjectsClient:
930
930
 
931
931
  - results: typing.Dict[str, typing.List[LocalEval]]. The eval results.
932
932
  ---
933
- from platform.client import AsyncPlatformApi
933
+ from platform.client import AsyncLlamaCloud
934
934
 
935
- client = AsyncPlatformApi(
935
+ client = AsyncLlamaCloud(
936
936
  token="YOUR_TOKEN",
937
937
  base_url="https://yourhost.com/path/to/api",
938
938
  )
@@ -968,9 +968,9 @@ class AsyncProjectsClient:
968
968
  Parameters:
969
969
  - project_id: str.
970
970
  ---
971
- from platform.client import AsyncPlatformApi
971
+ from platform.client import AsyncLlamaCloud
972
972
 
973
- client = AsyncPlatformApi(
973
+ client = AsyncLlamaCloud(
974
974
  token="YOUR_TOKEN",
975
975
  base_url="https://yourhost.com/path/to/api",
976
976
  )
@@ -1001,9 +1001,9 @@ class AsyncProjectsClient:
1001
1001
  Parameters:
1002
1002
  - project_id: str.
1003
1003
  ---
1004
- from platform.client import AsyncPlatformApi
1004
+ from platform.client import AsyncLlamaCloud
1005
1005
 
1006
- client = AsyncPlatformApi(
1006
+ client = AsyncLlamaCloud(
1007
1007
  token="YOUR_TOKEN",
1008
1008
  base_url="https://yourhost.com/path/to/api",
1009
1009
  )
@@ -1038,9 +1038,9 @@ class AsyncProjectsClient:
1038
1038
 
1039
1039
  - local_eval_set_id: str.
1040
1040
  ---
1041
- from platform.client import AsyncPlatformApi
1041
+ from platform.client import AsyncLlamaCloud
1042
1042
 
1043
- client = AsyncPlatformApi(
1043
+ client = AsyncLlamaCloud(
1044
1044
  token="YOUR_TOKEN",
1045
1045
  base_url="https://yourhost.com/path/to/api",
1046
1046
  )
@@ -1075,9 +1075,9 @@ class AsyncProjectsClient:
1075
1075
  Parameters:
1076
1076
  - project_id: str.
1077
1077
  ---
1078
- from platform.client import AsyncPlatformApi
1078
+ from platform.client import AsyncLlamaCloud
1079
1079
 
1080
- client = AsyncPlatformApi(
1080
+ client = AsyncLlamaCloud(
1081
1081
  token="YOUR_TOKEN",
1082
1082
  base_url="https://yourhost.com/path/to/api",
1083
1083
  )
@@ -1111,9 +1111,9 @@ class AsyncProjectsClient:
1111
1111
  - request: PromptMixinPrompts.
1112
1112
  ---
1113
1113
  from platform import PromptMixinPrompts
1114
- from platform.client import AsyncPlatformApi
1114
+ from platform.client import AsyncLlamaCloud
1115
1115
 
1116
- client = AsyncPlatformApi(
1116
+ client = AsyncLlamaCloud(
1117
1117
  token="YOUR_TOKEN",
1118
1118
  base_url="https://yourhost.com/path/to/api",
1119
1119
  )
@@ -1157,9 +1157,9 @@ class AsyncProjectsClient:
1157
1157
  - request: PromptMixinPrompts.
1158
1158
  ---
1159
1159
  from platform import PromptMixinPrompts
1160
- from platform.client import AsyncPlatformApi
1160
+ from platform.client import AsyncLlamaCloud
1161
1161
 
1162
- client = AsyncPlatformApi(
1162
+ client = AsyncLlamaCloud(
1163
1163
  token="YOUR_TOKEN",
1164
1164
  base_url="https://yourhost.com/path/to/api",
1165
1165
  )
@@ -1201,9 +1201,9 @@ class AsyncProjectsClient:
1201
1201
 
1202
1202
  - prompt_set_id: str.
1203
1203
  ---
1204
- from platform.client import AsyncPlatformApi
1204
+ from platform.client import AsyncLlamaCloud
1205
1205
 
1206
- client = AsyncPlatformApi(
1206
+ client = AsyncLlamaCloud(
1207
1207
  token="YOUR_TOKEN",
1208
1208
  base_url="https://yourhost.com/path/to/api",
1209
1209
  )
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: llama-cloud
3
- Version: 0.0.1
3
+ Version: 0.0.2
4
4
  Summary:
5
5
  Author: Logan Markewich
6
6
  Author-email: logan@runllama.ai
@@ -1,5 +1,5 @@
1
1
  llama_cloud/__init__.py,sha256=6HaLTVk7FR6IzYhV_u93-G3K6-XFGg40piOUmghj34k,7561
2
- llama_cloud/client.py,sha256=-f2FNVTcKoksesG4FgGgaA470slOnEjRfkAyJUJJ1Dk,3809
2
+ llama_cloud/client.py,sha256=RCaROySFECX9t10dsg8lBbTLlds0QD_x2zU9x9HfKYA,3807
3
3
  llama_cloud/core/__init__.py,sha256=QJS3CJ2TYP2E1Tge0CS6Z7r8LTNzJHQVX1hD3558eP0,519
4
4
  llama_cloud/core/api_error.py,sha256=RE8LELok2QCjABadECTvtDp7qejA1VmINCh6TbqPwSE,426
5
5
  llama_cloud/core/client_wrapper.py,sha256=xmj0jCdQ0ySzbSqHUWOkpRRy069y74I_HuXkWltcsVM,1507
@@ -10,38 +10,38 @@ llama_cloud/errors/__init__.py,sha256=pbbVUFtB9LCocA1RMWMMF_RKjsy5YkOKX5BAuE49w6
10
10
  llama_cloud/errors/unprocessable_entity_error.py,sha256=FvR7XPlV3Xx5nu8HNlmLhBRdk4so_gCHjYT5PyZe6sM,313
11
11
  llama_cloud/resources/__init__.py,sha256=oNfkOfXcF9JLlasW7zdX5eo3KNrfzgpfQlCpeyS8Gnw,1016
12
12
  llama_cloud/resources/api_keys/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
13
- llama_cloud/resources/api_keys/client.py,sha256=Dwobvv_XUVpFtab0gYYEKMcsM30CcotimUWLVgZPCQ4,11680
13
+ llama_cloud/resources/api_keys/client.py,sha256=7i4XzUJTACS_NzJL3OcC_fo5Kq3SMpau0S8g4u9v3N0,11664
14
14
  llama_cloud/resources/billing/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
15
- llama_cloud/resources/billing/client.py,sha256=_mLTnRIQYMF6W4Szs2syLwC6xn5F2-gaNjpbrpBZFfA,9458
15
+ llama_cloud/resources/billing/client.py,sha256=SVP1KgVWhyacXWhSdvGJ_4Bq4GqsEfBSXGngW7yaPJ8,9446
16
16
  llama_cloud/resources/component_definitions/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
17
- llama_cloud/resources/component_definitions/client.py,sha256=fUCmmLdq_fT6uXfGAN_eYaELxsrmjWO-obXQ9ryrzDo,7893
17
+ llama_cloud/resources/component_definitions/client.py,sha256=eLpV6pOVx20tCUW90iWkDYEnJqMgEamgMWMw47tw4VU,7881
18
18
  llama_cloud/resources/data_sinks/__init__.py,sha256=nsMEyxkVilxvQGSdJi0Z0yKZoTaTWewZIGJNoMwNDsw,205
19
- llama_cloud/resources/data_sinks/client.py,sha256=fVaZnBs45W2fjMYYfXwdl4mtNM3A-23Nqyt0WiqIJAI,20267
19
+ llama_cloud/resources/data_sinks/client.py,sha256=MRlyR6rGSZDuRENubstMDq79PQbi8s5svjdHPascikk,20243
20
20
  llama_cloud/resources/data_sinks/types/__init__.py,sha256=M9AO57_TUUgjUcGOhxcROql5U7UbJDbEm7aQj3YqU2I,269
21
21
  llama_cloud/resources/data_sinks/types/data_sink_update_component.py,sha256=TjBOpvPvUIyi-NT1Gv1vShMoe-jzDKc8UYaFfo7XOO8,249
22
22
  llama_cloud/resources/data_sinks/types/data_sink_update_component_one.py,sha256=wNaRFihU7fW2nhLALvqcRZngS9-rudxn_dgmh_z6FRs,639
23
23
  llama_cloud/resources/data_sources/__init__.py,sha256=CCs8ur4fvszPjy0GpTWmMjUAx0WykNgKDKFDNbkYLeM,289
24
- llama_cloud/resources/data_sources/client.py,sha256=TWnJVFLDrGoDaNy7bvYXc_pUprl8fwrMhmEtakP498Y,21479
24
+ llama_cloud/resources/data_sources/client.py,sha256=m6cu5eMJyoatzqNY7vjMa0SGj53sRZVY8pIKPyGevR0,21455
25
25
  llama_cloud/resources/data_sources/types/__init__.py,sha256=iOdDXvAM6w80PR62JCscsTOwzDIXHHcG_Ypv18DEdic,410
26
26
  llama_cloud/resources/data_sources/types/data_source_update_component.py,sha256=8MoJgdjYmN5WqntDpMXX34WJsf-Wsn0gYw_0t9SOTTA,257
27
27
  llama_cloud/resources/data_sources/types/data_source_update_component_one.py,sha256=jfHjlwkUonW0Z73XhJ3w0BZpmptuXU205FWXS1Ucf44,742
28
28
  llama_cloud/resources/data_sources/types/data_source_update_custom_metadata_value.py,sha256=3aFC-p8MSxjhOu2nFtqk0pixj6RqNqcFnbOYngUdZUk,215
29
29
  llama_cloud/resources/deprecated/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
30
- llama_cloud/resources/deprecated/client.py,sha256=wLNml7pI11v8u6zU_IBmZzNXsRxQ0Oi-lQXlPwwUWqI,37336
30
+ llama_cloud/resources/deprecated/client.py,sha256=hSGJara04Yg4RcGEKeZ2RjYggLz1OurrBTmZFXt6bfk,37288
31
31
  llama_cloud/resources/evals/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
32
- llama_cloud/resources/evals/client.py,sha256=4-CHlhLOqKGsygu2doMN-NJMOg6pVrZW0I4Ndv5GNeY,28623
32
+ llama_cloud/resources/evals/client.py,sha256=uCxe3pPXHhNb3uOMzPEugE-SEoxc45SJ6LevaDaIFuk,28583
33
33
  llama_cloud/resources/files/__init__.py,sha256=aZpyTj6KpZvA5XVwmuo1sIlRs7ba98btxVBZ6j5vIsI,155
34
- llama_cloud/resources/files/client.py,sha256=boSn_73Oj6yU9prCmK0qJSugprNKoVgVzOdMb0bfXw8,22770
34
+ llama_cloud/resources/files/client.py,sha256=6b8eM0GfAFQXkKzufpDOfyq19alY9wfcHc8vUm4n7dQ,22746
35
35
  llama_cloud/resources/files/types/__init__.py,sha256=ZWnnYWuDYZSfUJc7Jv3HyovzijdB--DTK4YB-uPcDsA,181
36
36
  llama_cloud/resources/files/types/file_create_resource_info_value.py,sha256=R7Y-CJf7fnbvIqE3xOI5XOrmPwLbVJLC7zpxMu8Zopk,201
37
37
  llama_cloud/resources/parsing/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
38
- llama_cloud/resources/parsing/client.py,sha256=fX12fzfihzpQdBAIpqmVLkoz4pnx_KZCPARPkEBF2Kw,37336
38
+ llama_cloud/resources/parsing/client.py,sha256=2_9U7zElV58GDVeh9eCI-VAYRe5d0CKej_C60_4JEq0,37288
39
39
  llama_cloud/resources/pipelines/__init__.py,sha256=H7yaFIN62vjuhU3TOKzzuf8qpxZRgw1xVa-eyig-2YU,175
40
- llama_cloud/resources/pipelines/client.py,sha256=rJhk3HmFVsYC6Bcs5hlIM7sSgZSViCQ-_1VBgo4jsU0,103757
40
+ llama_cloud/resources/pipelines/client.py,sha256=_ExQjfJ28fPZ5gQXOev8Hi0YhKL6xXuZs5IFVNhM7ZM,106484
41
41
  llama_cloud/resources/pipelines/types/__init__.py,sha256=xuT4OBPLrRfEe-E3UVdJvRjl9jTp7tNBK_YzZBb6Kj8,212
42
42
  llama_cloud/resources/pipelines/types/pipeline_file_update_custom_metadata_value.py,sha256=trI48WLxPcAqV9207Q6-3cj1nl4EGlZpw7En56ZsPgg,217
43
43
  llama_cloud/resources/projects/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
44
- llama_cloud/resources/projects/client.py,sha256=PolxtjYfE4d7sXGkQF_uxTh_WmRVToWnSqwrQQsNvUk,48418
44
+ llama_cloud/resources/projects/client.py,sha256=Ky2bq1O0cPBorT6r2jA_Z_YfbwDE4K0TRWkqVkWh_Hg,48354
45
45
  llama_cloud/types/__init__.py,sha256=xKnYxctZU1sCSVv6_GNg4w6t_o9vJIBTcD9Lqgu32Ts,10347
46
46
  llama_cloud/types/api_key.py,sha256=zEZG3fvKSlKeGkLOuRiMJRh_uv3fED_F3it_21rMHPo,1217
47
47
  llama_cloud/types/azure_open_ai_embedding.py,sha256=zCk8TBWeDHM635nsSmRLSs-GFQWyHo8v7Ovv-8rtgGQ,3346
@@ -167,7 +167,7 @@ llama_cloud/types/token_text_splitter.py,sha256=DxBaqoExtsGeYvUdZPP_LYpEo0_-Oj31
167
167
  llama_cloud/types/transformation_category_names.py,sha256=0xjYe-mDW9OKbTGqL5fSbTvqsfrG4LDu_stW_ubVLl4,582
168
168
  llama_cloud/types/validation_error.py,sha256=2JhGNJouo8QpfrMBoT_JCwYSn1nFN2Nnq0p9uPLDH-U,992
169
169
  llama_cloud/types/validation_error_loc_item.py,sha256=LAtjCHIllWRBFXvAZ5QZpp7CPXjdtN9EB7HrLVo6EP0,128
170
- llama_cloud-0.0.1.dist-info/LICENSE,sha256=_iNqtPcw1Ue7dZKwOwgPtbegMUkWVy15hC7bffAdNmY,1067
171
- llama_cloud-0.0.1.dist-info/METADATA,sha256=aolJTigfLCa_Mi8EXzWa6ON3dz0sG3KvJ8LH7WCURDw,750
172
- llama_cloud-0.0.1.dist-info/WHEEL,sha256=FMvqSimYX_P7y0a7UY-_Mc83r5zkBZsCYPm7Lr0Bsq4,88
173
- llama_cloud-0.0.1.dist-info/RECORD,,
170
+ llama_cloud-0.0.2.dist-info/LICENSE,sha256=_iNqtPcw1Ue7dZKwOwgPtbegMUkWVy15hC7bffAdNmY,1067
171
+ llama_cloud-0.0.2.dist-info/METADATA,sha256=trCof17Q1p97KJ_fY_RoyqWbgo9fyZhOmtienxOsdno,750
172
+ llama_cloud-0.0.2.dist-info/WHEEL,sha256=FMvqSimYX_P7y0a7UY-_Mc83r5zkBZsCYPm7Lr0Bsq4,88
173
+ llama_cloud-0.0.2.dist-info/RECORD,,