llama-cloud 0.1.38__py3-none-any.whl → 0.1.39__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of llama-cloud might be problematic. Click here for more details.

@@ -39,7 +39,6 @@ class RetrieversClient:
39
39
  name: typing.Optional[str] = None,
40
40
  project_id: typing.Optional[str] = None,
41
41
  organization_id: typing.Optional[str] = None,
42
- project_id: typing.Optional[str] = None,
43
42
  ) -> typing.List[Retriever]:
44
43
  """
45
44
  List Retrievers for a project.
@@ -50,8 +49,6 @@ class RetrieversClient:
50
49
  - project_id: typing.Optional[str].
51
50
 
52
51
  - organization_id: typing.Optional[str].
53
-
54
- - project_id: typing.Optional[str].
55
52
  ---
56
53
  from llama_cloud.client import LlamaCloud
57
54
 
@@ -64,7 +61,7 @@ class RetrieversClient:
64
61
  "GET",
65
62
  urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/retrievers"),
66
63
  params=remove_none_from_dict({"name": name, "project_id": project_id, "organization_id": organization_id}),
67
- headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
64
+ headers=self._client_wrapper.get_headers(),
68
65
  timeout=60,
69
66
  )
70
67
  if 200 <= _response.status_code < 300:
@@ -83,7 +80,6 @@ class RetrieversClient:
83
80
  project_id: typing.Optional[str] = None,
84
81
  organization_id: typing.Optional[str] = None,
85
82
  request: RetrieverCreate,
86
- project_id: typing.Optional[str] = None,
87
83
  ) -> Retriever:
88
84
  """
89
85
  Create a new Retriever.
@@ -94,8 +90,6 @@ class RetrieversClient:
94
90
  - organization_id: typing.Optional[str].
95
91
 
96
92
  - request: RetrieverCreate.
97
-
98
- - project_id: typing.Optional[str].
99
93
  ---
100
94
  from llama_cloud import RetrieverCreate
101
95
  from llama_cloud.client import LlamaCloud
@@ -114,7 +108,7 @@ class RetrieversClient:
114
108
  urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/retrievers"),
115
109
  params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
116
110
  json=jsonable_encoder(request),
117
- headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
111
+ headers=self._client_wrapper.get_headers(),
118
112
  timeout=60,
119
113
  )
120
114
  if 200 <= _response.status_code < 300:
@@ -133,7 +127,6 @@ class RetrieversClient:
133
127
  project_id: typing.Optional[str] = None,
134
128
  organization_id: typing.Optional[str] = None,
135
129
  request: RetrieverCreate,
136
- project_id: typing.Optional[str] = None,
137
130
  ) -> Retriever:
138
131
  """
139
132
  Upsert a new Retriever.
@@ -144,8 +137,6 @@ class RetrieversClient:
144
137
  - organization_id: typing.Optional[str].
145
138
 
146
139
  - request: RetrieverCreate.
147
-
148
- - project_id: typing.Optional[str].
149
140
  ---
150
141
  from llama_cloud import RetrieverCreate
151
142
  from llama_cloud.client import LlamaCloud
@@ -164,7 +155,7 @@ class RetrieversClient:
164
155
  urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/retrievers"),
165
156
  params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
166
157
  json=jsonable_encoder(request),
167
- headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
158
+ headers=self._client_wrapper.get_headers(),
168
159
  timeout=60,
169
160
  )
170
161
  if 200 <= _response.status_code < 300:
@@ -183,7 +174,6 @@ class RetrieversClient:
183
174
  *,
184
175
  project_id: typing.Optional[str] = None,
185
176
  organization_id: typing.Optional[str] = None,
186
- project_id: typing.Optional[str] = None,
187
177
  ) -> Retriever:
188
178
  """
189
179
  Get a Retriever by ID.
@@ -194,8 +184,6 @@ class RetrieversClient:
194
184
  - project_id: typing.Optional[str].
195
185
 
196
186
  - organization_id: typing.Optional[str].
197
-
198
- - project_id: typing.Optional[str].
199
187
  ---
200
188
  from llama_cloud.client import LlamaCloud
201
189
 
@@ -210,7 +198,7 @@ class RetrieversClient:
210
198
  "GET",
211
199
  urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", f"api/v1/retrievers/{retriever_id}"),
212
200
  params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
213
- headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
201
+ headers=self._client_wrapper.get_headers(),
214
202
  timeout=60,
215
203
  )
216
204
  if 200 <= _response.status_code < 300:
@@ -313,7 +301,6 @@ class RetrieversClient:
313
301
  rerank_top_n: typing.Optional[int] = OMIT,
314
302
  rerank_config: typing.Optional[ReRankConfig] = OMIT,
315
303
  query: str,
316
- project_id: typing.Optional[str] = None,
317
304
  ) -> CompositeRetrievalResult:
318
305
  """
319
306
  Retrieve data using a Retriever.
@@ -332,8 +319,6 @@ class RetrieversClient:
332
319
  - rerank_config: typing.Optional[ReRankConfig]. The rerank configuration for composite retrieval.
333
320
 
334
321
  - query: str. The query to retrieve against.
335
-
336
- - project_id: typing.Optional[str].
337
322
  ---
338
323
  from llama_cloud import CompositeRetrievalMode, ReRankConfig, ReRankerType
339
324
  from llama_cloud.client import LlamaCloud
@@ -364,7 +349,7 @@ class RetrieversClient:
364
349
  ),
365
350
  params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
366
351
  json=jsonable_encoder(_request),
367
- headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
352
+ headers=self._client_wrapper.get_headers(),
368
353
  timeout=60,
369
354
  )
370
355
  if 200 <= _response.status_code < 300:
@@ -387,7 +372,6 @@ class RetrieversClient:
387
372
  rerank_config: typing.Optional[ReRankConfig] = OMIT,
388
373
  query: str,
389
374
  pipelines: typing.Optional[typing.List[RetrieverPipeline]] = OMIT,
390
- project_id: typing.Optional[str] = None,
391
375
  ) -> CompositeRetrievalResult:
392
376
  """
393
377
  Retrieve data using specified pipelines without creating a persistent retriever.
@@ -406,8 +390,6 @@ class RetrieversClient:
406
390
  - query: str. The query to retrieve against.
407
391
 
408
392
  - pipelines: typing.Optional[typing.List[RetrieverPipeline]]. The pipelines to use for retrieval.
409
-
410
- - project_id: typing.Optional[str].
411
393
  ---
412
394
  from llama_cloud import CompositeRetrievalMode, ReRankConfig, ReRankerType
413
395
  from llama_cloud.client import LlamaCloud
@@ -437,7 +419,7 @@ class RetrieversClient:
437
419
  urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/retrievers/retrieve"),
438
420
  params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
439
421
  json=jsonable_encoder(_request),
440
- headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
422
+ headers=self._client_wrapper.get_headers(),
441
423
  timeout=60,
442
424
  )
443
425
  if 200 <= _response.status_code < 300:
@@ -461,7 +443,6 @@ class AsyncRetrieversClient:
461
443
  name: typing.Optional[str] = None,
462
444
  project_id: typing.Optional[str] = None,
463
445
  organization_id: typing.Optional[str] = None,
464
- project_id: typing.Optional[str] = None,
465
446
  ) -> typing.List[Retriever]:
466
447
  """
467
448
  List Retrievers for a project.
@@ -472,8 +453,6 @@ class AsyncRetrieversClient:
472
453
  - project_id: typing.Optional[str].
473
454
 
474
455
  - organization_id: typing.Optional[str].
475
-
476
- - project_id: typing.Optional[str].
477
456
  ---
478
457
  from llama_cloud.client import AsyncLlamaCloud
479
458
 
@@ -486,7 +465,7 @@ class AsyncRetrieversClient:
486
465
  "GET",
487
466
  urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/retrievers"),
488
467
  params=remove_none_from_dict({"name": name, "project_id": project_id, "organization_id": organization_id}),
489
- headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
468
+ headers=self._client_wrapper.get_headers(),
490
469
  timeout=60,
491
470
  )
492
471
  if 200 <= _response.status_code < 300:
@@ -505,7 +484,6 @@ class AsyncRetrieversClient:
505
484
  project_id: typing.Optional[str] = None,
506
485
  organization_id: typing.Optional[str] = None,
507
486
  request: RetrieverCreate,
508
- project_id: typing.Optional[str] = None,
509
487
  ) -> Retriever:
510
488
  """
511
489
  Create a new Retriever.
@@ -516,8 +494,6 @@ class AsyncRetrieversClient:
516
494
  - organization_id: typing.Optional[str].
517
495
 
518
496
  - request: RetrieverCreate.
519
-
520
- - project_id: typing.Optional[str].
521
497
  ---
522
498
  from llama_cloud import RetrieverCreate
523
499
  from llama_cloud.client import AsyncLlamaCloud
@@ -536,7 +512,7 @@ class AsyncRetrieversClient:
536
512
  urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/retrievers"),
537
513
  params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
538
514
  json=jsonable_encoder(request),
539
- headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
515
+ headers=self._client_wrapper.get_headers(),
540
516
  timeout=60,
541
517
  )
542
518
  if 200 <= _response.status_code < 300:
@@ -555,7 +531,6 @@ class AsyncRetrieversClient:
555
531
  project_id: typing.Optional[str] = None,
556
532
  organization_id: typing.Optional[str] = None,
557
533
  request: RetrieverCreate,
558
- project_id: typing.Optional[str] = None,
559
534
  ) -> Retriever:
560
535
  """
561
536
  Upsert a new Retriever.
@@ -566,8 +541,6 @@ class AsyncRetrieversClient:
566
541
  - organization_id: typing.Optional[str].
567
542
 
568
543
  - request: RetrieverCreate.
569
-
570
- - project_id: typing.Optional[str].
571
544
  ---
572
545
  from llama_cloud import RetrieverCreate
573
546
  from llama_cloud.client import AsyncLlamaCloud
@@ -586,7 +559,7 @@ class AsyncRetrieversClient:
586
559
  urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/retrievers"),
587
560
  params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
588
561
  json=jsonable_encoder(request),
589
- headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
562
+ headers=self._client_wrapper.get_headers(),
590
563
  timeout=60,
591
564
  )
592
565
  if 200 <= _response.status_code < 300:
@@ -605,7 +578,6 @@ class AsyncRetrieversClient:
605
578
  *,
606
579
  project_id: typing.Optional[str] = None,
607
580
  organization_id: typing.Optional[str] = None,
608
- project_id: typing.Optional[str] = None,
609
581
  ) -> Retriever:
610
582
  """
611
583
  Get a Retriever by ID.
@@ -616,8 +588,6 @@ class AsyncRetrieversClient:
616
588
  - project_id: typing.Optional[str].
617
589
 
618
590
  - organization_id: typing.Optional[str].
619
-
620
- - project_id: typing.Optional[str].
621
591
  ---
622
592
  from llama_cloud.client import AsyncLlamaCloud
623
593
 
@@ -632,7 +602,7 @@ class AsyncRetrieversClient:
632
602
  "GET",
633
603
  urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", f"api/v1/retrievers/{retriever_id}"),
634
604
  params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
635
- headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
605
+ headers=self._client_wrapper.get_headers(),
636
606
  timeout=60,
637
607
  )
638
608
  if 200 <= _response.status_code < 300:
@@ -735,7 +705,6 @@ class AsyncRetrieversClient:
735
705
  rerank_top_n: typing.Optional[int] = OMIT,
736
706
  rerank_config: typing.Optional[ReRankConfig] = OMIT,
737
707
  query: str,
738
- project_id: typing.Optional[str] = None,
739
708
  ) -> CompositeRetrievalResult:
740
709
  """
741
710
  Retrieve data using a Retriever.
@@ -754,8 +723,6 @@ class AsyncRetrieversClient:
754
723
  - rerank_config: typing.Optional[ReRankConfig]. The rerank configuration for composite retrieval.
755
724
 
756
725
  - query: str. The query to retrieve against.
757
-
758
- - project_id: typing.Optional[str].
759
726
  ---
760
727
  from llama_cloud import CompositeRetrievalMode, ReRankConfig, ReRankerType
761
728
  from llama_cloud.client import AsyncLlamaCloud
@@ -786,7 +753,7 @@ class AsyncRetrieversClient:
786
753
  ),
787
754
  params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
788
755
  json=jsonable_encoder(_request),
789
- headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
756
+ headers=self._client_wrapper.get_headers(),
790
757
  timeout=60,
791
758
  )
792
759
  if 200 <= _response.status_code < 300:
@@ -809,7 +776,6 @@ class AsyncRetrieversClient:
809
776
  rerank_config: typing.Optional[ReRankConfig] = OMIT,
810
777
  query: str,
811
778
  pipelines: typing.Optional[typing.List[RetrieverPipeline]] = OMIT,
812
- project_id: typing.Optional[str] = None,
813
779
  ) -> CompositeRetrievalResult:
814
780
  """
815
781
  Retrieve data using specified pipelines without creating a persistent retriever.
@@ -828,8 +794,6 @@ class AsyncRetrieversClient:
828
794
  - query: str. The query to retrieve against.
829
795
 
830
796
  - pipelines: typing.Optional[typing.List[RetrieverPipeline]]. The pipelines to use for retrieval.
831
-
832
- - project_id: typing.Optional[str].
833
797
  ---
834
798
  from llama_cloud import CompositeRetrievalMode, ReRankConfig, ReRankerType
835
799
  from llama_cloud.client import AsyncLlamaCloud
@@ -859,7 +823,7 @@ class AsyncRetrieversClient:
859
823
  urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/retrievers/retrieve"),
860
824
  params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
861
825
  json=jsonable_encoder(_request),
862
- headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
826
+ headers=self._client_wrapper.get_headers(),
863
827
  timeout=60,
864
828
  )
865
829
  if 200 <= _response.status_code < 300:
llama_cloud/types/file.py CHANGED
@@ -25,7 +25,7 @@ class File(pydantic.BaseModel):
25
25
  created_at: typing.Optional[dt.datetime]
26
26
  updated_at: typing.Optional[dt.datetime]
27
27
  name: str
28
- external_file_id: str = pydantic.Field(description="The ID of the file in the external system")
28
+ external_file_id: typing.Optional[str]
29
29
  file_size: typing.Optional[int]
30
30
  file_type: typing.Optional[str]
31
31
  project_id: str = pydantic.Field(description="The ID of the project that the file belongs to")
@@ -31,14 +31,14 @@ class PipelineFile(pydantic.BaseModel):
31
31
  external_file_id: typing.Optional[str]
32
32
  file_size: typing.Optional[int]
33
33
  file_type: typing.Optional[str]
34
- project_id: str = pydantic.Field(description="The ID of the project that the file belongs to")
34
+ project_id: typing.Optional[str]
35
35
  last_modified_at: typing.Optional[dt.datetime]
36
- resource_info: typing.Optional[typing.Dict[str, typing.Optional[PipelineFileResourceInfoValue]]]
37
- permission_info: typing.Optional[typing.Dict[str, typing.Optional[PipelineFilePermissionInfoValue]]]
38
- data_source_id: typing.Optional[str]
39
36
  file_id: typing.Optional[str]
40
37
  pipeline_id: str = pydantic.Field(description="The ID of the pipeline that the file is associated with")
38
+ resource_info: typing.Optional[typing.Dict[str, typing.Optional[PipelineFileResourceInfoValue]]]
39
+ permission_info: typing.Optional[typing.Dict[str, typing.Optional[PipelineFilePermissionInfoValue]]]
41
40
  custom_metadata: typing.Optional[typing.Dict[str, typing.Optional[PipelineFileCustomMetadataValue]]]
41
+ data_source_id: typing.Optional[str]
42
42
  config_hash: typing.Optional[typing.Dict[str, typing.Optional[PipelineFileConfigHashValue]]]
43
43
  indexed_page_count: typing.Optional[int]
44
44
  status: typing.Optional[PipelineFileStatus]
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: llama-cloud
3
- Version: 0.1.38
3
+ Version: 0.1.39
4
4
  Summary:
5
5
  License: MIT
6
6
  Author: Logan Markewich
@@ -15,36 +15,36 @@ llama_cloud/resources/admin/client.py,sha256=YIYy9kU1_xaE0gkpmZZbCgLzZj6XSrAUplS
15
15
  llama_cloud/resources/agent_deployments/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
16
16
  llama_cloud/resources/agent_deployments/client.py,sha256=3EOzOjmRs4KISgJ566enq3FCuN3YtskjO0OHqQGtkQ0,6122
17
17
  llama_cloud/resources/alpha/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
18
- llama_cloud/resources/alpha/client.py,sha256=9n--BGk6HStU79xHPsmAIgWUMj3m8m8l3jAG9fDUK4E,4624
18
+ llama_cloud/resources/alpha/client.py,sha256=d5cRIUykNpnVryuxWBPUpmo-2L1vMIDeZIF3DvTIx7E,4322
19
19
  llama_cloud/resources/beta/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
20
- llama_cloud/resources/beta/client.py,sha256=ItS2GfQe7EcchBq8uhJavBTetxDRYHDiyC_D5AwqCpY,66950
20
+ llama_cloud/resources/beta/client.py,sha256=_GNkHQxyZxhZOkLIRzfCw6PexQx-E8r_7R-3Wd9Y0uE,63128
21
21
  llama_cloud/resources/chat_apps/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
22
- llama_cloud/resources/chat_apps/client.py,sha256=UMQFuJ1Ia-p8URR_eSUa18heK8jQ3j2pUJB_ply4P-E,24551
22
+ llama_cloud/resources/chat_apps/client.py,sha256=orSI8rpQbUwVEToolEeiEi5Qe--suXFvfu6D9JDii5I,23595
23
23
  llama_cloud/resources/classifier/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
24
- llama_cloud/resources/classifier/client.py,sha256=XUK4JIMszQcRpPgxzavvCWr-4zQeoyaJLIK5ABSEHug,18826
24
+ llama_cloud/resources/classifier/client.py,sha256=Q1vdRTSe-QnZyFucTlxHIRmdCGa-PFf7XUazik7s59E,17618
25
25
  llama_cloud/resources/data_sinks/__init__.py,sha256=ZHUjn3HbKhq_7QS1q74r2m5RGKF5lxcvF2P6pGvpcis,147
26
- llama_cloud/resources/data_sinks/client.py,sha256=6CEYttEJI3FtJ_OBOYvk5mYwAqYzRAKctJzjNrMS5Is,21595
26
+ llama_cloud/resources/data_sinks/client.py,sha256=GpD6FhbGqkg2oUToyMG6J8hPxG_iG7W5ZJRo0qg3yzk,20639
27
27
  llama_cloud/resources/data_sinks/types/__init__.py,sha256=M1aTcufJwiEZo9B0KmYj9PfkSd6I1ooFt9tpIRGwgg8,168
28
28
  llama_cloud/resources/data_sinks/types/data_sink_update_component.py,sha256=ynPdEg844hZaD6EcAK0jrMY_vogtvmLTZ7FZSwWcor8,912
29
29
  llama_cloud/resources/data_sources/__init__.py,sha256=McURkcNBGHXH1hmRDRmZI1dRzJrekCTHZsgv03r2oZI,227
30
- llama_cloud/resources/data_sources/client.py,sha256=BB0jIbfprU23rejGs8OZDO6G-E6A0RVck7v_IKENa_A,22811
30
+ llama_cloud/resources/data_sources/client.py,sha256=SZFm8bW5nkaXringdSnmxHqvVjKM7cNNOtqVXjgTKhc,21855
31
31
  llama_cloud/resources/data_sources/types/__init__.py,sha256=Cd5xEECTzXqQSfJALfJPSjudlSLeb3RENeJVi8vwPbM,303
32
32
  llama_cloud/resources/data_sources/types/data_source_update_component.py,sha256=_jQY6FhcvenWdzi27SK1bSY8muXKLRkXlVrTqEWgKKc,1159
33
33
  llama_cloud/resources/data_sources/types/data_source_update_custom_metadata_value.py,sha256=3aFC-p8MSxjhOu2nFtqk0pixj6RqNqcFnbOYngUdZUk,215
34
34
  llama_cloud/resources/embedding_model_configs/__init__.py,sha256=cXDtKKq-gj7yjFjdQ5GrGyPs-T5tRV_0JjUMGlAbdUs,1115
35
- llama_cloud/resources/embedding_model_configs/client.py,sha256=Sy6B8clDN7eByYC7AP4hnX1Pt5StkHC0swznDBTT11k,19329
35
+ llama_cloud/resources/embedding_model_configs/client.py,sha256=2JDvZJtSger9QJ8luPct-2zvwjaJAR8VcKsTZ1wgYTE,17769
36
36
  llama_cloud/resources/embedding_model_configs/types/__init__.py,sha256=6-rcDwJhw_0shz3CjrPvlYBYXJJ1bLn-PpplhOsQ79w,1156
37
37
  llama_cloud/resources/embedding_model_configs/types/embedding_model_config_create_embedding_config.py,sha256=SQCHJk0AmBbKS5XKdcEJxhDhIMLQCmCI13IHC28v7vQ,3054
38
38
  llama_cloud/resources/evals/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
39
39
  llama_cloud/resources/evals/client.py,sha256=v2AyeQV0hVgC6xoP2gJNgneJMaeXALV1hIeirYGxlPw,3242
40
40
  llama_cloud/resources/files/__init__.py,sha256=Ws53l-S3kyAGFinYPOb9WpN84DtbFn6gLYZtI2akBLQ,169
41
- llama_cloud/resources/files/client.py,sha256=jSeVm7WCI5fPeLW8xlLKfqw2H5WyQvezgOjlTiDY-rc,61136
41
+ llama_cloud/resources/files/client.py,sha256=Crd0IR0cV5fld4jUGAHE8VsIbw7vCYrOIyBTSwDyitA,56242
42
42
  llama_cloud/resources/files/types/__init__.py,sha256=ZZuDQsYsxmQ9VwpfN7oqftzGRnFTR2EMYdCa7zARo4g,204
43
43
  llama_cloud/resources/files/types/file_create_from_url_resource_info_value.py,sha256=Wc8wFgujOO5pZvbbh2TMMzpa37GKZd14GYNJ9bdq7BE,214
44
44
  llama_cloud/resources/jobs/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
45
- llama_cloud/resources/jobs/client.py,sha256=imoGq17ICMxdJ1cfiNGSkkHqYMmBTjYJ_4uNSa_Q8AQ,6504
45
+ llama_cloud/resources/jobs/client.py,sha256=b2R_Oj2OCtcv-IIJNz9aq42hDgrOk_huqTSJhTB9VaA,6202
46
46
  llama_cloud/resources/llama_extract/__init__.py,sha256=V6VZ8hQXwAuvOOZyk43nnbINoDQqEr03AjKQPhYKluk,997
47
- llama_cloud/resources/llama_extract/client.py,sha256=CT5_Nv6Z3aLejsr6yX2Ic9FlZZ55zjjrt8mp2R9v4k8,86496
47
+ llama_cloud/resources/llama_extract/client.py,sha256=B_qhVsk-Qs81qrFOVgWqcvelSB3TLWFJCibnn--3BjE,83096
48
48
  llama_cloud/resources/llama_extract/types/__init__.py,sha256=2Iu4w5LXZY2Govr1RzahIfY0b84y658SQjMDtj7rH_0,1497
49
49
  llama_cloud/resources/llama_extract/types/extract_agent_create_data_schema.py,sha256=zB31hJQ8hKaIsPkfTWiX5hqsPVFMyyeWEDZ_Aq237jo,305
50
50
  llama_cloud/resources/llama_extract/types/extract_agent_create_data_schema_zero_value.py,sha256=xoyXH3f0Y5beMWBxmtXSz6QoB_df_-0QBsYdjBhZnGw,217
@@ -57,24 +57,24 @@ llama_cloud/resources/llama_extract/types/extract_schema_validate_request_data_s
57
57
  llama_cloud/resources/llama_extract/types/extract_stateless_request_data_schema.py,sha256=lBblR9zgjJsbWL-2bDisCj7EQiX6aky6GQ4tuMr3LtU,325
58
58
  llama_cloud/resources/llama_extract/types/extract_stateless_request_data_schema_zero_value.py,sha256=4-ONLmkrEP36ZH0qRXp3sbXCtLVNQQX4dLXFeF4u47g,222
59
59
  llama_cloud/resources/organizations/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
60
- llama_cloud/resources/organizations/client.py,sha256=57FNu1BQOmps5VXA_YZuYax3XZcbDcIAo1f0jXKzwWA,56991
60
+ llama_cloud/resources/organizations/client.py,sha256=RoN-nkN7VeRZnrrElXhaPrgQFzGMHgNY41_XpbCXP0g,56623
61
61
  llama_cloud/resources/parsing/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
62
- llama_cloud/resources/parsing/client.py,sha256=mPLm3eROPEk8rLzssT0fm9rTcQJ2-ZxlXLNAOzqJ1T8,89599
62
+ llama_cloud/resources/parsing/client.py,sha256=K9qYu5v99OqbMmeeeCioQDUnT085A5keCencSwES-gc,88995
63
63
  llama_cloud/resources/pipelines/__init__.py,sha256=zyvVEOF_krvEZkCIj_kZoMKfhDqHo_R32a1mv9CriQc,1193
64
- llama_cloud/resources/pipelines/client.py,sha256=OY9ck8_raZNrg_BumBqEyW7zBy_FRAbUvh75qG7Soos,135946
64
+ llama_cloud/resources/pipelines/client.py,sha256=VAqAm0oY_nXGkMPqXuzPEHS9kPtpuOE5sxfyqlzXuSI,134738
65
65
  llama_cloud/resources/pipelines/types/__init__.py,sha256=C68NQ5QzA0dFXf9oePFFGmV1vn96jcAp-QAznSgoRYQ,1375
66
66
  llama_cloud/resources/pipelines/types/pipeline_file_update_custom_metadata_value.py,sha256=trI48WLxPcAqV9207Q6-3cj1nl4EGlZpw7En56ZsPgg,217
67
67
  llama_cloud/resources/pipelines/types/pipeline_update_embedding_config.py,sha256=c8FF64fDrBMX_2RX4uY3CjbNc0Ss_AUJ4Eqs-KeV4Wc,2874
68
68
  llama_cloud/resources/pipelines/types/pipeline_update_transform_config.py,sha256=KbkyULMv-qeS3qRd31ia6pd5rOdypS0o2UL42NRcA7E,321
69
69
  llama_cloud/resources/pipelines/types/retrieval_params_search_filters_inference_schema_value.py,sha256=hZWXYlTib0af85ECcerC4xD-bUQe8rG3Q6G1jFTMQcI,228
70
70
  llama_cloud/resources/projects/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
71
- llama_cloud/resources/projects/client.py,sha256=nRwK-d56jBHMWsc0QtERnaHNlaIZi7ppbTwLjSc_A08,27158
71
+ llama_cloud/resources/projects/client.py,sha256=PF36iWtSa5amUt3q56YwLypOZjclIXSubCRv9NttpLs,25404
72
72
  llama_cloud/resources/reports/__init__.py,sha256=cruYbQ1bIuJbRpkfaQY7ajUEslffjd7KzvzMzbtPH94,217
73
- llama_cloud/resources/reports/client.py,sha256=2YjrKgOoCs75Tw4y-JbN7OOL5yMsiybrtDzi1Upc3IE,50558
73
+ llama_cloud/resources/reports/client.py,sha256=kHjtXVVc1Xi3T1GyBvSW5K4mTdr6xQwZA3vw-liRKBg,46736
74
74
  llama_cloud/resources/reports/types/__init__.py,sha256=LfwDYrI4RcQu-o42iAe7HkcwHww2YU90lOonBPTmZIk,291
75
75
  llama_cloud/resources/reports/types/update_report_plan_api_v_1_reports_report_id_plan_patch_request_action.py,sha256=Qh-MSeRvDBfNb5hoLELivv1pLtrYVf52WVoP7G8V34A,807
76
76
  llama_cloud/resources/retrievers/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
77
- llama_cloud/resources/retrievers/client.py,sha256=pG9G6tgcohh0xHviaaZoNuQ6JJUhWq2uwrTK0-d3p_4,33938
77
+ llama_cloud/resources/retrievers/client.py,sha256=z2LhmA-cZVFzr9P6loeCZYnJbvSIk0QitFeVFp-IyZk,32126
78
78
  llama_cloud/resources/users/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
79
79
  llama_cloud/resources/users/client.py,sha256=A2s8e2syQHkkSwPz-Lrt_Zxp1K-8nqJqj5EafE6NWYs,5545
80
80
  llama_cloud/types/__init__.py,sha256=HSHS4fNorQ2l79VYOegTeh4KG7LnIzpSvUOhTELh3pQ,33920
@@ -194,7 +194,7 @@ llama_cloud/types/extract_state.py,sha256=TNeVAXXKZaiM2srlbQlzRSn4_TDpR4xyT_yQhJ
194
194
  llama_cloud/types/extract_target.py,sha256=Gt-FNqblzcjdfq1hxsqEjWWu-HNLXdKy4w98nog52Ms,478
195
195
  llama_cloud/types/fail_page_mode.py,sha256=n4fgPpiEB5siPoEg0Sux4COg7ElNybjshxDoUihZwRU,786
196
196
  llama_cloud/types/failure_handling_config.py,sha256=EmAQW0qm7-JTSYFwhmIWxqkVNWym_AyAJIMEmeI9Cqc,1216
197
- llama_cloud/types/file.py,sha256=rQXitPRKOYw91nK5qOZ0vpOmIx_MCpRb0g78d9dQs6w,1822
197
+ llama_cloud/types/file.py,sha256=sXdF-cdHL3k1-DPIxAjYpb-kNHzcOAV_earVoYITzUA,1765
198
198
  llama_cloud/types/file_classification.py,sha256=jKzAc_3rg0Usyf3TNr-bI5HZn9zGIj9vYH90RKoDtiY,1418
199
199
  llama_cloud/types/file_count_by_status_response.py,sha256=WuorbZvKjDs9Ql1hUiQu4gN5iCm8d6fr92KLyHpRvQU,1356
200
200
  llama_cloud/types/file_create.py,sha256=eLUC50CzXOdAR_P2mBtX_R7kGteIVbP1V3LzuP1s0Xs,1629
@@ -308,7 +308,7 @@ llama_cloud/types/pipeline_data_source_custom_metadata_value.py,sha256=8n3r60sxM
308
308
  llama_cloud/types/pipeline_data_source_status.py,sha256=BD4xoftwp9lWC8EjJTnf3boIG_AyzjLPuP4qJxGhmcc,1039
309
309
  llama_cloud/types/pipeline_deployment.py,sha256=eVBrz032aPb2cqtIIVYT5MTHQvBNm89XazoNrRWVugo,1356
310
310
  llama_cloud/types/pipeline_embedding_config.py,sha256=7NJzlabQLFUFsvj7fye-oKLPasaXCWJBm-XuLxy-xmQ,3112
311
- llama_cloud/types/pipeline_file.py,sha256=4t4Xbszsbcjik58NINofsA4nrxmAI-pVu1LyqTiALgk,2572
311
+ llama_cloud/types/pipeline_file.py,sha256=zoE1A4pdD7S4cgDtR_aVToQ08JDt_siUJTYsdayG8s4,2510
312
312
  llama_cloud/types/pipeline_file_config_hash_value.py,sha256=4lvLnDpzNAHdiMkGJTTNDTu3p3H7Nxw5MR1Mzte7-_M,201
313
313
  llama_cloud/types/pipeline_file_create.py,sha256=yoMIzWED0ktKerE48kgzInBa3d0aNGO5JjTtDTDAn4A,1310
314
314
  llama_cloud/types/pipeline_file_create_custom_metadata_value.py,sha256=olVj5yhQFx1QqWO1Wv9d6AtL-YyYO9_OYtOfcD2ZeGY,217
@@ -398,7 +398,7 @@ llama_cloud/types/vertex_embedding_mode.py,sha256=yY23FjuWU_DkXjBb3JoKV4SCMqel2B
398
398
  llama_cloud/types/vertex_text_embedding.py,sha256=-C4fNCYfFl36ATdBMGFVPpiHIKxjk0KB1ERA2Ec20aU,1932
399
399
  llama_cloud/types/webhook_configuration.py,sha256=_Xm15whrWoKNBuCoO5y_NunA-ByhCAYK87LnC4W-Pzg,1350
400
400
  llama_cloud/types/webhook_configuration_webhook_events_item.py,sha256=OL3moFO_6hsKZYSBQBsSHmWA0NgLcLJgBPZfABwT60c,2544
401
- llama_cloud-0.1.38.dist-info/LICENSE,sha256=_iNqtPcw1Ue7dZKwOwgPtbegMUkWVy15hC7bffAdNmY,1067
402
- llama_cloud-0.1.38.dist-info/METADATA,sha256=ptOad5pDHjoeNoQMXsprngoskyNHjb05K-Q6qgPDqx0,1143
403
- llama_cloud-0.1.38.dist-info/WHEEL,sha256=sP946D7jFCHeNz5Iq4fL4Lu-PrWrFsgfLXbbkciIZwg,88
404
- llama_cloud-0.1.38.dist-info/RECORD,,
401
+ llama_cloud-0.1.39.dist-info/LICENSE,sha256=_iNqtPcw1Ue7dZKwOwgPtbegMUkWVy15hC7bffAdNmY,1067
402
+ llama_cloud-0.1.39.dist-info/METADATA,sha256=WNoZded1_WK8ATxqn2o6TaLGwsn2uijfjy07RnZd8bI,1143
403
+ llama_cloud-0.1.39.dist-info/WHEEL,sha256=sP946D7jFCHeNz5Iq4fL4Lu-PrWrFsgfLXbbkciIZwg,88
404
+ llama_cloud-0.1.39.dist-info/RECORD,,