llama-cloud 0.1.36__py3-none-any.whl → 0.1.38__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of llama-cloud might be problematic. Click here for more details.

Files changed (47) hide show
  1. llama_cloud/__init__.py +20 -4
  2. llama_cloud/client.py +3 -0
  3. llama_cloud/resources/__init__.py +3 -3
  4. llama_cloud/resources/admin/client.py +57 -0
  5. llama_cloud/resources/alpha/__init__.py +2 -0
  6. llama_cloud/resources/alpha/client.py +118 -0
  7. llama_cloud/resources/beta/client.py +576 -20
  8. llama_cloud/resources/chat_apps/client.py +32 -8
  9. llama_cloud/resources/classifier/client.py +139 -11
  10. llama_cloud/resources/data_sinks/client.py +32 -8
  11. llama_cloud/resources/data_sources/client.py +32 -8
  12. llama_cloud/resources/data_sources/types/data_source_update_component.py +2 -0
  13. llama_cloud/resources/embedding_model_configs/client.py +48 -12
  14. llama_cloud/resources/files/__init__.py +2 -2
  15. llama_cloud/resources/files/client.py +189 -113
  16. llama_cloud/resources/files/types/__init__.py +1 -3
  17. llama_cloud/resources/jobs/client.py +12 -6
  18. llama_cloud/resources/llama_extract/client.py +138 -32
  19. llama_cloud/resources/organizations/client.py +18 -4
  20. llama_cloud/resources/parsing/client.py +16 -4
  21. llama_cloud/resources/pipelines/client.py +32 -8
  22. llama_cloud/resources/projects/client.py +78 -18
  23. llama_cloud/resources/reports/client.py +126 -30
  24. llama_cloud/resources/retrievers/client.py +48 -12
  25. llama_cloud/types/__init__.py +20 -2
  26. llama_cloud/types/agent_deployment_summary.py +1 -0
  27. llama_cloud/types/classify_job.py +2 -0
  28. llama_cloud/types/cloud_jira_data_source_v_2.py +52 -0
  29. llama_cloud/types/cloud_jira_data_source_v_2_api_version.py +21 -0
  30. llama_cloud/types/configurable_data_source_names.py +4 -0
  31. llama_cloud/types/data_source_component.py +2 -0
  32. llama_cloud/types/data_source_create_component.py +2 -0
  33. llama_cloud/types/data_source_reader_version_metadata_reader_version.py +9 -1
  34. llama_cloud/types/file_create.py +41 -0
  35. llama_cloud/types/{classify_job_with_status.py → file_filter.py} +8 -15
  36. llama_cloud/types/file_query_response.py +38 -0
  37. llama_cloud/types/llama_extract_mode_availability.py +37 -0
  38. llama_cloud/types/llama_extract_mode_availability_status.py +17 -0
  39. llama_cloud/types/paginated_response_classify_job.py +34 -0
  40. llama_cloud/types/pipeline_data_source_component.py +2 -0
  41. llama_cloud/types/usage_response_active_alerts_item.py +4 -0
  42. {llama_cloud-0.1.36.dist-info → llama_cloud-0.1.38.dist-info}/METADATA +2 -1
  43. {llama_cloud-0.1.36.dist-info → llama_cloud-0.1.38.dist-info}/RECORD +47 -38
  44. {llama_cloud-0.1.36.dist-info → llama_cloud-0.1.38.dist-info}/WHEEL +1 -1
  45. /llama_cloud/{resources/files/types → types}/file_create_permission_info_value.py +0 -0
  46. /llama_cloud/{resources/files/types → types}/file_create_resource_info_value.py +0 -0
  47. {llama_cloud-0.1.36.dist-info → llama_cloud-0.1.38.dist-info}/LICENSE +0 -0
@@ -39,6 +39,7 @@ class RetrieversClient:
39
39
  name: typing.Optional[str] = None,
40
40
  project_id: typing.Optional[str] = None,
41
41
  organization_id: typing.Optional[str] = None,
42
+ project_id: typing.Optional[str] = None,
42
43
  ) -> typing.List[Retriever]:
43
44
  """
44
45
  List Retrievers for a project.
@@ -49,6 +50,8 @@ class RetrieversClient:
49
50
  - project_id: typing.Optional[str].
50
51
 
51
52
  - organization_id: typing.Optional[str].
53
+
54
+ - project_id: typing.Optional[str].
52
55
  ---
53
56
  from llama_cloud.client import LlamaCloud
54
57
 
@@ -61,7 +64,7 @@ class RetrieversClient:
61
64
  "GET",
62
65
  urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/retrievers"),
63
66
  params=remove_none_from_dict({"name": name, "project_id": project_id, "organization_id": organization_id}),
64
- headers=self._client_wrapper.get_headers(),
67
+ headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
65
68
  timeout=60,
66
69
  )
67
70
  if 200 <= _response.status_code < 300:
@@ -80,6 +83,7 @@ class RetrieversClient:
80
83
  project_id: typing.Optional[str] = None,
81
84
  organization_id: typing.Optional[str] = None,
82
85
  request: RetrieverCreate,
86
+ project_id: typing.Optional[str] = None,
83
87
  ) -> Retriever:
84
88
  """
85
89
  Create a new Retriever.
@@ -90,6 +94,8 @@ class RetrieversClient:
90
94
  - organization_id: typing.Optional[str].
91
95
 
92
96
  - request: RetrieverCreate.
97
+
98
+ - project_id: typing.Optional[str].
93
99
  ---
94
100
  from llama_cloud import RetrieverCreate
95
101
  from llama_cloud.client import LlamaCloud
@@ -108,7 +114,7 @@ class RetrieversClient:
108
114
  urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/retrievers"),
109
115
  params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
110
116
  json=jsonable_encoder(request),
111
- headers=self._client_wrapper.get_headers(),
117
+ headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
112
118
  timeout=60,
113
119
  )
114
120
  if 200 <= _response.status_code < 300:
@@ -127,6 +133,7 @@ class RetrieversClient:
127
133
  project_id: typing.Optional[str] = None,
128
134
  organization_id: typing.Optional[str] = None,
129
135
  request: RetrieverCreate,
136
+ project_id: typing.Optional[str] = None,
130
137
  ) -> Retriever:
131
138
  """
132
139
  Upsert a new Retriever.
@@ -137,6 +144,8 @@ class RetrieversClient:
137
144
  - organization_id: typing.Optional[str].
138
145
 
139
146
  - request: RetrieverCreate.
147
+
148
+ - project_id: typing.Optional[str].
140
149
  ---
141
150
  from llama_cloud import RetrieverCreate
142
151
  from llama_cloud.client import LlamaCloud
@@ -155,7 +164,7 @@ class RetrieversClient:
155
164
  urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/retrievers"),
156
165
  params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
157
166
  json=jsonable_encoder(request),
158
- headers=self._client_wrapper.get_headers(),
167
+ headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
159
168
  timeout=60,
160
169
  )
161
170
  if 200 <= _response.status_code < 300:
@@ -174,6 +183,7 @@ class RetrieversClient:
174
183
  *,
175
184
  project_id: typing.Optional[str] = None,
176
185
  organization_id: typing.Optional[str] = None,
186
+ project_id: typing.Optional[str] = None,
177
187
  ) -> Retriever:
178
188
  """
179
189
  Get a Retriever by ID.
@@ -184,6 +194,8 @@ class RetrieversClient:
184
194
  - project_id: typing.Optional[str].
185
195
 
186
196
  - organization_id: typing.Optional[str].
197
+
198
+ - project_id: typing.Optional[str].
187
199
  ---
188
200
  from llama_cloud.client import LlamaCloud
189
201
 
@@ -198,7 +210,7 @@ class RetrieversClient:
198
210
  "GET",
199
211
  urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", f"api/v1/retrievers/{retriever_id}"),
200
212
  params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
201
- headers=self._client_wrapper.get_headers(),
213
+ headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
202
214
  timeout=60,
203
215
  )
204
216
  if 200 <= _response.status_code < 300:
@@ -301,6 +313,7 @@ class RetrieversClient:
301
313
  rerank_top_n: typing.Optional[int] = OMIT,
302
314
  rerank_config: typing.Optional[ReRankConfig] = OMIT,
303
315
  query: str,
316
+ project_id: typing.Optional[str] = None,
304
317
  ) -> CompositeRetrievalResult:
305
318
  """
306
319
  Retrieve data using a Retriever.
@@ -319,6 +332,8 @@ class RetrieversClient:
319
332
  - rerank_config: typing.Optional[ReRankConfig]. The rerank configuration for composite retrieval.
320
333
 
321
334
  - query: str. The query to retrieve against.
335
+
336
+ - project_id: typing.Optional[str].
322
337
  ---
323
338
  from llama_cloud import CompositeRetrievalMode, ReRankConfig, ReRankerType
324
339
  from llama_cloud.client import LlamaCloud
@@ -349,7 +364,7 @@ class RetrieversClient:
349
364
  ),
350
365
  params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
351
366
  json=jsonable_encoder(_request),
352
- headers=self._client_wrapper.get_headers(),
367
+ headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
353
368
  timeout=60,
354
369
  )
355
370
  if 200 <= _response.status_code < 300:
@@ -372,6 +387,7 @@ class RetrieversClient:
372
387
  rerank_config: typing.Optional[ReRankConfig] = OMIT,
373
388
  query: str,
374
389
  pipelines: typing.Optional[typing.List[RetrieverPipeline]] = OMIT,
390
+ project_id: typing.Optional[str] = None,
375
391
  ) -> CompositeRetrievalResult:
376
392
  """
377
393
  Retrieve data using specified pipelines without creating a persistent retriever.
@@ -390,6 +406,8 @@ class RetrieversClient:
390
406
  - query: str. The query to retrieve against.
391
407
 
392
408
  - pipelines: typing.Optional[typing.List[RetrieverPipeline]]. The pipelines to use for retrieval.
409
+
410
+ - project_id: typing.Optional[str].
393
411
  ---
394
412
  from llama_cloud import CompositeRetrievalMode, ReRankConfig, ReRankerType
395
413
  from llama_cloud.client import LlamaCloud
@@ -419,7 +437,7 @@ class RetrieversClient:
419
437
  urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/retrievers/retrieve"),
420
438
  params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
421
439
  json=jsonable_encoder(_request),
422
- headers=self._client_wrapper.get_headers(),
440
+ headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
423
441
  timeout=60,
424
442
  )
425
443
  if 200 <= _response.status_code < 300:
@@ -443,6 +461,7 @@ class AsyncRetrieversClient:
443
461
  name: typing.Optional[str] = None,
444
462
  project_id: typing.Optional[str] = None,
445
463
  organization_id: typing.Optional[str] = None,
464
+ project_id: typing.Optional[str] = None,
446
465
  ) -> typing.List[Retriever]:
447
466
  """
448
467
  List Retrievers for a project.
@@ -453,6 +472,8 @@ class AsyncRetrieversClient:
453
472
  - project_id: typing.Optional[str].
454
473
 
455
474
  - organization_id: typing.Optional[str].
475
+
476
+ - project_id: typing.Optional[str].
456
477
  ---
457
478
  from llama_cloud.client import AsyncLlamaCloud
458
479
 
@@ -465,7 +486,7 @@ class AsyncRetrieversClient:
465
486
  "GET",
466
487
  urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/retrievers"),
467
488
  params=remove_none_from_dict({"name": name, "project_id": project_id, "organization_id": organization_id}),
468
- headers=self._client_wrapper.get_headers(),
489
+ headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
469
490
  timeout=60,
470
491
  )
471
492
  if 200 <= _response.status_code < 300:
@@ -484,6 +505,7 @@ class AsyncRetrieversClient:
484
505
  project_id: typing.Optional[str] = None,
485
506
  organization_id: typing.Optional[str] = None,
486
507
  request: RetrieverCreate,
508
+ project_id: typing.Optional[str] = None,
487
509
  ) -> Retriever:
488
510
  """
489
511
  Create a new Retriever.
@@ -494,6 +516,8 @@ class AsyncRetrieversClient:
494
516
  - organization_id: typing.Optional[str].
495
517
 
496
518
  - request: RetrieverCreate.
519
+
520
+ - project_id: typing.Optional[str].
497
521
  ---
498
522
  from llama_cloud import RetrieverCreate
499
523
  from llama_cloud.client import AsyncLlamaCloud
@@ -512,7 +536,7 @@ class AsyncRetrieversClient:
512
536
  urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/retrievers"),
513
537
  params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
514
538
  json=jsonable_encoder(request),
515
- headers=self._client_wrapper.get_headers(),
539
+ headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
516
540
  timeout=60,
517
541
  )
518
542
  if 200 <= _response.status_code < 300:
@@ -531,6 +555,7 @@ class AsyncRetrieversClient:
531
555
  project_id: typing.Optional[str] = None,
532
556
  organization_id: typing.Optional[str] = None,
533
557
  request: RetrieverCreate,
558
+ project_id: typing.Optional[str] = None,
534
559
  ) -> Retriever:
535
560
  """
536
561
  Upsert a new Retriever.
@@ -541,6 +566,8 @@ class AsyncRetrieversClient:
541
566
  - organization_id: typing.Optional[str].
542
567
 
543
568
  - request: RetrieverCreate.
569
+
570
+ - project_id: typing.Optional[str].
544
571
  ---
545
572
  from llama_cloud import RetrieverCreate
546
573
  from llama_cloud.client import AsyncLlamaCloud
@@ -559,7 +586,7 @@ class AsyncRetrieversClient:
559
586
  urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/retrievers"),
560
587
  params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
561
588
  json=jsonable_encoder(request),
562
- headers=self._client_wrapper.get_headers(),
589
+ headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
563
590
  timeout=60,
564
591
  )
565
592
  if 200 <= _response.status_code < 300:
@@ -578,6 +605,7 @@ class AsyncRetrieversClient:
578
605
  *,
579
606
  project_id: typing.Optional[str] = None,
580
607
  organization_id: typing.Optional[str] = None,
608
+ project_id: typing.Optional[str] = None,
581
609
  ) -> Retriever:
582
610
  """
583
611
  Get a Retriever by ID.
@@ -588,6 +616,8 @@ class AsyncRetrieversClient:
588
616
  - project_id: typing.Optional[str].
589
617
 
590
618
  - organization_id: typing.Optional[str].
619
+
620
+ - project_id: typing.Optional[str].
591
621
  ---
592
622
  from llama_cloud.client import AsyncLlamaCloud
593
623
 
@@ -602,7 +632,7 @@ class AsyncRetrieversClient:
602
632
  "GET",
603
633
  urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", f"api/v1/retrievers/{retriever_id}"),
604
634
  params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
605
- headers=self._client_wrapper.get_headers(),
635
+ headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
606
636
  timeout=60,
607
637
  )
608
638
  if 200 <= _response.status_code < 300:
@@ -705,6 +735,7 @@ class AsyncRetrieversClient:
705
735
  rerank_top_n: typing.Optional[int] = OMIT,
706
736
  rerank_config: typing.Optional[ReRankConfig] = OMIT,
707
737
  query: str,
738
+ project_id: typing.Optional[str] = None,
708
739
  ) -> CompositeRetrievalResult:
709
740
  """
710
741
  Retrieve data using a Retriever.
@@ -723,6 +754,8 @@ class AsyncRetrieversClient:
723
754
  - rerank_config: typing.Optional[ReRankConfig]. The rerank configuration for composite retrieval.
724
755
 
725
756
  - query: str. The query to retrieve against.
757
+
758
+ - project_id: typing.Optional[str].
726
759
  ---
727
760
  from llama_cloud import CompositeRetrievalMode, ReRankConfig, ReRankerType
728
761
  from llama_cloud.client import AsyncLlamaCloud
@@ -753,7 +786,7 @@ class AsyncRetrieversClient:
753
786
  ),
754
787
  params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
755
788
  json=jsonable_encoder(_request),
756
- headers=self._client_wrapper.get_headers(),
789
+ headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
757
790
  timeout=60,
758
791
  )
759
792
  if 200 <= _response.status_code < 300:
@@ -776,6 +809,7 @@ class AsyncRetrieversClient:
776
809
  rerank_config: typing.Optional[ReRankConfig] = OMIT,
777
810
  query: str,
778
811
  pipelines: typing.Optional[typing.List[RetrieverPipeline]] = OMIT,
812
+ project_id: typing.Optional[str] = None,
779
813
  ) -> CompositeRetrievalResult:
780
814
  """
781
815
  Retrieve data using specified pipelines without creating a persistent retriever.
@@ -794,6 +828,8 @@ class AsyncRetrieversClient:
794
828
  - query: str. The query to retrieve against.
795
829
 
796
830
  - pipelines: typing.Optional[typing.List[RetrieverPipeline]]. The pipelines to use for retrieval.
831
+
832
+ - project_id: typing.Optional[str].
797
833
  ---
798
834
  from llama_cloud import CompositeRetrievalMode, ReRankConfig, ReRankerType
799
835
  from llama_cloud.client import AsyncLlamaCloud
@@ -823,7 +859,7 @@ class AsyncRetrieversClient:
823
859
  urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/retrievers/retrieve"),
824
860
  params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
825
861
  json=jsonable_encoder(_request),
826
- headers=self._client_wrapper.get_headers(),
862
+ headers=remove_none_from_dict({**self._client_wrapper.get_headers(), "Project-Id": project_id}),
827
863
  timeout=60,
828
864
  )
829
865
  if 200 <= _response.status_code < 300:
@@ -44,7 +44,6 @@ from .classification_result import ClassificationResult
44
44
  from .classifier_rule import ClassifierRule
45
45
  from .classify_job import ClassifyJob
46
46
  from .classify_job_results import ClassifyJobResults
47
- from .classify_job_with_status import ClassifyJobWithStatus
48
47
  from .classify_parsing_configuration import ClassifyParsingConfiguration
49
48
  from .cloud_astra_db_vector_store import CloudAstraDbVectorStore
50
49
  from .cloud_az_storage_blob_data_source import CloudAzStorageBlobDataSource
@@ -54,6 +53,8 @@ from .cloud_confluence_data_source import CloudConfluenceDataSource
54
53
  from .cloud_document import CloudDocument
55
54
  from .cloud_document_create import CloudDocumentCreate
56
55
  from .cloud_jira_data_source import CloudJiraDataSource
56
+ from .cloud_jira_data_source_v_2 import CloudJiraDataSourceV2
57
+ from .cloud_jira_data_source_v_2_api_version import CloudJiraDataSourceV2ApiVersion
57
58
  from .cloud_milvus_vector_store import CloudMilvusVectorStore
58
59
  from .cloud_mongo_db_atlas_vector_search import CloudMongoDbAtlasVectorSearch
59
60
  from .cloud_notion_page_data_source import CloudNotionPageDataSource
@@ -148,10 +149,15 @@ from .failure_handling_config import FailureHandlingConfig
148
149
  from .file import File
149
150
  from .file_classification import FileClassification
150
151
  from .file_count_by_status_response import FileCountByStatusResponse
152
+ from .file_create import FileCreate
153
+ from .file_create_permission_info_value import FileCreatePermissionInfoValue
154
+ from .file_create_resource_info_value import FileCreateResourceInfoValue
151
155
  from .file_data import FileData
156
+ from .file_filter import FileFilter
152
157
  from .file_id_presigned_url import FileIdPresignedUrl
153
158
  from .file_parse_public import FileParsePublic
154
159
  from .file_permission_info_value import FilePermissionInfoValue
160
+ from .file_query_response import FileQueryResponse
155
161
  from .file_resource_info_value import FileResourceInfoValue
156
162
  from .file_store_info_response import FileStoreInfoResponse
157
163
  from .file_store_info_response_status import FileStoreInfoResponseStatus
@@ -193,6 +199,8 @@ from .job_record_with_usage_metrics import JobRecordWithUsageMetrics
193
199
  from .l_lama_parse_transform_config import LLamaParseTransformConfig
194
200
  from .legacy_parse_job_config import LegacyParseJobConfig
195
201
  from .license_info_response import LicenseInfoResponse
202
+ from .llama_extract_mode_availability import LlamaExtractModeAvailability
203
+ from .llama_extract_mode_availability_status import LlamaExtractModeAvailabilityStatus
196
204
  from .llama_extract_settings import LlamaExtractSettings
197
205
  from .llama_index_core_base_llms_types_chat_message import LlamaIndexCoreBaseLlmsTypesChatMessage
198
206
  from .llama_index_core_base_llms_types_chat_message_blocks_item import (
@@ -239,6 +247,7 @@ from .paginated_list_pipeline_files_response import PaginatedListPipelineFilesRe
239
247
  from .paginated_report_response import PaginatedReportResponse
240
248
  from .paginated_response_agent_data import PaginatedResponseAgentData
241
249
  from .paginated_response_aggregate_group import PaginatedResponseAggregateGroup
250
+ from .paginated_response_classify_job import PaginatedResponseClassifyJob
242
251
  from .paginated_response_quota_configuration import PaginatedResponseQuotaConfiguration
243
252
  from .parse_job_config import ParseJobConfig
244
253
  from .parse_job_config_priority import ParseJobConfigPriority
@@ -430,7 +439,6 @@ __all__ = [
430
439
  "ClassifierRule",
431
440
  "ClassifyJob",
432
441
  "ClassifyJobResults",
433
- "ClassifyJobWithStatus",
434
442
  "ClassifyParsingConfiguration",
435
443
  "CloudAstraDbVectorStore",
436
444
  "CloudAzStorageBlobDataSource",
@@ -440,6 +448,8 @@ __all__ = [
440
448
  "CloudDocument",
441
449
  "CloudDocumentCreate",
442
450
  "CloudJiraDataSource",
451
+ "CloudJiraDataSourceV2",
452
+ "CloudJiraDataSourceV2ApiVersion",
443
453
  "CloudMilvusVectorStore",
444
454
  "CloudMongoDbAtlasVectorSearch",
445
455
  "CloudNotionPageDataSource",
@@ -530,10 +540,15 @@ __all__ = [
530
540
  "File",
531
541
  "FileClassification",
532
542
  "FileCountByStatusResponse",
543
+ "FileCreate",
544
+ "FileCreatePermissionInfoValue",
545
+ "FileCreateResourceInfoValue",
533
546
  "FileData",
547
+ "FileFilter",
534
548
  "FileIdPresignedUrl",
535
549
  "FileParsePublic",
536
550
  "FilePermissionInfoValue",
551
+ "FileQueryResponse",
537
552
  "FileResourceInfoValue",
538
553
  "FileStoreInfoResponse",
539
554
  "FileStoreInfoResponseStatus",
@@ -573,6 +588,8 @@ __all__ = [
573
588
  "LLamaParseTransformConfig",
574
589
  "LegacyParseJobConfig",
575
590
  "LicenseInfoResponse",
591
+ "LlamaExtractModeAvailability",
592
+ "LlamaExtractModeAvailabilityStatus",
576
593
  "LlamaExtractSettings",
577
594
  "LlamaIndexCoreBaseLlmsTypesChatMessage",
578
595
  "LlamaIndexCoreBaseLlmsTypesChatMessageBlocksItem",
@@ -617,6 +634,7 @@ __all__ = [
617
634
  "PaginatedReportResponse",
618
635
  "PaginatedResponseAgentData",
619
636
  "PaginatedResponseAggregateGroup",
637
+ "PaginatedResponseClassifyJob",
620
638
  "PaginatedResponseQuotaConfiguration",
621
639
  "ParseJobConfig",
622
640
  "ParseJobConfigPriority",
@@ -23,6 +23,7 @@ class AgentDeploymentSummary(pydantic.BaseModel):
23
23
  display_name: str = pydantic.Field(description="Display name of the deployed app")
24
24
  created_at: dt.datetime = pydantic.Field(description="Timestamp when the app deployment was created")
25
25
  updated_at: dt.datetime = pydantic.Field(description="Timestamp when the app deployment was last updated")
26
+ api_key_id: typing.Optional[str]
26
27
 
27
28
  def json(self, **kwargs: typing.Any) -> str:
28
29
  kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
@@ -6,6 +6,7 @@ import typing
6
6
  from ..core.datetime_utils import serialize_datetime
7
7
  from .classifier_rule import ClassifierRule
8
8
  from .classify_parsing_configuration import ClassifyParsingConfiguration
9
+ from .status_enum import StatusEnum
9
10
 
10
11
  try:
11
12
  import pydantic
@@ -27,6 +28,7 @@ class ClassifyJob(pydantic.BaseModel):
27
28
  rules: typing.List[ClassifierRule] = pydantic.Field(description="The rules to classify the files")
28
29
  user_id: str = pydantic.Field(description="The ID of the user")
29
30
  project_id: str = pydantic.Field(description="The ID of the project")
31
+ status: StatusEnum = pydantic.Field(description="The status of the classify job")
30
32
  parsing_configuration: typing.Optional[ClassifyParsingConfiguration] = pydantic.Field(
31
33
  description="The configuration for the parsing job"
32
34
  )
@@ -0,0 +1,52 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import datetime as dt
4
+ import typing
5
+
6
+ from ..core.datetime_utils import serialize_datetime
7
+ from .cloud_jira_data_source_v_2_api_version import CloudJiraDataSourceV2ApiVersion
8
+
9
+ try:
10
+ import pydantic
11
+ if pydantic.__version__.startswith("1."):
12
+ raise ImportError
13
+ import pydantic.v1 as pydantic # type: ignore
14
+ except ImportError:
15
+ import pydantic # type: ignore
16
+
17
+
18
+ class CloudJiraDataSourceV2(pydantic.BaseModel):
19
+ """
20
+ Cloud Jira Data Source integrating JiraReaderV2.
21
+ """
22
+
23
+ supports_access_control: typing.Optional[bool]
24
+ email: typing.Optional[str]
25
+ api_token: typing.Optional[str]
26
+ server_url: str = pydantic.Field(description="The server url for Jira Cloud.")
27
+ cloud_id: typing.Optional[str]
28
+ authentication_mechanism: str = pydantic.Field(description="Type of Authentication for connecting to Jira APIs.")
29
+ api_version: typing.Optional[CloudJiraDataSourceV2ApiVersion] = pydantic.Field(
30
+ description="Jira REST API version to use (2 or 3). 3 supports Atlassian Document Format (ADF)."
31
+ )
32
+ query: str = pydantic.Field(description="JQL (Jira Query Language) query to search.")
33
+ fields: typing.Optional[typing.List[str]]
34
+ expand: typing.Optional[str]
35
+ requests_per_minute: typing.Optional[int]
36
+ get_permissions: typing.Optional[bool] = pydantic.Field(
37
+ description="Whether to fetch project role permissions and issue-level security"
38
+ )
39
+ class_name: typing.Optional[str]
40
+
41
+ def json(self, **kwargs: typing.Any) -> str:
42
+ kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
43
+ return super().json(**kwargs_with_defaults)
44
+
45
+ def dict(self, **kwargs: typing.Any) -> typing.Dict[str, typing.Any]:
46
+ kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
47
+ return super().dict(**kwargs_with_defaults)
48
+
49
+ class Config:
50
+ frozen = True
51
+ smart_union = True
52
+ json_encoders = {dt.datetime: serialize_datetime}
@@ -0,0 +1,21 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import enum
4
+ import typing
5
+
6
+ T_Result = typing.TypeVar("T_Result")
7
+
8
+
9
+ class CloudJiraDataSourceV2ApiVersion(str, enum.Enum):
10
+ """
11
+ Jira REST API version to use (2 or 3). 3 supports Atlassian Document Format (ADF).
12
+ """
13
+
14
+ TWO = "2"
15
+ THREE = "3"
16
+
17
+ def visit(self, two: typing.Callable[[], T_Result], three: typing.Callable[[], T_Result]) -> T_Result:
18
+ if self is CloudJiraDataSourceV2ApiVersion.TWO:
19
+ return two()
20
+ if self is CloudJiraDataSourceV2ApiVersion.THREE:
21
+ return three()
@@ -16,6 +16,7 @@ class ConfigurableDataSourceNames(str, enum.Enum):
16
16
  NOTION_PAGE = "NOTION_PAGE"
17
17
  CONFLUENCE = "CONFLUENCE"
18
18
  JIRA = "JIRA"
19
+ JIRA_V_2 = "JIRA_V2"
19
20
  BOX = "BOX"
20
21
 
21
22
  def visit(
@@ -29,6 +30,7 @@ class ConfigurableDataSourceNames(str, enum.Enum):
29
30
  notion_page: typing.Callable[[], T_Result],
30
31
  confluence: typing.Callable[[], T_Result],
31
32
  jira: typing.Callable[[], T_Result],
33
+ jira_v_2: typing.Callable[[], T_Result],
32
34
  box: typing.Callable[[], T_Result],
33
35
  ) -> T_Result:
34
36
  if self is ConfigurableDataSourceNames.S_3:
@@ -49,5 +51,7 @@ class ConfigurableDataSourceNames(str, enum.Enum):
49
51
  return confluence()
50
52
  if self is ConfigurableDataSourceNames.JIRA:
51
53
  return jira()
54
+ if self is ConfigurableDataSourceNames.JIRA_V_2:
55
+ return jira_v_2()
52
56
  if self is ConfigurableDataSourceNames.BOX:
53
57
  return box()
@@ -6,6 +6,7 @@ from .cloud_az_storage_blob_data_source import CloudAzStorageBlobDataSource
6
6
  from .cloud_box_data_source import CloudBoxDataSource
7
7
  from .cloud_confluence_data_source import CloudConfluenceDataSource
8
8
  from .cloud_jira_data_source import CloudJiraDataSource
9
+ from .cloud_jira_data_source_v_2 import CloudJiraDataSourceV2
9
10
  from .cloud_notion_page_data_source import CloudNotionPageDataSource
10
11
  from .cloud_one_drive_data_source import CloudOneDriveDataSource
11
12
  from .cloud_s_3_data_source import CloudS3DataSource
@@ -22,5 +23,6 @@ DataSourceComponent = typing.Union[
22
23
  CloudNotionPageDataSource,
23
24
  CloudConfluenceDataSource,
24
25
  CloudJiraDataSource,
26
+ CloudJiraDataSourceV2,
25
27
  CloudBoxDataSource,
26
28
  ]
@@ -6,6 +6,7 @@ from .cloud_az_storage_blob_data_source import CloudAzStorageBlobDataSource
6
6
  from .cloud_box_data_source import CloudBoxDataSource
7
7
  from .cloud_confluence_data_source import CloudConfluenceDataSource
8
8
  from .cloud_jira_data_source import CloudJiraDataSource
9
+ from .cloud_jira_data_source_v_2 import CloudJiraDataSourceV2
9
10
  from .cloud_notion_page_data_source import CloudNotionPageDataSource
10
11
  from .cloud_one_drive_data_source import CloudOneDriveDataSource
11
12
  from .cloud_s_3_data_source import CloudS3DataSource
@@ -22,5 +23,6 @@ DataSourceCreateComponent = typing.Union[
22
23
  CloudNotionPageDataSource,
23
24
  CloudConfluenceDataSource,
24
25
  CloudJiraDataSource,
26
+ CloudJiraDataSourceV2,
25
27
  CloudBoxDataSource,
26
28
  ]
@@ -9,9 +9,17 @@ T_Result = typing.TypeVar("T_Result")
9
9
  class DataSourceReaderVersionMetadataReaderVersion(str, enum.Enum):
10
10
  ONE_0 = "1.0"
11
11
  TWO_0 = "2.0"
12
+ TWO_1 = "2.1"
12
13
 
13
- def visit(self, one_0: typing.Callable[[], T_Result], two_0: typing.Callable[[], T_Result]) -> T_Result:
14
+ def visit(
15
+ self,
16
+ one_0: typing.Callable[[], T_Result],
17
+ two_0: typing.Callable[[], T_Result],
18
+ two_1: typing.Callable[[], T_Result],
19
+ ) -> T_Result:
14
20
  if self is DataSourceReaderVersionMetadataReaderVersion.ONE_0:
15
21
  return one_0()
16
22
  if self is DataSourceReaderVersionMetadataReaderVersion.TWO_0:
17
23
  return two_0()
24
+ if self is DataSourceReaderVersionMetadataReaderVersion.TWO_1:
25
+ return two_1()
@@ -0,0 +1,41 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import datetime as dt
4
+ import typing
5
+
6
+ from ..core.datetime_utils import serialize_datetime
7
+ from .file_create_permission_info_value import FileCreatePermissionInfoValue
8
+ from .file_create_resource_info_value import FileCreateResourceInfoValue
9
+
10
+ try:
11
+ import pydantic
12
+ if pydantic.__version__.startswith("1."):
13
+ raise ImportError
14
+ import pydantic.v1 as pydantic # type: ignore
15
+ except ImportError:
16
+ import pydantic # type: ignore
17
+
18
+
19
+ class FileCreate(pydantic.BaseModel):
20
+ name: str = pydantic.Field(
21
+ description="Name that will be used for created file. If possible, always include the file extension in the name."
22
+ )
23
+ external_file_id: typing.Optional[str]
24
+ file_size: typing.Optional[int]
25
+ last_modified_at: typing.Optional[dt.datetime]
26
+ resource_info: typing.Optional[typing.Dict[str, typing.Optional[FileCreateResourceInfoValue]]]
27
+ permission_info: typing.Optional[typing.Dict[str, typing.Optional[FileCreatePermissionInfoValue]]]
28
+ data_source_id: typing.Optional[str]
29
+
30
+ def json(self, **kwargs: typing.Any) -> str:
31
+ kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
32
+ return super().json(**kwargs_with_defaults)
33
+
34
+ def dict(self, **kwargs: typing.Any) -> typing.Dict[str, typing.Any]:
35
+ kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
36
+ return super().dict(**kwargs_with_defaults)
37
+
38
+ class Config:
39
+ frozen = True
40
+ smart_union = True
41
+ json_encoders = {dt.datetime: serialize_datetime}
@@ -4,9 +4,6 @@ import datetime as dt
4
4
  import typing
5
5
 
6
6
  from ..core.datetime_utils import serialize_datetime
7
- from .classifier_rule import ClassifierRule
8
- from .classify_parsing_configuration import ClassifyParsingConfiguration
9
- from .status_enum import StatusEnum
10
7
 
11
8
  try:
12
9
  import pydantic
@@ -17,21 +14,17 @@ except ImportError:
17
14
  import pydantic # type: ignore
18
15
 
19
16
 
20
- class ClassifyJobWithStatus(pydantic.BaseModel):
17
+ class FileFilter(pydantic.BaseModel):
21
18
  """
22
- A classify job with status.
19
+ Filter parameters for file queries.
23
20
  """
24
21
 
25
- id: str = pydantic.Field(description="Unique identifier")
26
- created_at: typing.Optional[dt.datetime]
27
- updated_at: typing.Optional[dt.datetime]
28
- rules: typing.List[ClassifierRule] = pydantic.Field(description="The rules to classify the files")
29
- user_id: str = pydantic.Field(description="The ID of the user")
30
- project_id: str = pydantic.Field(description="The ID of the project")
31
- parsing_configuration: typing.Optional[ClassifyParsingConfiguration] = pydantic.Field(
32
- description="The configuration for the parsing job"
33
- )
34
- status: StatusEnum = pydantic.Field(description="The status of the classify job")
22
+ project_id: typing.Optional[str]
23
+ file_ids: typing.Optional[typing.List[str]]
24
+ file_name: typing.Optional[str]
25
+ data_source_id: typing.Optional[str]
26
+ external_file_id: typing.Optional[str]
27
+ only_manually_uploaded: typing.Optional[bool]
35
28
 
36
29
  def json(self, **kwargs: typing.Any) -> str:
37
30
  kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}