llama-cloud 0.1.30__py3-none-any.whl → 0.1.31__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of llama-cloud might be problematic. Click here for more details.
- llama_cloud/__init__.py +26 -14
- llama_cloud/client.py +0 -3
- llama_cloud/resources/__init__.py +0 -2
- llama_cloud/resources/beta/client.py +602 -0
- llama_cloud/resources/parsing/client.py +8 -0
- llama_cloud/resources/pipelines/client.py +64 -0
- llama_cloud/types/__init__.py +26 -12
- llama_cloud/types/{model_configuration.py → agent_data.py} +8 -7
- llama_cloud/types/agent_deployment_summary.py +1 -1
- llama_cloud/types/{message.py → aggregate_group.py} +8 -9
- llama_cloud/types/base_plan.py +3 -0
- llama_cloud/types/filter_operation.py +46 -0
- llama_cloud/types/filter_operation_eq.py +6 -0
- llama_cloud/types/filter_operation_gt.py +6 -0
- llama_cloud/types/filter_operation_gte.py +6 -0
- llama_cloud/types/filter_operation_includes_item.py +6 -0
- llama_cloud/types/filter_operation_lt.py +6 -0
- llama_cloud/types/filter_operation_lte.py +6 -0
- llama_cloud/types/input_message.py +2 -2
- llama_cloud/types/legacy_parse_job_config.py +3 -0
- llama_cloud/types/llama_index_core_base_llms_types_chat_message.py +2 -2
- llama_cloud/types/llama_parse_parameters.py +1 -0
- llama_cloud/types/{llama_index_core_base_llms_types_message_role.py → message_role.py} +9 -9
- llama_cloud/types/{text_content_block.py → paginated_response_agent_data.py} +5 -5
- llama_cloud/types/paginated_response_aggregate_group.py +34 -0
- llama_cloud/types/parse_job_config.py +1 -0
- llama_cloud/types/playground_session.py +2 -2
- llama_cloud/types/role.py +0 -1
- llama_cloud/types/{app_schema_chat_chat_message.py → src_app_schema_chat_chat_message.py} +3 -3
- llama_cloud/types/user_organization_role.py +0 -1
- {llama_cloud-0.1.30.dist-info → llama_cloud-0.1.31.dist-info}/METADATA +1 -1
- {llama_cloud-0.1.30.dist-info → llama_cloud-0.1.31.dist-info}/RECORD +34 -29
- llama_cloud/resources/responses/__init__.py +0 -2
- llama_cloud/resources/responses/client.py +0 -137
- llama_cloud/types/app_schema_responses_message_role.py +0 -33
- {llama_cloud-0.1.30.dist-info → llama_cloud-0.1.31.dist-info}/LICENSE +0 -0
- {llama_cloud-0.1.30.dist-info → llama_cloud-0.1.31.dist-info}/WHEEL +0 -0
|
@@ -1,5 +1,5 @@
|
|
|
1
|
-
llama_cloud/__init__.py,sha256=
|
|
2
|
-
llama_cloud/client.py,sha256=
|
|
1
|
+
llama_cloud/__init__.py,sha256=T-HghZZ4yA4QPgXeEvHQsmp5o8o1K2amrf7SftKYwE4,25511
|
|
2
|
+
llama_cloud/client.py,sha256=VNO5-JE1H0zWJudlDA9GJ2N6qEKQvxN5Q5QgVNTQPSI,5893
|
|
3
3
|
llama_cloud/core/__init__.py,sha256=QJS3CJ2TYP2E1Tge0CS6Z7r8LTNzJHQVX1hD3558eP0,519
|
|
4
4
|
llama_cloud/core/api_error.py,sha256=RE8LELok2QCjABadECTvtDp7qejA1VmINCh6TbqPwSE,426
|
|
5
5
|
llama_cloud/core/client_wrapper.py,sha256=xmj0jCdQ0ySzbSqHUWOkpRRy069y74I_HuXkWltcsVM,1507
|
|
@@ -9,11 +9,11 @@ llama_cloud/core/remove_none_from_dict.py,sha256=8m91FC3YuVem0Gm9_sXhJ2tGvP33owJ
|
|
|
9
9
|
llama_cloud/environment.py,sha256=feTjOebeFZMrBdnHat4RE5aHlpt-sJm4NhK4ntV1htI,167
|
|
10
10
|
llama_cloud/errors/__init__.py,sha256=pbbVUFtB9LCocA1RMWMMF_RKjsy5YkOKX5BAuE49w6g,170
|
|
11
11
|
llama_cloud/errors/unprocessable_entity_error.py,sha256=FvR7XPlV3Xx5nu8HNlmLhBRdk4so_gCHjYT5PyZe6sM,313
|
|
12
|
-
llama_cloud/resources/__init__.py,sha256=
|
|
12
|
+
llama_cloud/resources/__init__.py,sha256=cFMt4FZb8n6SMbRXYzYqIR-PlJbO7C-jX4iBeCym_8E,4179
|
|
13
13
|
llama_cloud/resources/admin/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
14
14
|
llama_cloud/resources/admin/client.py,sha256=mzA_ezCjugKNmvWCMWEF0Z0k86ErACWov1VtPV1J2tU,3678
|
|
15
15
|
llama_cloud/resources/beta/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
16
|
-
llama_cloud/resources/beta/client.py,sha256=
|
|
16
|
+
llama_cloud/resources/beta/client.py,sha256=uJO08z4WF3I_tVyZEu0SiwfeSx3iQaTUPZkoh6Pevs8,39144
|
|
17
17
|
llama_cloud/resources/chat_apps/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
18
18
|
llama_cloud/resources/chat_apps/client.py,sha256=orSI8rpQbUwVEToolEeiEi5Qe--suXFvfu6D9JDii5I,23595
|
|
19
19
|
llama_cloud/resources/data_sinks/__init__.py,sha256=ZHUjn3HbKhq_7QS1q74r2m5RGKF5lxcvF2P6pGvpcis,147
|
|
@@ -55,9 +55,9 @@ llama_cloud/resources/llama_extract/types/extract_schema_validate_request_data_s
|
|
|
55
55
|
llama_cloud/resources/organizations/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
56
56
|
llama_cloud/resources/organizations/client.py,sha256=CdrdNdB9R-bOsNqZ4Jbm1BzG1RafXMFjuCsrVYf2OrE,56567
|
|
57
57
|
llama_cloud/resources/parsing/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
58
|
-
llama_cloud/resources/parsing/client.py,sha256=
|
|
58
|
+
llama_cloud/resources/parsing/client.py,sha256=EHrQKjOl_VPPbcbaXi5TSah8HBf7ooHijhMF7IEzBMg,88117
|
|
59
59
|
llama_cloud/resources/pipelines/__init__.py,sha256=zyvVEOF_krvEZkCIj_kZoMKfhDqHo_R32a1mv9CriQc,1193
|
|
60
|
-
llama_cloud/resources/pipelines/client.py,sha256=
|
|
60
|
+
llama_cloud/resources/pipelines/client.py,sha256=VAqAm0oY_nXGkMPqXuzPEHS9kPtpuOE5sxfyqlzXuSI,134738
|
|
61
61
|
llama_cloud/resources/pipelines/types/__init__.py,sha256=C68NQ5QzA0dFXf9oePFFGmV1vn96jcAp-QAznSgoRYQ,1375
|
|
62
62
|
llama_cloud/resources/pipelines/types/pipeline_file_update_custom_metadata_value.py,sha256=trI48WLxPcAqV9207Q6-3cj1nl4EGlZpw7En56ZsPgg,217
|
|
63
63
|
llama_cloud/resources/pipelines/types/pipeline_update_embedding_config.py,sha256=c8FF64fDrBMX_2RX4uY3CjbNc0Ss_AUJ4Eqs-KeV4Wc,2874
|
|
@@ -69,23 +69,21 @@ llama_cloud/resources/reports/__init__.py,sha256=cruYbQ1bIuJbRpkfaQY7ajUEslffjd7
|
|
|
69
69
|
llama_cloud/resources/reports/client.py,sha256=kHjtXVVc1Xi3T1GyBvSW5K4mTdr6xQwZA3vw-liRKBg,46736
|
|
70
70
|
llama_cloud/resources/reports/types/__init__.py,sha256=LfwDYrI4RcQu-o42iAe7HkcwHww2YU90lOonBPTmZIk,291
|
|
71
71
|
llama_cloud/resources/reports/types/update_report_plan_api_v_1_reports_report_id_plan_patch_request_action.py,sha256=Qh-MSeRvDBfNb5hoLELivv1pLtrYVf52WVoP7G8V34A,807
|
|
72
|
-
llama_cloud/resources/responses/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
73
|
-
llama_cloud/resources/responses/client.py,sha256=ard4U9yZcD89pJ_hyYqeRDIfQYaX2WGl36OK7re8q3U,5481
|
|
74
72
|
llama_cloud/resources/retrievers/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
|
75
73
|
llama_cloud/resources/retrievers/client.py,sha256=z2LhmA-cZVFzr9P6loeCZYnJbvSIk0QitFeVFp-IyZk,32126
|
|
76
|
-
llama_cloud/types/__init__.py,sha256=
|
|
74
|
+
llama_cloud/types/__init__.py,sha256=ZYnUvMdFPye-wlq-XeyWUmhtVeLpi8c0UR0vSemiHP4,30490
|
|
77
75
|
llama_cloud/types/advanced_mode_transform_config.py,sha256=4xCXye0_cPmVS1F8aNTx81sIaEPjQH9kiCCAIoqUzlI,1502
|
|
78
76
|
llama_cloud/types/advanced_mode_transform_config_chunking_config.py,sha256=wYbJnWLpeQDfhmDZz-wJfYzD1iGT5Jcxb9ga3mzUuvk,1983
|
|
79
77
|
llama_cloud/types/advanced_mode_transform_config_segmentation_config.py,sha256=anNGq0F5-IlbIW3kpC8OilzLJnUq5tdIcWHnRnmlYsg,1303
|
|
78
|
+
llama_cloud/types/agent_data.py,sha256=Onaoc1QeIn3Il-8r1vgEzqvef92gHclCO7AC4kucEMI,1220
|
|
80
79
|
llama_cloud/types/agent_deployment_list.py,sha256=7PWm2GHumo8CfqKU8fDRTJVDV4QQh8My1dhvBPO2zaA,1120
|
|
81
|
-
llama_cloud/types/agent_deployment_summary.py,sha256=
|
|
82
|
-
llama_cloud/types/
|
|
83
|
-
llama_cloud/types/app_schema_responses_message_role.py,sha256=Wod45VMOCo-z6DWNqCOOJAaiFuZHjHCPtgqKPlVt2fI,972
|
|
80
|
+
llama_cloud/types/agent_deployment_summary.py,sha256=YEZxnNvTGYHz3zV6eGldVKfcy5S_IM-KlcOzDUqTfiU,1605
|
|
81
|
+
llama_cloud/types/aggregate_group.py,sha256=LybxFl_1snA9VgG6f7sogwO7kYAwH_I88pkYc0oMOH0,1164
|
|
84
82
|
llama_cloud/types/audio_block.py,sha256=9JIGjZ8GU3C7ICv6XdNVN6_gWXyF18TJPaDuM9OUoMU,1071
|
|
85
83
|
llama_cloud/types/auto_transform_config.py,sha256=HVeHZM75DMRznScqLTfrMwcZwIdyWPuaEYbPewnHqwc,1168
|
|
86
84
|
llama_cloud/types/azure_open_ai_embedding.py,sha256=MeDqZoPYFN7Nv_imY9cfqDU9SPlEyAY4HcQZ4PF5X3g,2264
|
|
87
85
|
llama_cloud/types/azure_open_ai_embedding_config.py,sha256=o1zZhzcGElH3SeixFErrm7P_WFHQ6LvrLem_nKJWunw,1170
|
|
88
|
-
llama_cloud/types/base_plan.py,sha256=
|
|
86
|
+
llama_cloud/types/base_plan.py,sha256=kuRJi-OxFHbKAxoQWe08IG45_i8xL67WeOZFCGWkOHI,2049
|
|
89
87
|
llama_cloud/types/base_plan_metronome_plan_type.py,sha256=I3g_dVoWWztbmpWpYmseDqQSbwtlLUl2vS01tfgMjEA,499
|
|
90
88
|
llama_cloud/types/base_plan_name.py,sha256=keHQaw9YV9ghsWnGfnHrLtB4qNz0v4TWX4_MoO3flRM,1926
|
|
91
89
|
llama_cloud/types/base_plan_plan_frequency.py,sha256=idUZlDaSdMrMZ2lQ1ytBWM4QyduIZu6Gt2eLU0LVqH4,684
|
|
@@ -186,6 +184,13 @@ llama_cloud/types/file_parse_public.py,sha256=sshZ0BcjHMGpuz4ylSurv0K_3ejfPrUGGy
|
|
|
186
184
|
llama_cloud/types/file_permission_info_value.py,sha256=RyQlNbhvIKS87Ywu7XUaw5jDToZX64M9Wqzu1U_q2Us,197
|
|
187
185
|
llama_cloud/types/file_resource_info_value.py,sha256=g6T6ELeLK9jgcvX6r-EuAl_4JkwnyqdS0RRoabMReSU,195
|
|
188
186
|
llama_cloud/types/filter_condition.py,sha256=YEc-NaZbMha4oZVSKerZ6-gNYriNOZmTHTRMKX-9Ju0,678
|
|
187
|
+
llama_cloud/types/filter_operation.py,sha256=lzyF_LQ-bT_wubU2bSbV6q2oncCE3mypz3D6qkAR86U,1663
|
|
188
|
+
llama_cloud/types/filter_operation_eq.py,sha256=7UQkjycQvUFBvd1KRWfNacXAEgp2eGG6XNej0EikP1M,165
|
|
189
|
+
llama_cloud/types/filter_operation_gt.py,sha256=ueeaTBhCGM0xUWLjdFei55ecbtbR3jFuiAtXrinFNDk,165
|
|
190
|
+
llama_cloud/types/filter_operation_gte.py,sha256=A_8I_-EpBNqcX_KbwMdhXI0Kno3WCwZnPofSRJxECpU,166
|
|
191
|
+
llama_cloud/types/filter_operation_includes_item.py,sha256=kwI0NjIZVUfaNU3BBue-AAEkPl_42_GjE_CR0OwZV5Y,175
|
|
192
|
+
llama_cloud/types/filter_operation_lt.py,sha256=Njv9OnuI3tzo88EAMhsVN8BvuzR1164GQP4SggbZe1U,165
|
|
193
|
+
llama_cloud/types/filter_operation_lte.py,sha256=5Evci2M4XfkkWMlY746t52OiTYiO9SaIJ72QDPu2G7U,166
|
|
189
194
|
llama_cloud/types/filter_operator.py,sha256=tY_DWFVOoLrqDc-soJcSFvUL-MsltK6iLSK7IKK-TPs,2439
|
|
190
195
|
llama_cloud/types/free_credits_usage.py,sha256=TPktesYpM5gVeBXPbRFun19XaPJo-dIu0Xbrg-iX8qE,1052
|
|
191
196
|
llama_cloud/types/gemini_embedding.py,sha256=n9vuxFbXt_VNuaZvp7BlkFWmGMgehpJz_ICacIafdYw,1418
|
|
@@ -196,20 +201,19 @@ llama_cloud/types/hugging_face_inference_api_embedding_config.py,sha256=EFHhuPCx
|
|
|
196
201
|
llama_cloud/types/hugging_face_inference_api_embedding_token.py,sha256=A7-_YryBcsP4G5uRyJ9acao3XwX5-YC3NRndTeDAPj4,144
|
|
197
202
|
llama_cloud/types/image_block.py,sha256=Bccrsm1-B2hUzObP7Oy1H7IVnurixfTpL03i-yqfZp0,1112
|
|
198
203
|
llama_cloud/types/ingestion_error_response.py,sha256=8u0cyT44dnpkNeUKemTvJMUqi_WyPcYQKP_DMTqaFPY,1259
|
|
199
|
-
llama_cloud/types/input_message.py,sha256=
|
|
204
|
+
llama_cloud/types/input_message.py,sha256=Ym6-tX6CMWKuHfxRtyM2y16kqSS3BzHged9rFRFkX0g,1346
|
|
200
205
|
llama_cloud/types/job_name_mapping.py,sha256=2dQFQlVHoeSlkyEKSEJv0M3PzJf7hMvkuABj3vMY7ys,1617
|
|
201
206
|
llama_cloud/types/job_names.py,sha256=WacongwoJygg_gCyYjPsOVv3cmVtRaX633JNgFxy-d8,3915
|
|
202
207
|
llama_cloud/types/job_record.py,sha256=Z6sF9AruZJo-kTRgNufAWS3WK1yaEqop6kox1GpBYy4,2219
|
|
203
208
|
llama_cloud/types/job_record_parameters.py,sha256=Oqxp5y0owPfjLc_NR7AYE8P3zM2PJo36N9olbyNl7AA,3425
|
|
204
209
|
llama_cloud/types/job_record_with_usage_metrics.py,sha256=iNV2do5TB_0e3PoOz_DJyAaM6Cn9G8KG-dGPGgEs5SY,1198
|
|
205
210
|
llama_cloud/types/l_lama_parse_transform_config.py,sha256=YQRJZvKh1Ee2FUyW_N0nqYJoW599qBgH3JCH9SH6YLo,1249
|
|
206
|
-
llama_cloud/types/legacy_parse_job_config.py,sha256=
|
|
211
|
+
llama_cloud/types/legacy_parse_job_config.py,sha256=zZJFYnquo51NbEXhw-yhpRjIuaJNgg-T_fAI2J7-hrM,12660
|
|
207
212
|
llama_cloud/types/license_info_response.py,sha256=fE9vcWO8k92SBqb_wOyBu_16C61s72utA-SifEi9iBc,1192
|
|
208
213
|
llama_cloud/types/llama_extract_settings.py,sha256=Y60XxsxVHUtX-ZjC0tyNzsaDIj_ojxYC1iy2w4vti54,2532
|
|
209
|
-
llama_cloud/types/llama_index_core_base_llms_types_chat_message.py,sha256=
|
|
214
|
+
llama_cloud/types/llama_index_core_base_llms_types_chat_message.py,sha256=NelHo-T-ebVMhRKsqE_xV8AJW4c7o6lS0uEQnPsmTwg,1365
|
|
210
215
|
llama_cloud/types/llama_index_core_base_llms_types_chat_message_blocks_item.py,sha256=-aL8fh-w2Xf4uQs_LHzb3q6LL_onLAcVzCR5yMI4qJw,1571
|
|
211
|
-
llama_cloud/types/
|
|
212
|
-
llama_cloud/types/llama_parse_parameters.py,sha256=SiSqreBFW5hGf7gVuXdITwW1ugxv03L5VpNQqoeI6Pk,6260
|
|
216
|
+
llama_cloud/types/llama_parse_parameters.py,sha256=pgWWbeaoC8p01_c0bC--ksHjJt-_A7QUhQdCW4MAVIQ,6325
|
|
213
217
|
llama_cloud/types/llama_parse_parameters_priority.py,sha256=EFRudtaID_s8rLKlfW8O8O9TDbpZdniIidK-xchhfRI,830
|
|
214
218
|
llama_cloud/types/llama_parse_supported_file_extensions.py,sha256=B_0N3f8Aq59W9FbsH50mGBUiyWTIXQjHFl739uAyaQw,11207
|
|
215
219
|
llama_cloud/types/llm_model_data.py,sha256=6rrycqGwlK3LZ2S-WtgmeomithdLhDCgwBBZQ5KLaso,1300
|
|
@@ -217,13 +221,12 @@ llama_cloud/types/llm_parameters.py,sha256=RTKYt09lm9a1MlnBfYuTP2x_Ww4byUNNc1TqI
|
|
|
217
221
|
llama_cloud/types/load_files_job_config.py,sha256=R5sFgFmV__0mqLUuD7dkFoBJHG2ZLw5px9zRapvYcpE,1069
|
|
218
222
|
llama_cloud/types/managed_ingestion_status.py,sha256=3KVlcurpEBOPAesBUS5pSYLoQVIyZUlr90Mmv-uALHE,1290
|
|
219
223
|
llama_cloud/types/managed_ingestion_status_response.py,sha256=rdNpjNbQswF-6JG1e-EU374TP6Pjlxl0p7HJyNmuxTI,1373
|
|
220
|
-
llama_cloud/types/message.py,sha256=RnahpUQR7s_QOu7pOdS7GNkZe8rJaPLC-wTJbzhSQV8,1373
|
|
221
224
|
llama_cloud/types/message_annotation.py,sha256=n4F9w4LxwmGvgXDk6E8YPTMu_g0yEjZhZ_eNFXdS_bc,1017
|
|
225
|
+
llama_cloud/types/message_role.py,sha256=9MpXT9drR33TyT1-NiqB3uGbuxvWwtoOdSmKQE9HmJI,1359
|
|
222
226
|
llama_cloud/types/metadata_filter.py,sha256=LX2fGsUb4wvF5bj9iWO6IPQGi3i0L2Lb4cE6igeeX9Y,1438
|
|
223
227
|
llama_cloud/types/metadata_filter_value.py,sha256=ij721gXNI7zbgsuDl9-AqBcXg2WDuVZhYS5F5YqekEs,188
|
|
224
228
|
llama_cloud/types/metadata_filters.py,sha256=uSf6sB4oQu6WzMPNFG6Tc4euqEiYcj_X14Y5JWt9xVE,1315
|
|
225
229
|
llama_cloud/types/metadata_filters_filters_item.py,sha256=e8KhD2q6Qc2_aK6r5CvyxC0oWVYO4F4vBIcB9eMEPPM,246
|
|
226
|
-
llama_cloud/types/model_configuration.py,sha256=JD_KSml2EB1EpwuuJUdwZeN4_aO7mTd76tQ16zK2vuU,1370
|
|
227
230
|
llama_cloud/types/node_relationship.py,sha256=2e2PqWm0LOTiImvtsyiuaAPNIl0BItjSrQZTJv65GRA,1209
|
|
228
231
|
llama_cloud/types/none_chunking_config.py,sha256=D062t314Vp-s4n9h8wNgsYfElI4PonPKmihvjEmaqdA,952
|
|
229
232
|
llama_cloud/types/none_segmentation_config.py,sha256=j3jUA6E8uFtwDMEu4TFG3Q4ZGCGiuUfUW9AMO1NNqXU,956
|
|
@@ -242,7 +245,9 @@ llama_cloud/types/paginated_jobs_history_with_metrics.py,sha256=Bxy6N0x0FARJhgwN
|
|
|
242
245
|
llama_cloud/types/paginated_list_cloud_documents_response.py,sha256=MsjS0SWlT0syELDck4x2sxxR3_NC1e6QTdepgVmK9aY,1341
|
|
243
246
|
llama_cloud/types/paginated_list_pipeline_files_response.py,sha256=2TKR2oHSQRyLMqWz1qQBSIvz-ZJb8U_94367lwOJ2S4,1317
|
|
244
247
|
llama_cloud/types/paginated_report_response.py,sha256=o79QhQi9r0HZZrhvRlA6WGjxtyPuxN0xONhwXSwxtcs,1104
|
|
245
|
-
llama_cloud/types/
|
|
248
|
+
llama_cloud/types/paginated_response_agent_data.py,sha256=u6Y-Cq9qjGF5tskMOQChUNqyI91Tk-uQ6vQdi69cs80,1159
|
|
249
|
+
llama_cloud/types/paginated_response_aggregate_group.py,sha256=1ajZLZJLU6-GuQ_PPsEVRFZ6bm9he807F_F_DmB2HlQ,1179
|
|
250
|
+
llama_cloud/types/parse_job_config.py,sha256=gLRQOaPgTfQuaNzriYtjDPucSFXt1AWyG19tGfzoy5M,6788
|
|
246
251
|
llama_cloud/types/parse_job_config_priority.py,sha256=__-gVv1GzktVCYZVyl6zeDt0pAZwYl-mxM0xkIHPEro,800
|
|
247
252
|
llama_cloud/types/parse_plan_level.py,sha256=GBkDS19qfHseBa17EXfuTPNT4GNv5alyPrWEvWji3GY,528
|
|
248
253
|
llama_cloud/types/parser_languages.py,sha256=Ps3IlaSt6tyxEI657N3-vZL96r2puk8wsf31cWnO-SI,10840
|
|
@@ -286,7 +291,7 @@ llama_cloud/types/pipeline_status.py,sha256=aC340nhfuPSrFVZOH_DhgYHWe985J3WNHrwv
|
|
|
286
291
|
llama_cloud/types/pipeline_transform_config.py,sha256=zMr-ePLKGjbaScxbAHaSwYBL7rrNibVlnn0cbgElDfU,824
|
|
287
292
|
llama_cloud/types/pipeline_type.py,sha256=tTqrhxHP5xd7W2dQGD0e5FOv886nwJssyaVlXpWrtRo,551
|
|
288
293
|
llama_cloud/types/plan_limits.py,sha256=WAbDbRl8gsQxvhmuVB0YT8mry-0uKg6c66uivyppdQU,2056
|
|
289
|
-
llama_cloud/types/playground_session.py,sha256=
|
|
294
|
+
llama_cloud/types/playground_session.py,sha256=BZZk9F_FVuMPcCE5dVNACPqHKIvyWGSkbRrrQOweaaw,1868
|
|
290
295
|
llama_cloud/types/pooling.py,sha256=5Fr6c8rx9SDWwWzEvD78suob2d79ktodUtLUAUHMbP8,651
|
|
291
296
|
llama_cloud/types/preset_composite_retrieval_params.py,sha256=yEf1pk4Wz5J6SxgB8elklwuyVDCRSZqfWC6x3hJUS4Q,1366
|
|
292
297
|
llama_cloud/types/preset_retrieval_params.py,sha256=TcyljefpspJSveMR9L5DQHlqW4jZeexBsXus_LkHkJA,2365
|
|
@@ -322,17 +327,17 @@ llama_cloud/types/retrieve_results.py,sha256=rHArmu05K3NvIQepHX5nsVOfcMsZj3MaIcP
|
|
|
322
327
|
llama_cloud/types/retriever.py,sha256=ZItPsorL8x1XjtJT49ZodaMqU8h2GfwlB4U4cgnfZkM,1626
|
|
323
328
|
llama_cloud/types/retriever_create.py,sha256=WyUR9DRzu3Q9tzKEeXCdQuzCY6WKi9ADJkZea9rqvxU,1286
|
|
324
329
|
llama_cloud/types/retriever_pipeline.py,sha256=F1pZDxg8JdQXRHE6ciFezd7a-Wv5bHplPcGDED-J4b0,1330
|
|
325
|
-
llama_cloud/types/role.py,sha256=
|
|
330
|
+
llama_cloud/types/role.py,sha256=4pbyLVNPleDd624cDcOhu9y1WvqC0J0gmNirTOW97iA,1342
|
|
326
331
|
llama_cloud/types/schema_relax_mode.py,sha256=v4or6dYTvWvBBNtEd2ZSaUAb1706I0Zuh-Xztm-zx_0,635
|
|
327
332
|
llama_cloud/types/semantic_chunking_config.py,sha256=dFDniTVWpRc7UcmVFvljUoyL5Ztd-l-YrHII7U-yM-k,1053
|
|
328
333
|
llama_cloud/types/sentence_chunking_config.py,sha256=NA9xidK5ICxJPkEMQZWNcsV0Hw9Co_bzRWeYe4uSh9I,1116
|
|
334
|
+
llama_cloud/types/src_app_schema_chat_chat_message.py,sha256=ddMQXZybeExPVFMNe8FWghyXXWktsujpZ_0Xmou3Zz8,1596
|
|
329
335
|
llama_cloud/types/status_enum.py,sha256=cUBIlys89E8PUzmVqqawu7qTDF0aRqBwiijOmRDPvx0,1018
|
|
330
336
|
llama_cloud/types/struct_mode.py,sha256=ROicwjXfFmgVU8_xSVxJlnFUzRNKG5VIEF1wYg9uOPU,1020
|
|
331
337
|
llama_cloud/types/struct_parse_conf.py,sha256=WlL8y0IBvdzGsDtFUlEZLzoUODwmOWAJi0viS9unL18,2297
|
|
332
338
|
llama_cloud/types/supported_llm_model.py,sha256=hubSopFICVNEegbJbtbpK6zRHwFPwUNtrw_NAw_3bfg,1380
|
|
333
339
|
llama_cloud/types/supported_llm_model_names.py,sha256=PXL0gA1lc0GJNzZHnjOscoxHpPW787A8Adh-2egAKo8,2512
|
|
334
340
|
llama_cloud/types/text_block.py,sha256=X154sQkSyposXuRcEWNp_tWcDQ-AI6q_-MfJUN5exP8,958
|
|
335
|
-
llama_cloud/types/text_content_block.py,sha256=MKMBMhJS7Tr-Vmr4MhhDgH8pO6r-_g_8bjWYT8LxitA,1130
|
|
336
341
|
llama_cloud/types/text_node.py,sha256=Tq3QmuKC5cIHvC9wAtvhsXl1g2sACs2yJwQ0Uko8GSU,2846
|
|
337
342
|
llama_cloud/types/text_node_relationships_value.py,sha256=qmXURTk1Xg7ZDzRSSV1uDEel0AXRLohND5ioezibHY0,217
|
|
338
343
|
llama_cloud/types/text_node_with_score.py,sha256=k-KYWO_mgJBvO6xUfOD5W6v1Ku9E586_HsvDoQbLfuQ,1229
|
|
@@ -345,7 +350,7 @@ llama_cloud/types/user_job_record.py,sha256=mJHdokJsemXJOwM2l7fsW3X0SlwSNcy7yHbc
|
|
|
345
350
|
llama_cloud/types/user_organization.py,sha256=yKewpOrMcB-CbujGNTjkX6QiWYr5HVsRIFQ-WX8kp2I,1729
|
|
346
351
|
llama_cloud/types/user_organization_create.py,sha256=Zj57s9xuYVnLW2p8i4j2QORL-G1y7Ab3avXE1baERQY,1189
|
|
347
352
|
llama_cloud/types/user_organization_delete.py,sha256=bEfgQMdTd6oAMZXtvSm5BhZahG1wAVDBXZ8e7V9UN7w,1159
|
|
348
|
-
llama_cloud/types/user_organization_role.py,sha256=
|
|
353
|
+
llama_cloud/types/user_organization_role.py,sha256=Tcfu9QISF5nRpo9jvboHzX-Yfg6b676UNfdjzjUIgAs,1448
|
|
349
354
|
llama_cloud/types/validation_error.py,sha256=yZDLtjUHDY5w82Ra6CW0H9sLAr18R0RY1UNgJKR72DQ,1084
|
|
350
355
|
llama_cloud/types/validation_error_loc_item.py,sha256=LAtjCHIllWRBFXvAZ5QZpp7CPXjdtN9EB7HrLVo6EP0,128
|
|
351
356
|
llama_cloud/types/vertex_ai_embedding_config.py,sha256=DvQk2xMJFmo54MEXTzoM4KSADyhGm_ygmFyx6wIcQdw,1159
|
|
@@ -353,7 +358,7 @@ llama_cloud/types/vertex_embedding_mode.py,sha256=yY23FjuWU_DkXjBb3JoKV4SCMqel2B
|
|
|
353
358
|
llama_cloud/types/vertex_text_embedding.py,sha256=-C4fNCYfFl36ATdBMGFVPpiHIKxjk0KB1ERA2Ec20aU,1932
|
|
354
359
|
llama_cloud/types/webhook_configuration.py,sha256=_Xm15whrWoKNBuCoO5y_NunA-ByhCAYK87LnC4W-Pzg,1350
|
|
355
360
|
llama_cloud/types/webhook_configuration_webhook_events_item.py,sha256=LTfOwphnoYUQYwsHGTlCxoVU_PseIRAbmQJRBdyXnbg,1519
|
|
356
|
-
llama_cloud-0.1.
|
|
357
|
-
llama_cloud-0.1.
|
|
358
|
-
llama_cloud-0.1.
|
|
359
|
-
llama_cloud-0.1.
|
|
361
|
+
llama_cloud-0.1.31.dist-info/LICENSE,sha256=_iNqtPcw1Ue7dZKwOwgPtbegMUkWVy15hC7bffAdNmY,1067
|
|
362
|
+
llama_cloud-0.1.31.dist-info/METADATA,sha256=8CeVAzwOJTcdsTWT_WakSrnOErE56P3ZK70PxR5g5b8,1194
|
|
363
|
+
llama_cloud-0.1.31.dist-info/WHEEL,sha256=b4K_helf-jlQoXBBETfwnf4B04YC67LOev0jo4fX5m8,88
|
|
364
|
+
llama_cloud-0.1.31.dist-info/RECORD,,
|
|
@@ -1,137 +0,0 @@
|
|
|
1
|
-
# This file was auto-generated by Fern from our API Definition.
|
|
2
|
-
|
|
3
|
-
import typing
|
|
4
|
-
import urllib.parse
|
|
5
|
-
from json.decoder import JSONDecodeError
|
|
6
|
-
|
|
7
|
-
from ...core.api_error import ApiError
|
|
8
|
-
from ...core.client_wrapper import AsyncClientWrapper, SyncClientWrapper
|
|
9
|
-
from ...core.jsonable_encoder import jsonable_encoder
|
|
10
|
-
from ...core.remove_none_from_dict import remove_none_from_dict
|
|
11
|
-
from ...errors.unprocessable_entity_error import UnprocessableEntityError
|
|
12
|
-
from ...types.http_validation_error import HttpValidationError
|
|
13
|
-
from ...types.message import Message
|
|
14
|
-
from ...types.model_configuration import ModelConfiguration
|
|
15
|
-
|
|
16
|
-
try:
|
|
17
|
-
import pydantic
|
|
18
|
-
if pydantic.__version__.startswith("1."):
|
|
19
|
-
raise ImportError
|
|
20
|
-
import pydantic.v1 as pydantic # type: ignore
|
|
21
|
-
except ImportError:
|
|
22
|
-
import pydantic # type: ignore
|
|
23
|
-
|
|
24
|
-
# this is used as the default value for optional parameters
|
|
25
|
-
OMIT = typing.cast(typing.Any, ...)
|
|
26
|
-
|
|
27
|
-
|
|
28
|
-
class ResponsesClient:
|
|
29
|
-
def __init__(self, *, client_wrapper: SyncClientWrapper):
|
|
30
|
-
self._client_wrapper = client_wrapper
|
|
31
|
-
|
|
32
|
-
def generate_response(
|
|
33
|
-
self,
|
|
34
|
-
*,
|
|
35
|
-
project_id: typing.Optional[str] = None,
|
|
36
|
-
organization_id: typing.Optional[str] = None,
|
|
37
|
-
messages: typing.List[Message],
|
|
38
|
-
model_configuration: ModelConfiguration,
|
|
39
|
-
) -> typing.Any:
|
|
40
|
-
"""
|
|
41
|
-
EXPERIMENTAL - SSE endpoint for basic response generation (dummy stream).
|
|
42
|
-
|
|
43
|
-
Parameters:
|
|
44
|
-
- project_id: typing.Optional[str].
|
|
45
|
-
|
|
46
|
-
- organization_id: typing.Optional[str].
|
|
47
|
-
|
|
48
|
-
- messages: typing.List[Message]. List of messages in the conversation
|
|
49
|
-
|
|
50
|
-
- model_configuration: ModelConfiguration. Configuration for the model to use in the response
|
|
51
|
-
---
|
|
52
|
-
from llama_cloud import ModelConfiguration, SupportedLlmModelNames
|
|
53
|
-
from llama_cloud.client import LlamaCloud
|
|
54
|
-
|
|
55
|
-
client = LlamaCloud(
|
|
56
|
-
token="YOUR_TOKEN",
|
|
57
|
-
)
|
|
58
|
-
client.responses.generate_response(
|
|
59
|
-
messages=[],
|
|
60
|
-
model_configuration=ModelConfiguration(
|
|
61
|
-
model_name=SupportedLlmModelNames.GPT_4_O,
|
|
62
|
-
),
|
|
63
|
-
)
|
|
64
|
-
"""
|
|
65
|
-
_response = self._client_wrapper.httpx_client.request(
|
|
66
|
-
"POST",
|
|
67
|
-
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/responses/generate"),
|
|
68
|
-
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
69
|
-
json=jsonable_encoder({"messages": messages, "model_configuration": model_configuration}),
|
|
70
|
-
headers=self._client_wrapper.get_headers(),
|
|
71
|
-
timeout=60,
|
|
72
|
-
)
|
|
73
|
-
if 200 <= _response.status_code < 300:
|
|
74
|
-
return pydantic.parse_obj_as(typing.Any, _response.json()) # type: ignore
|
|
75
|
-
if _response.status_code == 422:
|
|
76
|
-
raise UnprocessableEntityError(pydantic.parse_obj_as(HttpValidationError, _response.json())) # type: ignore
|
|
77
|
-
try:
|
|
78
|
-
_response_json = _response.json()
|
|
79
|
-
except JSONDecodeError:
|
|
80
|
-
raise ApiError(status_code=_response.status_code, body=_response.text)
|
|
81
|
-
raise ApiError(status_code=_response.status_code, body=_response_json)
|
|
82
|
-
|
|
83
|
-
|
|
84
|
-
class AsyncResponsesClient:
|
|
85
|
-
def __init__(self, *, client_wrapper: AsyncClientWrapper):
|
|
86
|
-
self._client_wrapper = client_wrapper
|
|
87
|
-
|
|
88
|
-
async def generate_response(
|
|
89
|
-
self,
|
|
90
|
-
*,
|
|
91
|
-
project_id: typing.Optional[str] = None,
|
|
92
|
-
organization_id: typing.Optional[str] = None,
|
|
93
|
-
messages: typing.List[Message],
|
|
94
|
-
model_configuration: ModelConfiguration,
|
|
95
|
-
) -> typing.Any:
|
|
96
|
-
"""
|
|
97
|
-
EXPERIMENTAL - SSE endpoint for basic response generation (dummy stream).
|
|
98
|
-
|
|
99
|
-
Parameters:
|
|
100
|
-
- project_id: typing.Optional[str].
|
|
101
|
-
|
|
102
|
-
- organization_id: typing.Optional[str].
|
|
103
|
-
|
|
104
|
-
- messages: typing.List[Message]. List of messages in the conversation
|
|
105
|
-
|
|
106
|
-
- model_configuration: ModelConfiguration. Configuration for the model to use in the response
|
|
107
|
-
---
|
|
108
|
-
from llama_cloud import ModelConfiguration, SupportedLlmModelNames
|
|
109
|
-
from llama_cloud.client import AsyncLlamaCloud
|
|
110
|
-
|
|
111
|
-
client = AsyncLlamaCloud(
|
|
112
|
-
token="YOUR_TOKEN",
|
|
113
|
-
)
|
|
114
|
-
await client.responses.generate_response(
|
|
115
|
-
messages=[],
|
|
116
|
-
model_configuration=ModelConfiguration(
|
|
117
|
-
model_name=SupportedLlmModelNames.GPT_4_O,
|
|
118
|
-
),
|
|
119
|
-
)
|
|
120
|
-
"""
|
|
121
|
-
_response = await self._client_wrapper.httpx_client.request(
|
|
122
|
-
"POST",
|
|
123
|
-
urllib.parse.urljoin(f"{self._client_wrapper.get_base_url()}/", "api/v1/responses/generate"),
|
|
124
|
-
params=remove_none_from_dict({"project_id": project_id, "organization_id": organization_id}),
|
|
125
|
-
json=jsonable_encoder({"messages": messages, "model_configuration": model_configuration}),
|
|
126
|
-
headers=self._client_wrapper.get_headers(),
|
|
127
|
-
timeout=60,
|
|
128
|
-
)
|
|
129
|
-
if 200 <= _response.status_code < 300:
|
|
130
|
-
return pydantic.parse_obj_as(typing.Any, _response.json()) # type: ignore
|
|
131
|
-
if _response.status_code == 422:
|
|
132
|
-
raise UnprocessableEntityError(pydantic.parse_obj_as(HttpValidationError, _response.json())) # type: ignore
|
|
133
|
-
try:
|
|
134
|
-
_response_json = _response.json()
|
|
135
|
-
except JSONDecodeError:
|
|
136
|
-
raise ApiError(status_code=_response.status_code, body=_response.text)
|
|
137
|
-
raise ApiError(status_code=_response.status_code, body=_response_json)
|
|
@@ -1,33 +0,0 @@
|
|
|
1
|
-
# This file was auto-generated by Fern from our API Definition.
|
|
2
|
-
|
|
3
|
-
import enum
|
|
4
|
-
import typing
|
|
5
|
-
|
|
6
|
-
T_Result = typing.TypeVar("T_Result")
|
|
7
|
-
|
|
8
|
-
|
|
9
|
-
class AppSchemaResponsesMessageRole(str, enum.Enum):
|
|
10
|
-
"""
|
|
11
|
-
Enum representing the role of a message in a conversation.
|
|
12
|
-
|
|
13
|
-
- system: The system message that sets the context or instructions.
|
|
14
|
-
- user: The user's message in the conversation.
|
|
15
|
-
- assistant: The AI assistant's response in the conversation.
|
|
16
|
-
"""
|
|
17
|
-
|
|
18
|
-
SYSTEM = "system"
|
|
19
|
-
USER = "user"
|
|
20
|
-
ASSISTANT = "assistant"
|
|
21
|
-
|
|
22
|
-
def visit(
|
|
23
|
-
self,
|
|
24
|
-
system: typing.Callable[[], T_Result],
|
|
25
|
-
user: typing.Callable[[], T_Result],
|
|
26
|
-
assistant: typing.Callable[[], T_Result],
|
|
27
|
-
) -> T_Result:
|
|
28
|
-
if self is AppSchemaResponsesMessageRole.SYSTEM:
|
|
29
|
-
return system()
|
|
30
|
-
if self is AppSchemaResponsesMessageRole.USER:
|
|
31
|
-
return user()
|
|
32
|
-
if self is AppSchemaResponsesMessageRole.ASSISTANT:
|
|
33
|
-
return assistant()
|
|
File without changes
|
|
File without changes
|