lionagi 0.4.0__py3-none-any.whl → 0.5.1__py3-none-any.whl
Sign up to get free protection for your applications and to get access to all the features.
- lionagi/__init__.py +14 -46
- lionagi/core/__init__.py +3 -1
- lionagi/core/_class_registry.py +69 -0
- lionagi/core/action/__init__.py +3 -13
- lionagi/core/action/action_manager.py +287 -0
- lionagi/core/action/base.py +109 -0
- lionagi/core/action/function_calling.py +127 -92
- lionagi/core/action/tool.py +172 -70
- lionagi/core/action/types.py +16 -0
- lionagi/core/communication/__init__.py +3 -0
- lionagi/core/communication/action_request.py +163 -0
- lionagi/core/communication/action_response.py +149 -0
- lionagi/core/communication/assistant_response.py +161 -0
- lionagi/core/communication/base_mail.py +49 -0
- lionagi/core/communication/instruction.py +376 -0
- lionagi/core/communication/message.py +286 -0
- lionagi/core/communication/message_manager.py +530 -0
- lionagi/core/communication/system.py +116 -0
- lionagi/core/communication/templates/README.md +28 -0
- lionagi/core/communication/templates/action_request.jinja2 +5 -0
- lionagi/core/communication/templates/action_response.jinja2 +9 -0
- lionagi/core/communication/templates/assistant_response.jinja2 +2 -0
- lionagi/core/communication/templates/instruction_message.jinja2 +61 -0
- lionagi/core/communication/templates/system_message.jinja2 +11 -0
- lionagi/core/communication/templates/tool_schemas.jinja2 +7 -0
- lionagi/core/communication/types.py +27 -0
- lionagi/core/communication/utils.py +254 -0
- lionagi/core/forms/__init__.py +3 -0
- lionagi/core/forms/base.py +232 -0
- lionagi/core/forms/form.py +791 -0
- lionagi/core/forms/report.py +321 -0
- lionagi/core/forms/types.py +13 -0
- lionagi/core/forms/utils.py +26 -0
- lionagi/core/generic/__init__.py +3 -6
- lionagi/core/generic/component.py +422 -0
- lionagi/core/generic/edge.py +143 -101
- lionagi/core/generic/element.py +195 -0
- lionagi/core/generic/graph.py +297 -180
- lionagi/core/generic/log.py +151 -0
- lionagi/core/generic/log_manager.py +320 -0
- lionagi/core/generic/node.py +7 -229
- lionagi/core/generic/pile.py +1017 -0
- lionagi/core/generic/progression.py +388 -0
- lionagi/core/generic/types.py +23 -0
- lionagi/core/generic/utils.py +50 -0
- lionagi/core/models/__init__.py +5 -0
- lionagi/core/models/base.py +85 -0
- lionagi/core/models/field_model.py +122 -0
- lionagi/core/models/new_model_params.py +195 -0
- lionagi/core/models/note.py +351 -0
- lionagi/core/models/operable_model.py +392 -0
- lionagi/core/models/schema_model.py +50 -0
- lionagi/core/models/types.py +10 -0
- lionagi/core/session/__init__.py +3 -0
- lionagi/core/session/branch.py +115 -415
- lionagi/core/session/branch_mixins.py +545 -0
- lionagi/core/session/session.py +122 -257
- lionagi/core/session/types.py +8 -0
- lionagi/core/typing/__init__.py +9 -0
- lionagi/core/typing/concepts.py +132 -0
- lionagi/core/typing/config.py +15 -0
- lionagi/core/typing/id.py +221 -0
- lionagi/core/typing/pydantic_.py +33 -0
- lionagi/core/typing/typing_.py +54 -0
- lionagi/integrations/__init__.py +0 -1
- lionagi/integrations/anthropic_/AnthropicModel.py +268 -0
- lionagi/integrations/anthropic_/AnthropicService.py +117 -0
- lionagi/integrations/anthropic_/__init__.py +3 -0
- lionagi/integrations/anthropic_/anthropic_max_output_token_data.yaml +7 -0
- lionagi/integrations/anthropic_/anthropic_price_data.yaml +14 -0
- lionagi/integrations/anthropic_/api_endpoints/__init__.py +3 -0
- lionagi/integrations/anthropic_/api_endpoints/api_request.py +277 -0
- lionagi/integrations/anthropic_/api_endpoints/data_models.py +40 -0
- lionagi/integrations/anthropic_/api_endpoints/match_response.py +119 -0
- lionagi/integrations/anthropic_/api_endpoints/messages/__init__.py +3 -0
- lionagi/integrations/anthropic_/api_endpoints/messages/request/__init__.py +3 -0
- lionagi/integrations/anthropic_/api_endpoints/messages/request/message_models.py +14 -0
- lionagi/integrations/anthropic_/api_endpoints/messages/request/request_body.py +74 -0
- lionagi/integrations/anthropic_/api_endpoints/messages/response/content_models.py +32 -0
- lionagi/integrations/anthropic_/api_endpoints/messages/response/response_body.py +101 -0
- lionagi/integrations/anthropic_/api_endpoints/messages/response/usage_models.py +25 -0
- lionagi/integrations/anthropic_/version.py +5 -0
- lionagi/integrations/groq_/GroqModel.py +318 -0
- lionagi/integrations/groq_/GroqService.py +151 -0
- lionagi/integrations/groq_/__init__.py +3 -0
- lionagi/integrations/groq_/api_endpoints/data_models.py +187 -0
- lionagi/integrations/groq_/api_endpoints/groq_request.py +288 -0
- lionagi/integrations/groq_/api_endpoints/match_response.py +106 -0
- lionagi/integrations/groq_/api_endpoints/response_utils.py +105 -0
- lionagi/integrations/groq_/groq_max_output_token_data.yaml +21 -0
- lionagi/integrations/groq_/groq_price_data.yaml +58 -0
- lionagi/integrations/groq_/groq_rate_limits.yaml +105 -0
- lionagi/integrations/groq_/version.py +5 -0
- lionagi/integrations/litellm_/__init__.py +3 -0
- lionagi/integrations/litellm_/imodel.py +73 -0
- lionagi/integrations/ollama_/OllamaModel.py +244 -0
- lionagi/integrations/ollama_/OllamaService.py +142 -0
- lionagi/integrations/ollama_/__init__.py +3 -0
- lionagi/integrations/ollama_/api_endpoints/__init__.py +3 -0
- lionagi/integrations/ollama_/api_endpoints/api_request.py +179 -0
- lionagi/integrations/ollama_/api_endpoints/chat_completion/__init__.py +3 -0
- lionagi/integrations/ollama_/api_endpoints/chat_completion/message_models.py +31 -0
- lionagi/integrations/ollama_/api_endpoints/chat_completion/request_body.py +46 -0
- lionagi/integrations/ollama_/api_endpoints/chat_completion/response_body.py +67 -0
- lionagi/integrations/ollama_/api_endpoints/chat_completion/tool_models.py +49 -0
- lionagi/integrations/ollama_/api_endpoints/completion/request_body.py +72 -0
- lionagi/integrations/ollama_/api_endpoints/completion/response_body.py +59 -0
- lionagi/integrations/ollama_/api_endpoints/data_models.py +15 -0
- lionagi/integrations/ollama_/api_endpoints/embedding/request_body.py +33 -0
- lionagi/integrations/ollama_/api_endpoints/embedding/response_body.py +29 -0
- lionagi/integrations/ollama_/api_endpoints/match_data_model.py +62 -0
- lionagi/integrations/ollama_/api_endpoints/match_response.py +190 -0
- lionagi/integrations/ollama_/api_endpoints/model/__init__.py +3 -0
- lionagi/integrations/ollama_/api_endpoints/model/copy_model.py +13 -0
- lionagi/integrations/ollama_/api_endpoints/model/create_model.py +28 -0
- lionagi/integrations/ollama_/api_endpoints/model/delete_model.py +11 -0
- lionagi/integrations/ollama_/api_endpoints/model/list_model.py +60 -0
- lionagi/integrations/ollama_/api_endpoints/model/pull_model.py +34 -0
- lionagi/integrations/ollama_/api_endpoints/model/push_model.py +35 -0
- lionagi/integrations/ollama_/api_endpoints/model/show_model.py +36 -0
- lionagi/integrations/ollama_/api_endpoints/option_models.py +68 -0
- lionagi/integrations/openai_/OpenAIModel.py +418 -0
- lionagi/integrations/openai_/OpenAIService.py +426 -0
- lionagi/integrations/openai_/api_endpoints/__init__.py +3 -0
- lionagi/integrations/openai_/api_endpoints/api_request.py +277 -0
- lionagi/integrations/openai_/api_endpoints/audio/__init__.py +9 -0
- lionagi/integrations/openai_/api_endpoints/audio/speech_models.py +34 -0
- lionagi/integrations/openai_/api_endpoints/audio/transcription_models.py +136 -0
- lionagi/integrations/openai_/api_endpoints/audio/translation_models.py +41 -0
- lionagi/integrations/openai_/api_endpoints/audio/types.py +41 -0
- lionagi/integrations/openai_/api_endpoints/batch/__init__.py +17 -0
- lionagi/integrations/openai_/api_endpoints/batch/batch_models.py +146 -0
- lionagi/integrations/openai_/api_endpoints/batch/cancel_batch.py +7 -0
- lionagi/integrations/openai_/api_endpoints/batch/create_batch.py +26 -0
- lionagi/integrations/openai_/api_endpoints/batch/list_batch.py +37 -0
- lionagi/integrations/openai_/api_endpoints/batch/request_object_models.py +65 -0
- lionagi/integrations/openai_/api_endpoints/batch/retrieve_batch.py +7 -0
- lionagi/integrations/openai_/api_endpoints/batch/types.py +4 -0
- lionagi/integrations/openai_/api_endpoints/chat_completions/__init__.py +1 -0
- lionagi/integrations/openai_/api_endpoints/chat_completions/request/__init__.py +39 -0
- lionagi/integrations/openai_/api_endpoints/chat_completions/request/message_models.py +121 -0
- lionagi/integrations/openai_/api_endpoints/chat_completions/request/request_body.py +221 -0
- lionagi/integrations/openai_/api_endpoints/chat_completions/request/response_format.py +71 -0
- lionagi/integrations/openai_/api_endpoints/chat_completions/request/stream_options.py +14 -0
- lionagi/integrations/openai_/api_endpoints/chat_completions/request/tool_choice_models.py +17 -0
- lionagi/integrations/openai_/api_endpoints/chat_completions/request/tool_models.py +54 -0
- lionagi/integrations/openai_/api_endpoints/chat_completions/request/types.py +18 -0
- lionagi/integrations/openai_/api_endpoints/chat_completions/response/choice_models.py +62 -0
- lionagi/integrations/openai_/api_endpoints/chat_completions/response/function_models.py +16 -0
- lionagi/integrations/openai_/api_endpoints/chat_completions/response/log_prob_models.py +47 -0
- lionagi/integrations/openai_/api_endpoints/chat_completions/response/message_models.py +25 -0
- lionagi/integrations/openai_/api_endpoints/chat_completions/response/response_body.py +99 -0
- lionagi/integrations/openai_/api_endpoints/chat_completions/response/types.py +8 -0
- lionagi/integrations/openai_/api_endpoints/chat_completions/response/usage_models.py +24 -0
- lionagi/integrations/openai_/api_endpoints/chat_completions/util.py +46 -0
- lionagi/integrations/openai_/api_endpoints/data_models.py +23 -0
- lionagi/integrations/openai_/api_endpoints/embeddings/__init__.py +3 -0
- lionagi/integrations/openai_/api_endpoints/embeddings/request_body.py +79 -0
- lionagi/integrations/openai_/api_endpoints/embeddings/response_body.py +67 -0
- lionagi/integrations/openai_/api_endpoints/files/__init__.py +11 -0
- lionagi/integrations/openai_/api_endpoints/files/delete_file.py +20 -0
- lionagi/integrations/openai_/api_endpoints/files/file_models.py +56 -0
- lionagi/integrations/openai_/api_endpoints/files/list_files.py +27 -0
- lionagi/integrations/openai_/api_endpoints/files/retrieve_file.py +9 -0
- lionagi/integrations/openai_/api_endpoints/files/upload_file.py +38 -0
- lionagi/integrations/openai_/api_endpoints/fine_tuning/__init__.py +37 -0
- lionagi/integrations/openai_/api_endpoints/fine_tuning/cancel_jobs.py +9 -0
- lionagi/integrations/openai_/api_endpoints/fine_tuning/create_jobs.py +133 -0
- lionagi/integrations/openai_/api_endpoints/fine_tuning/fine_tuning_job_checkpoint_models.py +58 -0
- lionagi/integrations/openai_/api_endpoints/fine_tuning/fine_tuning_job_event_models.py +31 -0
- lionagi/integrations/openai_/api_endpoints/fine_tuning/fine_tuning_job_models.py +140 -0
- lionagi/integrations/openai_/api_endpoints/fine_tuning/list_fine_tuning_checkpoints.py +51 -0
- lionagi/integrations/openai_/api_endpoints/fine_tuning/list_fine_tuning_events.py +42 -0
- lionagi/integrations/openai_/api_endpoints/fine_tuning/list_fine_tuning_jobs.py +31 -0
- lionagi/integrations/openai_/api_endpoints/fine_tuning/retrieve_jobs.py +9 -0
- lionagi/integrations/openai_/api_endpoints/fine_tuning/training_format.py +30 -0
- lionagi/integrations/openai_/api_endpoints/images/__init__.py +9 -0
- lionagi/integrations/openai_/api_endpoints/images/image_edit_models.py +69 -0
- lionagi/integrations/openai_/api_endpoints/images/image_models.py +56 -0
- lionagi/integrations/openai_/api_endpoints/images/image_variation_models.py +56 -0
- lionagi/integrations/openai_/api_endpoints/images/response_body.py +30 -0
- lionagi/integrations/openai_/api_endpoints/match_data_model.py +197 -0
- lionagi/integrations/openai_/api_endpoints/match_response.py +336 -0
- lionagi/integrations/openai_/api_endpoints/models/__init__.py +7 -0
- lionagi/integrations/openai_/api_endpoints/models/delete_fine_tuned_model.py +17 -0
- lionagi/integrations/openai_/api_endpoints/models/models_models.py +31 -0
- lionagi/integrations/openai_/api_endpoints/models/retrieve_model.py +9 -0
- lionagi/integrations/openai_/api_endpoints/moderations/__init__.py +3 -0
- lionagi/integrations/openai_/api_endpoints/moderations/request_body.py +20 -0
- lionagi/integrations/openai_/api_endpoints/moderations/response_body.py +139 -0
- lionagi/integrations/openai_/api_endpoints/uploads/__init__.py +19 -0
- lionagi/integrations/openai_/api_endpoints/uploads/add_upload_part.py +11 -0
- lionagi/integrations/openai_/api_endpoints/uploads/cancel_upload.py +7 -0
- lionagi/integrations/openai_/api_endpoints/uploads/complete_upload.py +18 -0
- lionagi/integrations/openai_/api_endpoints/uploads/create_upload.py +17 -0
- lionagi/integrations/openai_/api_endpoints/uploads/uploads_models.py +52 -0
- lionagi/integrations/openai_/image_token_calculator/image_token_calculator.py +92 -0
- lionagi/integrations/openai_/image_token_calculator/openai_image_token_data.yaml +15 -0
- lionagi/integrations/openai_/openai_max_output_token_data.yaml +12 -0
- lionagi/integrations/openai_/openai_price_data.yaml +26 -0
- lionagi/integrations/openai_/version.py +1 -0
- lionagi/integrations/pandas_/__init__.py +24 -0
- lionagi/integrations/pandas_/extend_df.py +61 -0
- lionagi/integrations/pandas_/read.py +103 -0
- lionagi/integrations/pandas_/remove_rows.py +61 -0
- lionagi/integrations/pandas_/replace_keywords.py +65 -0
- lionagi/integrations/pandas_/save.py +131 -0
- lionagi/integrations/pandas_/search_keywords.py +69 -0
- lionagi/integrations/pandas_/to_df.py +196 -0
- lionagi/integrations/pandas_/update_cells.py +54 -0
- lionagi/integrations/perplexity_/PerplexityModel.py +269 -0
- lionagi/integrations/perplexity_/PerplexityService.py +113 -0
- lionagi/integrations/perplexity_/__init__.py +3 -0
- lionagi/integrations/perplexity_/api_endpoints/api_request.py +171 -0
- lionagi/integrations/perplexity_/api_endpoints/chat_completions/request/request_body.py +121 -0
- lionagi/integrations/perplexity_/api_endpoints/chat_completions/response/response_body.py +146 -0
- lionagi/integrations/perplexity_/api_endpoints/data_models.py +63 -0
- lionagi/integrations/perplexity_/api_endpoints/match_response.py +26 -0
- lionagi/integrations/perplexity_/perplexity_max_output_token_data.yaml +3 -0
- lionagi/integrations/perplexity_/perplexity_price_data.yaml +10 -0
- lionagi/integrations/perplexity_/version.py +1 -0
- lionagi/integrations/pydantic_/__init__.py +8 -0
- lionagi/integrations/pydantic_/break_down_annotation.py +81 -0
- lionagi/integrations/pydantic_/new_model.py +208 -0
- lionagi/integrations/services.py +17 -0
- lionagi/libs/__init__.py +0 -55
- lionagi/libs/compress/models.py +62 -0
- lionagi/libs/compress/utils.py +81 -0
- lionagi/libs/constants.py +98 -0
- lionagi/libs/file/chunk.py +265 -0
- lionagi/libs/file/file_ops.py +114 -0
- lionagi/libs/file/params.py +212 -0
- lionagi/libs/file/path.py +301 -0
- lionagi/libs/file/process.py +139 -0
- lionagi/libs/file/save.py +90 -0
- lionagi/libs/file/types.py +22 -0
- lionagi/libs/func/async_calls/__init__.py +21 -0
- lionagi/libs/func/async_calls/alcall.py +157 -0
- lionagi/libs/func/async_calls/bcall.py +82 -0
- lionagi/libs/func/async_calls/mcall.py +134 -0
- lionagi/libs/func/async_calls/pcall.py +149 -0
- lionagi/libs/func/async_calls/rcall.py +185 -0
- lionagi/libs/func/async_calls/tcall.py +114 -0
- lionagi/libs/func/async_calls/ucall.py +85 -0
- lionagi/libs/func/decorators.py +277 -0
- lionagi/libs/func/lcall.py +57 -0
- lionagi/libs/func/params.py +64 -0
- lionagi/libs/func/throttle.py +119 -0
- lionagi/libs/func/types.py +39 -0
- lionagi/libs/func/utils.py +96 -0
- lionagi/libs/package/imports.py +162 -0
- lionagi/libs/package/management.py +58 -0
- lionagi/libs/package/params.py +26 -0
- lionagi/libs/package/system.py +18 -0
- lionagi/libs/package/types.py +26 -0
- lionagi/libs/parse/__init__.py +1 -0
- lionagi/libs/parse/flatten/__init__.py +9 -0
- lionagi/libs/parse/flatten/flatten.py +168 -0
- lionagi/libs/parse/flatten/params.py +52 -0
- lionagi/libs/parse/flatten/unflatten.py +79 -0
- lionagi/libs/parse/json/__init__.py +27 -0
- lionagi/libs/parse/json/as_readable.py +104 -0
- lionagi/libs/parse/json/extract.py +102 -0
- lionagi/libs/parse/json/parse.py +179 -0
- lionagi/libs/parse/json/schema.py +227 -0
- lionagi/libs/parse/json/to_json.py +71 -0
- lionagi/libs/parse/nested/__init__.py +33 -0
- lionagi/libs/parse/nested/nfilter.py +55 -0
- lionagi/libs/parse/nested/nget.py +40 -0
- lionagi/libs/parse/nested/ninsert.py +103 -0
- lionagi/libs/parse/nested/nmerge.py +155 -0
- lionagi/libs/parse/nested/npop.py +66 -0
- lionagi/libs/parse/nested/nset.py +89 -0
- lionagi/libs/parse/nested/to_flat_list.py +64 -0
- lionagi/libs/parse/nested/utils.py +185 -0
- lionagi/libs/parse/string_parse/__init__.py +11 -0
- lionagi/libs/parse/string_parse/code_block.py +73 -0
- lionagi/libs/parse/string_parse/docstring.py +179 -0
- lionagi/libs/parse/string_parse/function_.py +92 -0
- lionagi/libs/parse/type_convert/__init__.py +19 -0
- lionagi/libs/parse/type_convert/params.py +145 -0
- lionagi/libs/parse/type_convert/to_dict.py +333 -0
- lionagi/libs/parse/type_convert/to_list.py +186 -0
- lionagi/libs/parse/type_convert/to_num.py +358 -0
- lionagi/libs/parse/type_convert/to_str.py +195 -0
- lionagi/libs/parse/types.py +9 -0
- lionagi/libs/parse/validate/__init__.py +14 -0
- lionagi/libs/parse/validate/boolean.py +96 -0
- lionagi/libs/parse/validate/keys.py +150 -0
- lionagi/libs/parse/validate/mapping.py +109 -0
- lionagi/libs/parse/validate/params.py +62 -0
- lionagi/libs/parse/xml/__init__.py +10 -0
- lionagi/libs/parse/xml/convert.py +56 -0
- lionagi/libs/parse/xml/parser.py +93 -0
- lionagi/libs/string_similarity/__init__.py +32 -0
- lionagi/libs/string_similarity/algorithms.py +219 -0
- lionagi/libs/string_similarity/matcher.py +102 -0
- lionagi/libs/string_similarity/utils.py +15 -0
- lionagi/libs/utils.py +255 -0
- lionagi/operations/__init__.py +3 -6
- lionagi/operations/brainstorm/__init__.py +3 -0
- lionagi/operations/brainstorm/brainstorm.py +204 -0
- lionagi/operations/brainstorm/prompt.py +1 -0
- lionagi/operations/plan/__init__.py +3 -0
- lionagi/operations/plan/plan.py +172 -0
- lionagi/operations/plan/prompt.py +21 -0
- lionagi/operations/select/__init__.py +3 -0
- lionagi/operations/select/prompt.py +1 -0
- lionagi/operations/select/select.py +100 -0
- lionagi/operations/select/utils.py +107 -0
- lionagi/operations/utils.py +35 -0
- lionagi/protocols/adapters/adapter.py +79 -0
- lionagi/protocols/adapters/json_adapter.py +43 -0
- lionagi/protocols/adapters/pandas_adapter.py +96 -0
- lionagi/protocols/configs/__init__.py +15 -0
- lionagi/protocols/configs/branch_config.py +86 -0
- lionagi/protocols/configs/id_config.py +15 -0
- lionagi/protocols/configs/imodel_config.py +73 -0
- lionagi/protocols/configs/log_config.py +93 -0
- lionagi/protocols/configs/retry_config.py +29 -0
- lionagi/protocols/operatives/__init__.py +15 -0
- lionagi/protocols/operatives/action.py +181 -0
- lionagi/protocols/operatives/instruct.py +196 -0
- lionagi/protocols/operatives/operative.py +182 -0
- lionagi/protocols/operatives/prompts.py +232 -0
- lionagi/protocols/operatives/reason.py +56 -0
- lionagi/protocols/operatives/step.py +217 -0
- lionagi/protocols/registries/_component_registry.py +19 -0
- lionagi/protocols/registries/_pile_registry.py +26 -0
- lionagi/service/__init__.py +13 -0
- lionagi/service/complete_request_info.py +11 -0
- lionagi/service/imodel.py +125 -0
- lionagi/service/rate_limiter.py +108 -0
- lionagi/service/service.py +41 -0
- lionagi/service/service_match_util.py +131 -0
- lionagi/service/service_util.py +72 -0
- lionagi/service/token_calculator.py +51 -0
- lionagi/settings.py +136 -0
- lionagi/strategies/base.py +53 -0
- lionagi/strategies/concurrent.py +71 -0
- lionagi/strategies/concurrent_chunk.py +43 -0
- lionagi/strategies/concurrent_sequential_chunk.py +104 -0
- lionagi/strategies/params.py +128 -0
- lionagi/strategies/sequential.py +23 -0
- lionagi/strategies/sequential_chunk.py +89 -0
- lionagi/strategies/sequential_concurrent_chunk.py +100 -0
- lionagi/strategies/types.py +21 -0
- lionagi/strategies/utils.py +49 -0
- lionagi/version.py +1 -1
- lionagi-0.5.1.dist-info/METADATA +545 -0
- lionagi-0.5.1.dist-info/RECORD +373 -0
- {lionagi-0.4.0.dist-info → lionagi-0.5.1.dist-info}/WHEEL +1 -1
- lionagi/core/_setting/_setting.py +0 -59
- lionagi/core/action/README.md +0 -20
- lionagi/core/action/manual.py +0 -1
- lionagi/core/action/node.py +0 -94
- lionagi/core/action/tool_manager.py +0 -342
- lionagi/core/agent/README.md +0 -1
- lionagi/core/agent/base_agent.py +0 -82
- lionagi/core/agent/eval/README.md +0 -1
- lionagi/core/agent/eval/evaluator.py +0 -1
- lionagi/core/agent/eval/vote.py +0 -40
- lionagi/core/agent/learn/learner.py +0 -59
- lionagi/core/agent/plan/unit_template.py +0 -1
- lionagi/core/collections/README.md +0 -23
- lionagi/core/collections/__init__.py +0 -16
- lionagi/core/collections/_logger.py +0 -312
- lionagi/core/collections/abc/README.md +0 -63
- lionagi/core/collections/abc/__init__.py +0 -53
- lionagi/core/collections/abc/component.py +0 -620
- lionagi/core/collections/abc/concepts.py +0 -277
- lionagi/core/collections/abc/exceptions.py +0 -136
- lionagi/core/collections/abc/util.py +0 -45
- lionagi/core/collections/exchange.py +0 -146
- lionagi/core/collections/flow.py +0 -416
- lionagi/core/collections/model.py +0 -465
- lionagi/core/collections/pile.py +0 -1232
- lionagi/core/collections/progression.py +0 -221
- lionagi/core/collections/util.py +0 -73
- lionagi/core/director/README.md +0 -1
- lionagi/core/director/direct.py +0 -298
- lionagi/core/director/director.py +0 -2
- lionagi/core/director/operations/select.py +0 -3
- lionagi/core/director/operations/utils.py +0 -6
- lionagi/core/engine/branch_engine.py +0 -361
- lionagi/core/engine/instruction_map_engine.py +0 -213
- lionagi/core/engine/sandbox_.py +0 -16
- lionagi/core/engine/script_engine.py +0 -89
- lionagi/core/executor/base_executor.py +0 -97
- lionagi/core/executor/graph_executor.py +0 -335
- lionagi/core/executor/neo4j_executor.py +0 -394
- lionagi/core/generic/README.md +0 -0
- lionagi/core/generic/edge_condition.py +0 -17
- lionagi/core/generic/hyperedge.py +0 -1
- lionagi/core/generic/tree.py +0 -49
- lionagi/core/generic/tree_node.py +0 -85
- lionagi/core/mail/__init__.py +0 -11
- lionagi/core/mail/mail.py +0 -26
- lionagi/core/mail/mail_manager.py +0 -185
- lionagi/core/mail/package.py +0 -49
- lionagi/core/mail/start_mail.py +0 -36
- lionagi/core/message/__init__.py +0 -18
- lionagi/core/message/action_request.py +0 -114
- lionagi/core/message/action_response.py +0 -121
- lionagi/core/message/assistant_response.py +0 -80
- lionagi/core/message/instruction.py +0 -194
- lionagi/core/message/message.py +0 -86
- lionagi/core/message/system.py +0 -71
- lionagi/core/message/util.py +0 -274
- lionagi/core/report/__init__.py +0 -4
- lionagi/core/report/base.py +0 -201
- lionagi/core/report/form.py +0 -212
- lionagi/core/report/report.py +0 -150
- lionagi/core/report/util.py +0 -15
- lionagi/core/rule/_default.py +0 -17
- lionagi/core/rule/action.py +0 -87
- lionagi/core/rule/base.py +0 -234
- lionagi/core/rule/boolean.py +0 -56
- lionagi/core/rule/choice.py +0 -48
- lionagi/core/rule/mapping.py +0 -82
- lionagi/core/rule/number.py +0 -73
- lionagi/core/rule/rulebook.py +0 -45
- lionagi/core/rule/string.py +0 -43
- lionagi/core/rule/util.py +0 -0
- lionagi/core/session/directive_mixin.py +0 -307
- lionagi/core/structure/__init__.py +0 -1
- lionagi/core/structure/chain.py +0 -1
- lionagi/core/structure/forest.py +0 -1
- lionagi/core/structure/graph.py +0 -1
- lionagi/core/structure/tree.py +0 -1
- lionagi/core/unit/__init__.py +0 -4
- lionagi/core/unit/parallel_unit.py +0 -234
- lionagi/core/unit/template/action.py +0 -65
- lionagi/core/unit/template/base.py +0 -35
- lionagi/core/unit/template/plan.py +0 -69
- lionagi/core/unit/template/predict.py +0 -95
- lionagi/core/unit/template/score.py +0 -108
- lionagi/core/unit/template/select.py +0 -91
- lionagi/core/unit/unit.py +0 -452
- lionagi/core/unit/unit_form.py +0 -290
- lionagi/core/unit/unit_mixin.py +0 -1166
- lionagi/core/unit/util.py +0 -103
- lionagi/core/validator/validator.py +0 -376
- lionagi/core/work/work.py +0 -59
- lionagi/core/work/work_edge.py +0 -102
- lionagi/core/work/work_function.py +0 -114
- lionagi/core/work/work_function_node.py +0 -50
- lionagi/core/work/work_queue.py +0 -90
- lionagi/core/work/work_task.py +0 -151
- lionagi/core/work/worker.py +0 -410
- lionagi/core/work/worker_engine.py +0 -208
- lionagi/core/work/worklog.py +0 -108
- lionagi/experimental/compressor/base.py +0 -47
- lionagi/experimental/compressor/llm_compressor.py +0 -265
- lionagi/experimental/compressor/llm_summarizer.py +0 -61
- lionagi/experimental/compressor/util.py +0 -70
- lionagi/experimental/directive/README.md +0 -1
- lionagi/experimental/directive/__init__.py +0 -19
- lionagi/experimental/directive/parser/base_parser.py +0 -294
- lionagi/experimental/directive/parser/base_syntax.txt +0 -200
- lionagi/experimental/directive/template/base_template.py +0 -71
- lionagi/experimental/directive/template/schema.py +0 -36
- lionagi/experimental/directive/tokenizer.py +0 -59
- lionagi/experimental/evaluator/README.md +0 -1
- lionagi/experimental/evaluator/ast_evaluator.py +0 -119
- lionagi/experimental/evaluator/base_evaluator.py +0 -213
- lionagi/experimental/knowledge/__init__.py +0 -0
- lionagi/experimental/knowledge/base.py +0 -10
- lionagi/experimental/knowledge/graph.py +0 -0
- lionagi/experimental/memory/__init__.py +0 -0
- lionagi/experimental/strategies/__init__.py +0 -0
- lionagi/experimental/strategies/base.py +0 -1
- lionagi/integrations/bridge/__init__.py +0 -4
- lionagi/integrations/bridge/autogen_/__init__.py +0 -0
- lionagi/integrations/bridge/autogen_/autogen_.py +0 -127
- lionagi/integrations/bridge/langchain_/__init__.py +0 -0
- lionagi/integrations/bridge/langchain_/documents.py +0 -138
- lionagi/integrations/bridge/langchain_/langchain_bridge.py +0 -68
- lionagi/integrations/bridge/llamaindex_/__init__.py +0 -0
- lionagi/integrations/bridge/llamaindex_/index.py +0 -36
- lionagi/integrations/bridge/llamaindex_/llama_index_bridge.py +0 -108
- lionagi/integrations/bridge/llamaindex_/llama_pack.py +0 -256
- lionagi/integrations/bridge/llamaindex_/node_parser.py +0 -92
- lionagi/integrations/bridge/llamaindex_/reader.py +0 -201
- lionagi/integrations/bridge/llamaindex_/textnode.py +0 -59
- lionagi/integrations/bridge/pydantic_/__init__.py +0 -0
- lionagi/integrations/bridge/pydantic_/pydantic_bridge.py +0 -7
- lionagi/integrations/bridge/transformers_/__init__.py +0 -0
- lionagi/integrations/bridge/transformers_/install_.py +0 -39
- lionagi/integrations/chunker/__init__.py +0 -0
- lionagi/integrations/chunker/chunk.py +0 -314
- lionagi/integrations/config/__init__.py +0 -4
- lionagi/integrations/config/mlx_configs.py +0 -1
- lionagi/integrations/config/oai_configs.py +0 -154
- lionagi/integrations/config/ollama_configs.py +0 -1
- lionagi/integrations/config/openrouter_configs.py +0 -74
- lionagi/integrations/langchain_/__init__.py +0 -0
- lionagi/integrations/llamaindex_/__init__.py +0 -0
- lionagi/integrations/loader/__init__.py +0 -0
- lionagi/integrations/loader/load.py +0 -257
- lionagi/integrations/loader/load_util.py +0 -214
- lionagi/integrations/provider/__init__.py +0 -11
- lionagi/integrations/provider/_mapping.py +0 -47
- lionagi/integrations/provider/litellm.py +0 -53
- lionagi/integrations/provider/mistralai.py +0 -1
- lionagi/integrations/provider/mlx_service.py +0 -55
- lionagi/integrations/provider/oai.py +0 -196
- lionagi/integrations/provider/ollama.py +0 -55
- lionagi/integrations/provider/openrouter.py +0 -170
- lionagi/integrations/provider/services.py +0 -138
- lionagi/integrations/provider/transformers.py +0 -108
- lionagi/integrations/storage/__init__.py +0 -3
- lionagi/integrations/storage/neo4j.py +0 -681
- lionagi/integrations/storage/storage_util.py +0 -302
- lionagi/integrations/storage/structure_excel.py +0 -291
- lionagi/integrations/storage/to_csv.py +0 -70
- lionagi/integrations/storage/to_excel.py +0 -91
- lionagi/libs/ln_api.py +0 -944
- lionagi/libs/ln_async.py +0 -208
- lionagi/libs/ln_context.py +0 -37
- lionagi/libs/ln_convert.py +0 -671
- lionagi/libs/ln_dataframe.py +0 -187
- lionagi/libs/ln_func_call.py +0 -1328
- lionagi/libs/ln_image.py +0 -114
- lionagi/libs/ln_knowledge_graph.py +0 -422
- lionagi/libs/ln_nested.py +0 -822
- lionagi/libs/ln_parse.py +0 -750
- lionagi/libs/ln_queue.py +0 -107
- lionagi/libs/ln_tokenize.py +0 -179
- lionagi/libs/ln_validate.py +0 -299
- lionagi/libs/special_tokens.py +0 -172
- lionagi/libs/sys_util.py +0 -710
- lionagi/lions/__init__.py +0 -0
- lionagi/lions/coder/__init__.py +0 -0
- lionagi/lions/coder/add_feature.py +0 -20
- lionagi/lions/coder/base_prompts.py +0 -22
- lionagi/lions/coder/code_form.py +0 -15
- lionagi/lions/coder/coder.py +0 -184
- lionagi/lions/coder/util.py +0 -101
- lionagi/lions/director/__init__.py +0 -0
- lionagi/lions/judge/__init__.py +0 -0
- lionagi/lions/judge/config.py +0 -8
- lionagi/lions/judge/data/__init__.py +0 -0
- lionagi/lions/judge/data/sample_codes.py +0 -526
- lionagi/lions/judge/data/sample_rurbic.py +0 -48
- lionagi/lions/judge/forms/__init__.py +0 -0
- lionagi/lions/judge/forms/code_analysis_form.py +0 -126
- lionagi/lions/judge/rubric.py +0 -34
- lionagi/lions/judge/services/__init__.py +0 -0
- lionagi/lions/judge/services/judge_code.py +0 -49
- lionagi/lions/researcher/__init__.py +0 -0
- lionagi/lions/researcher/data_source/__init__.py +0 -0
- lionagi/lions/researcher/data_source/finhub_.py +0 -192
- lionagi/lions/researcher/data_source/google_.py +0 -207
- lionagi/lions/researcher/data_source/wiki_.py +0 -98
- lionagi/lions/researcher/data_source/yfinance_.py +0 -21
- lionagi/operations/brainstorm.py +0 -87
- lionagi/operations/config.py +0 -6
- lionagi/operations/rank.py +0 -102
- lionagi/operations/score.py +0 -144
- lionagi/operations/select.py +0 -141
- lionagi-0.4.0.dist-info/METADATA +0 -241
- lionagi-0.4.0.dist-info/RECORD +0 -249
- /lionagi/{core/_setting → integrations/anthropic_/api_endpoints/messages/response}/__init__.py +0 -0
- /lionagi/{core/agent → integrations/groq_/api_endpoints}/__init__.py +0 -0
- /lionagi/{core/agent/eval → integrations/ollama_/api_endpoints/completion}/__init__.py +0 -0
- /lionagi/{core/agent/learn → integrations/ollama_/api_endpoints/embedding}/__init__.py +0 -0
- /lionagi/{core/agent/plan → integrations/openai_}/__init__.py +0 -0
- /lionagi/{core/director → integrations/openai_/api_endpoints/chat_completions/response}/__init__.py +0 -0
- /lionagi/{core/director/operations → integrations/openai_/image_token_calculator}/__init__.py +0 -0
- /lionagi/{core/engine → integrations/perplexity_/api_endpoints}/__init__.py +0 -0
- /lionagi/{core/executor → integrations/perplexity_/api_endpoints/chat_completions}/__init__.py +0 -0
- /lionagi/{core/generic/registry/component_registry → integrations/perplexity_/api_endpoints/chat_completions/request}/__init__.py +0 -0
- /lionagi/{core/rule → integrations/perplexity_/api_endpoints/chat_completions/response}/__init__.py +0 -0
- /lionagi/{core/unit/template → libs/compress}/__init__.py +0 -0
- /lionagi/{core/validator → libs/file}/__init__.py +0 -0
- /lionagi/{core/work → libs/func}/__init__.py +0 -0
- /lionagi/{experimental → libs/package}/__init__.py +0 -0
- /lionagi/{core/agent/plan/plan.py → libs/parse/params.py} +0 -0
- /lionagi/{experimental/compressor → protocols}/__init__.py +0 -0
- /lionagi/{experimental/directive/parser → protocols/adapters}/__init__.py +0 -0
- /lionagi/{experimental/directive/template → protocols/registries}/__init__.py +0 -0
- /lionagi/{experimental/evaluator → strategies}/__init__.py +0 -0
- {lionagi-0.4.0.dist-info → lionagi-0.5.1.dist-info/licenses}/LICENSE +0 -0
lionagi/core/work/worklog.py
DELETED
@@ -1,108 +0,0 @@
|
|
1
|
-
from lionagi.core.collections import Pile, pile, progression
|
2
|
-
from lionagi.core.collections.abc import Progressable
|
3
|
-
from lionagi.core.work.work import Work, WorkStatus
|
4
|
-
from lionagi.core.work.work_queue import WorkQueue
|
5
|
-
|
6
|
-
|
7
|
-
class WorkLog(Progressable):
|
8
|
-
"""
|
9
|
-
A class representing a log of work items.
|
10
|
-
|
11
|
-
Attributes:
|
12
|
-
pile (Pile): A pile containing work items.
|
13
|
-
pending (Progression): A progression of pending work items.
|
14
|
-
queue (WorkQueue): A queue to manage the execution of work items.
|
15
|
-
"""
|
16
|
-
|
17
|
-
def __init__(self, capacity=10, workpile=None, refresh_time=1):
|
18
|
-
"""
|
19
|
-
Initializes a new instance of WorkLog.
|
20
|
-
|
21
|
-
Args:
|
22
|
-
capacity (int): The capacity of the work queue batch processing.
|
23
|
-
workpile (Pile, optional): An optional pile of initial work items.
|
24
|
-
refresh_time (int, optional): The time interval to refresh the work log queue.
|
25
|
-
Defaults to 1.
|
26
|
-
"""
|
27
|
-
self.pile = (
|
28
|
-
workpile
|
29
|
-
if workpile and isinstance(workpile, Pile)
|
30
|
-
else pile({}, Work)
|
31
|
-
)
|
32
|
-
self.pending = progression(workpile) if workpile else progression()
|
33
|
-
self.queue = WorkQueue(capacity=capacity, refresh_time=refresh_time)
|
34
|
-
|
35
|
-
async def append(self, work: Work):
|
36
|
-
"""
|
37
|
-
Appends a new work item to the log.
|
38
|
-
|
39
|
-
Args:
|
40
|
-
work (Work): The work item to append.
|
41
|
-
"""
|
42
|
-
self.pile.append(work)
|
43
|
-
self.pending.append(work)
|
44
|
-
|
45
|
-
async def forward(self):
|
46
|
-
"""
|
47
|
-
Forwards pending work items to the queue.
|
48
|
-
"""
|
49
|
-
while len(self.pending) > 0:
|
50
|
-
work: Work = self.pile[self.pending.popleft()]
|
51
|
-
await self.queue.enqueue(work)
|
52
|
-
|
53
|
-
async def stop(self):
|
54
|
-
"""
|
55
|
-
Stops the work queue.
|
56
|
-
"""
|
57
|
-
await self.queue.stop()
|
58
|
-
|
59
|
-
@property
|
60
|
-
def pending_work(self):
|
61
|
-
"""
|
62
|
-
Retrieves the pile of pending work items.
|
63
|
-
|
64
|
-
Returns:
|
65
|
-
Pile: A pile of pending work items.
|
66
|
-
"""
|
67
|
-
return pile([i for i in self.pile if i.status == WorkStatus.PENDING])
|
68
|
-
|
69
|
-
@property
|
70
|
-
def stopped(self):
|
71
|
-
"""
|
72
|
-
Checks if the work queue is stopped.
|
73
|
-
|
74
|
-
Returns:
|
75
|
-
bool: True if the work queue is stopped, else False.
|
76
|
-
"""
|
77
|
-
return self.queue.stopped
|
78
|
-
|
79
|
-
@property
|
80
|
-
def completed_work(self):
|
81
|
-
"""
|
82
|
-
Retrieves the pile of completed work items.
|
83
|
-
|
84
|
-
Returns:
|
85
|
-
Pile: A pile of completed work items.
|
86
|
-
"""
|
87
|
-
return pile([i for i in self.pile if i.status == WorkStatus.COMPLETED])
|
88
|
-
|
89
|
-
def __contains__(self, work):
|
90
|
-
"""
|
91
|
-
Checks if a work item is in the pile.
|
92
|
-
|
93
|
-
Args:
|
94
|
-
work (Work): The work item to check.
|
95
|
-
|
96
|
-
Returns:
|
97
|
-
bool: True if the work item is in the pile, else False.
|
98
|
-
"""
|
99
|
-
return work in self.pile
|
100
|
-
|
101
|
-
def __iter__(self):
|
102
|
-
"""
|
103
|
-
Returns an iterator over the work pile.
|
104
|
-
|
105
|
-
Returns:
|
106
|
-
Iterator: An iterator over the work pile.
|
107
|
-
"""
|
108
|
-
return iter(self.pile)
|
@@ -1,47 +0,0 @@
|
|
1
|
-
from abc import ABC
|
2
|
-
|
3
|
-
from lionagi.core.collections import iModel
|
4
|
-
|
5
|
-
|
6
|
-
class TokenCompressor(ABC):
|
7
|
-
"""
|
8
|
-
NOTICE:
|
9
|
-
The token compressor system is inspired by LLMLingua.
|
10
|
-
https://github.com/microsoft/LLMLingua
|
11
|
-
|
12
|
-
MIT License
|
13
|
-
Copyright (c) Microsoft Corporation.
|
14
|
-
|
15
|
-
Permission is hereby granted, free of charge, to any person obtaining a copy
|
16
|
-
of this software and associated documentation files (the "Software"), to deal
|
17
|
-
in the Software without restriction, including without limitation the rights
|
18
|
-
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
|
19
|
-
copies of the Software, and to permit persons to whom the Software is
|
20
|
-
furnished to do so, subject to the following conditions:
|
21
|
-
|
22
|
-
Authors:
|
23
|
-
Huiqiang Jiang, Qianhui Wu, Chin-Yew Lin, Yuqing Yang, Lili Qiu
|
24
|
-
@inproceedings{jiang-etal-2023-llmlingua,
|
25
|
-
title = "{LLML}ingua: Compressing Prompts for Accelerated Inference of Large Language Models",
|
26
|
-
author = "Huiqiang Jiang and Qianhui Wu and Chin-Yew Lin and Yuqing Yang and Lili Qiu",
|
27
|
-
booktitle = "Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing",
|
28
|
-
month = dec,
|
29
|
-
year = "2023",
|
30
|
-
publisher = "Association for Computational Linguistics",
|
31
|
-
url = "https://aclanthology.org/2023.emnlp-main.825",
|
32
|
-
doi = "10.18653/v1/2023.emnlp-main.825",
|
33
|
-
pages = "13358--13376",
|
34
|
-
}
|
35
|
-
|
36
|
-
LionAGI Modifications:
|
37
|
-
- Only borrowed the concept of token compression via perplexity
|
38
|
-
- Removed the dependency on the LLMLingua library
|
39
|
-
- use logprobs from GPT model to calculate perplexity
|
40
|
-
- added async ability to the functions
|
41
|
-
- used lionagi existing iModel class for API calls
|
42
|
-
"""
|
43
|
-
|
44
|
-
def __init__(self, imodel: iModel, tokenizer=None, splitter=None):
|
45
|
-
self.imodel = imodel
|
46
|
-
self.tokenizer = tokenizer
|
47
|
-
self.splitter = splitter
|
@@ -1,265 +0,0 @@
|
|
1
|
-
import asyncio
|
2
|
-
from time import time
|
3
|
-
|
4
|
-
import numpy as np
|
5
|
-
|
6
|
-
from lionagi import alcall
|
7
|
-
from lionagi.core.collections import iModel
|
8
|
-
from lionagi.libs.ln_convert import to_list
|
9
|
-
from lionagi.libs.ln_tokenize import TokenizeUtil
|
10
|
-
|
11
|
-
from .base import TokenCompressor
|
12
|
-
|
13
|
-
# inspired by LLMLingua, MIT License, Copyright (c) Microsoft Corporation.
|
14
|
-
# https://github.com/microsoft/LLMLingua
|
15
|
-
|
16
|
-
|
17
|
-
class LLMCompressor(TokenCompressor):
|
18
|
-
|
19
|
-
def __init__(
|
20
|
-
self,
|
21
|
-
imodel: iModel = None,
|
22
|
-
system_msg=None,
|
23
|
-
tokenizer=None, # must be a callable or object with a tokenize method
|
24
|
-
splitter=None, # must be a callable or object with a split/chunk/segment method
|
25
|
-
target_ratio=0.2,
|
26
|
-
n_samples=5, # the cumulative samples to take in each perplexity calculation
|
27
|
-
chunk_size=64,
|
28
|
-
max_tokens_per_sample=80,
|
29
|
-
min_compression_score=0, # (0-1) the minimum score to consider for compression, 0 means all
|
30
|
-
split_overlap=0,
|
31
|
-
split_threshold=0,
|
32
|
-
verbose=True,
|
33
|
-
):
|
34
|
-
imodel = imodel or iModel(model="gpt-3.5-turbo", temperature=0.3)
|
35
|
-
super().__init__(imodel=imodel, tokenizer=tokenizer, splitter=splitter)
|
36
|
-
self.system_msg = (
|
37
|
-
system_msg
|
38
|
-
or "Concisely summarize and compress the information for storage:"
|
39
|
-
)
|
40
|
-
self.target_ratio = target_ratio
|
41
|
-
self.n_samples = n_samples
|
42
|
-
self.chunk_size = chunk_size
|
43
|
-
self.max_tokens_per_sample = max_tokens_per_sample
|
44
|
-
self.min_compression_score = min_compression_score
|
45
|
-
self.verbose = verbose
|
46
|
-
self.split_overlap = split_overlap
|
47
|
-
self.split_threshold = split_threshold
|
48
|
-
|
49
|
-
def tokenize(self, text, encoding_name=None, return_byte=False, **kwargs):
|
50
|
-
"""
|
51
|
-
by default you can use `encoding_name` to be one of,
|
52
|
-
['gpt2', 'r50k_base', 'p50k_base', 'p50k_edit', 'cl100k_base', 'o200k_base']
|
53
|
-
|
54
|
-
or you can use `encoding_model` that tiktoken supports in their mapping such as "gpt-4o"
|
55
|
-
"""
|
56
|
-
if not self.tokenizer:
|
57
|
-
return TokenizeUtil.tokenize(
|
58
|
-
text,
|
59
|
-
encoding_model=self.imodel.iModel_name,
|
60
|
-
encoding_name=encoding_name,
|
61
|
-
return_byte=return_byte,
|
62
|
-
)
|
63
|
-
|
64
|
-
if hasattr(self.tokenizer, "tokenize"):
|
65
|
-
return self.tokenizer.tokenize(text, **kwargs)
|
66
|
-
|
67
|
-
return self.tokenizer(text, **kwargs)
|
68
|
-
|
69
|
-
def split(
|
70
|
-
self,
|
71
|
-
text,
|
72
|
-
chunk_size=None,
|
73
|
-
overlap=None,
|
74
|
-
threshold=None,
|
75
|
-
by_chars=False,
|
76
|
-
return_tokens=False,
|
77
|
-
return_byte=False,
|
78
|
-
**kwargs,
|
79
|
-
):
|
80
|
-
if not self.splitter:
|
81
|
-
splitter = (
|
82
|
-
TokenizeUtil.chunk_by_chars
|
83
|
-
if by_chars
|
84
|
-
else TokenizeUtil.chunk_by_tokens
|
85
|
-
)
|
86
|
-
return splitter(
|
87
|
-
text,
|
88
|
-
chunk_size or self.chunk_size,
|
89
|
-
overlap or self.split_overlap,
|
90
|
-
threshold or self.split_threshold,
|
91
|
-
return_tokens=return_tokens,
|
92
|
-
return_byte=return_byte,
|
93
|
-
)
|
94
|
-
|
95
|
-
a = [
|
96
|
-
getattr(self.splitter, i, None)
|
97
|
-
for i in ["split", "chunk", "segment"]
|
98
|
-
if i is not None
|
99
|
-
][0]
|
100
|
-
a = getattr(self.splitter, a)
|
101
|
-
return a(text, **kwargs)
|
102
|
-
|
103
|
-
async def rank_by_pplex(
|
104
|
-
self,
|
105
|
-
items: list,
|
106
|
-
initial_text=None,
|
107
|
-
cumulative=False,
|
108
|
-
n_samples=None,
|
109
|
-
**kwargs,
|
110
|
-
):
|
111
|
-
"""
|
112
|
-
rank a list of items according to their perplexity
|
113
|
-
an item can be a single token or a list of tokens
|
114
|
-
|
115
|
-
kwargs: additional arguments to pass to the model
|
116
|
-
"""
|
117
|
-
|
118
|
-
async def _get_item_perplexity(item):
|
119
|
-
item = item if isinstance(item, list) else [item]
|
120
|
-
item = (
|
121
|
-
item[: self.max_tokens_per_sample]
|
122
|
-
if len(item) > self.max_tokens_per_sample
|
123
|
-
else item
|
124
|
-
)
|
125
|
-
return await self.imodel.compute_perplexity(
|
126
|
-
initial_context=initial_text,
|
127
|
-
tokens=item,
|
128
|
-
n_samples=n_samples or self.n_samples,
|
129
|
-
system_msg=self.system_msg,
|
130
|
-
**kwargs,
|
131
|
-
)
|
132
|
-
|
133
|
-
if not isinstance(items, list):
|
134
|
-
items = self.tokenize(items)
|
135
|
-
|
136
|
-
if len(items) == 1:
|
137
|
-
return [items] # no need to rank a single item
|
138
|
-
|
139
|
-
_segments = []
|
140
|
-
_context = initial_text or ""
|
141
|
-
_task = []
|
142
|
-
|
143
|
-
if cumulative:
|
144
|
-
for i in items:
|
145
|
-
if isinstance(i, list):
|
146
|
-
_context += " " + " ".join(i).strip()
|
147
|
-
else:
|
148
|
-
_context += " " + i.strip()
|
149
|
-
|
150
|
-
_segments.append(_context)
|
151
|
-
else:
|
152
|
-
_segments = items
|
153
|
-
|
154
|
-
for i in _segments:
|
155
|
-
_task.append(asyncio.create_task(_get_item_perplexity(i)))
|
156
|
-
|
157
|
-
results = await asyncio.gather(*_task)
|
158
|
-
results = [(item, pplex) for item, pplex in zip(items, results)]
|
159
|
-
return sorted(results, key=lambda x: x[1]["logprobs"], reverse=True)
|
160
|
-
|
161
|
-
async def compress(
|
162
|
-
self,
|
163
|
-
text,
|
164
|
-
target_ratio=None,
|
165
|
-
initial_text=None,
|
166
|
-
cumulative=False,
|
167
|
-
split_kwargs=None,
|
168
|
-
split_overlap=None,
|
169
|
-
split_threshold=None,
|
170
|
-
rank_by="perplexity",
|
171
|
-
min_compression_score=None,
|
172
|
-
verbose=True,
|
173
|
-
**kwargs,
|
174
|
-
):
|
175
|
-
start = time()
|
176
|
-
if split_kwargs is None:
|
177
|
-
split_kwargs = {}
|
178
|
-
split_kwargs["chunk_size"] = self.max_tokens_per_sample
|
179
|
-
split_kwargs["overlap"] = split_overlap or 0
|
180
|
-
split_kwargs["threshold"] = split_threshold or 0
|
181
|
-
|
182
|
-
len_tokens = len(self.tokenize(text))
|
183
|
-
|
184
|
-
items = self.split(text, return_tokens=True, **split_kwargs)
|
185
|
-
|
186
|
-
if rank_by == "perplexity":
|
187
|
-
ranked_items = await self.rank_by_pplex(
|
188
|
-
items=items,
|
189
|
-
initial_text=initial_text,
|
190
|
-
cumulative=cumulative,
|
191
|
-
**kwargs,
|
192
|
-
)
|
193
|
-
|
194
|
-
prompt_tokens = sum(
|
195
|
-
[i[1]["num_prompt_tokens"] for i in ranked_items]
|
196
|
-
)
|
197
|
-
|
198
|
-
num_completion_tokens = sum(
|
199
|
-
[i[1]["num_completion_tokens"] for i in ranked_items]
|
200
|
-
)
|
201
|
-
|
202
|
-
price = (
|
203
|
-
prompt_tokens * 0.5 / 1000000
|
204
|
-
+ num_completion_tokens * 1.5 / 1000000
|
205
|
-
)
|
206
|
-
|
207
|
-
selected_items = self.select_by_pplex(
|
208
|
-
ranked_items=ranked_items,
|
209
|
-
target_compression_ratio=target_ratio or self.target_ratio,
|
210
|
-
original_length=len_tokens,
|
211
|
-
min_pplex=min_compression_score or self.min_compression_score,
|
212
|
-
)
|
213
|
-
|
214
|
-
if verbose:
|
215
|
-
msg = ""
|
216
|
-
msg += f"Original Token number: {len_tokens}\n"
|
217
|
-
|
218
|
-
def _f(i):
|
219
|
-
if isinstance(i, str):
|
220
|
-
i = self.tokenize(i)
|
221
|
-
|
222
|
-
if isinstance(i, list):
|
223
|
-
return len(to_list(i, dropna=True, flatten=True))
|
224
|
-
|
225
|
-
len_ = sum([_f(i) for i in selected_items])
|
226
|
-
msg += f"Selected Token number: {len_}\n"
|
227
|
-
msg += f"Token Compression Ratio: {len_ / len_tokens:.03f}\n"
|
228
|
-
msg += f"Compression Time: {time() - start:.04f} seconds\n"
|
229
|
-
msg += f"Compression Model: {self.imodel.iModel_name}\n"
|
230
|
-
msg += f"Compression Method: {rank_by}\n"
|
231
|
-
msg += f"Compression Usage: ${price:.05f}\n"
|
232
|
-
print(msg)
|
233
|
-
|
234
|
-
a = "".join([i.strip() for i in selected_items]).strip()
|
235
|
-
a = a.replace("\n\n", "")
|
236
|
-
return a
|
237
|
-
|
238
|
-
raise ValueError(f"Ranking method {rank_by} is not supported")
|
239
|
-
|
240
|
-
def select_by_pplex(
|
241
|
-
self,
|
242
|
-
ranked_items,
|
243
|
-
target_compression_ratio,
|
244
|
-
original_length,
|
245
|
-
min_pplex=None,
|
246
|
-
):
|
247
|
-
min_pplex = min_pplex or 0
|
248
|
-
|
249
|
-
desired_length = int(original_length * target_compression_ratio)
|
250
|
-
|
251
|
-
items = []
|
252
|
-
current_length = 0
|
253
|
-
|
254
|
-
for item, info in ranked_items:
|
255
|
-
if info["perplexity"] > min_pplex:
|
256
|
-
item = self.tokenize(item) if isinstance(item, str) else item
|
257
|
-
item = item if isinstance(item, list) else [item]
|
258
|
-
item = to_list(item, dropna=True, flatten=True)
|
259
|
-
if current_length + len(item) > desired_length:
|
260
|
-
break
|
261
|
-
else:
|
262
|
-
current_length += len(item)
|
263
|
-
items.append("".join(item))
|
264
|
-
|
265
|
-
return items
|
@@ -1,61 +0,0 @@
|
|
1
|
-
# from lionagi.core.collections import iModel
|
2
|
-
# from .base import TokenCompressor
|
3
|
-
|
4
|
-
|
5
|
-
# class LLMSummarizer(TokenCompressor):
|
6
|
-
|
7
|
-
# def __init__(
|
8
|
-
# self, imodel: iModel = None, system_msg=None, tokenizer=None, splitter=None,
|
9
|
-
# max_tokens=25, target_ratio=0.3
|
10
|
-
# ):
|
11
|
-
# imodel = imodel or iModel(model="gpt-3.5-turbo", max_tokens=max_tokens)
|
12
|
-
# super().__init__(imodel=imodel, tokenizer=tokenizer, splitter=splitter)
|
13
|
-
# self.system_msg = (
|
14
|
-
# system_msg
|
15
|
-
# or "Summarize the following sentence to be concise and informative:"
|
16
|
-
# )
|
17
|
-
# self.target_ratio = target_ratio
|
18
|
-
|
19
|
-
# async def summarize_sentence(self, sentence, **kwargs):
|
20
|
-
# messages = [
|
21
|
-
# {"role": "system", "content": self.system_msg},
|
22
|
-
# {"role": "user", "content": sentence},
|
23
|
-
# ]
|
24
|
-
# response = await self.imodel.call_chat_completion(messages, **kwargs)
|
25
|
-
# return response["choices"][0]["message"]["content"]
|
26
|
-
|
27
|
-
# def tokenize(self, text):
|
28
|
-
# tokenize_func = self.tokenizer or tokenize
|
29
|
-
# return tokenize_func(text)
|
30
|
-
|
31
|
-
# def split(self, text):
|
32
|
-
# split_func = self.splitter or split_into_segments
|
33
|
-
# return split_func(text)
|
34
|
-
|
35
|
-
# # Function to enforce maximum sentence length
|
36
|
-
# def enforce_max_sentence_length(self, sentence, max_words=25):
|
37
|
-
# words = self.tokenize(sentence)
|
38
|
-
# if len(words) > max_words:
|
39
|
-
# sentence = ' '.join(words[:max_words])
|
40
|
-
# return sentence
|
41
|
-
|
42
|
-
# async def summarize_text(self, text, max_length_per_sentence=25, target_ratio=None, **kwargs):
|
43
|
-
# sentences = self.split(text)
|
44
|
-
# summarized = await alcall(
|
45
|
-
# sentences, self.summarize_sentence, **kwargs
|
46
|
-
# )
|
47
|
-
# summarized = [
|
48
|
-
# self.enforce_max_sentence_length(sentence, max_length_per_sentence)
|
49
|
-
# for sentence in summarized
|
50
|
-
# ]
|
51
|
-
|
52
|
-
# original_length = len(self.tokenize(text))
|
53
|
-
# summarized_length = len(self.tokenize(' '.join(summarized)))
|
54
|
-
# current_ratio = summarized_length / original_length
|
55
|
-
|
56
|
-
# target_ratio = target_ratio or self.target_ratio
|
57
|
-
# if current_ratio > target_ratio:
|
58
|
-
# words_to_remove = int((current_ratio - target_ratio) * original_length)
|
59
|
-
# return ' '.join(summarized[:-words_to_remove])
|
60
|
-
|
61
|
-
# return ' '.join(summarized)
|
@@ -1,70 +0,0 @@
|
|
1
|
-
# import asyncio
|
2
|
-
# from lionagi import alcall
|
3
|
-
# from lionagi.libs.ln_convert import to_list
|
4
|
-
# import numpy as np
|
5
|
-
|
6
|
-
# def split_into_segments(text):
|
7
|
-
# segments = text.split(".") # Splitting by period followed by a space
|
8
|
-
# return [segment.strip() for segment in segments if segment]
|
9
|
-
|
10
|
-
# # Tokenize the segment
|
11
|
-
# def tokenize(segment):
|
12
|
-
# tokens = segment.split() # Simple space-based tokenization
|
13
|
-
# return tokens
|
14
|
-
|
15
|
-
# async def calculate_perplexity(system_msg: str, imodel, tokens, initial_context=None, **kwargs):
|
16
|
-
# _tasks = []
|
17
|
-
# _context = initial_context or ""
|
18
|
-
# for i in range(len(tokens)):
|
19
|
-
# _context += " " + tokens[i]
|
20
|
-
# messages = [
|
21
|
-
# {"role": "system", "content": system_msg},
|
22
|
-
# {"role": "user", "content": _context},
|
23
|
-
# ]
|
24
|
-
# task = asyncio.create_task(
|
25
|
-
# imodel.call_chat_completion(
|
26
|
-
# messages=messages, logprobs=True, max_tokens=1, **kwargs
|
27
|
-
# )
|
28
|
-
# )
|
29
|
-
# _tasks.append(task)
|
30
|
-
|
31
|
-
# results = await asyncio.gather(*_tasks)
|
32
|
-
# logprobs = [
|
33
|
-
# result[1]["choices"][0]["logprobs"]["content"] for result in results
|
34
|
-
# ]
|
35
|
-
# logprobs = to_list(logprobs, flatten=True, dropna=True)
|
36
|
-
# logprobs = [lprob_["logprob"] for lprob_ in logprobs]
|
37
|
-
# return np.exp(np.mean(logprobs))
|
38
|
-
|
39
|
-
# async def rank_by_perplexity(
|
40
|
-
# text: str | list[str] = None, # if list we assume they are already well split
|
41
|
-
# initial_text=None,
|
42
|
-
|
43
|
-
# segments,
|
44
|
-
# initial_text=None,
|
45
|
-
# cumulative=False,
|
46
|
-
# **kwargs
|
47
|
-
# ):
|
48
|
-
# _segments = []
|
49
|
-
# _context = initial_text or ""
|
50
|
-
# _task = []
|
51
|
-
|
52
|
-
# if cumulative:
|
53
|
-
# for i in range(1, len(segments)):
|
54
|
-
# _context += " " + segments[i - 1]
|
55
|
-
# _segments.append(_context)
|
56
|
-
# else:
|
57
|
-
# _segments = segments
|
58
|
-
|
59
|
-
# for i in segments:
|
60
|
-
# _task.append(asyncio.create_task(
|
61
|
-
# calculate_perplexity(
|
62
|
-
# self.system_msg, self.imodel, self.tokenize(i), **kwargs)
|
63
|
-
# )
|
64
|
-
# )
|
65
|
-
# segment_perplexities = await asyncio.gather(*_task)
|
66
|
-
|
67
|
-
# return {
|
68
|
-
# segment: perplexity
|
69
|
-
# for segment, perplexity in zip(segments, segment_perplexities)
|
70
|
-
# }
|
@@ -1 +0,0 @@
|
|
1
|
-
TODO
|
@@ -1,19 +0,0 @@
|
|
1
|
-
# from ..form.predict import predict
|
2
|
-
# from .select import select
|
3
|
-
# from ..form.score import score
|
4
|
-
# from ..form.react import react
|
5
|
-
# from .vote import vote
|
6
|
-
# from ..form.plan import plan
|
7
|
-
# from .cot import chain_of_thoughts, chain_of_react
|
8
|
-
|
9
|
-
|
10
|
-
# __all__ = [
|
11
|
-
# "predict",
|
12
|
-
# "select",
|
13
|
-
# "score",
|
14
|
-
# "vote",
|
15
|
-
# "react",
|
16
|
-
# "plan",
|
17
|
-
# "chain_of_thoughts",
|
18
|
-
# "chain_of_react",
|
19
|
-
# ]
|