edsl 0.1.14__py3-none-any.whl → 0.1.40__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- edsl/Base.py +348 -38
- edsl/BaseDiff.py +260 -0
- edsl/TemplateLoader.py +24 -0
- edsl/__init__.py +46 -10
- edsl/__version__.py +1 -0
- edsl/agents/Agent.py +842 -144
- edsl/agents/AgentList.py +521 -25
- edsl/agents/Invigilator.py +250 -374
- edsl/agents/InvigilatorBase.py +257 -0
- edsl/agents/PromptConstructor.py +272 -0
- edsl/agents/QuestionInstructionPromptBuilder.py +128 -0
- edsl/agents/QuestionTemplateReplacementsBuilder.py +137 -0
- edsl/agents/descriptors.py +43 -13
- edsl/agents/prompt_helpers.py +129 -0
- edsl/agents/question_option_processor.py +172 -0
- edsl/auto/AutoStudy.py +130 -0
- edsl/auto/StageBase.py +243 -0
- edsl/auto/StageGenerateSurvey.py +178 -0
- edsl/auto/StageLabelQuestions.py +125 -0
- edsl/auto/StagePersona.py +61 -0
- edsl/auto/StagePersonaDimensionValueRanges.py +88 -0
- edsl/auto/StagePersonaDimensionValues.py +74 -0
- edsl/auto/StagePersonaDimensions.py +69 -0
- edsl/auto/StageQuestions.py +74 -0
- edsl/auto/SurveyCreatorPipeline.py +21 -0
- edsl/auto/utilities.py +218 -0
- edsl/base/Base.py +279 -0
- edsl/config.py +121 -104
- edsl/conversation/Conversation.py +290 -0
- edsl/conversation/car_buying.py +59 -0
- edsl/conversation/chips.py +95 -0
- edsl/conversation/mug_negotiation.py +81 -0
- edsl/conversation/next_speaker_utilities.py +93 -0
- edsl/coop/CoopFunctionsMixin.py +15 -0
- edsl/coop/ExpectedParrotKeyHandler.py +125 -0
- edsl/coop/PriceFetcher.py +54 -0
- edsl/coop/__init__.py +1 -0
- edsl/coop/coop.py +1029 -134
- edsl/coop/utils.py +131 -0
- edsl/data/Cache.py +560 -89
- edsl/data/CacheEntry.py +230 -0
- edsl/data/CacheHandler.py +168 -0
- edsl/data/RemoteCacheSync.py +186 -0
- edsl/data/SQLiteDict.py +292 -0
- edsl/data/__init__.py +5 -3
- edsl/data/orm.py +6 -33
- edsl/data_transfer_models.py +74 -27
- edsl/enums.py +165 -8
- edsl/exceptions/BaseException.py +21 -0
- edsl/exceptions/__init__.py +52 -46
- edsl/exceptions/agents.py +33 -15
- edsl/exceptions/cache.py +5 -0
- edsl/exceptions/coop.py +8 -0
- edsl/exceptions/general.py +34 -0
- edsl/exceptions/inference_services.py +5 -0
- edsl/exceptions/jobs.py +15 -0
- edsl/exceptions/language_models.py +46 -1
- edsl/exceptions/questions.py +80 -5
- edsl/exceptions/results.py +16 -5
- edsl/exceptions/scenarios.py +29 -0
- edsl/exceptions/surveys.py +13 -10
- edsl/inference_services/AnthropicService.py +106 -0
- edsl/inference_services/AvailableModelCacheHandler.py +184 -0
- edsl/inference_services/AvailableModelFetcher.py +215 -0
- edsl/inference_services/AwsBedrock.py +118 -0
- edsl/inference_services/AzureAI.py +215 -0
- edsl/inference_services/DeepInfraService.py +18 -0
- edsl/inference_services/GoogleService.py +143 -0
- edsl/inference_services/GroqService.py +20 -0
- edsl/inference_services/InferenceServiceABC.py +80 -0
- edsl/inference_services/InferenceServicesCollection.py +138 -0
- edsl/inference_services/MistralAIService.py +120 -0
- edsl/inference_services/OllamaService.py +18 -0
- edsl/inference_services/OpenAIService.py +236 -0
- edsl/inference_services/PerplexityService.py +160 -0
- edsl/inference_services/ServiceAvailability.py +135 -0
- edsl/inference_services/TestService.py +90 -0
- edsl/inference_services/TogetherAIService.py +172 -0
- edsl/inference_services/data_structures.py +134 -0
- edsl/inference_services/models_available_cache.py +118 -0
- edsl/inference_services/rate_limits_cache.py +25 -0
- edsl/inference_services/registry.py +41 -0
- edsl/inference_services/write_available.py +10 -0
- edsl/jobs/AnswerQuestionFunctionConstructor.py +223 -0
- edsl/jobs/Answers.py +21 -20
- edsl/jobs/FetchInvigilator.py +47 -0
- edsl/jobs/InterviewTaskManager.py +98 -0
- edsl/jobs/InterviewsConstructor.py +50 -0
- edsl/jobs/Jobs.py +684 -204
- edsl/jobs/JobsChecks.py +172 -0
- edsl/jobs/JobsComponentConstructor.py +189 -0
- edsl/jobs/JobsPrompts.py +270 -0
- edsl/jobs/JobsRemoteInferenceHandler.py +311 -0
- edsl/jobs/JobsRemoteInferenceLogger.py +239 -0
- edsl/jobs/RequestTokenEstimator.py +30 -0
- edsl/jobs/async_interview_runner.py +138 -0
- edsl/jobs/buckets/BucketCollection.py +104 -0
- edsl/jobs/buckets/ModelBuckets.py +65 -0
- edsl/jobs/buckets/TokenBucket.py +283 -0
- edsl/jobs/buckets/TokenBucketAPI.py +211 -0
- edsl/jobs/buckets/TokenBucketClient.py +191 -0
- edsl/jobs/check_survey_scenario_compatibility.py +85 -0
- edsl/jobs/data_structures.py +120 -0
- edsl/jobs/decorators.py +35 -0
- edsl/jobs/interviews/Interview.py +392 -0
- edsl/jobs/interviews/InterviewExceptionCollection.py +99 -0
- edsl/jobs/interviews/InterviewExceptionEntry.py +186 -0
- edsl/jobs/interviews/InterviewStatistic.py +63 -0
- edsl/jobs/interviews/InterviewStatisticsCollection.py +25 -0
- edsl/jobs/interviews/InterviewStatusDictionary.py +78 -0
- edsl/jobs/interviews/InterviewStatusLog.py +92 -0
- edsl/jobs/interviews/ReportErrors.py +66 -0
- edsl/jobs/interviews/interview_status_enum.py +9 -0
- edsl/jobs/jobs_status_enums.py +9 -0
- edsl/jobs/loggers/HTMLTableJobLogger.py +304 -0
- edsl/jobs/results_exceptions_handler.py +98 -0
- edsl/jobs/runners/JobsRunnerAsyncio.py +151 -110
- edsl/jobs/runners/JobsRunnerStatus.py +298 -0
- edsl/jobs/tasks/QuestionTaskCreator.py +244 -0
- edsl/jobs/tasks/TaskCreators.py +64 -0
- edsl/jobs/tasks/TaskHistory.py +470 -0
- edsl/jobs/tasks/TaskStatusLog.py +23 -0
- edsl/jobs/tasks/task_status_enum.py +161 -0
- edsl/jobs/tokens/InterviewTokenUsage.py +27 -0
- edsl/jobs/tokens/TokenUsage.py +34 -0
- edsl/language_models/ComputeCost.py +63 -0
- edsl/language_models/LanguageModel.py +507 -386
- edsl/language_models/ModelList.py +164 -0
- edsl/language_models/PriceManager.py +127 -0
- edsl/language_models/RawResponseHandler.py +106 -0
- edsl/language_models/RegisterLanguageModelsMeta.py +184 -0
- edsl/language_models/__init__.py +1 -8
- edsl/language_models/fake_openai_call.py +15 -0
- edsl/language_models/fake_openai_service.py +61 -0
- edsl/language_models/key_management/KeyLookup.py +63 -0
- edsl/language_models/key_management/KeyLookupBuilder.py +273 -0
- edsl/language_models/key_management/KeyLookupCollection.py +38 -0
- edsl/language_models/key_management/__init__.py +0 -0
- edsl/language_models/key_management/models.py +131 -0
- edsl/language_models/model.py +256 -0
- edsl/language_models/repair.py +109 -41
- edsl/language_models/utilities.py +65 -0
- edsl/notebooks/Notebook.py +263 -0
- edsl/notebooks/NotebookToLaTeX.py +142 -0
- edsl/notebooks/__init__.py +1 -0
- edsl/prompts/Prompt.py +222 -93
- edsl/prompts/__init__.py +1 -1
- edsl/questions/ExceptionExplainer.py +77 -0
- edsl/questions/HTMLQuestion.py +103 -0
- edsl/questions/QuestionBase.py +518 -0
- edsl/questions/QuestionBasePromptsMixin.py +221 -0
- edsl/questions/QuestionBudget.py +164 -67
- edsl/questions/QuestionCheckBox.py +281 -62
- edsl/questions/QuestionDict.py +343 -0
- edsl/questions/QuestionExtract.py +136 -50
- edsl/questions/QuestionFreeText.py +79 -55
- edsl/questions/QuestionFunctional.py +138 -41
- edsl/questions/QuestionList.py +184 -57
- edsl/questions/QuestionMatrix.py +265 -0
- edsl/questions/QuestionMultipleChoice.py +293 -69
- edsl/questions/QuestionNumerical.py +109 -56
- edsl/questions/QuestionRank.py +244 -49
- edsl/questions/Quick.py +41 -0
- edsl/questions/SimpleAskMixin.py +74 -0
- edsl/questions/__init__.py +9 -6
- edsl/questions/{AnswerValidatorMixin.py → answer_validator_mixin.py} +153 -38
- edsl/questions/compose_questions.py +13 -7
- edsl/questions/data_structures.py +20 -0
- edsl/questions/decorators.py +21 -0
- edsl/questions/derived/QuestionLikertFive.py +28 -26
- edsl/questions/derived/QuestionLinearScale.py +41 -28
- edsl/questions/derived/QuestionTopK.py +34 -26
- edsl/questions/derived/QuestionYesNo.py +40 -27
- edsl/questions/descriptors.py +228 -74
- edsl/questions/loop_processor.py +149 -0
- edsl/questions/prompt_templates/question_budget.jinja +13 -0
- edsl/questions/prompt_templates/question_checkbox.jinja +32 -0
- edsl/questions/prompt_templates/question_extract.jinja +11 -0
- edsl/questions/prompt_templates/question_free_text.jinja +3 -0
- edsl/questions/prompt_templates/question_linear_scale.jinja +11 -0
- edsl/questions/prompt_templates/question_list.jinja +17 -0
- edsl/questions/prompt_templates/question_multiple_choice.jinja +33 -0
- edsl/questions/prompt_templates/question_numerical.jinja +37 -0
- edsl/questions/question_base_gen_mixin.py +168 -0
- edsl/questions/question_registry.py +130 -46
- edsl/questions/register_questions_meta.py +71 -0
- edsl/questions/response_validator_abc.py +188 -0
- edsl/questions/response_validator_factory.py +34 -0
- edsl/questions/settings.py +5 -2
- edsl/questions/templates/__init__.py +0 -0
- edsl/questions/templates/budget/__init__.py +0 -0
- edsl/questions/templates/budget/answering_instructions.jinja +7 -0
- edsl/questions/templates/budget/question_presentation.jinja +7 -0
- edsl/questions/templates/checkbox/__init__.py +0 -0
- edsl/questions/templates/checkbox/answering_instructions.jinja +10 -0
- edsl/questions/templates/checkbox/question_presentation.jinja +22 -0
- edsl/questions/templates/dict/__init__.py +0 -0
- edsl/questions/templates/dict/answering_instructions.jinja +21 -0
- edsl/questions/templates/dict/question_presentation.jinja +1 -0
- edsl/questions/templates/extract/__init__.py +0 -0
- edsl/questions/templates/extract/answering_instructions.jinja +7 -0
- edsl/questions/templates/extract/question_presentation.jinja +1 -0
- edsl/questions/templates/free_text/__init__.py +0 -0
- edsl/questions/templates/free_text/answering_instructions.jinja +0 -0
- edsl/questions/templates/free_text/question_presentation.jinja +1 -0
- edsl/questions/templates/likert_five/__init__.py +0 -0
- edsl/questions/templates/likert_five/answering_instructions.jinja +10 -0
- edsl/questions/templates/likert_five/question_presentation.jinja +12 -0
- edsl/questions/templates/linear_scale/__init__.py +0 -0
- edsl/questions/templates/linear_scale/answering_instructions.jinja +5 -0
- edsl/questions/templates/linear_scale/question_presentation.jinja +5 -0
- edsl/questions/templates/list/__init__.py +0 -0
- edsl/questions/templates/list/answering_instructions.jinja +4 -0
- edsl/questions/templates/list/question_presentation.jinja +5 -0
- edsl/questions/templates/matrix/__init__.py +1 -0
- edsl/questions/templates/matrix/answering_instructions.jinja +5 -0
- edsl/questions/templates/matrix/question_presentation.jinja +20 -0
- edsl/questions/templates/multiple_choice/__init__.py +0 -0
- edsl/questions/templates/multiple_choice/answering_instructions.jinja +9 -0
- edsl/questions/templates/multiple_choice/html.jinja +0 -0
- edsl/questions/templates/multiple_choice/question_presentation.jinja +12 -0
- edsl/questions/templates/numerical/__init__.py +0 -0
- edsl/questions/templates/numerical/answering_instructions.jinja +7 -0
- edsl/questions/templates/numerical/question_presentation.jinja +7 -0
- edsl/questions/templates/rank/__init__.py +0 -0
- edsl/questions/templates/rank/answering_instructions.jinja +11 -0
- edsl/questions/templates/rank/question_presentation.jinja +15 -0
- edsl/questions/templates/top_k/__init__.py +0 -0
- edsl/questions/templates/top_k/answering_instructions.jinja +8 -0
- edsl/questions/templates/top_k/question_presentation.jinja +22 -0
- edsl/questions/templates/yes_no/__init__.py +0 -0
- edsl/questions/templates/yes_no/answering_instructions.jinja +6 -0
- edsl/questions/templates/yes_no/question_presentation.jinja +12 -0
- edsl/results/CSSParameterizer.py +108 -0
- edsl/results/Dataset.py +550 -19
- edsl/results/DatasetExportMixin.py +594 -0
- edsl/results/DatasetTree.py +295 -0
- edsl/results/MarkdownToDocx.py +122 -0
- edsl/results/MarkdownToPDF.py +111 -0
- edsl/results/Result.py +477 -173
- edsl/results/Results.py +987 -269
- edsl/results/ResultsExportMixin.py +28 -125
- edsl/results/ResultsGGMixin.py +83 -15
- edsl/results/TableDisplay.py +125 -0
- edsl/results/TextEditor.py +50 -0
- edsl/results/__init__.py +1 -1
- edsl/results/file_exports.py +252 -0
- edsl/results/results_fetch_mixin.py +33 -0
- edsl/results/results_selector.py +145 -0
- edsl/results/results_tools_mixin.py +98 -0
- edsl/results/smart_objects.py +96 -0
- edsl/results/table_data_class.py +12 -0
- edsl/results/table_display.css +78 -0
- edsl/results/table_renderers.py +118 -0
- edsl/results/tree_explore.py +115 -0
- edsl/scenarios/ConstructDownloadLink.py +109 -0
- edsl/scenarios/DocumentChunker.py +102 -0
- edsl/scenarios/DocxScenario.py +16 -0
- edsl/scenarios/FileStore.py +543 -0
- edsl/scenarios/PdfExtractor.py +40 -0
- edsl/scenarios/Scenario.py +431 -62
- edsl/scenarios/ScenarioHtmlMixin.py +65 -0
- edsl/scenarios/ScenarioList.py +1415 -45
- edsl/scenarios/ScenarioListExportMixin.py +45 -0
- edsl/scenarios/ScenarioListPdfMixin.py +239 -0
- edsl/scenarios/__init__.py +2 -0
- edsl/scenarios/directory_scanner.py +96 -0
- edsl/scenarios/file_methods.py +85 -0
- edsl/scenarios/handlers/__init__.py +13 -0
- edsl/scenarios/handlers/csv.py +49 -0
- edsl/scenarios/handlers/docx.py +76 -0
- edsl/scenarios/handlers/html.py +37 -0
- edsl/scenarios/handlers/json.py +111 -0
- edsl/scenarios/handlers/latex.py +5 -0
- edsl/scenarios/handlers/md.py +51 -0
- edsl/scenarios/handlers/pdf.py +68 -0
- edsl/scenarios/handlers/png.py +39 -0
- edsl/scenarios/handlers/pptx.py +105 -0
- edsl/scenarios/handlers/py.py +294 -0
- edsl/scenarios/handlers/sql.py +313 -0
- edsl/scenarios/handlers/sqlite.py +149 -0
- edsl/scenarios/handlers/txt.py +33 -0
- edsl/scenarios/scenario_join.py +131 -0
- edsl/scenarios/scenario_selector.py +156 -0
- edsl/shared.py +1 -0
- edsl/study/ObjectEntry.py +173 -0
- edsl/study/ProofOfWork.py +113 -0
- edsl/study/SnapShot.py +80 -0
- edsl/study/Study.py +521 -0
- edsl/study/__init__.py +4 -0
- edsl/surveys/ConstructDAG.py +92 -0
- edsl/surveys/DAG.py +92 -11
- edsl/surveys/EditSurvey.py +221 -0
- edsl/surveys/InstructionHandler.py +100 -0
- edsl/surveys/Memory.py +9 -4
- edsl/surveys/MemoryManagement.py +72 -0
- edsl/surveys/MemoryPlan.py +156 -35
- edsl/surveys/Rule.py +221 -74
- edsl/surveys/RuleCollection.py +241 -61
- edsl/surveys/RuleManager.py +172 -0
- edsl/surveys/Simulator.py +75 -0
- edsl/surveys/Survey.py +1079 -339
- edsl/surveys/SurveyCSS.py +273 -0
- edsl/surveys/SurveyExportMixin.py +235 -40
- edsl/surveys/SurveyFlowVisualization.py +181 -0
- edsl/surveys/SurveyQualtricsImport.py +284 -0
- edsl/surveys/SurveyToApp.py +141 -0
- edsl/surveys/__init__.py +4 -2
- edsl/surveys/base.py +19 -3
- edsl/surveys/descriptors.py +17 -6
- edsl/surveys/instructions/ChangeInstruction.py +48 -0
- edsl/surveys/instructions/Instruction.py +56 -0
- edsl/surveys/instructions/InstructionCollection.py +82 -0
- edsl/surveys/instructions/__init__.py +0 -0
- edsl/templates/error_reporting/base.html +24 -0
- edsl/templates/error_reporting/exceptions_by_model.html +35 -0
- edsl/templates/error_reporting/exceptions_by_question_name.html +17 -0
- edsl/templates/error_reporting/exceptions_by_type.html +17 -0
- edsl/templates/error_reporting/interview_details.html +116 -0
- edsl/templates/error_reporting/interviews.html +19 -0
- edsl/templates/error_reporting/overview.html +5 -0
- edsl/templates/error_reporting/performance_plot.html +2 -0
- edsl/templates/error_reporting/report.css +74 -0
- edsl/templates/error_reporting/report.html +118 -0
- edsl/templates/error_reporting/report.js +25 -0
- edsl/tools/__init__.py +1 -0
- edsl/tools/clusters.py +192 -0
- edsl/tools/embeddings.py +27 -0
- edsl/tools/embeddings_plotting.py +118 -0
- edsl/tools/plotting.py +112 -0
- edsl/tools/summarize.py +18 -0
- edsl/utilities/PrettyList.py +56 -0
- edsl/utilities/SystemInfo.py +5 -0
- edsl/utilities/__init__.py +21 -20
- edsl/utilities/ast_utilities.py +3 -0
- edsl/utilities/data/Registry.py +2 -0
- edsl/utilities/decorators.py +41 -0
- edsl/utilities/gcp_bucket/__init__.py +0 -0
- edsl/utilities/gcp_bucket/cloud_storage.py +96 -0
- edsl/utilities/interface.py +310 -60
- edsl/utilities/is_notebook.py +18 -0
- edsl/utilities/is_valid_variable_name.py +11 -0
- edsl/utilities/naming_utilities.py +263 -0
- edsl/utilities/remove_edsl_version.py +24 -0
- edsl/utilities/repair_functions.py +28 -0
- edsl/utilities/restricted_python.py +70 -0
- edsl/utilities/utilities.py +203 -13
- edsl-0.1.40.dist-info/METADATA +111 -0
- edsl-0.1.40.dist-info/RECORD +362 -0
- {edsl-0.1.14.dist-info → edsl-0.1.40.dist-info}/WHEEL +1 -1
- edsl/agents/AgentListExportMixin.py +0 -24
- edsl/coop/old.py +0 -31
- edsl/data/Database.py +0 -141
- edsl/data/crud.py +0 -121
- edsl/jobs/Interview.py +0 -417
- edsl/jobs/JobsRunner.py +0 -63
- edsl/jobs/JobsRunnerStatusMixin.py +0 -115
- edsl/jobs/base.py +0 -47
- edsl/jobs/buckets.py +0 -166
- edsl/jobs/runners/JobsRunnerDryRun.py +0 -19
- edsl/jobs/runners/JobsRunnerStreaming.py +0 -54
- edsl/jobs/task_management.py +0 -218
- edsl/jobs/token_tracking.py +0 -78
- edsl/language_models/DeepInfra.py +0 -69
- edsl/language_models/OpenAI.py +0 -98
- edsl/language_models/model_interfaces/GeminiPro.py +0 -66
- edsl/language_models/model_interfaces/LanguageModelOpenAIFour.py +0 -8
- edsl/language_models/model_interfaces/LanguageModelOpenAIThreeFiveTurbo.py +0 -8
- edsl/language_models/model_interfaces/LlamaTwo13B.py +0 -21
- edsl/language_models/model_interfaces/LlamaTwo70B.py +0 -21
- edsl/language_models/model_interfaces/Mixtral8x7B.py +0 -24
- edsl/language_models/registry.py +0 -81
- edsl/language_models/schemas.py +0 -15
- edsl/language_models/unused/ReplicateBase.py +0 -83
- edsl/prompts/QuestionInstructionsBase.py +0 -6
- edsl/prompts/library/agent_instructions.py +0 -29
- edsl/prompts/library/agent_persona.py +0 -17
- edsl/prompts/library/question_budget.py +0 -26
- edsl/prompts/library/question_checkbox.py +0 -32
- edsl/prompts/library/question_extract.py +0 -19
- edsl/prompts/library/question_freetext.py +0 -14
- edsl/prompts/library/question_linear_scale.py +0 -20
- edsl/prompts/library/question_list.py +0 -22
- edsl/prompts/library/question_multiple_choice.py +0 -44
- edsl/prompts/library/question_numerical.py +0 -31
- edsl/prompts/library/question_rank.py +0 -21
- edsl/prompts/prompt_config.py +0 -33
- edsl/prompts/registry.py +0 -185
- edsl/questions/Question.py +0 -240
- edsl/report/InputOutputDataTypes.py +0 -134
- edsl/report/RegressionMixin.py +0 -28
- edsl/report/ReportOutputs.py +0 -1228
- edsl/report/ResultsFetchMixin.py +0 -106
- edsl/report/ResultsOutputMixin.py +0 -14
- edsl/report/demo.ipynb +0 -645
- edsl/results/ResultsDBMixin.py +0 -184
- edsl/surveys/SurveyFlowVisualizationMixin.py +0 -92
- edsl/trackers/Tracker.py +0 -91
- edsl/trackers/TrackerAPI.py +0 -196
- edsl/trackers/TrackerTasks.py +0 -70
- edsl/utilities/pastebin.py +0 -141
- edsl-0.1.14.dist-info/METADATA +0 -69
- edsl-0.1.14.dist-info/RECORD +0 -141
- /edsl/{language_models/model_interfaces → inference_services}/__init__.py +0 -0
- /edsl/{report/__init__.py → jobs/runners/JobsRunnerStatusData.py} +0 -0
- /edsl/{trackers/__init__.py → language_models/ServiceDataSources.py} +0 -0
- {edsl-0.1.14.dist-info → edsl-0.1.40.dist-info}/LICENSE +0 -0
edsl/jobs/JobsChecks.py
ADDED
@@ -0,0 +1,172 @@
|
|
1
|
+
import os
|
2
|
+
from edsl.exceptions.general import MissingAPIKeyError
|
3
|
+
|
4
|
+
|
5
|
+
class JobsChecks:
|
6
|
+
def __init__(self, jobs):
|
7
|
+
"""Checks a Jobs object for missing API keys and other requirements."""
|
8
|
+
self.jobs = jobs
|
9
|
+
|
10
|
+
def check_api_keys(self) -> None:
|
11
|
+
from edsl.language_models.model import Model
|
12
|
+
|
13
|
+
if len(self.jobs.models) == 0:
|
14
|
+
models = [Model()]
|
15
|
+
else:
|
16
|
+
models = self.jobs.models
|
17
|
+
|
18
|
+
for model in models: # + [Model()]:
|
19
|
+
if not model.has_valid_api_key():
|
20
|
+
raise MissingAPIKeyError(
|
21
|
+
model_name=str(model.model),
|
22
|
+
inference_service=model._inference_service_,
|
23
|
+
)
|
24
|
+
|
25
|
+
def get_missing_api_keys(self) -> set:
|
26
|
+
"""
|
27
|
+
Returns a list of the api keys that a user needs to run this job, but does not currently have in their .env file.
|
28
|
+
"""
|
29
|
+
missing_api_keys = set()
|
30
|
+
|
31
|
+
from edsl.language_models.model import Model
|
32
|
+
from edsl.enums import service_to_api_keyname
|
33
|
+
|
34
|
+
for model in self.jobs.models + [Model()]:
|
35
|
+
if not model.has_valid_api_key():
|
36
|
+
key_name = service_to_api_keyname.get(
|
37
|
+
model._inference_service_, "NOT FOUND"
|
38
|
+
)
|
39
|
+
missing_api_keys.add(key_name)
|
40
|
+
|
41
|
+
return missing_api_keys
|
42
|
+
|
43
|
+
def user_has_ep_api_key(self) -> bool:
|
44
|
+
"""
|
45
|
+
Returns True if the user has an EXPECTED_PARROT_API_KEY in their env.
|
46
|
+
|
47
|
+
Otherwise, returns False.
|
48
|
+
"""
|
49
|
+
|
50
|
+
coop_api_key = os.getenv("EXPECTED_PARROT_API_KEY")
|
51
|
+
|
52
|
+
if coop_api_key is not None:
|
53
|
+
return True
|
54
|
+
else:
|
55
|
+
return False
|
56
|
+
|
57
|
+
def user_has_all_model_keys(self):
|
58
|
+
"""
|
59
|
+
Returns True if the user has all model keys required to run their job.
|
60
|
+
|
61
|
+
Otherwise, returns False.
|
62
|
+
"""
|
63
|
+
|
64
|
+
try:
|
65
|
+
self.check_api_keys()
|
66
|
+
return True
|
67
|
+
except MissingAPIKeyError:
|
68
|
+
return False
|
69
|
+
except Exception:
|
70
|
+
raise
|
71
|
+
|
72
|
+
def needs_external_llms(self) -> bool:
|
73
|
+
"""
|
74
|
+
Returns True if the job needs external LLMs to run.
|
75
|
+
|
76
|
+
Otherwise, returns False.
|
77
|
+
"""
|
78
|
+
# These cases are necessary to skip the API key check during doctests
|
79
|
+
|
80
|
+
# Accounts for Results.example()
|
81
|
+
all_agents_answer_questions_directly = len(self.jobs.agents) > 0 and all(
|
82
|
+
[hasattr(a, "answer_question_directly") for a in self.jobs.agents]
|
83
|
+
)
|
84
|
+
|
85
|
+
# Accounts for InterviewExceptionEntry.example()
|
86
|
+
only_model_is_test = set([m.model for m in self.jobs.models]) == set(["test"])
|
87
|
+
|
88
|
+
# Accounts for Survey.__call__
|
89
|
+
all_questions_are_functional = set(
|
90
|
+
[q.question_type for q in self.jobs.survey.questions]
|
91
|
+
) == set(["functional"])
|
92
|
+
|
93
|
+
if (
|
94
|
+
all_agents_answer_questions_directly
|
95
|
+
or only_model_is_test
|
96
|
+
or all_questions_are_functional
|
97
|
+
):
|
98
|
+
return False
|
99
|
+
else:
|
100
|
+
return True
|
101
|
+
|
102
|
+
def needs_key_process(self):
|
103
|
+
"""
|
104
|
+
A User needs the key process when:
|
105
|
+
1. They don't have all the model keys
|
106
|
+
2. They don't have the EP API
|
107
|
+
3. They need external LLMs to run the job
|
108
|
+
"""
|
109
|
+
return (
|
110
|
+
not self.user_has_all_model_keys()
|
111
|
+
and not self.user_has_ep_api_key()
|
112
|
+
and self.needs_external_llms()
|
113
|
+
)
|
114
|
+
|
115
|
+
def status(self) -> dict:
|
116
|
+
"""
|
117
|
+
Returns a dictionary with the status of the job checks.
|
118
|
+
"""
|
119
|
+
return {
|
120
|
+
"user_has_ep_api_key": self.user_has_ep_api_key(),
|
121
|
+
"user_has_all_model_keys": self.user_has_all_model_keys(),
|
122
|
+
"needs_external_llms": self.needs_external_llms(),
|
123
|
+
"needs_key_process": self.needs_key_process(),
|
124
|
+
}
|
125
|
+
|
126
|
+
def key_process(self):
|
127
|
+
import secrets
|
128
|
+
from dotenv import load_dotenv
|
129
|
+
from edsl.config import CONFIG
|
130
|
+
from edsl.coop.coop import Coop
|
131
|
+
from edsl.utilities.utilities import write_api_key_to_env
|
132
|
+
|
133
|
+
missing_api_keys = self.get_missing_api_keys()
|
134
|
+
|
135
|
+
edsl_auth_token = secrets.token_urlsafe(16)
|
136
|
+
|
137
|
+
print("You're missing some of the API keys needed to run this job:")
|
138
|
+
for api_key in missing_api_keys:
|
139
|
+
print(f" 🔑 {api_key}")
|
140
|
+
print(
|
141
|
+
"\nYou can either add the missing keys to your .env file, or use remote inference."
|
142
|
+
)
|
143
|
+
print("Remote inference allows you to run jobs on our server.")
|
144
|
+
|
145
|
+
coop = Coop()
|
146
|
+
coop._display_login_url(
|
147
|
+
edsl_auth_token=edsl_auth_token,
|
148
|
+
link_description="\n🚀 To use remote inference, sign up at the following link:",
|
149
|
+
)
|
150
|
+
|
151
|
+
print(
|
152
|
+
"\nOnce you log in, we will automatically retrieve your Expected Parrot API key and continue your job remotely."
|
153
|
+
)
|
154
|
+
|
155
|
+
api_key = coop._poll_for_api_key(edsl_auth_token)
|
156
|
+
|
157
|
+
if api_key is None:
|
158
|
+
print("\nTimed out waiting for login. Please try again.")
|
159
|
+
return
|
160
|
+
|
161
|
+
path_to_env = write_api_key_to_env(api_key)
|
162
|
+
print("\n✨ API key retrieved and written to .env file at the following path:")
|
163
|
+
print(f" {path_to_env}")
|
164
|
+
|
165
|
+
# Retrieve API key so we can continue running the job
|
166
|
+
load_dotenv()
|
167
|
+
|
168
|
+
|
169
|
+
if __name__ == "__main__":
|
170
|
+
import doctest
|
171
|
+
|
172
|
+
doctest.testmod()
|
@@ -0,0 +1,189 @@
|
|
1
|
+
from typing import Union, Sequence, TYPE_CHECKING
|
2
|
+
|
3
|
+
if TYPE_CHECKING:
|
4
|
+
from edsl.agents.Agent import Agent
|
5
|
+
from edsl.language_models.LanguageModel import LanguageModel
|
6
|
+
from edsl.scenarios.Scenario import Scenario
|
7
|
+
from edsl.jobs.Jobs import Jobs
|
8
|
+
|
9
|
+
|
10
|
+
class JobsComponentConstructor:
|
11
|
+
"Handles the creation of Agents, Scenarios, and LanguageModels in a job."
|
12
|
+
|
13
|
+
def __init__(self, jobs: "Jobs"):
|
14
|
+
self.jobs = jobs
|
15
|
+
|
16
|
+
def by(
|
17
|
+
self,
|
18
|
+
*args: Union[
|
19
|
+
"Agent",
|
20
|
+
"Scenario",
|
21
|
+
"LanguageModel",
|
22
|
+
Sequence[Union["Agent", "Scenario", "LanguageModel"]],
|
23
|
+
],
|
24
|
+
) -> "Jobs":
|
25
|
+
"""
|
26
|
+
Add Agents, Scenarios and LanguageModels to a job.
|
27
|
+
|
28
|
+
:param args: objects or a sequence (list, tuple, ...) of objects of the same type
|
29
|
+
|
30
|
+
If no objects of this type exist in the Jobs instance, it stores the new objects as a list in the corresponding attribute.
|
31
|
+
Otherwise, it combines the new objects with existing objects using the object's `__add__` method.
|
32
|
+
|
33
|
+
This 'by' is intended to create a fluent interface.
|
34
|
+
|
35
|
+
>>> from edsl.surveys import Survey
|
36
|
+
>>> from edsl.questions import QuestionFreeText
|
37
|
+
>>> q = QuestionFreeText(question_name="name", question_text="What is your name?")
|
38
|
+
>>> from edsl.jobs import Jobs
|
39
|
+
>>> j = Jobs(survey = Survey(questions=[q]))
|
40
|
+
>>> j
|
41
|
+
Jobs(survey=Survey(...), agents=AgentList([]), models=ModelList([]), scenarios=ScenarioList([]))
|
42
|
+
>>> from edsl import Agent; a = Agent(traits = {"status": "Sad"})
|
43
|
+
>>> j.by(a).agents
|
44
|
+
AgentList([Agent(traits = {'status': 'Sad'})])
|
45
|
+
|
46
|
+
|
47
|
+
Notes:
|
48
|
+
- all objects must implement the 'get_value', 'set_value', and `__add__` methods
|
49
|
+
- agents: traits of new agents are combined with traits of existing agents. New and existing agents should not have overlapping traits, and do not increase the # agents in the instance
|
50
|
+
- scenarios: traits of new scenarios are combined with traits of old existing. New scenarios will overwrite overlapping traits, and do not increase the number of scenarios in the instance
|
51
|
+
- models: new models overwrite old models.
|
52
|
+
"""
|
53
|
+
from edsl.results.Dataset import Dataset
|
54
|
+
|
55
|
+
if isinstance(
|
56
|
+
args[0], Dataset
|
57
|
+
): # let the user use a Dataset as if it were a ScenarioList
|
58
|
+
args = args[0].to_scenario_list()
|
59
|
+
|
60
|
+
passed_objects = self._turn_args_to_list(
|
61
|
+
args
|
62
|
+
) # objects can also be passed comma-separated
|
63
|
+
|
64
|
+
current_objects, objects_key = self._get_current_objects_of_this_type(
|
65
|
+
passed_objects[0]
|
66
|
+
)
|
67
|
+
|
68
|
+
if not current_objects:
|
69
|
+
new_objects = passed_objects
|
70
|
+
else:
|
71
|
+
new_objects = self._merge_objects(passed_objects, current_objects)
|
72
|
+
|
73
|
+
setattr(self.jobs, objects_key, new_objects) # update the job object
|
74
|
+
return self.jobs
|
75
|
+
|
76
|
+
@staticmethod
|
77
|
+
def _turn_args_to_list(args):
|
78
|
+
"""Return a list of the first argument if it is a sequence, otherwise returns a list of all the arguments.
|
79
|
+
|
80
|
+
Example:
|
81
|
+
|
82
|
+
>>> JobsComponentConstructor._turn_args_to_list([1,2,3])
|
83
|
+
[1, 2, 3]
|
84
|
+
|
85
|
+
"""
|
86
|
+
|
87
|
+
def did_user_pass_a_sequence(args):
|
88
|
+
"""Return True if the user passed a sequence, False otherwise.
|
89
|
+
|
90
|
+
Example:
|
91
|
+
|
92
|
+
>>> did_user_pass_a_sequence([1,2,3])
|
93
|
+
True
|
94
|
+
|
95
|
+
>>> did_user_pass_a_sequence(1)
|
96
|
+
False
|
97
|
+
"""
|
98
|
+
return len(args) == 1 and isinstance(args[0], Sequence)
|
99
|
+
|
100
|
+
if did_user_pass_a_sequence(args):
|
101
|
+
container_class = JobsComponentConstructor._get_container_class(args[0][0])
|
102
|
+
return container_class(args[0])
|
103
|
+
else:
|
104
|
+
container_class = JobsComponentConstructor._get_container_class(args[0])
|
105
|
+
return container_class(args)
|
106
|
+
|
107
|
+
def _get_current_objects_of_this_type(
|
108
|
+
self, object: Union["Agent", "Scenario", "LanguageModel"]
|
109
|
+
) -> tuple[list, str]:
|
110
|
+
from edsl.agents.Agent import Agent
|
111
|
+
from edsl.scenarios.Scenario import Scenario
|
112
|
+
from edsl.language_models.LanguageModel import LanguageModel
|
113
|
+
|
114
|
+
"""Return the current objects of the same type as the first argument.
|
115
|
+
|
116
|
+
>>> from edsl.jobs import Jobs
|
117
|
+
>>> j = JobsComponentConstructor(Jobs.example())
|
118
|
+
>>> j._get_current_objects_of_this_type(j.agents[0])
|
119
|
+
(AgentList([Agent(traits = {'status': 'Joyful'}), Agent(traits = {'status': 'Sad'})]), 'agents')
|
120
|
+
"""
|
121
|
+
class_to_key = {
|
122
|
+
Agent: "agents",
|
123
|
+
Scenario: "scenarios",
|
124
|
+
LanguageModel: "models",
|
125
|
+
}
|
126
|
+
for class_type in class_to_key:
|
127
|
+
if isinstance(object, class_type) or issubclass(
|
128
|
+
object.__class__, class_type
|
129
|
+
):
|
130
|
+
key = class_to_key[class_type]
|
131
|
+
break
|
132
|
+
else:
|
133
|
+
raise ValueError(
|
134
|
+
f"First argument must be an Agent, Scenario, or LanguageModel, not {object}"
|
135
|
+
)
|
136
|
+
current_objects = getattr(self.jobs, key, None)
|
137
|
+
return current_objects, key
|
138
|
+
|
139
|
+
@staticmethod
|
140
|
+
def _get_empty_container_object(object):
|
141
|
+
from edsl.agents.AgentList import AgentList
|
142
|
+
from edsl.scenarios.ScenarioList import ScenarioList
|
143
|
+
|
144
|
+
return {"Agent": AgentList([]), "Scenario": ScenarioList([])}.get(
|
145
|
+
object.__class__.__name__, []
|
146
|
+
)
|
147
|
+
|
148
|
+
@staticmethod
|
149
|
+
def _merge_objects(passed_objects, current_objects) -> list:
|
150
|
+
"""
|
151
|
+
Combine all the existing objects with the new objects.
|
152
|
+
|
153
|
+
For example, if the user passes in 3 agents,
|
154
|
+
and there are 2 existing agents, this will create 6 new agents
|
155
|
+
>>> from edsl.jobs import Jobs
|
156
|
+
>>> JobsComponentConstructor(Jobs(survey = []))._merge_objects([1,2,3], [4,5,6])
|
157
|
+
[5, 6, 7, 6, 7, 8, 7, 8, 9]
|
158
|
+
"""
|
159
|
+
new_objects = JobsComponentConstructor._get_empty_container_object(
|
160
|
+
passed_objects[0]
|
161
|
+
)
|
162
|
+
for current_object in current_objects:
|
163
|
+
for new_object in passed_objects:
|
164
|
+
new_objects.append(current_object + new_object)
|
165
|
+
return new_objects
|
166
|
+
|
167
|
+
@staticmethod
|
168
|
+
def _get_container_class(object):
|
169
|
+
from edsl.agents.AgentList import AgentList
|
170
|
+
from edsl.agents.Agent import Agent
|
171
|
+
from edsl.scenarios.Scenario import Scenario
|
172
|
+
from edsl.scenarios.ScenarioList import ScenarioList
|
173
|
+
from edsl.language_models.ModelList import ModelList
|
174
|
+
|
175
|
+
if isinstance(object, Agent):
|
176
|
+
return AgentList
|
177
|
+
elif isinstance(object, Scenario):
|
178
|
+
return ScenarioList
|
179
|
+
elif isinstance(object, ModelList):
|
180
|
+
return ModelList
|
181
|
+
else:
|
182
|
+
return list
|
183
|
+
|
184
|
+
|
185
|
+
if __name__ == "__main__":
|
186
|
+
"""Run the module's doctests."""
|
187
|
+
import doctest
|
188
|
+
|
189
|
+
doctest.testmod(optionflags=doctest.ELLIPSIS)
|
edsl/jobs/JobsPrompts.py
ADDED
@@ -0,0 +1,270 @@
|
|
1
|
+
from typing import List, TYPE_CHECKING
|
2
|
+
|
3
|
+
from edsl.results.Dataset import Dataset
|
4
|
+
|
5
|
+
if TYPE_CHECKING:
|
6
|
+
from edsl.jobs import Jobs
|
7
|
+
|
8
|
+
# from edsl.jobs.interviews.Interview import Interview
|
9
|
+
# from edsl.results.Dataset import Dataset
|
10
|
+
# from edsl.agents.AgentList import AgentList
|
11
|
+
# from edsl.scenarios.ScenarioList import ScenarioList
|
12
|
+
# from edsl.surveys.Survey import Survey
|
13
|
+
|
14
|
+
from edsl.jobs.FetchInvigilator import FetchInvigilator
|
15
|
+
|
16
|
+
|
17
|
+
class JobsPrompts:
|
18
|
+
def __init__(self, jobs: "Jobs"):
|
19
|
+
self.interviews = jobs.interviews()
|
20
|
+
self.agents = jobs.agents
|
21
|
+
self.scenarios = jobs.scenarios
|
22
|
+
self.survey = jobs.survey
|
23
|
+
self._price_lookup = None
|
24
|
+
|
25
|
+
@property
|
26
|
+
def price_lookup(self):
|
27
|
+
if self._price_lookup is None:
|
28
|
+
from edsl.coop.coop import Coop
|
29
|
+
|
30
|
+
c = Coop()
|
31
|
+
self._price_lookup = c.fetch_prices()
|
32
|
+
return self._price_lookup
|
33
|
+
|
34
|
+
def prompts(self) -> "Dataset":
|
35
|
+
"""Return a Dataset of prompts that will be used.
|
36
|
+
|
37
|
+
>>> from edsl.jobs import Jobs
|
38
|
+
>>> Jobs.example().prompts()
|
39
|
+
Dataset(...)
|
40
|
+
"""
|
41
|
+
interviews = self.interviews
|
42
|
+
interview_indices = []
|
43
|
+
question_names = []
|
44
|
+
user_prompts = []
|
45
|
+
system_prompts = []
|
46
|
+
scenario_indices = []
|
47
|
+
agent_indices = []
|
48
|
+
models = []
|
49
|
+
costs = []
|
50
|
+
|
51
|
+
for interview_index, interview in enumerate(interviews):
|
52
|
+
invigilators = [
|
53
|
+
FetchInvigilator(interview)(question)
|
54
|
+
for question in interview.survey.questions
|
55
|
+
]
|
56
|
+
for _, invigilator in enumerate(invigilators):
|
57
|
+
prompts = invigilator.get_prompts()
|
58
|
+
user_prompt = prompts["user_prompt"]
|
59
|
+
system_prompt = prompts["system_prompt"]
|
60
|
+
user_prompts.append(user_prompt)
|
61
|
+
system_prompts.append(system_prompt)
|
62
|
+
agent_index = self.agents.index(invigilator.agent)
|
63
|
+
agent_indices.append(agent_index)
|
64
|
+
interview_indices.append(interview_index)
|
65
|
+
scenario_index = self.scenarios.index(invigilator.scenario)
|
66
|
+
scenario_indices.append(scenario_index)
|
67
|
+
models.append(invigilator.model.model)
|
68
|
+
question_names.append(invigilator.question.question_name)
|
69
|
+
|
70
|
+
prompt_cost = self.estimate_prompt_cost(
|
71
|
+
system_prompt=system_prompt,
|
72
|
+
user_prompt=user_prompt,
|
73
|
+
price_lookup=self.price_lookup,
|
74
|
+
inference_service=invigilator.model._inference_service_,
|
75
|
+
model=invigilator.model.model,
|
76
|
+
)
|
77
|
+
costs.append(prompt_cost["cost_usd"])
|
78
|
+
|
79
|
+
d = Dataset(
|
80
|
+
[
|
81
|
+
{"user_prompt": user_prompts},
|
82
|
+
{"system_prompt": system_prompts},
|
83
|
+
{"interview_index": interview_indices},
|
84
|
+
{"question_name": question_names},
|
85
|
+
{"scenario_index": scenario_indices},
|
86
|
+
{"agent_index": agent_indices},
|
87
|
+
{"model": models},
|
88
|
+
{"estimated_cost": costs},
|
89
|
+
]
|
90
|
+
)
|
91
|
+
return d
|
92
|
+
|
93
|
+
@staticmethod
|
94
|
+
def estimate_prompt_cost(
|
95
|
+
system_prompt: str,
|
96
|
+
user_prompt: str,
|
97
|
+
price_lookup: dict,
|
98
|
+
inference_service: str,
|
99
|
+
model: str,
|
100
|
+
) -> dict:
|
101
|
+
"""Estimates the cost of a prompt. Takes piping into account."""
|
102
|
+
import math
|
103
|
+
|
104
|
+
def get_piping_multiplier(prompt: str):
|
105
|
+
"""Returns 2 if a prompt includes Jinja braces, and 1 otherwise."""
|
106
|
+
|
107
|
+
if "{{" in prompt and "}}" in prompt:
|
108
|
+
return 2
|
109
|
+
return 1
|
110
|
+
|
111
|
+
# Look up prices per token
|
112
|
+
key = (inference_service, model)
|
113
|
+
|
114
|
+
try:
|
115
|
+
relevant_prices = price_lookup[key]
|
116
|
+
|
117
|
+
service_input_token_price = float(
|
118
|
+
relevant_prices["input"]["service_stated_token_price"]
|
119
|
+
)
|
120
|
+
service_input_token_qty = float(
|
121
|
+
relevant_prices["input"]["service_stated_token_qty"]
|
122
|
+
)
|
123
|
+
input_price_per_token = service_input_token_price / service_input_token_qty
|
124
|
+
|
125
|
+
service_output_token_price = float(
|
126
|
+
relevant_prices["output"]["service_stated_token_price"]
|
127
|
+
)
|
128
|
+
service_output_token_qty = float(
|
129
|
+
relevant_prices["output"]["service_stated_token_qty"]
|
130
|
+
)
|
131
|
+
output_price_per_token = (
|
132
|
+
service_output_token_price / service_output_token_qty
|
133
|
+
)
|
134
|
+
|
135
|
+
except KeyError:
|
136
|
+
# A KeyError is likely to occur if we cannot retrieve prices (the price_lookup dict is empty)
|
137
|
+
# Use a sensible default
|
138
|
+
|
139
|
+
import warnings
|
140
|
+
|
141
|
+
warnings.warn(
|
142
|
+
"Price data could not be retrieved. Using default estimates for input and output token prices. Input: $0.15 / 1M tokens; Output: $0.60 / 1M tokens"
|
143
|
+
)
|
144
|
+
input_price_per_token = 0.00000015 # $0.15 / 1M tokens
|
145
|
+
output_price_per_token = 0.00000060 # $0.60 / 1M tokens
|
146
|
+
|
147
|
+
# Compute the number of characters (double if the question involves piping)
|
148
|
+
user_prompt_chars = len(str(user_prompt)) * get_piping_multiplier(
|
149
|
+
str(user_prompt)
|
150
|
+
)
|
151
|
+
system_prompt_chars = len(str(system_prompt)) * get_piping_multiplier(
|
152
|
+
str(system_prompt)
|
153
|
+
)
|
154
|
+
|
155
|
+
# Convert into tokens (1 token approx. equals 4 characters)
|
156
|
+
input_tokens = (user_prompt_chars + system_prompt_chars) // 4
|
157
|
+
|
158
|
+
output_tokens = math.ceil(0.75 * input_tokens)
|
159
|
+
|
160
|
+
cost = (
|
161
|
+
input_tokens * input_price_per_token
|
162
|
+
+ output_tokens * output_price_per_token
|
163
|
+
)
|
164
|
+
|
165
|
+
return {
|
166
|
+
"input_tokens": input_tokens,
|
167
|
+
"output_tokens": output_tokens,
|
168
|
+
"cost_usd": cost,
|
169
|
+
}
|
170
|
+
|
171
|
+
def estimate_job_cost_from_external_prices(
|
172
|
+
self, price_lookup: dict, iterations: int = 1
|
173
|
+
) -> dict:
|
174
|
+
"""
|
175
|
+
Estimates the cost of a job according to the following assumptions:
|
176
|
+
|
177
|
+
- 1 token = 4 characters.
|
178
|
+
- For each prompt, output tokens = input tokens * 0.75, rounded up to the nearest integer.
|
179
|
+
|
180
|
+
price_lookup is an external pricing dictionary.
|
181
|
+
"""
|
182
|
+
|
183
|
+
import pandas as pd
|
184
|
+
|
185
|
+
interviews = self.interviews
|
186
|
+
data = []
|
187
|
+
for interview in interviews:
|
188
|
+
invigilators = [
|
189
|
+
FetchInvigilator(interview)(question)
|
190
|
+
for question in self.survey.questions
|
191
|
+
]
|
192
|
+
for invigilator in invigilators:
|
193
|
+
prompts = invigilator.get_prompts()
|
194
|
+
|
195
|
+
# By this point, agent and scenario data has already been added to the prompts
|
196
|
+
user_prompt = prompts["user_prompt"]
|
197
|
+
system_prompt = prompts["system_prompt"]
|
198
|
+
inference_service = invigilator.model._inference_service_
|
199
|
+
model = invigilator.model.model
|
200
|
+
|
201
|
+
prompt_cost = self.estimate_prompt_cost(
|
202
|
+
system_prompt=system_prompt,
|
203
|
+
user_prompt=user_prompt,
|
204
|
+
price_lookup=price_lookup,
|
205
|
+
inference_service=inference_service,
|
206
|
+
model=model,
|
207
|
+
)
|
208
|
+
|
209
|
+
data.append(
|
210
|
+
{
|
211
|
+
"user_prompt": user_prompt,
|
212
|
+
"system_prompt": system_prompt,
|
213
|
+
"estimated_input_tokens": prompt_cost["input_tokens"],
|
214
|
+
"estimated_output_tokens": prompt_cost["output_tokens"],
|
215
|
+
"estimated_cost_usd": prompt_cost["cost_usd"],
|
216
|
+
"inference_service": inference_service,
|
217
|
+
"model": model,
|
218
|
+
}
|
219
|
+
)
|
220
|
+
|
221
|
+
df = pd.DataFrame.from_records(data)
|
222
|
+
|
223
|
+
df = (
|
224
|
+
df.groupby(["inference_service", "model"])
|
225
|
+
.agg(
|
226
|
+
{
|
227
|
+
"estimated_cost_usd": "sum",
|
228
|
+
"estimated_input_tokens": "sum",
|
229
|
+
"estimated_output_tokens": "sum",
|
230
|
+
}
|
231
|
+
)
|
232
|
+
.reset_index()
|
233
|
+
)
|
234
|
+
df["estimated_cost_usd"] = df["estimated_cost_usd"] * iterations
|
235
|
+
df["estimated_input_tokens"] = df["estimated_input_tokens"] * iterations
|
236
|
+
df["estimated_output_tokens"] = df["estimated_output_tokens"] * iterations
|
237
|
+
|
238
|
+
estimated_costs_by_model = df.to_dict("records")
|
239
|
+
|
240
|
+
estimated_total_cost = sum(
|
241
|
+
model["estimated_cost_usd"] for model in estimated_costs_by_model
|
242
|
+
)
|
243
|
+
estimated_total_input_tokens = sum(
|
244
|
+
model["estimated_input_tokens"] for model in estimated_costs_by_model
|
245
|
+
)
|
246
|
+
estimated_total_output_tokens = sum(
|
247
|
+
model["estimated_output_tokens"] for model in estimated_costs_by_model
|
248
|
+
)
|
249
|
+
|
250
|
+
output = {
|
251
|
+
"estimated_total_cost_usd": estimated_total_cost,
|
252
|
+
"estimated_total_input_tokens": estimated_total_input_tokens,
|
253
|
+
"estimated_total_output_tokens": estimated_total_output_tokens,
|
254
|
+
"model_costs": estimated_costs_by_model,
|
255
|
+
}
|
256
|
+
|
257
|
+
return output
|
258
|
+
|
259
|
+
def estimate_job_cost(self, iterations: int = 1) -> dict:
|
260
|
+
"""
|
261
|
+
Estimates the cost of a job according to the following assumptions:
|
262
|
+
|
263
|
+
- 1 token = 4 characters.
|
264
|
+
- For each prompt, output tokens = input tokens * 0.75, rounded up to the nearest integer.
|
265
|
+
|
266
|
+
Fetches prices from Coop.
|
267
|
+
"""
|
268
|
+
return self.estimate_job_cost_from_external_prices(
|
269
|
+
price_lookup=self.price_lookup, iterations=iterations
|
270
|
+
)
|