edsl 0.1.39.dev3__py3-none-any.whl → 0.1.39.dev4__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- edsl/Base.py +413 -332
- edsl/BaseDiff.py +260 -260
- edsl/TemplateLoader.py +24 -24
- edsl/__init__.py +57 -49
- edsl/__version__.py +1 -1
- edsl/agents/Agent.py +1071 -867
- edsl/agents/AgentList.py +551 -413
- edsl/agents/Invigilator.py +284 -233
- edsl/agents/InvigilatorBase.py +257 -270
- edsl/agents/PromptConstructor.py +272 -354
- edsl/agents/QuestionInstructionPromptBuilder.py +128 -0
- edsl/agents/QuestionTemplateReplacementsBuilder.py +137 -0
- edsl/agents/__init__.py +2 -3
- edsl/agents/descriptors.py +99 -99
- edsl/agents/prompt_helpers.py +129 -129
- edsl/agents/question_option_processor.py +172 -0
- edsl/auto/AutoStudy.py +130 -117
- edsl/auto/StageBase.py +243 -230
- edsl/auto/StageGenerateSurvey.py +178 -178
- edsl/auto/StageLabelQuestions.py +125 -125
- edsl/auto/StagePersona.py +61 -61
- edsl/auto/StagePersonaDimensionValueRanges.py +88 -88
- edsl/auto/StagePersonaDimensionValues.py +74 -74
- edsl/auto/StagePersonaDimensions.py +69 -69
- edsl/auto/StageQuestions.py +74 -73
- edsl/auto/SurveyCreatorPipeline.py +21 -21
- edsl/auto/utilities.py +218 -224
- edsl/base/Base.py +279 -279
- edsl/config.py +177 -157
- edsl/conversation/Conversation.py +290 -290
- edsl/conversation/car_buying.py +59 -58
- edsl/conversation/chips.py +95 -95
- edsl/conversation/mug_negotiation.py +81 -81
- edsl/conversation/next_speaker_utilities.py +93 -93
- edsl/coop/CoopFunctionsMixin.py +15 -0
- edsl/coop/ExpectedParrotKeyHandler.py +125 -0
- edsl/coop/PriceFetcher.py +54 -54
- edsl/coop/__init__.py +2 -2
- edsl/coop/coop.py +1106 -1028
- edsl/coop/utils.py +131 -131
- edsl/data/Cache.py +573 -555
- edsl/data/CacheEntry.py +230 -233
- edsl/data/CacheHandler.py +168 -149
- edsl/data/RemoteCacheSync.py +186 -78
- edsl/data/SQLiteDict.py +292 -292
- edsl/data/__init__.py +5 -4
- edsl/data/hack.py +10 -0
- edsl/data/orm.py +10 -10
- edsl/data_transfer_models.py +74 -73
- edsl/enums.py +202 -175
- edsl/exceptions/BaseException.py +21 -21
- edsl/exceptions/__init__.py +54 -54
- edsl/exceptions/agents.py +54 -42
- edsl/exceptions/cache.py +5 -5
- edsl/exceptions/configuration.py +16 -16
- edsl/exceptions/coop.py +10 -10
- edsl/exceptions/data.py +14 -14
- edsl/exceptions/general.py +34 -34
- edsl/exceptions/inference_services.py +5 -0
- edsl/exceptions/jobs.py +33 -33
- edsl/exceptions/language_models.py +63 -63
- edsl/exceptions/prompts.py +15 -15
- edsl/exceptions/questions.py +109 -91
- edsl/exceptions/results.py +29 -29
- edsl/exceptions/scenarios.py +29 -22
- edsl/exceptions/surveys.py +37 -37
- edsl/inference_services/AnthropicService.py +106 -87
- edsl/inference_services/AvailableModelCacheHandler.py +184 -0
- edsl/inference_services/AvailableModelFetcher.py +215 -0
- edsl/inference_services/AwsBedrock.py +118 -120
- edsl/inference_services/AzureAI.py +215 -217
- edsl/inference_services/DeepInfraService.py +18 -18
- edsl/inference_services/GoogleService.py +143 -148
- edsl/inference_services/GroqService.py +20 -20
- edsl/inference_services/InferenceServiceABC.py +80 -147
- edsl/inference_services/InferenceServicesCollection.py +138 -97
- edsl/inference_services/MistralAIService.py +120 -123
- edsl/inference_services/OllamaService.py +18 -18
- edsl/inference_services/OpenAIService.py +236 -224
- edsl/inference_services/PerplexityService.py +160 -163
- edsl/inference_services/ServiceAvailability.py +135 -0
- edsl/inference_services/TestService.py +90 -89
- edsl/inference_services/TogetherAIService.py +172 -170
- edsl/inference_services/data_structures.py +134 -0
- edsl/inference_services/models_available_cache.py +118 -118
- edsl/inference_services/rate_limits_cache.py +25 -25
- edsl/inference_services/registry.py +41 -41
- edsl/inference_services/write_available.py +10 -10
- edsl/jobs/AnswerQuestionFunctionConstructor.py +223 -0
- edsl/jobs/Answers.py +43 -56
- edsl/jobs/FetchInvigilator.py +47 -0
- edsl/jobs/InterviewTaskManager.py +98 -0
- edsl/jobs/InterviewsConstructor.py +50 -0
- edsl/jobs/Jobs.py +823 -898
- edsl/jobs/JobsChecks.py +172 -147
- edsl/jobs/JobsComponentConstructor.py +189 -0
- edsl/jobs/JobsPrompts.py +270 -268
- edsl/jobs/JobsRemoteInferenceHandler.py +311 -239
- edsl/jobs/JobsRemoteInferenceLogger.py +239 -0
- edsl/jobs/RequestTokenEstimator.py +30 -0
- edsl/jobs/__init__.py +1 -1
- edsl/jobs/async_interview_runner.py +138 -0
- edsl/jobs/buckets/BucketCollection.py +104 -63
- edsl/jobs/buckets/ModelBuckets.py +65 -65
- edsl/jobs/buckets/TokenBucket.py +283 -251
- edsl/jobs/buckets/TokenBucketAPI.py +211 -0
- edsl/jobs/buckets/TokenBucketClient.py +191 -0
- edsl/jobs/check_survey_scenario_compatibility.py +85 -0
- edsl/jobs/data_structures.py +120 -0
- edsl/jobs/decorators.py +35 -0
- edsl/jobs/interviews/Interview.py +396 -661
- edsl/jobs/interviews/InterviewExceptionCollection.py +99 -99
- edsl/jobs/interviews/InterviewExceptionEntry.py +186 -186
- edsl/jobs/interviews/InterviewStatistic.py +63 -63
- edsl/jobs/interviews/InterviewStatisticsCollection.py +25 -25
- edsl/jobs/interviews/InterviewStatusDictionary.py +78 -78
- edsl/jobs/interviews/InterviewStatusLog.py +92 -92
- edsl/jobs/interviews/ReportErrors.py +66 -66
- edsl/jobs/interviews/interview_status_enum.py +9 -9
- edsl/jobs/jobs_status_enums.py +9 -0
- edsl/jobs/loggers/HTMLTableJobLogger.py +304 -0
- edsl/jobs/results_exceptions_handler.py +98 -0
- edsl/jobs/runners/JobsRunnerAsyncio.py +151 -466
- edsl/jobs/runners/JobsRunnerStatus.py +297 -330
- edsl/jobs/tasks/QuestionTaskCreator.py +244 -242
- edsl/jobs/tasks/TaskCreators.py +64 -64
- edsl/jobs/tasks/TaskHistory.py +470 -450
- edsl/jobs/tasks/TaskStatusLog.py +23 -23
- edsl/jobs/tasks/task_status_enum.py +161 -163
- edsl/jobs/tokens/InterviewTokenUsage.py +27 -27
- edsl/jobs/tokens/TokenUsage.py +34 -34
- edsl/language_models/ComputeCost.py +63 -0
- edsl/language_models/LanguageModel.py +626 -668
- edsl/language_models/ModelList.py +164 -155
- edsl/language_models/PriceManager.py +127 -0
- edsl/language_models/RawResponseHandler.py +106 -0
- edsl/language_models/RegisterLanguageModelsMeta.py +184 -184
- edsl/language_models/ServiceDataSources.py +0 -0
- edsl/language_models/__init__.py +2 -3
- edsl/language_models/fake_openai_call.py +15 -15
- edsl/language_models/fake_openai_service.py +61 -61
- edsl/language_models/key_management/KeyLookup.py +63 -0
- edsl/language_models/key_management/KeyLookupBuilder.py +273 -0
- edsl/language_models/key_management/KeyLookupCollection.py +38 -0
- edsl/language_models/key_management/__init__.py +0 -0
- edsl/language_models/key_management/models.py +131 -0
- edsl/language_models/model.py +256 -0
- edsl/language_models/repair.py +156 -156
- edsl/language_models/utilities.py +65 -64
- edsl/notebooks/Notebook.py +263 -258
- edsl/notebooks/NotebookToLaTeX.py +142 -0
- edsl/notebooks/__init__.py +1 -1
- edsl/prompts/Prompt.py +352 -362
- edsl/prompts/__init__.py +2 -2
- edsl/questions/ExceptionExplainer.py +77 -0
- edsl/questions/HTMLQuestion.py +103 -0
- edsl/questions/QuestionBase.py +518 -664
- edsl/questions/QuestionBasePromptsMixin.py +221 -217
- edsl/questions/QuestionBudget.py +227 -227
- edsl/questions/QuestionCheckBox.py +359 -359
- edsl/questions/QuestionExtract.py +180 -182
- edsl/questions/QuestionFreeText.py +113 -114
- edsl/questions/QuestionFunctional.py +166 -166
- edsl/questions/QuestionList.py +223 -231
- edsl/questions/QuestionMatrix.py +265 -0
- edsl/questions/QuestionMultipleChoice.py +330 -286
- edsl/questions/QuestionNumerical.py +151 -153
- edsl/questions/QuestionRank.py +314 -324
- edsl/questions/Quick.py +41 -41
- edsl/questions/SimpleAskMixin.py +74 -73
- edsl/questions/__init__.py +27 -26
- edsl/questions/{AnswerValidatorMixin.py → answer_validator_mixin.py} +334 -289
- edsl/questions/compose_questions.py +98 -98
- edsl/questions/data_structures.py +20 -0
- edsl/questions/decorators.py +21 -21
- edsl/questions/derived/QuestionLikertFive.py +76 -76
- edsl/questions/derived/QuestionLinearScale.py +90 -87
- edsl/questions/derived/QuestionTopK.py +93 -93
- edsl/questions/derived/QuestionYesNo.py +82 -82
- edsl/questions/descriptors.py +427 -413
- edsl/questions/loop_processor.py +149 -0
- edsl/questions/prompt_templates/question_budget.jinja +13 -13
- edsl/questions/prompt_templates/question_checkbox.jinja +32 -32
- edsl/questions/prompt_templates/question_extract.jinja +11 -11
- edsl/questions/prompt_templates/question_free_text.jinja +3 -3
- edsl/questions/prompt_templates/question_linear_scale.jinja +11 -11
- edsl/questions/prompt_templates/question_list.jinja +17 -17
- edsl/questions/prompt_templates/question_multiple_choice.jinja +33 -33
- edsl/questions/prompt_templates/question_numerical.jinja +36 -36
- edsl/questions/{QuestionBaseGenMixin.py → question_base_gen_mixin.py} +168 -161
- edsl/questions/question_registry.py +177 -177
- edsl/questions/{RegisterQuestionsMeta.py → register_questions_meta.py} +71 -71
- edsl/questions/{ResponseValidatorABC.py → response_validator_abc.py} +188 -174
- edsl/questions/response_validator_factory.py +34 -0
- edsl/questions/settings.py +12 -12
- edsl/questions/templates/budget/answering_instructions.jinja +7 -7
- edsl/questions/templates/budget/question_presentation.jinja +7 -7
- edsl/questions/templates/checkbox/answering_instructions.jinja +10 -10
- edsl/questions/templates/checkbox/question_presentation.jinja +22 -22
- edsl/questions/templates/extract/answering_instructions.jinja +7 -7
- edsl/questions/templates/likert_five/answering_instructions.jinja +10 -10
- edsl/questions/templates/likert_five/question_presentation.jinja +11 -11
- edsl/questions/templates/linear_scale/answering_instructions.jinja +5 -5
- edsl/questions/templates/linear_scale/question_presentation.jinja +5 -5
- edsl/questions/templates/list/answering_instructions.jinja +3 -3
- edsl/questions/templates/list/question_presentation.jinja +5 -5
- edsl/questions/templates/matrix/__init__.py +1 -0
- edsl/questions/templates/matrix/answering_instructions.jinja +5 -0
- edsl/questions/templates/matrix/question_presentation.jinja +20 -0
- edsl/questions/templates/multiple_choice/answering_instructions.jinja +9 -9
- edsl/questions/templates/multiple_choice/question_presentation.jinja +11 -11
- edsl/questions/templates/numerical/answering_instructions.jinja +6 -6
- edsl/questions/templates/numerical/question_presentation.jinja +6 -6
- edsl/questions/templates/rank/answering_instructions.jinja +11 -11
- edsl/questions/templates/rank/question_presentation.jinja +15 -15
- edsl/questions/templates/top_k/answering_instructions.jinja +8 -8
- edsl/questions/templates/top_k/question_presentation.jinja +22 -22
- edsl/questions/templates/yes_no/answering_instructions.jinja +6 -6
- edsl/questions/templates/yes_no/question_presentation.jinja +11 -11
- edsl/results/CSSParameterizer.py +108 -108
- edsl/results/Dataset.py +587 -424
- edsl/results/DatasetExportMixin.py +594 -731
- edsl/results/DatasetTree.py +295 -275
- edsl/results/MarkdownToDocx.py +122 -0
- edsl/results/MarkdownToPDF.py +111 -0
- edsl/results/Result.py +557 -465
- edsl/results/Results.py +1183 -1165
- edsl/results/ResultsExportMixin.py +45 -43
- edsl/results/ResultsGGMixin.py +121 -121
- edsl/results/TableDisplay.py +125 -198
- edsl/results/TextEditor.py +50 -0
- edsl/results/__init__.py +2 -2
- edsl/results/file_exports.py +252 -0
- edsl/results/{ResultsFetchMixin.py → results_fetch_mixin.py} +33 -33
- edsl/results/{Selector.py → results_selector.py} +145 -135
- edsl/results/{ResultsToolsMixin.py → results_tools_mixin.py} +98 -98
- edsl/results/smart_objects.py +96 -0
- edsl/results/table_data_class.py +12 -0
- edsl/results/table_display.css +77 -77
- edsl/results/table_renderers.py +118 -0
- edsl/results/tree_explore.py +115 -115
- edsl/scenarios/ConstructDownloadLink.py +109 -0
- edsl/scenarios/DocumentChunker.py +102 -0
- edsl/scenarios/DocxScenario.py +16 -0
- edsl/scenarios/FileStore.py +511 -632
- edsl/scenarios/PdfExtractor.py +40 -0
- edsl/scenarios/Scenario.py +498 -601
- edsl/scenarios/ScenarioHtmlMixin.py +65 -64
- edsl/scenarios/ScenarioList.py +1458 -1287
- edsl/scenarios/ScenarioListExportMixin.py +45 -52
- edsl/scenarios/ScenarioListPdfMixin.py +239 -261
- edsl/scenarios/__init__.py +3 -4
- edsl/scenarios/directory_scanner.py +96 -0
- edsl/scenarios/file_methods.py +85 -0
- edsl/scenarios/handlers/__init__.py +13 -0
- edsl/scenarios/handlers/csv.py +38 -0
- edsl/scenarios/handlers/docx.py +76 -0
- edsl/scenarios/handlers/html.py +37 -0
- edsl/scenarios/handlers/json.py +111 -0
- edsl/scenarios/handlers/latex.py +5 -0
- edsl/scenarios/handlers/md.py +51 -0
- edsl/scenarios/handlers/pdf.py +68 -0
- edsl/scenarios/handlers/png.py +39 -0
- edsl/scenarios/handlers/pptx.py +105 -0
- edsl/scenarios/handlers/py.py +294 -0
- edsl/scenarios/handlers/sql.py +313 -0
- edsl/scenarios/handlers/sqlite.py +149 -0
- edsl/scenarios/handlers/txt.py +33 -0
- edsl/scenarios/{ScenarioJoin.py → scenario_join.py} +131 -127
- edsl/scenarios/scenario_selector.py +156 -0
- edsl/shared.py +1 -1
- edsl/study/ObjectEntry.py +173 -173
- edsl/study/ProofOfWork.py +113 -113
- edsl/study/SnapShot.py +80 -80
- edsl/study/Study.py +521 -528
- edsl/study/__init__.py +4 -4
- edsl/surveys/ConstructDAG.py +92 -0
- edsl/surveys/DAG.py +148 -148
- edsl/surveys/EditSurvey.py +221 -0
- edsl/surveys/InstructionHandler.py +100 -0
- edsl/surveys/Memory.py +31 -31
- edsl/surveys/MemoryManagement.py +72 -0
- edsl/surveys/MemoryPlan.py +244 -244
- edsl/surveys/Rule.py +327 -326
- edsl/surveys/RuleCollection.py +385 -387
- edsl/surveys/RuleManager.py +172 -0
- edsl/surveys/Simulator.py +75 -0
- edsl/surveys/Survey.py +1280 -1801
- edsl/surveys/SurveyCSS.py +273 -261
- edsl/surveys/SurveyExportMixin.py +259 -259
- edsl/surveys/{SurveyFlowVisualizationMixin.py → SurveyFlowVisualization.py} +181 -179
- edsl/surveys/SurveyQualtricsImport.py +284 -284
- edsl/surveys/SurveyToApp.py +141 -0
- edsl/surveys/__init__.py +5 -3
- edsl/surveys/base.py +53 -53
- edsl/surveys/descriptors.py +60 -56
- edsl/surveys/instructions/ChangeInstruction.py +48 -49
- edsl/surveys/instructions/Instruction.py +56 -65
- edsl/surveys/instructions/InstructionCollection.py +82 -77
- edsl/templates/error_reporting/base.html +23 -23
- edsl/templates/error_reporting/exceptions_by_model.html +34 -34
- edsl/templates/error_reporting/exceptions_by_question_name.html +16 -16
- edsl/templates/error_reporting/exceptions_by_type.html +16 -16
- edsl/templates/error_reporting/interview_details.html +115 -115
- edsl/templates/error_reporting/interviews.html +19 -19
- edsl/templates/error_reporting/overview.html +4 -4
- edsl/templates/error_reporting/performance_plot.html +1 -1
- edsl/templates/error_reporting/report.css +73 -73
- edsl/templates/error_reporting/report.html +117 -117
- edsl/templates/error_reporting/report.js +25 -25
- edsl/test_h +1 -0
- edsl/tools/__init__.py +1 -1
- edsl/tools/clusters.py +192 -192
- edsl/tools/embeddings.py +27 -27
- edsl/tools/embeddings_plotting.py +118 -118
- edsl/tools/plotting.py +112 -112
- edsl/tools/summarize.py +18 -18
- edsl/utilities/PrettyList.py +56 -0
- edsl/utilities/SystemInfo.py +28 -28
- edsl/utilities/__init__.py +22 -22
- edsl/utilities/ast_utilities.py +25 -25
- edsl/utilities/data/Registry.py +6 -6
- edsl/utilities/data/__init__.py +1 -1
- edsl/utilities/data/scooter_results.json +1 -1
- edsl/utilities/decorators.py +77 -77
- edsl/utilities/gcp_bucket/cloud_storage.py +96 -96
- edsl/utilities/gcp_bucket/example.py +50 -0
- edsl/utilities/interface.py +627 -627
- edsl/utilities/is_notebook.py +18 -0
- edsl/utilities/is_valid_variable_name.py +11 -0
- edsl/utilities/naming_utilities.py +263 -263
- edsl/utilities/remove_edsl_version.py +24 -0
- edsl/utilities/repair_functions.py +28 -28
- edsl/utilities/restricted_python.py +70 -70
- edsl/utilities/utilities.py +436 -424
- {edsl-0.1.39.dev3.dist-info → edsl-0.1.39.dev4.dist-info}/LICENSE +21 -21
- {edsl-0.1.39.dev3.dist-info → edsl-0.1.39.dev4.dist-info}/METADATA +13 -11
- edsl-0.1.39.dev4.dist-info/RECORD +361 -0
- edsl/language_models/KeyLookup.py +0 -30
- edsl/language_models/registry.py +0 -190
- edsl/language_models/unused/ReplicateBase.py +0 -83
- edsl/results/ResultsDBMixin.py +0 -238
- edsl-0.1.39.dev3.dist-info/RECORD +0 -277
- {edsl-0.1.39.dev3.dist-info → edsl-0.1.39.dev4.dist-info}/WHEEL +0 -0
@@ -1,668 +1,626 @@
|
|
1
|
-
"""This module contains the LanguageModel class, which is an abstract base class for all language models.
|
2
|
-
|
3
|
-
Terminology:
|
4
|
-
|
5
|
-
raw_response: The JSON response from the model. This has all the model meta-data about the call.
|
6
|
-
|
7
|
-
edsl_augmented_response: The JSON response from model, but augmented with EDSL-specific information,
|
8
|
-
such as the cache key, token usage, etc.
|
9
|
-
|
10
|
-
generated_tokens: The actual tokens generated by the model. This is the output that is used by the user.
|
11
|
-
edsl_answer_dict: The parsed JSON response from the model either {'answer': ...} or {'answer': ..., 'comment': ...}
|
12
|
-
|
13
|
-
"""
|
14
|
-
|
15
|
-
from __future__ import annotations
|
16
|
-
import warnings
|
17
|
-
from functools import wraps
|
18
|
-
import asyncio
|
19
|
-
import json
|
20
|
-
import os
|
21
|
-
from typing import (
|
22
|
-
Coroutine,
|
23
|
-
Any,
|
24
|
-
|
25
|
-
|
26
|
-
|
27
|
-
|
28
|
-
|
29
|
-
|
30
|
-
|
31
|
-
|
32
|
-
|
33
|
-
|
34
|
-
|
35
|
-
|
36
|
-
|
37
|
-
|
38
|
-
|
39
|
-
|
40
|
-
|
41
|
-
|
42
|
-
|
43
|
-
|
44
|
-
|
45
|
-
from edsl.
|
46
|
-
|
47
|
-
|
48
|
-
|
49
|
-
from edsl.
|
50
|
-
|
51
|
-
|
52
|
-
|
53
|
-
from edsl.
|
54
|
-
|
55
|
-
from edsl.
|
56
|
-
from edsl.
|
57
|
-
|
58
|
-
|
59
|
-
|
60
|
-
|
61
|
-
|
62
|
-
|
63
|
-
|
64
|
-
|
65
|
-
|
66
|
-
|
67
|
-
|
68
|
-
|
69
|
-
|
70
|
-
|
71
|
-
|
72
|
-
|
73
|
-
|
74
|
-
|
75
|
-
|
76
|
-
|
77
|
-
|
78
|
-
|
79
|
-
|
80
|
-
|
81
|
-
|
82
|
-
|
83
|
-
|
84
|
-
|
85
|
-
|
86
|
-
|
87
|
-
|
88
|
-
|
89
|
-
|
90
|
-
|
91
|
-
|
92
|
-
|
93
|
-
|
94
|
-
|
95
|
-
|
96
|
-
|
97
|
-
|
98
|
-
|
99
|
-
|
100
|
-
|
101
|
-
|
102
|
-
|
103
|
-
|
104
|
-
|
105
|
-
|
106
|
-
|
107
|
-
|
108
|
-
@
|
109
|
-
def
|
110
|
-
|
111
|
-
|
112
|
-
|
113
|
-
|
114
|
-
|
115
|
-
|
116
|
-
|
117
|
-
|
118
|
-
|
119
|
-
|
120
|
-
|
121
|
-
|
122
|
-
|
123
|
-
|
124
|
-
)
|
125
|
-
|
126
|
-
|
127
|
-
|
128
|
-
|
129
|
-
|
130
|
-
|
131
|
-
|
132
|
-
|
133
|
-
|
134
|
-
|
135
|
-
|
136
|
-
tpm
|
137
|
-
|
138
|
-
|
139
|
-
|
140
|
-
|
141
|
-
|
142
|
-
|
143
|
-
|
144
|
-
|
145
|
-
|
146
|
-
|
147
|
-
|
148
|
-
|
149
|
-
|
150
|
-
|
151
|
-
|
152
|
-
|
153
|
-
|
154
|
-
|
155
|
-
|
156
|
-
|
157
|
-
|
158
|
-
|
159
|
-
|
160
|
-
|
161
|
-
|
162
|
-
|
163
|
-
|
164
|
-
|
165
|
-
|
166
|
-
|
167
|
-
|
168
|
-
|
169
|
-
|
170
|
-
|
171
|
-
|
172
|
-
|
173
|
-
|
174
|
-
|
175
|
-
|
176
|
-
|
177
|
-
|
178
|
-
|
179
|
-
|
180
|
-
|
181
|
-
return self.
|
182
|
-
|
183
|
-
|
184
|
-
|
185
|
-
self
|
186
|
-
|
187
|
-
|
188
|
-
|
189
|
-
|
190
|
-
|
191
|
-
|
192
|
-
|
193
|
-
|
194
|
-
|
195
|
-
|
196
|
-
|
197
|
-
|
198
|
-
def
|
199
|
-
|
200
|
-
|
201
|
-
|
202
|
-
|
203
|
-
|
204
|
-
|
205
|
-
|
206
|
-
|
207
|
-
|
208
|
-
|
209
|
-
|
210
|
-
|
211
|
-
|
212
|
-
|
213
|
-
return self
|
214
|
-
|
215
|
-
|
216
|
-
|
217
|
-
|
218
|
-
|
219
|
-
|
220
|
-
|
221
|
-
|
222
|
-
|
223
|
-
|
224
|
-
|
225
|
-
|
226
|
-
|
227
|
-
|
228
|
-
|
229
|
-
|
230
|
-
|
231
|
-
|
232
|
-
|
233
|
-
|
234
|
-
|
235
|
-
|
236
|
-
|
237
|
-
|
238
|
-
|
239
|
-
|
240
|
-
|
241
|
-
|
242
|
-
|
243
|
-
|
244
|
-
|
245
|
-
>>>
|
246
|
-
|
247
|
-
|
248
|
-
"""
|
249
|
-
|
250
|
-
|
251
|
-
|
252
|
-
|
253
|
-
|
254
|
-
|
255
|
-
|
256
|
-
>>>
|
257
|
-
|
258
|
-
|
259
|
-
|
260
|
-
|
261
|
-
|
262
|
-
|
263
|
-
|
264
|
-
|
265
|
-
|
266
|
-
|
267
|
-
|
268
|
-
|
269
|
-
|
270
|
-
|
271
|
-
|
272
|
-
"""
|
273
|
-
|
274
|
-
|
275
|
-
|
276
|
-
|
277
|
-
|
278
|
-
|
279
|
-
|
280
|
-
|
281
|
-
|
282
|
-
|
283
|
-
|
284
|
-
|
285
|
-
|
286
|
-
|
287
|
-
|
288
|
-
|
289
|
-
|
290
|
-
|
291
|
-
|
292
|
-
|
293
|
-
|
294
|
-
|
295
|
-
|
296
|
-
|
297
|
-
|
298
|
-
|
299
|
-
|
300
|
-
|
301
|
-
|
302
|
-
|
303
|
-
|
304
|
-
|
305
|
-
|
306
|
-
|
307
|
-
|
308
|
-
|
309
|
-
|
310
|
-
|
311
|
-
|
312
|
-
|
313
|
-
|
314
|
-
|
315
|
-
|
316
|
-
|
317
|
-
>>>
|
318
|
-
|
319
|
-
|
320
|
-
|
321
|
-
|
322
|
-
"""
|
323
|
-
|
324
|
-
|
325
|
-
|
326
|
-
|
327
|
-
|
328
|
-
|
329
|
-
|
330
|
-
|
331
|
-
|
332
|
-
|
333
|
-
|
334
|
-
|
335
|
-
|
336
|
-
|
337
|
-
|
338
|
-
|
339
|
-
|
340
|
-
|
341
|
-
|
342
|
-
|
343
|
-
|
344
|
-
"""
|
345
|
-
|
346
|
-
|
347
|
-
|
348
|
-
|
349
|
-
|
350
|
-
|
351
|
-
|
352
|
-
|
353
|
-
|
354
|
-
|
355
|
-
|
356
|
-
|
357
|
-
|
358
|
-
|
359
|
-
|
360
|
-
|
361
|
-
|
362
|
-
|
363
|
-
|
364
|
-
|
365
|
-
|
366
|
-
|
367
|
-
|
368
|
-
|
369
|
-
|
370
|
-
|
371
|
-
|
372
|
-
|
373
|
-
|
374
|
-
|
375
|
-
|
376
|
-
|
377
|
-
|
378
|
-
|
379
|
-
|
380
|
-
|
381
|
-
|
382
|
-
|
383
|
-
|
384
|
-
|
385
|
-
|
386
|
-
|
387
|
-
|
388
|
-
|
389
|
-
|
390
|
-
|
391
|
-
|
392
|
-
|
393
|
-
|
394
|
-
|
395
|
-
|
396
|
-
|
397
|
-
|
398
|
-
|
399
|
-
|
400
|
-
|
401
|
-
|
402
|
-
|
403
|
-
|
404
|
-
|
405
|
-
|
406
|
-
|
407
|
-
|
408
|
-
|
409
|
-
|
410
|
-
|
411
|
-
|
412
|
-
|
413
|
-
|
414
|
-
|
415
|
-
|
416
|
-
|
417
|
-
|
418
|
-
|
419
|
-
|
420
|
-
|
421
|
-
|
422
|
-
|
423
|
-
|
424
|
-
|
425
|
-
|
426
|
-
|
427
|
-
|
428
|
-
|
429
|
-
|
430
|
-
|
431
|
-
|
432
|
-
|
433
|
-
|
434
|
-
|
435
|
-
|
436
|
-
|
437
|
-
|
438
|
-
|
439
|
-
|
440
|
-
|
441
|
-
|
442
|
-
|
443
|
-
|
444
|
-
|
445
|
-
|
446
|
-
|
447
|
-
|
448
|
-
|
449
|
-
"
|
450
|
-
|
451
|
-
|
452
|
-
|
453
|
-
|
454
|
-
|
455
|
-
|
456
|
-
|
457
|
-
|
458
|
-
|
459
|
-
|
460
|
-
)
|
461
|
-
|
462
|
-
|
463
|
-
|
464
|
-
|
465
|
-
|
466
|
-
|
467
|
-
|
468
|
-
|
469
|
-
|
470
|
-
|
471
|
-
|
472
|
-
|
473
|
-
|
474
|
-
|
475
|
-
|
476
|
-
|
477
|
-
|
478
|
-
|
479
|
-
|
480
|
-
|
481
|
-
|
482
|
-
|
483
|
-
|
484
|
-
|
485
|
-
|
486
|
-
|
487
|
-
|
488
|
-
|
489
|
-
|
490
|
-
|
491
|
-
):
|
492
|
-
"""
|
493
|
-
|
494
|
-
|
495
|
-
|
496
|
-
|
497
|
-
)
|
498
|
-
|
499
|
-
|
500
|
-
|
501
|
-
|
502
|
-
|
503
|
-
|
504
|
-
|
505
|
-
|
506
|
-
|
507
|
-
|
508
|
-
|
509
|
-
|
510
|
-
|
511
|
-
|
512
|
-
|
513
|
-
|
514
|
-
|
515
|
-
|
516
|
-
|
517
|
-
|
518
|
-
|
519
|
-
|
520
|
-
|
521
|
-
|
522
|
-
|
523
|
-
|
524
|
-
|
525
|
-
|
526
|
-
|
527
|
-
|
528
|
-
|
529
|
-
|
530
|
-
|
531
|
-
|
532
|
-
|
533
|
-
|
534
|
-
|
535
|
-
|
536
|
-
|
537
|
-
|
538
|
-
|
539
|
-
|
540
|
-
|
541
|
-
|
542
|
-
|
543
|
-
|
544
|
-
|
545
|
-
|
546
|
-
|
547
|
-
|
548
|
-
|
549
|
-
|
550
|
-
|
551
|
-
|
552
|
-
|
553
|
-
|
554
|
-
|
555
|
-
|
556
|
-
|
557
|
-
|
558
|
-
|
559
|
-
|
560
|
-
|
561
|
-
|
562
|
-
|
563
|
-
|
564
|
-
|
565
|
-
|
566
|
-
|
567
|
-
|
568
|
-
|
569
|
-
|
570
|
-
else:
|
571
|
-
|
572
|
-
|
573
|
-
|
574
|
-
|
575
|
-
|
576
|
-
|
577
|
-
|
578
|
-
|
579
|
-
|
580
|
-
|
581
|
-
|
582
|
-
|
583
|
-
|
584
|
-
|
585
|
-
|
586
|
-
|
587
|
-
|
588
|
-
|
589
|
-
|
590
|
-
|
591
|
-
|
592
|
-
|
593
|
-
|
594
|
-
|
595
|
-
|
596
|
-
|
597
|
-
|
598
|
-
|
599
|
-
|
600
|
-
|
601
|
-
|
602
|
-
|
603
|
-
|
604
|
-
|
605
|
-
|
606
|
-
|
607
|
-
|
608
|
-
|
609
|
-
|
610
|
-
|
611
|
-
|
612
|
-
|
613
|
-
|
614
|
-
|
615
|
-
|
616
|
-
|
617
|
-
|
618
|
-
|
619
|
-
|
620
|
-
|
621
|
-
|
622
|
-
|
623
|
-
|
624
|
-
|
625
|
-
|
626
|
-
|
627
|
-
by(m1, m2, m3) not by(m1).by(m2).by(m3)."""
|
628
|
-
)
|
629
|
-
return other_model or self
|
630
|
-
|
631
|
-
@classmethod
|
632
|
-
def example(
|
633
|
-
cls,
|
634
|
-
test_model: bool = False,
|
635
|
-
canned_response: str = "Hello world",
|
636
|
-
throw_exception: bool = False,
|
637
|
-
) -> LanguageModel:
|
638
|
-
"""Return a default instance of the class.
|
639
|
-
|
640
|
-
>>> from edsl.language_models import LanguageModel
|
641
|
-
>>> m = LanguageModel.example(test_model = True, canned_response = "WOWZA!")
|
642
|
-
>>> isinstance(m, LanguageModel)
|
643
|
-
True
|
644
|
-
>>> from edsl import QuestionFreeText
|
645
|
-
>>> q = QuestionFreeText(question_text = "What is your name?", question_name = 'example')
|
646
|
-
>>> q.by(m).run(cache = False, disable_remote_cache = True, disable_remote_inference = True).select('example').first()
|
647
|
-
'WOWZA!'
|
648
|
-
>>> m = LanguageModel.example(test_model = True, canned_response = "WOWZA!", throw_exception = True)
|
649
|
-
>>> r = q.by(m).run(cache = False, disable_remote_cache = True, disable_remote_inference = True, print_exceptions = True)
|
650
|
-
Exception report saved to ...
|
651
|
-
Also see: ...
|
652
|
-
"""
|
653
|
-
from edsl import Model
|
654
|
-
|
655
|
-
if test_model:
|
656
|
-
m = Model(
|
657
|
-
"test", canned_response=canned_response, throw_exception=throw_exception
|
658
|
-
)
|
659
|
-
return m
|
660
|
-
else:
|
661
|
-
return Model(skip_api_key_check=True)
|
662
|
-
|
663
|
-
|
664
|
-
if __name__ == "__main__":
|
665
|
-
"""Run the module's test suite."""
|
666
|
-
import doctest
|
667
|
-
|
668
|
-
doctest.testmod(optionflags=doctest.ELLIPSIS)
|
1
|
+
"""This module contains the LanguageModel class, which is an abstract base class for all language models.
|
2
|
+
|
3
|
+
Terminology:
|
4
|
+
|
5
|
+
raw_response: The JSON response from the model. This has all the model meta-data about the call.
|
6
|
+
|
7
|
+
edsl_augmented_response: The JSON response from model, but augmented with EDSL-specific information,
|
8
|
+
such as the cache key, token usage, etc.
|
9
|
+
|
10
|
+
generated_tokens: The actual tokens generated by the model. This is the output that is used by the user.
|
11
|
+
edsl_answer_dict: The parsed JSON response from the model either {'answer': ...} or {'answer': ..., 'comment': ...}
|
12
|
+
|
13
|
+
"""
|
14
|
+
|
15
|
+
from __future__ import annotations
|
16
|
+
import warnings
|
17
|
+
from functools import wraps
|
18
|
+
import asyncio
|
19
|
+
import json
|
20
|
+
import os
|
21
|
+
from typing import (
|
22
|
+
Coroutine,
|
23
|
+
Any,
|
24
|
+
Type,
|
25
|
+
Union,
|
26
|
+
List,
|
27
|
+
get_type_hints,
|
28
|
+
TypedDict,
|
29
|
+
Optional,
|
30
|
+
TYPE_CHECKING,
|
31
|
+
)
|
32
|
+
from abc import ABC, abstractmethod
|
33
|
+
|
34
|
+
from edsl.data_transfer_models import (
|
35
|
+
ModelResponse,
|
36
|
+
ModelInputs,
|
37
|
+
EDSLOutput,
|
38
|
+
AgentResponseDict,
|
39
|
+
)
|
40
|
+
|
41
|
+
if TYPE_CHECKING:
|
42
|
+
from edsl.data.Cache import Cache
|
43
|
+
from edsl.scenarios.FileStore import FileStore
|
44
|
+
from edsl.questions.QuestionBase import QuestionBase
|
45
|
+
from edsl.language_models.key_management.KeyLookup import KeyLookup
|
46
|
+
|
47
|
+
from edsl.enums import InferenceServiceType
|
48
|
+
|
49
|
+
from edsl.utilities.decorators import (
|
50
|
+
sync_wrapper,
|
51
|
+
jupyter_nb_handler,
|
52
|
+
)
|
53
|
+
from edsl.utilities.remove_edsl_version import remove_edsl_version
|
54
|
+
|
55
|
+
from edsl.Base import PersistenceMixin, RepresentationMixin
|
56
|
+
from edsl.language_models.RegisterLanguageModelsMeta import RegisterLanguageModelsMeta
|
57
|
+
|
58
|
+
from edsl.language_models.key_management.KeyLookupCollection import (
|
59
|
+
KeyLookupCollection,
|
60
|
+
)
|
61
|
+
|
62
|
+
from edsl.language_models.RawResponseHandler import RawResponseHandler
|
63
|
+
|
64
|
+
|
65
|
+
def handle_key_error(func):
|
66
|
+
"""Handle KeyError exceptions."""
|
67
|
+
|
68
|
+
@wraps(func)
|
69
|
+
def wrapper(*args, **kwargs):
|
70
|
+
try:
|
71
|
+
return func(*args, **kwargs)
|
72
|
+
assert True == False
|
73
|
+
except KeyError as e:
|
74
|
+
return f"""KeyError occurred: {e}. This is most likely because the model you are using
|
75
|
+
returned a JSON object we were not expecting."""
|
76
|
+
|
77
|
+
return wrapper
|
78
|
+
|
79
|
+
|
80
|
+
class classproperty:
|
81
|
+
def __init__(self, method):
|
82
|
+
self.method = method
|
83
|
+
|
84
|
+
def __get__(self, instance, cls):
|
85
|
+
return self.method(cls)
|
86
|
+
|
87
|
+
|
88
|
+
from edsl.Base import HashingMixin
|
89
|
+
|
90
|
+
|
91
|
+
class LanguageModel(
|
92
|
+
PersistenceMixin,
|
93
|
+
RepresentationMixin,
|
94
|
+
HashingMixin,
|
95
|
+
ABC,
|
96
|
+
metaclass=RegisterLanguageModelsMeta,
|
97
|
+
):
|
98
|
+
"""ABC for Language Models."""
|
99
|
+
|
100
|
+
_model_ = None
|
101
|
+
key_sequence = (
|
102
|
+
None # This should be something like ["choices", 0, "message", "content"]
|
103
|
+
)
|
104
|
+
|
105
|
+
DEFAULT_RPM = 100
|
106
|
+
DEFAULT_TPM = 1000
|
107
|
+
|
108
|
+
@classproperty
|
109
|
+
def response_handler(cls):
|
110
|
+
key_sequence = cls.key_sequence
|
111
|
+
usage_sequence = cls.usage_sequence if hasattr(cls, "usage_sequence") else None
|
112
|
+
return RawResponseHandler(key_sequence, usage_sequence)
|
113
|
+
|
114
|
+
def __init__(
|
115
|
+
self,
|
116
|
+
tpm: Optional[float] = None,
|
117
|
+
rpm: Optional[float] = None,
|
118
|
+
omit_system_prompt_if_empty_string: bool = True,
|
119
|
+
key_lookup: Optional["KeyLookup"] = None,
|
120
|
+
**kwargs,
|
121
|
+
):
|
122
|
+
"""Initialize the LanguageModel."""
|
123
|
+
self.model = getattr(self, "_model_", None)
|
124
|
+
default_parameters = getattr(self, "_parameters_", None)
|
125
|
+
parameters = self._overide_default_parameters(kwargs, default_parameters)
|
126
|
+
self.parameters = parameters
|
127
|
+
self.remote = False
|
128
|
+
self.omit_system_prompt_if_empty = omit_system_prompt_if_empty_string
|
129
|
+
|
130
|
+
self.key_lookup = self._set_key_lookup(key_lookup)
|
131
|
+
self.model_info = self.key_lookup.get(self._inference_service_)
|
132
|
+
|
133
|
+
if rpm is not None:
|
134
|
+
self._rpm = rpm
|
135
|
+
|
136
|
+
if tpm is not None:
|
137
|
+
self._tpm = tpm
|
138
|
+
|
139
|
+
for key, value in parameters.items():
|
140
|
+
setattr(self, key, value)
|
141
|
+
|
142
|
+
for key, value in kwargs.items():
|
143
|
+
if key not in parameters:
|
144
|
+
setattr(self, key, value)
|
145
|
+
|
146
|
+
if kwargs.get("skip_api_key_check", False):
|
147
|
+
# Skip the API key check. Sometimes this is useful for testing.
|
148
|
+
self._api_token = None
|
149
|
+
|
150
|
+
def _set_key_lookup(self, key_lookup: "KeyLookup") -> "KeyLookup":
|
151
|
+
"""Set the key lookup."""
|
152
|
+
if key_lookup is not None:
|
153
|
+
return key_lookup
|
154
|
+
else:
|
155
|
+
klc = KeyLookupCollection()
|
156
|
+
klc.add_key_lookup(fetch_order=("config", "env"))
|
157
|
+
return klc.get(("config", "env"))
|
158
|
+
|
159
|
+
def set_key_lookup(self, key_lookup: "KeyLookup") -> None:
|
160
|
+
"""Set the key lookup, later"""
|
161
|
+
if hasattr(self, "_api_token"):
|
162
|
+
del self._api_token
|
163
|
+
self.key_lookup = key_lookup
|
164
|
+
|
165
|
+
def ask_question(self, question: "QuestionBase") -> str:
|
166
|
+
"""Ask a question and return the response.
|
167
|
+
|
168
|
+
:param question: The question to ask.
|
169
|
+
"""
|
170
|
+
user_prompt = question.get_instructions().render(question.data).text
|
171
|
+
system_prompt = "You are a helpful agent pretending to be a human."
|
172
|
+
return self.execute_model_call(user_prompt, system_prompt)
|
173
|
+
|
174
|
+
@property
|
175
|
+
def rpm(self):
|
176
|
+
if not hasattr(self, "_rpm"):
|
177
|
+
if self.model_info is None:
|
178
|
+
self._rpm = self.DEFAULT_RPM
|
179
|
+
else:
|
180
|
+
self._rpm = self.model_info.rpm
|
181
|
+
return self._rpm
|
182
|
+
|
183
|
+
@property
|
184
|
+
def tpm(self):
|
185
|
+
if not hasattr(self, "_tpm"):
|
186
|
+
if self.model_info is None:
|
187
|
+
self._tpm = self.DEFAULT_TPM
|
188
|
+
else:
|
189
|
+
self._tpm = self.model_info.tpm
|
190
|
+
return self._tpm
|
191
|
+
|
192
|
+
# in case we want to override the default values
|
193
|
+
@tpm.setter
|
194
|
+
def tpm(self, value):
|
195
|
+
self._tpm = value
|
196
|
+
|
197
|
+
@rpm.setter
|
198
|
+
def rpm(self, value):
|
199
|
+
self._rpm = value
|
200
|
+
|
201
|
+
@property
|
202
|
+
def api_token(self) -> str:
|
203
|
+
if not hasattr(self, "_api_token"):
|
204
|
+
info = self.key_lookup.get(self._inference_service_, None)
|
205
|
+
if info is None:
|
206
|
+
raise ValueError(
|
207
|
+
f"No key found for service '{self._inference_service_}'"
|
208
|
+
)
|
209
|
+
self._api_token = info.api_token
|
210
|
+
return self._api_token
|
211
|
+
|
212
|
+
def __getitem__(self, key):
|
213
|
+
return getattr(self, key)
|
214
|
+
|
215
|
+
def hello(self, verbose=False):
|
216
|
+
"""Runs a simple test to check if the model is working."""
|
217
|
+
token = self.api_token
|
218
|
+
masked = token[: min(8, len(token))] + "..."
|
219
|
+
if verbose:
|
220
|
+
print(f"Current key is {masked}")
|
221
|
+
return self.execute_model_call(
|
222
|
+
user_prompt="Hello, model!", system_prompt="You are a helpful agent."
|
223
|
+
)
|
224
|
+
|
225
|
+
def has_valid_api_key(self) -> bool:
|
226
|
+
"""Check if the model has a valid API key.
|
227
|
+
|
228
|
+
>>> LanguageModel.example().has_valid_api_key() : # doctest: +SKIP
|
229
|
+
True
|
230
|
+
|
231
|
+
This method is used to check if the model has a valid API key.
|
232
|
+
"""
|
233
|
+
from edsl.enums import service_to_api_keyname
|
234
|
+
|
235
|
+
if self._model_ == "test":
|
236
|
+
return True
|
237
|
+
|
238
|
+
key_name = service_to_api_keyname.get(self._inference_service_, "NOT FOUND")
|
239
|
+
key_value = os.getenv(key_name)
|
240
|
+
return key_value is not None
|
241
|
+
|
242
|
+
def __hash__(self) -> str:
|
243
|
+
"""Allow the model to be used as a key in a dictionary.
|
244
|
+
|
245
|
+
>>> m = LanguageModel.example()
|
246
|
+
>>> hash(m)
|
247
|
+
1811901442659237949
|
248
|
+
"""
|
249
|
+
from edsl.utilities.utilities import dict_hash
|
250
|
+
|
251
|
+
return dict_hash(self.to_dict(add_edsl_version=False))
|
252
|
+
|
253
|
+
def __eq__(self, other) -> bool:
|
254
|
+
"""Check is two models are the same.
|
255
|
+
|
256
|
+
>>> m1 = LanguageModel.example()
|
257
|
+
>>> m2 = LanguageModel.example()
|
258
|
+
>>> m1 == m2
|
259
|
+
True
|
260
|
+
|
261
|
+
"""
|
262
|
+
return self.model == other.model and self.parameters == other.parameters
|
263
|
+
|
264
|
+
@staticmethod
|
265
|
+
def _overide_default_parameters(passed_parameter_dict, default_parameter_dict):
|
266
|
+
"""Return a dictionary of parameters, with passed parameters taking precedence over defaults.
|
267
|
+
|
268
|
+
>>> LanguageModel._overide_default_parameters(passed_parameter_dict={"temperature": 0.5}, default_parameter_dict={"temperature":0.9})
|
269
|
+
{'temperature': 0.5}
|
270
|
+
>>> LanguageModel._overide_default_parameters(passed_parameter_dict={"temperature": 0.5}, default_parameter_dict={"temperature":0.9, "max_tokens": 1000})
|
271
|
+
{'temperature': 0.5, 'max_tokens': 1000}
|
272
|
+
"""
|
273
|
+
# this is the case when data is loaded from a dict after serialization
|
274
|
+
if "parameters" in passed_parameter_dict:
|
275
|
+
passed_parameter_dict = passed_parameter_dict["parameters"]
|
276
|
+
return {
|
277
|
+
parameter_name: passed_parameter_dict.get(parameter_name, default_value)
|
278
|
+
for parameter_name, default_value in default_parameter_dict.items()
|
279
|
+
}
|
280
|
+
|
281
|
+
def __call__(self, user_prompt: str, system_prompt: str):
|
282
|
+
return self.execute_model_call(user_prompt, system_prompt)
|
283
|
+
|
284
|
+
@abstractmethod
|
285
|
+
async def async_execute_model_call(user_prompt: str, system_prompt: str):
|
286
|
+
"""Execute the model call and returns a coroutine."""
|
287
|
+
pass
|
288
|
+
|
289
|
+
async def remote_async_execute_model_call(
|
290
|
+
self, user_prompt: str, system_prompt: str
|
291
|
+
):
|
292
|
+
"""Execute the model call and returns the result as a coroutine, using Coop."""
|
293
|
+
from edsl.coop import Coop
|
294
|
+
|
295
|
+
client = Coop()
|
296
|
+
response_data = await client.remote_async_execute_model_call(
|
297
|
+
self.to_dict(), user_prompt, system_prompt
|
298
|
+
)
|
299
|
+
return response_data
|
300
|
+
|
301
|
+
@jupyter_nb_handler
|
302
|
+
def execute_model_call(self, *args, **kwargs) -> Coroutine:
|
303
|
+
"""Execute the model call and returns the result as a coroutine."""
|
304
|
+
|
305
|
+
async def main():
|
306
|
+
results = await asyncio.gather(
|
307
|
+
self.async_execute_model_call(*args, **kwargs)
|
308
|
+
)
|
309
|
+
return results[0] # Since there's only one task, return its result
|
310
|
+
|
311
|
+
return main()
|
312
|
+
|
313
|
+
@classmethod
|
314
|
+
def get_generated_token_string(cls, raw_response: dict[str, Any]) -> str:
|
315
|
+
"""Return the generated token string from the raw response.
|
316
|
+
|
317
|
+
>>> m = LanguageModel.example(test_model = True)
|
318
|
+
>>> raw_response = m.execute_model_call("Hello, model!", "You are a helpful agent.")
|
319
|
+
>>> m.get_generated_token_string(raw_response)
|
320
|
+
'Hello world'
|
321
|
+
|
322
|
+
"""
|
323
|
+
return cls.response_handler.get_generated_token_string(raw_response)
|
324
|
+
|
325
|
+
@classmethod
|
326
|
+
def get_usage_dict(cls, raw_response: dict[str, Any]) -> dict[str, Any]:
|
327
|
+
"""Return the usage dictionary from the raw response."""
|
328
|
+
return cls.response_handler.get_usage_dict(raw_response)
|
329
|
+
|
330
|
+
@classmethod
|
331
|
+
def parse_response(cls, raw_response: dict[str, Any]) -> EDSLOutput:
|
332
|
+
"""Parses the API response and returns the response text."""
|
333
|
+
return cls.response_handler.parse_response(raw_response)
|
334
|
+
|
335
|
+
async def _async_get_intended_model_call_outcome(
|
336
|
+
self,
|
337
|
+
user_prompt: str,
|
338
|
+
system_prompt: str,
|
339
|
+
cache: Cache,
|
340
|
+
iteration: int = 0,
|
341
|
+
files_list: Optional[List[FileStore]] = None,
|
342
|
+
invigilator=None,
|
343
|
+
) -> ModelResponse:
|
344
|
+
"""Handle caching of responses.
|
345
|
+
|
346
|
+
:param user_prompt: The user's prompt.
|
347
|
+
:param system_prompt: The system's prompt.
|
348
|
+
:param iteration: The iteration number.
|
349
|
+
:param cache: The cache to use.
|
350
|
+
:param files_list: The list of files to use.
|
351
|
+
:param invigilator: The invigilator to use.
|
352
|
+
|
353
|
+
If the cache isn't being used, it just returns a 'fresh' call to the LLM.
|
354
|
+
But if cache is being used, it first checks the database to see if the response is already there.
|
355
|
+
If it is, it returns the cached response, but again appends some tracking information.
|
356
|
+
If it isn't, it calls the LLM, saves the response to the database, and returns the response with tracking information.
|
357
|
+
|
358
|
+
If self.use_cache is True, then attempts to retrieve the response from the database;
|
359
|
+
if not in the DB, calls the LLM and writes the response to the DB.
|
360
|
+
|
361
|
+
>>> from edsl import Cache
|
362
|
+
>>> m = LanguageModel.example(test_model = True)
|
363
|
+
>>> m._get_intended_model_call_outcome(user_prompt = "Hello", system_prompt = "hello", cache = Cache())
|
364
|
+
ModelResponse(...)"""
|
365
|
+
|
366
|
+
if files_list:
|
367
|
+
files_hash = "+".join([str(hash(file)) for file in files_list])
|
368
|
+
user_prompt_with_hashes = user_prompt + f" {files_hash}"
|
369
|
+
else:
|
370
|
+
user_prompt_with_hashes = user_prompt
|
371
|
+
|
372
|
+
cache_call_params = {
|
373
|
+
"model": str(self.model),
|
374
|
+
"parameters": self.parameters,
|
375
|
+
"system_prompt": system_prompt,
|
376
|
+
"user_prompt": user_prompt_with_hashes,
|
377
|
+
"iteration": iteration,
|
378
|
+
}
|
379
|
+
cached_response, cache_key = cache.fetch(**cache_call_params)
|
380
|
+
|
381
|
+
if cache_used := cached_response is not None:
|
382
|
+
response = json.loads(cached_response)
|
383
|
+
else:
|
384
|
+
f = (
|
385
|
+
self.remote_async_execute_model_call
|
386
|
+
if hasattr(self, "remote") and self.remote
|
387
|
+
else self.async_execute_model_call
|
388
|
+
)
|
389
|
+
params = {
|
390
|
+
"user_prompt": user_prompt,
|
391
|
+
"system_prompt": system_prompt,
|
392
|
+
"files_list": files_list,
|
393
|
+
}
|
394
|
+
from edsl.config import CONFIG
|
395
|
+
|
396
|
+
TIMEOUT = float(CONFIG.get("EDSL_API_TIMEOUT"))
|
397
|
+
|
398
|
+
response = await asyncio.wait_for(f(**params), timeout=TIMEOUT)
|
399
|
+
new_cache_key = cache.store(
|
400
|
+
**cache_call_params, response=response
|
401
|
+
) # store the response in the cache
|
402
|
+
assert new_cache_key == cache_key # should be the same
|
403
|
+
|
404
|
+
cost = self.cost(response)
|
405
|
+
return ModelResponse(
|
406
|
+
response=response,
|
407
|
+
cache_used=cache_used,
|
408
|
+
cache_key=cache_key,
|
409
|
+
cached_response=cached_response,
|
410
|
+
cost=cost,
|
411
|
+
)
|
412
|
+
|
413
|
+
_get_intended_model_call_outcome = sync_wrapper(
|
414
|
+
_async_get_intended_model_call_outcome
|
415
|
+
)
|
416
|
+
|
417
|
+
def simple_ask(
|
418
|
+
self,
|
419
|
+
question: QuestionBase,
|
420
|
+
system_prompt="You are a helpful agent pretending to be a human.",
|
421
|
+
top_logprobs=2,
|
422
|
+
):
|
423
|
+
"""Ask a question and return the response."""
|
424
|
+
self.logprobs = True
|
425
|
+
self.top_logprobs = top_logprobs
|
426
|
+
return self.execute_model_call(
|
427
|
+
user_prompt=question.human_readable(), system_prompt=system_prompt
|
428
|
+
)
|
429
|
+
|
430
|
+
async def async_get_response(
|
431
|
+
self,
|
432
|
+
user_prompt: str,
|
433
|
+
system_prompt: str,
|
434
|
+
cache: Cache,
|
435
|
+
iteration: int = 1,
|
436
|
+
files_list: Optional[List[FileStore]] = None,
|
437
|
+
**kwargs,
|
438
|
+
) -> dict:
|
439
|
+
"""Get response, parse, and return as string.
|
440
|
+
|
441
|
+
:param user_prompt: The user's prompt.
|
442
|
+
:param system_prompt: The system's prompt.
|
443
|
+
:param cache: The cache to use.
|
444
|
+
:param iteration: The iteration number.
|
445
|
+
:param files_list: The list of files to use.
|
446
|
+
|
447
|
+
"""
|
448
|
+
params = {
|
449
|
+
"user_prompt": user_prompt,
|
450
|
+
"system_prompt": system_prompt,
|
451
|
+
"iteration": iteration,
|
452
|
+
"cache": cache,
|
453
|
+
"files_list": files_list,
|
454
|
+
}
|
455
|
+
if "invigilator" in kwargs:
|
456
|
+
params.update({"invigilator": kwargs["invigilator"]})
|
457
|
+
|
458
|
+
model_inputs = ModelInputs(user_prompt=user_prompt, system_prompt=system_prompt)
|
459
|
+
model_outputs: ModelResponse = (
|
460
|
+
await self._async_get_intended_model_call_outcome(**params)
|
461
|
+
)
|
462
|
+
edsl_dict: EDSLOutput = self.parse_response(model_outputs.response)
|
463
|
+
|
464
|
+
agent_response_dict = AgentResponseDict(
|
465
|
+
model_inputs=model_inputs,
|
466
|
+
model_outputs=model_outputs,
|
467
|
+
edsl_dict=edsl_dict,
|
468
|
+
)
|
469
|
+
return agent_response_dict
|
470
|
+
|
471
|
+
get_response = sync_wrapper(async_get_response)
|
472
|
+
|
473
|
+
def cost(self, raw_response: dict[str, Any]) -> Union[float, str]:
|
474
|
+
"""Return the dollar cost of a raw response.
|
475
|
+
|
476
|
+
:param raw_response: The raw response from the model.
|
477
|
+
"""
|
478
|
+
|
479
|
+
usage = self.get_usage_dict(raw_response)
|
480
|
+
from edsl.language_models.PriceManager import PriceManager
|
481
|
+
|
482
|
+
price_manger = PriceManager()
|
483
|
+
return price_manger.calculate_cost(
|
484
|
+
inference_service=self._inference_service_,
|
485
|
+
model=self.model,
|
486
|
+
usage=usage,
|
487
|
+
input_token_name=self.input_token_name,
|
488
|
+
output_token_name=self.output_token_name,
|
489
|
+
)
|
490
|
+
|
491
|
+
def to_dict(self, add_edsl_version: bool = True) -> dict[str, Any]:
|
492
|
+
"""Convert instance to a dictionary
|
493
|
+
|
494
|
+
:param add_edsl_version: Whether to add the EDSL version to the dictionary.
|
495
|
+
|
496
|
+
>>> m = LanguageModel.example()
|
497
|
+
>>> m.to_dict()
|
498
|
+
{'model': '...', 'parameters': {'temperature': ..., 'max_tokens': ..., 'top_p': ..., 'frequency_penalty': ..., 'presence_penalty': ..., 'logprobs': False, 'top_logprobs': ...}, 'edsl_version': '...', 'edsl_class_name': 'LanguageModel'}
|
499
|
+
"""
|
500
|
+
d = {
|
501
|
+
"model": self.model,
|
502
|
+
"parameters": self.parameters,
|
503
|
+
}
|
504
|
+
if add_edsl_version:
|
505
|
+
from edsl import __version__
|
506
|
+
|
507
|
+
d["edsl_version"] = __version__
|
508
|
+
d["edsl_class_name"] = self.__class__.__name__
|
509
|
+
return d
|
510
|
+
|
511
|
+
@classmethod
|
512
|
+
@remove_edsl_version
|
513
|
+
def from_dict(cls, data: dict) -> Type[LanguageModel]:
|
514
|
+
"""Convert dictionary to a LanguageModel child instance."""
|
515
|
+
from edsl.language_models.model import get_model_class
|
516
|
+
|
517
|
+
model_class = get_model_class(data["model"])
|
518
|
+
return model_class(**data)
|
519
|
+
|
520
|
+
def __repr__(self) -> str:
|
521
|
+
"""Return a representation of the object."""
|
522
|
+
param_string = ", ".join(
|
523
|
+
f"{key} = {value}" for key, value in self.parameters.items()
|
524
|
+
)
|
525
|
+
return (
|
526
|
+
f"Model(model_name = '{self.model}'"
|
527
|
+
+ (f", {param_string}" if param_string else "")
|
528
|
+
+ ")"
|
529
|
+
)
|
530
|
+
|
531
|
+
def __add__(self, other_model: Type[LanguageModel]) -> Type[LanguageModel]:
|
532
|
+
"""Combine two models into a single model (other_model takes precedence over self)."""
|
533
|
+
import warnings
|
534
|
+
|
535
|
+
warnings.warn(
|
536
|
+
f"""Warning: one model is replacing another. If you want to run both models, use a single `by` e.g.,
|
537
|
+
by(m1, m2, m3) not by(m1).by(m2).by(m3)."""
|
538
|
+
)
|
539
|
+
return other_model or self
|
540
|
+
|
541
|
+
@classmethod
|
542
|
+
def example(
|
543
|
+
cls,
|
544
|
+
test_model: bool = False,
|
545
|
+
canned_response: str = "Hello world",
|
546
|
+
throw_exception: bool = False,
|
547
|
+
) -> LanguageModel:
|
548
|
+
"""Return a default instance of the class.
|
549
|
+
|
550
|
+
>>> from edsl.language_models import LanguageModel
|
551
|
+
>>> m = LanguageModel.example(test_model = True, canned_response = "WOWZA!")
|
552
|
+
>>> isinstance(m, LanguageModel)
|
553
|
+
True
|
554
|
+
>>> from edsl import QuestionFreeText
|
555
|
+
>>> q = QuestionFreeText(question_text = "What is your name?", question_name = 'example')
|
556
|
+
>>> q.by(m).run(cache = False, disable_remote_cache = True, disable_remote_inference = True).select('example').first()
|
557
|
+
'WOWZA!'
|
558
|
+
>>> m = LanguageModel.example(test_model = True, canned_response = "WOWZA!", throw_exception = True)
|
559
|
+
>>> r = q.by(m).run(cache = False, disable_remote_cache = True, disable_remote_inference = True, print_exceptions = True)
|
560
|
+
Exception report saved to ...
|
561
|
+
Also see: ...
|
562
|
+
"""
|
563
|
+
from edsl.language_models.model import Model
|
564
|
+
|
565
|
+
if test_model:
|
566
|
+
m = Model(
|
567
|
+
"test", canned_response=canned_response, throw_exception=throw_exception
|
568
|
+
)
|
569
|
+
return m
|
570
|
+
else:
|
571
|
+
return Model(skip_api_key_check=True)
|
572
|
+
|
573
|
+
def from_cache(self, cache: "Cache") -> LanguageModel:
|
574
|
+
|
575
|
+
from copy import deepcopy
|
576
|
+
from types import MethodType
|
577
|
+
from edsl import Cache
|
578
|
+
|
579
|
+
new_instance = deepcopy(self)
|
580
|
+
print("Cache entries", len(cache))
|
581
|
+
new_instance.cache = Cache(
|
582
|
+
data={k: v for k, v in cache.items() if v.model == self.model}
|
583
|
+
)
|
584
|
+
print("Cache entries with same model", len(new_instance.cache))
|
585
|
+
|
586
|
+
new_instance.user_prompts = [
|
587
|
+
ce.user_prompt for ce in new_instance.cache.values()
|
588
|
+
]
|
589
|
+
new_instance.system_prompts = [
|
590
|
+
ce.system_prompt for ce in new_instance.cache.values()
|
591
|
+
]
|
592
|
+
|
593
|
+
async def async_execute_model_call(self, user_prompt: str, system_prompt: str):
|
594
|
+
cache_call_params = {
|
595
|
+
"model": str(self.model),
|
596
|
+
"parameters": self.parameters,
|
597
|
+
"system_prompt": system_prompt,
|
598
|
+
"user_prompt": user_prompt,
|
599
|
+
"iteration": 1,
|
600
|
+
}
|
601
|
+
cached_response, cache_key = cache.fetch(**cache_call_params)
|
602
|
+
response = json.loads(cached_response)
|
603
|
+
cost = 0
|
604
|
+
return ModelResponse(
|
605
|
+
response=response,
|
606
|
+
cache_used=True,
|
607
|
+
cache_key=cache_key,
|
608
|
+
cached_response=cached_response,
|
609
|
+
cost=cost,
|
610
|
+
)
|
611
|
+
|
612
|
+
# Bind the new method to the copied instance
|
613
|
+
setattr(
|
614
|
+
new_instance,
|
615
|
+
"async_execute_model_call",
|
616
|
+
MethodType(async_execute_model_call, new_instance),
|
617
|
+
)
|
618
|
+
|
619
|
+
return new_instance
|
620
|
+
|
621
|
+
|
622
|
+
if __name__ == "__main__":
|
623
|
+
"""Run the module's test suite."""
|
624
|
+
import doctest
|
625
|
+
|
626
|
+
doctest.testmod(optionflags=doctest.ELLIPSIS)
|