edsl 0.1.36.dev2__py3-none-any.whl → 0.1.36.dev6__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- edsl/Base.py +303 -298
- edsl/BaseDiff.py +260 -260
- edsl/TemplateLoader.py +24 -24
- edsl/__init__.py +47 -47
- edsl/__version__.py +1 -1
- edsl/agents/Agent.py +804 -800
- edsl/agents/AgentList.py +337 -337
- edsl/agents/Invigilator.py +222 -222
- edsl/agents/InvigilatorBase.py +294 -294
- edsl/agents/PromptConstructor.py +312 -311
- edsl/agents/__init__.py +3 -3
- edsl/agents/descriptors.py +86 -86
- edsl/agents/prompt_helpers.py +129 -129
- edsl/auto/AutoStudy.py +117 -117
- edsl/auto/StageBase.py +230 -230
- edsl/auto/StageGenerateSurvey.py +178 -178
- edsl/auto/StageLabelQuestions.py +125 -125
- edsl/auto/StagePersona.py +61 -61
- edsl/auto/StagePersonaDimensionValueRanges.py +88 -88
- edsl/auto/StagePersonaDimensionValues.py +74 -74
- edsl/auto/StagePersonaDimensions.py +69 -69
- edsl/auto/StageQuestions.py +73 -73
- edsl/auto/SurveyCreatorPipeline.py +21 -21
- edsl/auto/utilities.py +224 -224
- edsl/base/Base.py +289 -289
- edsl/config.py +149 -149
- edsl/conjure/AgentConstructionMixin.py +152 -152
- edsl/conjure/Conjure.py +62 -62
- edsl/conjure/InputData.py +659 -659
- edsl/conjure/InputDataCSV.py +48 -48
- edsl/conjure/InputDataMixinQuestionStats.py +182 -182
- edsl/conjure/InputDataPyRead.py +91 -91
- edsl/conjure/InputDataSPSS.py +8 -8
- edsl/conjure/InputDataStata.py +8 -8
- edsl/conjure/QuestionOptionMixin.py +76 -76
- edsl/conjure/QuestionTypeMixin.py +23 -23
- edsl/conjure/RawQuestion.py +65 -65
- edsl/conjure/SurveyResponses.py +7 -7
- edsl/conjure/__init__.py +9 -9
- edsl/conjure/naming_utilities.py +263 -263
- edsl/conjure/utilities.py +201 -201
- edsl/conversation/Conversation.py +238 -238
- edsl/conversation/car_buying.py +58 -58
- edsl/conversation/mug_negotiation.py +81 -81
- edsl/conversation/next_speaker_utilities.py +93 -93
- edsl/coop/PriceFetcher.py +54 -58
- edsl/coop/__init__.py +2 -2
- edsl/coop/coop.py +849 -815
- edsl/coop/utils.py +131 -131
- edsl/data/Cache.py +527 -527
- edsl/data/CacheEntry.py +228 -228
- edsl/data/CacheHandler.py +149 -149
- edsl/data/RemoteCacheSync.py +84 -0
- edsl/data/SQLiteDict.py +292 -292
- edsl/data/__init__.py +4 -4
- edsl/data/orm.py +10 -10
- edsl/data_transfer_models.py +73 -73
- edsl/enums.py +173 -173
- edsl/exceptions/__init__.py +50 -50
- edsl/exceptions/agents.py +40 -40
- edsl/exceptions/configuration.py +16 -16
- edsl/exceptions/coop.py +10 -2
- edsl/exceptions/data.py +14 -14
- edsl/exceptions/general.py +34 -34
- edsl/exceptions/jobs.py +33 -33
- edsl/exceptions/language_models.py +63 -63
- edsl/exceptions/prompts.py +15 -15
- edsl/exceptions/questions.py +91 -91
- edsl/exceptions/results.py +26 -26
- edsl/exceptions/surveys.py +34 -34
- edsl/inference_services/AnthropicService.py +87 -87
- edsl/inference_services/AwsBedrock.py +115 -115
- edsl/inference_services/AzureAI.py +217 -217
- edsl/inference_services/DeepInfraService.py +18 -18
- edsl/inference_services/GoogleService.py +156 -156
- edsl/inference_services/GroqService.py +20 -20
- edsl/inference_services/InferenceServiceABC.py +147 -119
- edsl/inference_services/InferenceServicesCollection.py +72 -68
- edsl/inference_services/MistralAIService.py +123 -123
- edsl/inference_services/OllamaService.py +18 -18
- edsl/inference_services/OpenAIService.py +224 -224
- edsl/inference_services/TestService.py +89 -89
- edsl/inference_services/TogetherAIService.py +170 -170
- edsl/inference_services/models_available_cache.py +118 -94
- edsl/inference_services/rate_limits_cache.py +25 -25
- edsl/inference_services/registry.py +39 -39
- edsl/inference_services/write_available.py +10 -10
- edsl/jobs/Answers.py +56 -56
- edsl/jobs/Jobs.py +1112 -1089
- edsl/jobs/__init__.py +1 -1
- edsl/jobs/buckets/BucketCollection.py +63 -63
- edsl/jobs/buckets/ModelBuckets.py +65 -65
- edsl/jobs/buckets/TokenBucket.py +248 -248
- edsl/jobs/interviews/Interview.py +651 -633
- edsl/jobs/interviews/InterviewExceptionCollection.py +99 -90
- edsl/jobs/interviews/InterviewExceptionEntry.py +182 -164
- edsl/jobs/interviews/InterviewStatistic.py +63 -63
- edsl/jobs/interviews/InterviewStatisticsCollection.py +25 -25
- edsl/jobs/interviews/InterviewStatusDictionary.py +78 -78
- edsl/jobs/interviews/InterviewStatusLog.py +92 -92
- edsl/jobs/interviews/ReportErrors.py +66 -66
- edsl/jobs/interviews/interview_status_enum.py +9 -9
- edsl/jobs/runners/JobsRunnerAsyncio.py +337 -343
- edsl/jobs/runners/JobsRunnerStatus.py +332 -332
- edsl/jobs/tasks/QuestionTaskCreator.py +242 -242
- edsl/jobs/tasks/TaskCreators.py +64 -64
- edsl/jobs/tasks/TaskHistory.py +441 -425
- edsl/jobs/tasks/TaskStatusLog.py +23 -23
- edsl/jobs/tasks/task_status_enum.py +163 -163
- edsl/jobs/tokens/InterviewTokenUsage.py +27 -27
- edsl/jobs/tokens/TokenUsage.py +34 -34
- edsl/language_models/LanguageModel.py +718 -718
- edsl/language_models/ModelList.py +102 -102
- edsl/language_models/RegisterLanguageModelsMeta.py +184 -184
- edsl/language_models/__init__.py +2 -2
- edsl/language_models/fake_openai_call.py +15 -15
- edsl/language_models/fake_openai_service.py +61 -61
- edsl/language_models/registry.py +137 -137
- edsl/language_models/repair.py +156 -156
- edsl/language_models/unused/ReplicateBase.py +83 -83
- edsl/language_models/utilities.py +64 -64
- edsl/notebooks/Notebook.py +259 -259
- edsl/notebooks/__init__.py +1 -1
- edsl/prompts/Prompt.py +358 -358
- edsl/prompts/__init__.py +2 -2
- edsl/questions/AnswerValidatorMixin.py +289 -289
- edsl/questions/QuestionBase.py +616 -616
- edsl/questions/QuestionBaseGenMixin.py +161 -161
- edsl/questions/QuestionBasePromptsMixin.py +266 -266
- edsl/questions/QuestionBudget.py +227 -227
- edsl/questions/QuestionCheckBox.py +359 -359
- edsl/questions/QuestionExtract.py +183 -183
- edsl/questions/QuestionFreeText.py +113 -113
- edsl/questions/QuestionFunctional.py +159 -155
- edsl/questions/QuestionList.py +231 -231
- edsl/questions/QuestionMultipleChoice.py +286 -286
- edsl/questions/QuestionNumerical.py +153 -153
- edsl/questions/QuestionRank.py +324 -324
- edsl/questions/Quick.py +41 -41
- edsl/questions/RegisterQuestionsMeta.py +71 -71
- edsl/questions/ResponseValidatorABC.py +174 -174
- edsl/questions/SimpleAskMixin.py +73 -73
- edsl/questions/__init__.py +26 -26
- edsl/questions/compose_questions.py +98 -98
- edsl/questions/decorators.py +21 -21
- edsl/questions/derived/QuestionLikertFive.py +76 -76
- edsl/questions/derived/QuestionLinearScale.py +87 -87
- edsl/questions/derived/QuestionTopK.py +91 -91
- edsl/questions/derived/QuestionYesNo.py +82 -82
- edsl/questions/descriptors.py +418 -418
- edsl/questions/prompt_templates/question_budget.jinja +13 -13
- edsl/questions/prompt_templates/question_checkbox.jinja +32 -32
- edsl/questions/prompt_templates/question_extract.jinja +11 -11
- edsl/questions/prompt_templates/question_free_text.jinja +3 -3
- edsl/questions/prompt_templates/question_linear_scale.jinja +11 -11
- edsl/questions/prompt_templates/question_list.jinja +17 -17
- edsl/questions/prompt_templates/question_multiple_choice.jinja +33 -33
- edsl/questions/prompt_templates/question_numerical.jinja +36 -36
- edsl/questions/question_registry.py +147 -147
- edsl/questions/settings.py +12 -12
- edsl/questions/templates/budget/answering_instructions.jinja +7 -7
- edsl/questions/templates/budget/question_presentation.jinja +7 -7
- edsl/questions/templates/checkbox/answering_instructions.jinja +10 -10
- edsl/questions/templates/checkbox/question_presentation.jinja +22 -22
- edsl/questions/templates/extract/answering_instructions.jinja +7 -7
- edsl/questions/templates/likert_five/answering_instructions.jinja +10 -10
- edsl/questions/templates/likert_five/question_presentation.jinja +11 -11
- edsl/questions/templates/linear_scale/answering_instructions.jinja +5 -5
- edsl/questions/templates/linear_scale/question_presentation.jinja +5 -5
- edsl/questions/templates/list/answering_instructions.jinja +3 -3
- edsl/questions/templates/list/question_presentation.jinja +5 -5
- edsl/questions/templates/multiple_choice/answering_instructions.jinja +9 -9
- edsl/questions/templates/multiple_choice/question_presentation.jinja +11 -11
- edsl/questions/templates/numerical/answering_instructions.jinja +6 -6
- edsl/questions/templates/numerical/question_presentation.jinja +6 -6
- edsl/questions/templates/rank/answering_instructions.jinja +11 -11
- edsl/questions/templates/rank/question_presentation.jinja +15 -15
- edsl/questions/templates/top_k/answering_instructions.jinja +8 -8
- edsl/questions/templates/top_k/question_presentation.jinja +22 -22
- edsl/questions/templates/yes_no/answering_instructions.jinja +6 -6
- edsl/questions/templates/yes_no/question_presentation.jinja +11 -11
- edsl/results/Dataset.py +293 -281
- edsl/results/DatasetExportMixin.py +693 -693
- edsl/results/DatasetTree.py +145 -145
- edsl/results/Result.py +433 -431
- edsl/results/Results.py +1158 -1146
- edsl/results/ResultsDBMixin.py +238 -238
- edsl/results/ResultsExportMixin.py +43 -43
- edsl/results/ResultsFetchMixin.py +33 -33
- edsl/results/ResultsGGMixin.py +121 -121
- edsl/results/ResultsToolsMixin.py +98 -98
- edsl/results/Selector.py +118 -118
- edsl/results/__init__.py +2 -2
- edsl/results/tree_explore.py +115 -115
- edsl/scenarios/FileStore.py +443 -443
- edsl/scenarios/Scenario.py +507 -496
- edsl/scenarios/ScenarioHtmlMixin.py +59 -59
- edsl/scenarios/ScenarioList.py +1101 -1101
- edsl/scenarios/ScenarioListExportMixin.py +52 -52
- edsl/scenarios/ScenarioListPdfMixin.py +261 -261
- edsl/scenarios/__init__.py +2 -2
- edsl/shared.py +1 -1
- edsl/study/ObjectEntry.py +173 -173
- edsl/study/ProofOfWork.py +113 -113
- edsl/study/SnapShot.py +80 -80
- edsl/study/Study.py +528 -528
- edsl/study/__init__.py +4 -4
- edsl/surveys/DAG.py +148 -148
- edsl/surveys/Memory.py +31 -31
- edsl/surveys/MemoryPlan.py +244 -244
- edsl/surveys/Rule.py +324 -324
- edsl/surveys/RuleCollection.py +387 -387
- edsl/surveys/Survey.py +1772 -1769
- edsl/surveys/SurveyCSS.py +261 -261
- edsl/surveys/SurveyExportMixin.py +259 -259
- edsl/surveys/SurveyFlowVisualizationMixin.py +121 -121
- edsl/surveys/SurveyQualtricsImport.py +284 -284
- edsl/surveys/__init__.py +3 -3
- edsl/surveys/base.py +53 -53
- edsl/surveys/descriptors.py +56 -56
- edsl/surveys/instructions/ChangeInstruction.py +47 -47
- edsl/surveys/instructions/Instruction.py +51 -34
- edsl/surveys/instructions/InstructionCollection.py +77 -77
- edsl/templates/error_reporting/base.html +23 -23
- edsl/templates/error_reporting/exceptions_by_model.html +34 -34
- edsl/templates/error_reporting/exceptions_by_question_name.html +16 -16
- edsl/templates/error_reporting/exceptions_by_type.html +16 -16
- edsl/templates/error_reporting/interview_details.html +115 -115
- edsl/templates/error_reporting/interviews.html +9 -9
- edsl/templates/error_reporting/overview.html +4 -4
- edsl/templates/error_reporting/performance_plot.html +1 -1
- edsl/templates/error_reporting/report.css +73 -73
- edsl/templates/error_reporting/report.html +117 -117
- edsl/templates/error_reporting/report.js +25 -25
- edsl/tools/__init__.py +1 -1
- edsl/tools/clusters.py +192 -192
- edsl/tools/embeddings.py +27 -27
- edsl/tools/embeddings_plotting.py +118 -118
- edsl/tools/plotting.py +112 -112
- edsl/tools/summarize.py +18 -18
- edsl/utilities/SystemInfo.py +28 -28
- edsl/utilities/__init__.py +22 -22
- edsl/utilities/ast_utilities.py +25 -25
- edsl/utilities/data/Registry.py +6 -6
- edsl/utilities/data/__init__.py +1 -1
- edsl/utilities/data/scooter_results.json +1 -1
- edsl/utilities/decorators.py +77 -77
- edsl/utilities/gcp_bucket/cloud_storage.py +96 -96
- edsl/utilities/interface.py +627 -627
- edsl/utilities/repair_functions.py +28 -28
- edsl/utilities/restricted_python.py +70 -70
- edsl/utilities/utilities.py +391 -391
- {edsl-0.1.36.dev2.dist-info → edsl-0.1.36.dev6.dist-info}/LICENSE +21 -21
- {edsl-0.1.36.dev2.dist-info → edsl-0.1.36.dev6.dist-info}/METADATA +1 -1
- edsl-0.1.36.dev6.dist-info/RECORD +279 -0
- edsl-0.1.36.dev2.dist-info/RECORD +0 -278
- {edsl-0.1.36.dev2.dist-info → edsl-0.1.36.dev6.dist-info}/WHEEL +0 -0
edsl/agents/Invigilator.py
CHANGED
@@ -1,222 +1,222 @@
|
|
1
|
-
"""Module for creating Invigilators, which are objects to administer a question to an Agent."""
|
2
|
-
|
3
|
-
from typing import Dict, Any, Optional
|
4
|
-
|
5
|
-
from edsl.prompts.Prompt import Prompt
|
6
|
-
from edsl.utilities.decorators import sync_wrapper, jupyter_nb_handler
|
7
|
-
|
8
|
-
# from edsl.prompts.registry import get_classes as prompt_lookup
|
9
|
-
from edsl.exceptions.questions import QuestionAnswerValidationError
|
10
|
-
from edsl.agents.InvigilatorBase import InvigilatorBase
|
11
|
-
from edsl.data_transfer_models import AgentResponseDict, EDSLResultObjectInput
|
12
|
-
from edsl.agents.PromptConstructor import PromptConstructor
|
13
|
-
|
14
|
-
|
15
|
-
class NotApplicable(str):
|
16
|
-
def __new__(cls):
|
17
|
-
instance = super().__new__(cls, "Not Applicable")
|
18
|
-
instance.literal = "Not Applicable"
|
19
|
-
return instance
|
20
|
-
|
21
|
-
|
22
|
-
class InvigilatorAI(InvigilatorBase):
|
23
|
-
"""An invigilator that uses an AI model to answer questions."""
|
24
|
-
|
25
|
-
def get_prompts(self) -> Dict[str, Prompt]:
|
26
|
-
"""Return the prompts used."""
|
27
|
-
return self.prompt_constructor.get_prompts()
|
28
|
-
|
29
|
-
async def async_answer_question(self) -> AgentResponseDict:
|
30
|
-
"""Answer a question using the AI model.
|
31
|
-
|
32
|
-
>>> i = InvigilatorAI.example()
|
33
|
-
>>> i.answer_question()
|
34
|
-
{'message': [{'text': 'SPAM!'}], 'usage': {'prompt_tokens': 1, 'completion_tokens': 1}}
|
35
|
-
"""
|
36
|
-
prompts = self.get_prompts()
|
37
|
-
params = {
|
38
|
-
"user_prompt": prompts["user_prompt"].text,
|
39
|
-
"system_prompt": prompts["system_prompt"].text,
|
40
|
-
}
|
41
|
-
if "encoded_image" in prompts:
|
42
|
-
params["encoded_image"] = prompts["encoded_image"]
|
43
|
-
if "files_list" in prompts:
|
44
|
-
params["files_list"] = prompts["files_list"]
|
45
|
-
|
46
|
-
params.update({"iteration": self.iteration, "cache": self.cache})
|
47
|
-
|
48
|
-
agent_response_dict: AgentResponseDict = await self.model.async_get_response(
|
49
|
-
**params
|
50
|
-
)
|
51
|
-
# store to self in case validation failure
|
52
|
-
self.raw_model_response = agent_response_dict.model_outputs.response
|
53
|
-
self.generated_tokens = agent_response_dict.edsl_dict.generated_tokens
|
54
|
-
|
55
|
-
return self.extract_edsl_result_entry_and_validate(agent_response_dict)
|
56
|
-
|
57
|
-
def _remove_from_cache(self, cache_key) -> None:
|
58
|
-
"""Remove an entry from the cache."""
|
59
|
-
if cache_key:
|
60
|
-
del self.cache.data[cache_key]
|
61
|
-
|
62
|
-
def determine_answer(self, raw_answer: str) -> Any:
|
63
|
-
question_dict = self.survey.question_names_to_questions()
|
64
|
-
# iterates through the current answers and updates the question_dict (which is all questions)
|
65
|
-
for other_question, answer in self.current_answers.items():
|
66
|
-
if other_question in question_dict:
|
67
|
-
question_dict[other_question].answer = answer
|
68
|
-
else:
|
69
|
-
# it might be a comment
|
70
|
-
if (
|
71
|
-
new_question := other_question.split("_comment")[0]
|
72
|
-
) in question_dict:
|
73
|
-
question_dict[new_question].comment = answer
|
74
|
-
|
75
|
-
combined_dict = {**question_dict, **self.scenario}
|
76
|
-
# sometimes the answer is a code, so we need to translate it
|
77
|
-
return self.question._translate_answer_code_to_answer(raw_answer, combined_dict)
|
78
|
-
|
79
|
-
def extract_edsl_result_entry_and_validate(
|
80
|
-
self, agent_response_dict: AgentResponseDict
|
81
|
-
) -> EDSLResultObjectInput:
|
82
|
-
edsl_dict = agent_response_dict.edsl_dict._asdict()
|
83
|
-
exception_occurred = None
|
84
|
-
validated = False
|
85
|
-
try:
|
86
|
-
# if the question has jinja parameters, it might be easier to make a new question
|
87
|
-
# with those all filled in & then validate that
|
88
|
-
# breakpoint()
|
89
|
-
if self.question.parameters:
|
90
|
-
prior_answers_dict = self.prompt_constructor.prior_answers_dict()
|
91
|
-
question_with_validators = self.question.render(
|
92
|
-
self.scenario | prior_answers_dict
|
93
|
-
)
|
94
|
-
question_with_validators.use_code = self.question.use_code
|
95
|
-
# if question_with_validators.parameters:
|
96
|
-
# raise ValueError(
|
97
|
-
# f"The question still has parameters after rendering: {question_with_validators}"
|
98
|
-
# )
|
99
|
-
else:
|
100
|
-
question_with_validators = self.question
|
101
|
-
|
102
|
-
# breakpoint()
|
103
|
-
validated_edsl_dict = question_with_validators._validate_answer(edsl_dict)
|
104
|
-
answer = self.determine_answer(validated_edsl_dict["answer"])
|
105
|
-
comment = validated_edsl_dict.get("comment", "")
|
106
|
-
validated = True
|
107
|
-
except QuestionAnswerValidationError as e:
|
108
|
-
answer = None
|
109
|
-
comment = "The response was not valid."
|
110
|
-
# if self.raise_validation_errors:
|
111
|
-
exception_occurred = e
|
112
|
-
except Exception as non_validation_error:
|
113
|
-
answer = None
|
114
|
-
comment = "Some other error occurred."
|
115
|
-
exception_occurred = non_validation_error
|
116
|
-
finally:
|
117
|
-
# even if validation failes, we still return the result
|
118
|
-
data = {
|
119
|
-
"answer": answer,
|
120
|
-
"comment": comment,
|
121
|
-
"generated_tokens": agent_response_dict.edsl_dict.generated_tokens,
|
122
|
-
"question_name": self.question.question_name,
|
123
|
-
"prompts": self.get_prompts(),
|
124
|
-
"cached_response": agent_response_dict.model_outputs.cached_response,
|
125
|
-
"raw_model_response": agent_response_dict.model_outputs.response,
|
126
|
-
"cache_used": agent_response_dict.model_outputs.cache_used,
|
127
|
-
"cache_key": agent_response_dict.model_outputs.cache_key,
|
128
|
-
"validated": validated,
|
129
|
-
"exception_occurred": exception_occurred,
|
130
|
-
"cost": agent_response_dict.model_outputs.cost,
|
131
|
-
}
|
132
|
-
result = EDSLResultObjectInput(**data)
|
133
|
-
return result
|
134
|
-
|
135
|
-
answer_question = sync_wrapper(async_answer_question)
|
136
|
-
|
137
|
-
|
138
|
-
class InvigilatorHuman(InvigilatorBase):
|
139
|
-
"""An invigilator for when a human is answering the question."""
|
140
|
-
|
141
|
-
validate_response: bool = False
|
142
|
-
translate_response: bool = False
|
143
|
-
|
144
|
-
async def async_answer_question(self, iteration: int = 0) -> AgentResponseDict:
|
145
|
-
"""Return the answer to the question."""
|
146
|
-
comment = "This is a real survey response from a human."
|
147
|
-
|
148
|
-
def __repr__(self):
|
149
|
-
return f"{self.literal}"
|
150
|
-
|
151
|
-
exception_occurred = None
|
152
|
-
validated = False
|
153
|
-
try:
|
154
|
-
answer = self.agent.answer_question_directly(self.question, self.scenario)
|
155
|
-
self.raw_model_response = answer
|
156
|
-
|
157
|
-
if self.validate_response:
|
158
|
-
_ = self.question._validate_answer({"answer": answer})
|
159
|
-
if self.translate_response:
|
160
|
-
answer = self.question._translate_answer_code_to_answer(
|
161
|
-
answer, self.scenario
|
162
|
-
)
|
163
|
-
validated = True
|
164
|
-
except QuestionAnswerValidationError as e:
|
165
|
-
answer = None
|
166
|
-
if self.raise_validation_errors:
|
167
|
-
exception_occurred = e
|
168
|
-
except Exception as e:
|
169
|
-
answer = None
|
170
|
-
if self.raise_validation_errors:
|
171
|
-
exception_occurred = e
|
172
|
-
finally:
|
173
|
-
data = {
|
174
|
-
"generated_tokens": NotApplicable(),
|
175
|
-
"question_name": self.question.question_name,
|
176
|
-
"prompts": self.get_prompts(),
|
177
|
-
"cached_response": NotApplicable(),
|
178
|
-
"raw_model_response": NotApplicable(),
|
179
|
-
"cache_used": NotApplicable(),
|
180
|
-
"cache_key": NotApplicable(),
|
181
|
-
"answer": answer,
|
182
|
-
"comment": comment,
|
183
|
-
"validated": validated,
|
184
|
-
"exception_occurred": exception_occurred,
|
185
|
-
}
|
186
|
-
return EDSLResultObjectInput(**data)
|
187
|
-
|
188
|
-
|
189
|
-
class InvigilatorFunctional(InvigilatorBase):
|
190
|
-
"""A Invigilator for when the question has a answer_question_directly function."""
|
191
|
-
|
192
|
-
async def async_answer_question(self, iteration: int = 0) -> AgentResponseDict:
|
193
|
-
"""Return the answer to the question."""
|
194
|
-
func = self.question.answer_question_directly
|
195
|
-
answer = func(scenario=self.scenario, agent_traits=self.agent.traits)
|
196
|
-
|
197
|
-
return EDSLResultObjectInput(
|
198
|
-
generated_tokens=str(answer),
|
199
|
-
question_name=self.question.question_name,
|
200
|
-
prompts=self.get_prompts(),
|
201
|
-
cached_response=NotApplicable(),
|
202
|
-
raw_model_response=NotApplicable(),
|
203
|
-
cache_used=NotApplicable(),
|
204
|
-
cache_key=NotApplicable(),
|
205
|
-
answer=answer["answer"],
|
206
|
-
comment="This is the result of a functional question.",
|
207
|
-
validated=True,
|
208
|
-
exception_occurred=None,
|
209
|
-
)
|
210
|
-
|
211
|
-
def get_prompts(self) -> Dict[str, Prompt]:
|
212
|
-
"""Return the prompts used."""
|
213
|
-
return {
|
214
|
-
"user_prompt": Prompt("NA"),
|
215
|
-
"system_prompt": Prompt("NA"),
|
216
|
-
}
|
217
|
-
|
218
|
-
|
219
|
-
if __name__ == "__main__":
|
220
|
-
import doctest
|
221
|
-
|
222
|
-
doctest.testmod(optionflags=doctest.ELLIPSIS)
|
1
|
+
"""Module for creating Invigilators, which are objects to administer a question to an Agent."""
|
2
|
+
|
3
|
+
from typing import Dict, Any, Optional
|
4
|
+
|
5
|
+
from edsl.prompts.Prompt import Prompt
|
6
|
+
from edsl.utilities.decorators import sync_wrapper, jupyter_nb_handler
|
7
|
+
|
8
|
+
# from edsl.prompts.registry import get_classes as prompt_lookup
|
9
|
+
from edsl.exceptions.questions import QuestionAnswerValidationError
|
10
|
+
from edsl.agents.InvigilatorBase import InvigilatorBase
|
11
|
+
from edsl.data_transfer_models import AgentResponseDict, EDSLResultObjectInput
|
12
|
+
from edsl.agents.PromptConstructor import PromptConstructor
|
13
|
+
|
14
|
+
|
15
|
+
class NotApplicable(str):
|
16
|
+
def __new__(cls):
|
17
|
+
instance = super().__new__(cls, "Not Applicable")
|
18
|
+
instance.literal = "Not Applicable"
|
19
|
+
return instance
|
20
|
+
|
21
|
+
|
22
|
+
class InvigilatorAI(InvigilatorBase):
|
23
|
+
"""An invigilator that uses an AI model to answer questions."""
|
24
|
+
|
25
|
+
def get_prompts(self) -> Dict[str, Prompt]:
|
26
|
+
"""Return the prompts used."""
|
27
|
+
return self.prompt_constructor.get_prompts()
|
28
|
+
|
29
|
+
async def async_answer_question(self) -> AgentResponseDict:
|
30
|
+
"""Answer a question using the AI model.
|
31
|
+
|
32
|
+
>>> i = InvigilatorAI.example()
|
33
|
+
>>> i.answer_question()
|
34
|
+
{'message': [{'text': 'SPAM!'}], 'usage': {'prompt_tokens': 1, 'completion_tokens': 1}}
|
35
|
+
"""
|
36
|
+
prompts = self.get_prompts()
|
37
|
+
params = {
|
38
|
+
"user_prompt": prompts["user_prompt"].text,
|
39
|
+
"system_prompt": prompts["system_prompt"].text,
|
40
|
+
}
|
41
|
+
if "encoded_image" in prompts:
|
42
|
+
params["encoded_image"] = prompts["encoded_image"]
|
43
|
+
if "files_list" in prompts:
|
44
|
+
params["files_list"] = prompts["files_list"]
|
45
|
+
|
46
|
+
params.update({"iteration": self.iteration, "cache": self.cache})
|
47
|
+
|
48
|
+
agent_response_dict: AgentResponseDict = await self.model.async_get_response(
|
49
|
+
**params
|
50
|
+
)
|
51
|
+
# store to self in case validation failure
|
52
|
+
self.raw_model_response = agent_response_dict.model_outputs.response
|
53
|
+
self.generated_tokens = agent_response_dict.edsl_dict.generated_tokens
|
54
|
+
|
55
|
+
return self.extract_edsl_result_entry_and_validate(agent_response_dict)
|
56
|
+
|
57
|
+
def _remove_from_cache(self, cache_key) -> None:
|
58
|
+
"""Remove an entry from the cache."""
|
59
|
+
if cache_key:
|
60
|
+
del self.cache.data[cache_key]
|
61
|
+
|
62
|
+
def determine_answer(self, raw_answer: str) -> Any:
|
63
|
+
question_dict = self.survey.question_names_to_questions()
|
64
|
+
# iterates through the current answers and updates the question_dict (which is all questions)
|
65
|
+
for other_question, answer in self.current_answers.items():
|
66
|
+
if other_question in question_dict:
|
67
|
+
question_dict[other_question].answer = answer
|
68
|
+
else:
|
69
|
+
# it might be a comment
|
70
|
+
if (
|
71
|
+
new_question := other_question.split("_comment")[0]
|
72
|
+
) in question_dict:
|
73
|
+
question_dict[new_question].comment = answer
|
74
|
+
|
75
|
+
combined_dict = {**question_dict, **self.scenario}
|
76
|
+
# sometimes the answer is a code, so we need to translate it
|
77
|
+
return self.question._translate_answer_code_to_answer(raw_answer, combined_dict)
|
78
|
+
|
79
|
+
def extract_edsl_result_entry_and_validate(
|
80
|
+
self, agent_response_dict: AgentResponseDict
|
81
|
+
) -> EDSLResultObjectInput:
|
82
|
+
edsl_dict = agent_response_dict.edsl_dict._asdict()
|
83
|
+
exception_occurred = None
|
84
|
+
validated = False
|
85
|
+
try:
|
86
|
+
# if the question has jinja parameters, it might be easier to make a new question
|
87
|
+
# with those all filled in & then validate that
|
88
|
+
# breakpoint()
|
89
|
+
if self.question.parameters:
|
90
|
+
prior_answers_dict = self.prompt_constructor.prior_answers_dict()
|
91
|
+
question_with_validators = self.question.render(
|
92
|
+
self.scenario | prior_answers_dict
|
93
|
+
)
|
94
|
+
question_with_validators.use_code = self.question.use_code
|
95
|
+
# if question_with_validators.parameters:
|
96
|
+
# raise ValueError(
|
97
|
+
# f"The question still has parameters after rendering: {question_with_validators}"
|
98
|
+
# )
|
99
|
+
else:
|
100
|
+
question_with_validators = self.question
|
101
|
+
|
102
|
+
# breakpoint()
|
103
|
+
validated_edsl_dict = question_with_validators._validate_answer(edsl_dict)
|
104
|
+
answer = self.determine_answer(validated_edsl_dict["answer"])
|
105
|
+
comment = validated_edsl_dict.get("comment", "")
|
106
|
+
validated = True
|
107
|
+
except QuestionAnswerValidationError as e:
|
108
|
+
answer = None
|
109
|
+
comment = "The response was not valid."
|
110
|
+
# if self.raise_validation_errors:
|
111
|
+
exception_occurred = e
|
112
|
+
except Exception as non_validation_error:
|
113
|
+
answer = None
|
114
|
+
comment = "Some other error occurred."
|
115
|
+
exception_occurred = non_validation_error
|
116
|
+
finally:
|
117
|
+
# even if validation failes, we still return the result
|
118
|
+
data = {
|
119
|
+
"answer": answer,
|
120
|
+
"comment": comment,
|
121
|
+
"generated_tokens": agent_response_dict.edsl_dict.generated_tokens,
|
122
|
+
"question_name": self.question.question_name,
|
123
|
+
"prompts": self.get_prompts(),
|
124
|
+
"cached_response": agent_response_dict.model_outputs.cached_response,
|
125
|
+
"raw_model_response": agent_response_dict.model_outputs.response,
|
126
|
+
"cache_used": agent_response_dict.model_outputs.cache_used,
|
127
|
+
"cache_key": agent_response_dict.model_outputs.cache_key,
|
128
|
+
"validated": validated,
|
129
|
+
"exception_occurred": exception_occurred,
|
130
|
+
"cost": agent_response_dict.model_outputs.cost,
|
131
|
+
}
|
132
|
+
result = EDSLResultObjectInput(**data)
|
133
|
+
return result
|
134
|
+
|
135
|
+
answer_question = sync_wrapper(async_answer_question)
|
136
|
+
|
137
|
+
|
138
|
+
class InvigilatorHuman(InvigilatorBase):
|
139
|
+
"""An invigilator for when a human is answering the question."""
|
140
|
+
|
141
|
+
validate_response: bool = False
|
142
|
+
translate_response: bool = False
|
143
|
+
|
144
|
+
async def async_answer_question(self, iteration: int = 0) -> AgentResponseDict:
|
145
|
+
"""Return the answer to the question."""
|
146
|
+
comment = "This is a real survey response from a human."
|
147
|
+
|
148
|
+
def __repr__(self):
|
149
|
+
return f"{self.literal}"
|
150
|
+
|
151
|
+
exception_occurred = None
|
152
|
+
validated = False
|
153
|
+
try:
|
154
|
+
answer = self.agent.answer_question_directly(self.question, self.scenario)
|
155
|
+
self.raw_model_response = answer
|
156
|
+
|
157
|
+
if self.validate_response:
|
158
|
+
_ = self.question._validate_answer({"answer": answer})
|
159
|
+
if self.translate_response:
|
160
|
+
answer = self.question._translate_answer_code_to_answer(
|
161
|
+
answer, self.scenario
|
162
|
+
)
|
163
|
+
validated = True
|
164
|
+
except QuestionAnswerValidationError as e:
|
165
|
+
answer = None
|
166
|
+
if self.raise_validation_errors:
|
167
|
+
exception_occurred = e
|
168
|
+
except Exception as e:
|
169
|
+
answer = None
|
170
|
+
if self.raise_validation_errors:
|
171
|
+
exception_occurred = e
|
172
|
+
finally:
|
173
|
+
data = {
|
174
|
+
"generated_tokens": NotApplicable(),
|
175
|
+
"question_name": self.question.question_name,
|
176
|
+
"prompts": self.get_prompts(),
|
177
|
+
"cached_response": NotApplicable(),
|
178
|
+
"raw_model_response": NotApplicable(),
|
179
|
+
"cache_used": NotApplicable(),
|
180
|
+
"cache_key": NotApplicable(),
|
181
|
+
"answer": answer,
|
182
|
+
"comment": comment,
|
183
|
+
"validated": validated,
|
184
|
+
"exception_occurred": exception_occurred,
|
185
|
+
}
|
186
|
+
return EDSLResultObjectInput(**data)
|
187
|
+
|
188
|
+
|
189
|
+
class InvigilatorFunctional(InvigilatorBase):
|
190
|
+
"""A Invigilator for when the question has a answer_question_directly function."""
|
191
|
+
|
192
|
+
async def async_answer_question(self, iteration: int = 0) -> AgentResponseDict:
|
193
|
+
"""Return the answer to the question."""
|
194
|
+
func = self.question.answer_question_directly
|
195
|
+
answer = func(scenario=self.scenario, agent_traits=self.agent.traits)
|
196
|
+
|
197
|
+
return EDSLResultObjectInput(
|
198
|
+
generated_tokens=str(answer),
|
199
|
+
question_name=self.question.question_name,
|
200
|
+
prompts=self.get_prompts(),
|
201
|
+
cached_response=NotApplicable(),
|
202
|
+
raw_model_response=NotApplicable(),
|
203
|
+
cache_used=NotApplicable(),
|
204
|
+
cache_key=NotApplicable(),
|
205
|
+
answer=answer["answer"],
|
206
|
+
comment="This is the result of a functional question.",
|
207
|
+
validated=True,
|
208
|
+
exception_occurred=None,
|
209
|
+
)
|
210
|
+
|
211
|
+
def get_prompts(self) -> Dict[str, Prompt]:
|
212
|
+
"""Return the prompts used."""
|
213
|
+
return {
|
214
|
+
"user_prompt": Prompt("NA"),
|
215
|
+
"system_prompt": Prompt("NA"),
|
216
|
+
}
|
217
|
+
|
218
|
+
|
219
|
+
if __name__ == "__main__":
|
220
|
+
import doctest
|
221
|
+
|
222
|
+
doctest.testmod(optionflags=doctest.ELLIPSIS)
|