edsl 0.1.39.dev3__py3-none-any.whl → 0.1.39.dev5__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (341) hide show
  1. edsl/Base.py +413 -332
  2. edsl/BaseDiff.py +260 -260
  3. edsl/TemplateLoader.py +24 -24
  4. edsl/__init__.py +57 -49
  5. edsl/__version__.py +1 -1
  6. edsl/agents/Agent.py +1071 -867
  7. edsl/agents/AgentList.py +551 -413
  8. edsl/agents/Invigilator.py +284 -233
  9. edsl/agents/InvigilatorBase.py +257 -270
  10. edsl/agents/PromptConstructor.py +272 -354
  11. edsl/agents/QuestionInstructionPromptBuilder.py +128 -0
  12. edsl/agents/QuestionTemplateReplacementsBuilder.py +137 -0
  13. edsl/agents/__init__.py +2 -3
  14. edsl/agents/descriptors.py +99 -99
  15. edsl/agents/prompt_helpers.py +129 -129
  16. edsl/agents/question_option_processor.py +172 -0
  17. edsl/auto/AutoStudy.py +130 -117
  18. edsl/auto/StageBase.py +243 -230
  19. edsl/auto/StageGenerateSurvey.py +178 -178
  20. edsl/auto/StageLabelQuestions.py +125 -125
  21. edsl/auto/StagePersona.py +61 -61
  22. edsl/auto/StagePersonaDimensionValueRanges.py +88 -88
  23. edsl/auto/StagePersonaDimensionValues.py +74 -74
  24. edsl/auto/StagePersonaDimensions.py +69 -69
  25. edsl/auto/StageQuestions.py +74 -73
  26. edsl/auto/SurveyCreatorPipeline.py +21 -21
  27. edsl/auto/utilities.py +218 -224
  28. edsl/base/Base.py +279 -279
  29. edsl/config.py +177 -157
  30. edsl/conversation/Conversation.py +290 -290
  31. edsl/conversation/car_buying.py +59 -58
  32. edsl/conversation/chips.py +95 -95
  33. edsl/conversation/mug_negotiation.py +81 -81
  34. edsl/conversation/next_speaker_utilities.py +93 -93
  35. edsl/coop/CoopFunctionsMixin.py +15 -0
  36. edsl/coop/ExpectedParrotKeyHandler.py +125 -0
  37. edsl/coop/PriceFetcher.py +54 -54
  38. edsl/coop/__init__.py +2 -2
  39. edsl/coop/coop.py +1106 -1028
  40. edsl/coop/utils.py +131 -131
  41. edsl/data/Cache.py +573 -555
  42. edsl/data/CacheEntry.py +230 -233
  43. edsl/data/CacheHandler.py +168 -149
  44. edsl/data/RemoteCacheSync.py +186 -78
  45. edsl/data/SQLiteDict.py +292 -292
  46. edsl/data/__init__.py +5 -4
  47. edsl/data/orm.py +10 -10
  48. edsl/data_transfer_models.py +74 -73
  49. edsl/enums.py +202 -175
  50. edsl/exceptions/BaseException.py +21 -21
  51. edsl/exceptions/__init__.py +54 -54
  52. edsl/exceptions/agents.py +54 -42
  53. edsl/exceptions/cache.py +5 -5
  54. edsl/exceptions/configuration.py +16 -16
  55. edsl/exceptions/coop.py +10 -10
  56. edsl/exceptions/data.py +14 -14
  57. edsl/exceptions/general.py +34 -34
  58. edsl/exceptions/inference_services.py +5 -0
  59. edsl/exceptions/jobs.py +33 -33
  60. edsl/exceptions/language_models.py +63 -63
  61. edsl/exceptions/prompts.py +15 -15
  62. edsl/exceptions/questions.py +109 -91
  63. edsl/exceptions/results.py +29 -29
  64. edsl/exceptions/scenarios.py +29 -22
  65. edsl/exceptions/surveys.py +37 -37
  66. edsl/inference_services/AnthropicService.py +106 -87
  67. edsl/inference_services/AvailableModelCacheHandler.py +184 -0
  68. edsl/inference_services/AvailableModelFetcher.py +215 -0
  69. edsl/inference_services/AwsBedrock.py +118 -120
  70. edsl/inference_services/AzureAI.py +215 -217
  71. edsl/inference_services/DeepInfraService.py +18 -18
  72. edsl/inference_services/GoogleService.py +143 -148
  73. edsl/inference_services/GroqService.py +20 -20
  74. edsl/inference_services/InferenceServiceABC.py +80 -147
  75. edsl/inference_services/InferenceServicesCollection.py +138 -97
  76. edsl/inference_services/MistralAIService.py +120 -123
  77. edsl/inference_services/OllamaService.py +18 -18
  78. edsl/inference_services/OpenAIService.py +236 -224
  79. edsl/inference_services/PerplexityService.py +160 -163
  80. edsl/inference_services/ServiceAvailability.py +135 -0
  81. edsl/inference_services/TestService.py +90 -89
  82. edsl/inference_services/TogetherAIService.py +172 -170
  83. edsl/inference_services/data_structures.py +134 -0
  84. edsl/inference_services/models_available_cache.py +118 -118
  85. edsl/inference_services/rate_limits_cache.py +25 -25
  86. edsl/inference_services/registry.py +41 -41
  87. edsl/inference_services/write_available.py +10 -10
  88. edsl/jobs/AnswerQuestionFunctionConstructor.py +223 -0
  89. edsl/jobs/Answers.py +43 -56
  90. edsl/jobs/FetchInvigilator.py +47 -0
  91. edsl/jobs/InterviewTaskManager.py +98 -0
  92. edsl/jobs/InterviewsConstructor.py +50 -0
  93. edsl/jobs/Jobs.py +823 -898
  94. edsl/jobs/JobsChecks.py +172 -147
  95. edsl/jobs/JobsComponentConstructor.py +189 -0
  96. edsl/jobs/JobsPrompts.py +270 -268
  97. edsl/jobs/JobsRemoteInferenceHandler.py +311 -239
  98. edsl/jobs/JobsRemoteInferenceLogger.py +239 -0
  99. edsl/jobs/RequestTokenEstimator.py +30 -0
  100. edsl/jobs/__init__.py +1 -1
  101. edsl/jobs/async_interview_runner.py +138 -0
  102. edsl/jobs/buckets/BucketCollection.py +104 -63
  103. edsl/jobs/buckets/ModelBuckets.py +65 -65
  104. edsl/jobs/buckets/TokenBucket.py +283 -251
  105. edsl/jobs/buckets/TokenBucketAPI.py +211 -0
  106. edsl/jobs/buckets/TokenBucketClient.py +191 -0
  107. edsl/jobs/check_survey_scenario_compatibility.py +85 -0
  108. edsl/jobs/data_structures.py +120 -0
  109. edsl/jobs/decorators.py +35 -0
  110. edsl/jobs/interviews/Interview.py +396 -661
  111. edsl/jobs/interviews/InterviewExceptionCollection.py +99 -99
  112. edsl/jobs/interviews/InterviewExceptionEntry.py +186 -186
  113. edsl/jobs/interviews/InterviewStatistic.py +63 -63
  114. edsl/jobs/interviews/InterviewStatisticsCollection.py +25 -25
  115. edsl/jobs/interviews/InterviewStatusDictionary.py +78 -78
  116. edsl/jobs/interviews/InterviewStatusLog.py +92 -92
  117. edsl/jobs/interviews/ReportErrors.py +66 -66
  118. edsl/jobs/interviews/interview_status_enum.py +9 -9
  119. edsl/jobs/jobs_status_enums.py +9 -0
  120. edsl/jobs/loggers/HTMLTableJobLogger.py +304 -0
  121. edsl/jobs/results_exceptions_handler.py +98 -0
  122. edsl/jobs/runners/JobsRunnerAsyncio.py +151 -466
  123. edsl/jobs/runners/JobsRunnerStatus.py +297 -330
  124. edsl/jobs/tasks/QuestionTaskCreator.py +244 -242
  125. edsl/jobs/tasks/TaskCreators.py +64 -64
  126. edsl/jobs/tasks/TaskHistory.py +470 -450
  127. edsl/jobs/tasks/TaskStatusLog.py +23 -23
  128. edsl/jobs/tasks/task_status_enum.py +161 -163
  129. edsl/jobs/tokens/InterviewTokenUsage.py +27 -27
  130. edsl/jobs/tokens/TokenUsage.py +34 -34
  131. edsl/language_models/ComputeCost.py +63 -0
  132. edsl/language_models/LanguageModel.py +626 -668
  133. edsl/language_models/ModelList.py +164 -155
  134. edsl/language_models/PriceManager.py +127 -0
  135. edsl/language_models/RawResponseHandler.py +106 -0
  136. edsl/language_models/RegisterLanguageModelsMeta.py +184 -184
  137. edsl/language_models/ServiceDataSources.py +0 -0
  138. edsl/language_models/__init__.py +2 -3
  139. edsl/language_models/fake_openai_call.py +15 -15
  140. edsl/language_models/fake_openai_service.py +61 -61
  141. edsl/language_models/key_management/KeyLookup.py +63 -0
  142. edsl/language_models/key_management/KeyLookupBuilder.py +273 -0
  143. edsl/language_models/key_management/KeyLookupCollection.py +38 -0
  144. edsl/language_models/key_management/__init__.py +0 -0
  145. edsl/language_models/key_management/models.py +131 -0
  146. edsl/language_models/model.py +256 -0
  147. edsl/language_models/repair.py +156 -156
  148. edsl/language_models/utilities.py +65 -64
  149. edsl/notebooks/Notebook.py +263 -258
  150. edsl/notebooks/NotebookToLaTeX.py +142 -0
  151. edsl/notebooks/__init__.py +1 -1
  152. edsl/prompts/Prompt.py +352 -362
  153. edsl/prompts/__init__.py +2 -2
  154. edsl/questions/ExceptionExplainer.py +77 -0
  155. edsl/questions/HTMLQuestion.py +103 -0
  156. edsl/questions/QuestionBase.py +518 -664
  157. edsl/questions/QuestionBasePromptsMixin.py +221 -217
  158. edsl/questions/QuestionBudget.py +227 -227
  159. edsl/questions/QuestionCheckBox.py +359 -359
  160. edsl/questions/QuestionExtract.py +180 -182
  161. edsl/questions/QuestionFreeText.py +113 -114
  162. edsl/questions/QuestionFunctional.py +166 -166
  163. edsl/questions/QuestionList.py +223 -231
  164. edsl/questions/QuestionMatrix.py +265 -0
  165. edsl/questions/QuestionMultipleChoice.py +330 -286
  166. edsl/questions/QuestionNumerical.py +151 -153
  167. edsl/questions/QuestionRank.py +314 -324
  168. edsl/questions/Quick.py +41 -41
  169. edsl/questions/SimpleAskMixin.py +74 -73
  170. edsl/questions/__init__.py +27 -26
  171. edsl/questions/{AnswerValidatorMixin.py → answer_validator_mixin.py} +334 -289
  172. edsl/questions/compose_questions.py +98 -98
  173. edsl/questions/data_structures.py +20 -0
  174. edsl/questions/decorators.py +21 -21
  175. edsl/questions/derived/QuestionLikertFive.py +76 -76
  176. edsl/questions/derived/QuestionLinearScale.py +90 -87
  177. edsl/questions/derived/QuestionTopK.py +93 -93
  178. edsl/questions/derived/QuestionYesNo.py +82 -82
  179. edsl/questions/descriptors.py +427 -413
  180. edsl/questions/loop_processor.py +149 -0
  181. edsl/questions/prompt_templates/question_budget.jinja +13 -13
  182. edsl/questions/prompt_templates/question_checkbox.jinja +32 -32
  183. edsl/questions/prompt_templates/question_extract.jinja +11 -11
  184. edsl/questions/prompt_templates/question_free_text.jinja +3 -3
  185. edsl/questions/prompt_templates/question_linear_scale.jinja +11 -11
  186. edsl/questions/prompt_templates/question_list.jinja +17 -17
  187. edsl/questions/prompt_templates/question_multiple_choice.jinja +33 -33
  188. edsl/questions/prompt_templates/question_numerical.jinja +36 -36
  189. edsl/questions/{QuestionBaseGenMixin.py → question_base_gen_mixin.py} +168 -161
  190. edsl/questions/question_registry.py +177 -177
  191. edsl/questions/{RegisterQuestionsMeta.py → register_questions_meta.py} +71 -71
  192. edsl/questions/{ResponseValidatorABC.py → response_validator_abc.py} +188 -174
  193. edsl/questions/response_validator_factory.py +34 -0
  194. edsl/questions/settings.py +12 -12
  195. edsl/questions/templates/budget/answering_instructions.jinja +7 -7
  196. edsl/questions/templates/budget/question_presentation.jinja +7 -7
  197. edsl/questions/templates/checkbox/answering_instructions.jinja +10 -10
  198. edsl/questions/templates/checkbox/question_presentation.jinja +22 -22
  199. edsl/questions/templates/extract/answering_instructions.jinja +7 -7
  200. edsl/questions/templates/likert_five/answering_instructions.jinja +10 -10
  201. edsl/questions/templates/likert_five/question_presentation.jinja +11 -11
  202. edsl/questions/templates/linear_scale/answering_instructions.jinja +5 -5
  203. edsl/questions/templates/linear_scale/question_presentation.jinja +5 -5
  204. edsl/questions/templates/list/answering_instructions.jinja +3 -3
  205. edsl/questions/templates/list/question_presentation.jinja +5 -5
  206. edsl/questions/templates/matrix/__init__.py +1 -0
  207. edsl/questions/templates/matrix/answering_instructions.jinja +5 -0
  208. edsl/questions/templates/matrix/question_presentation.jinja +20 -0
  209. edsl/questions/templates/multiple_choice/answering_instructions.jinja +9 -9
  210. edsl/questions/templates/multiple_choice/question_presentation.jinja +11 -11
  211. edsl/questions/templates/numerical/answering_instructions.jinja +6 -6
  212. edsl/questions/templates/numerical/question_presentation.jinja +6 -6
  213. edsl/questions/templates/rank/answering_instructions.jinja +11 -11
  214. edsl/questions/templates/rank/question_presentation.jinja +15 -15
  215. edsl/questions/templates/top_k/answering_instructions.jinja +8 -8
  216. edsl/questions/templates/top_k/question_presentation.jinja +22 -22
  217. edsl/questions/templates/yes_no/answering_instructions.jinja +6 -6
  218. edsl/questions/templates/yes_no/question_presentation.jinja +11 -11
  219. edsl/results/CSSParameterizer.py +108 -108
  220. edsl/results/Dataset.py +587 -424
  221. edsl/results/DatasetExportMixin.py +594 -731
  222. edsl/results/DatasetTree.py +295 -275
  223. edsl/results/MarkdownToDocx.py +122 -0
  224. edsl/results/MarkdownToPDF.py +111 -0
  225. edsl/results/Result.py +557 -465
  226. edsl/results/Results.py +1183 -1165
  227. edsl/results/ResultsExportMixin.py +45 -43
  228. edsl/results/ResultsGGMixin.py +121 -121
  229. edsl/results/TableDisplay.py +125 -198
  230. edsl/results/TextEditor.py +50 -0
  231. edsl/results/__init__.py +2 -2
  232. edsl/results/file_exports.py +252 -0
  233. edsl/results/{ResultsFetchMixin.py → results_fetch_mixin.py} +33 -33
  234. edsl/results/{Selector.py → results_selector.py} +145 -135
  235. edsl/results/{ResultsToolsMixin.py → results_tools_mixin.py} +98 -98
  236. edsl/results/smart_objects.py +96 -0
  237. edsl/results/table_data_class.py +12 -0
  238. edsl/results/table_display.css +77 -77
  239. edsl/results/table_renderers.py +118 -0
  240. edsl/results/tree_explore.py +115 -115
  241. edsl/scenarios/ConstructDownloadLink.py +109 -0
  242. edsl/scenarios/DocumentChunker.py +102 -0
  243. edsl/scenarios/DocxScenario.py +16 -0
  244. edsl/scenarios/FileStore.py +511 -632
  245. edsl/scenarios/PdfExtractor.py +40 -0
  246. edsl/scenarios/Scenario.py +498 -601
  247. edsl/scenarios/ScenarioHtmlMixin.py +65 -64
  248. edsl/scenarios/ScenarioList.py +1458 -1287
  249. edsl/scenarios/ScenarioListExportMixin.py +45 -52
  250. edsl/scenarios/ScenarioListPdfMixin.py +239 -261
  251. edsl/scenarios/__init__.py +3 -4
  252. edsl/scenarios/directory_scanner.py +96 -0
  253. edsl/scenarios/file_methods.py +85 -0
  254. edsl/scenarios/handlers/__init__.py +13 -0
  255. edsl/scenarios/handlers/csv.py +38 -0
  256. edsl/scenarios/handlers/docx.py +76 -0
  257. edsl/scenarios/handlers/html.py +37 -0
  258. edsl/scenarios/handlers/json.py +111 -0
  259. edsl/scenarios/handlers/latex.py +5 -0
  260. edsl/scenarios/handlers/md.py +51 -0
  261. edsl/scenarios/handlers/pdf.py +68 -0
  262. edsl/scenarios/handlers/png.py +39 -0
  263. edsl/scenarios/handlers/pptx.py +105 -0
  264. edsl/scenarios/handlers/py.py +294 -0
  265. edsl/scenarios/handlers/sql.py +313 -0
  266. edsl/scenarios/handlers/sqlite.py +149 -0
  267. edsl/scenarios/handlers/txt.py +33 -0
  268. edsl/scenarios/{ScenarioJoin.py → scenario_join.py} +131 -127
  269. edsl/scenarios/scenario_selector.py +156 -0
  270. edsl/shared.py +1 -1
  271. edsl/study/ObjectEntry.py +173 -173
  272. edsl/study/ProofOfWork.py +113 -113
  273. edsl/study/SnapShot.py +80 -80
  274. edsl/study/Study.py +521 -528
  275. edsl/study/__init__.py +4 -4
  276. edsl/surveys/ConstructDAG.py +92 -0
  277. edsl/surveys/DAG.py +148 -148
  278. edsl/surveys/EditSurvey.py +221 -0
  279. edsl/surveys/InstructionHandler.py +100 -0
  280. edsl/surveys/Memory.py +31 -31
  281. edsl/surveys/MemoryManagement.py +72 -0
  282. edsl/surveys/MemoryPlan.py +244 -244
  283. edsl/surveys/Rule.py +327 -326
  284. edsl/surveys/RuleCollection.py +385 -387
  285. edsl/surveys/RuleManager.py +172 -0
  286. edsl/surveys/Simulator.py +75 -0
  287. edsl/surveys/Survey.py +1280 -1801
  288. edsl/surveys/SurveyCSS.py +273 -261
  289. edsl/surveys/SurveyExportMixin.py +259 -259
  290. edsl/surveys/{SurveyFlowVisualizationMixin.py → SurveyFlowVisualization.py} +181 -179
  291. edsl/surveys/SurveyQualtricsImport.py +284 -284
  292. edsl/surveys/SurveyToApp.py +141 -0
  293. edsl/surveys/__init__.py +5 -3
  294. edsl/surveys/base.py +53 -53
  295. edsl/surveys/descriptors.py +60 -56
  296. edsl/surveys/instructions/ChangeInstruction.py +48 -49
  297. edsl/surveys/instructions/Instruction.py +56 -65
  298. edsl/surveys/instructions/InstructionCollection.py +82 -77
  299. edsl/templates/error_reporting/base.html +23 -23
  300. edsl/templates/error_reporting/exceptions_by_model.html +34 -34
  301. edsl/templates/error_reporting/exceptions_by_question_name.html +16 -16
  302. edsl/templates/error_reporting/exceptions_by_type.html +16 -16
  303. edsl/templates/error_reporting/interview_details.html +115 -115
  304. edsl/templates/error_reporting/interviews.html +19 -19
  305. edsl/templates/error_reporting/overview.html +4 -4
  306. edsl/templates/error_reporting/performance_plot.html +1 -1
  307. edsl/templates/error_reporting/report.css +73 -73
  308. edsl/templates/error_reporting/report.html +117 -117
  309. edsl/templates/error_reporting/report.js +25 -25
  310. edsl/tools/__init__.py +1 -1
  311. edsl/tools/clusters.py +192 -192
  312. edsl/tools/embeddings.py +27 -27
  313. edsl/tools/embeddings_plotting.py +118 -118
  314. edsl/tools/plotting.py +112 -112
  315. edsl/tools/summarize.py +18 -18
  316. edsl/utilities/PrettyList.py +56 -0
  317. edsl/utilities/SystemInfo.py +28 -28
  318. edsl/utilities/__init__.py +22 -22
  319. edsl/utilities/ast_utilities.py +25 -25
  320. edsl/utilities/data/Registry.py +6 -6
  321. edsl/utilities/data/__init__.py +1 -1
  322. edsl/utilities/data/scooter_results.json +1 -1
  323. edsl/utilities/decorators.py +77 -77
  324. edsl/utilities/gcp_bucket/cloud_storage.py +96 -96
  325. edsl/utilities/interface.py +627 -627
  326. edsl/utilities/is_notebook.py +18 -0
  327. edsl/utilities/is_valid_variable_name.py +11 -0
  328. edsl/utilities/naming_utilities.py +263 -263
  329. edsl/utilities/remove_edsl_version.py +24 -0
  330. edsl/utilities/repair_functions.py +28 -28
  331. edsl/utilities/restricted_python.py +70 -70
  332. edsl/utilities/utilities.py +436 -424
  333. {edsl-0.1.39.dev3.dist-info → edsl-0.1.39.dev5.dist-info}/LICENSE +21 -21
  334. {edsl-0.1.39.dev3.dist-info → edsl-0.1.39.dev5.dist-info}/METADATA +13 -11
  335. edsl-0.1.39.dev5.dist-info/RECORD +358 -0
  336. {edsl-0.1.39.dev3.dist-info → edsl-0.1.39.dev5.dist-info}/WHEEL +1 -1
  337. edsl/language_models/KeyLookup.py +0 -30
  338. edsl/language_models/registry.py +0 -190
  339. edsl/language_models/unused/ReplicateBase.py +0 -83
  340. edsl/results/ResultsDBMixin.py +0 -238
  341. edsl-0.1.39.dev3.dist-info/RECORD +0 -277
edsl/jobs/Jobs.py CHANGED
@@ -1,898 +1,823 @@
1
- # """The Jobs class is a collection of agents, scenarios and models and one survey."""
2
- from __future__ import annotations
3
- import warnings
4
- import requests
5
- from itertools import product
6
- from typing import Literal, Optional, Union, Sequence, Generator, TYPE_CHECKING
7
-
8
- from edsl.Base import Base
9
-
10
- from edsl.exceptions import MissingAPIKeyError
11
- from edsl.jobs.buckets.BucketCollection import BucketCollection
12
- from edsl.jobs.interviews.Interview import Interview
13
- from edsl.jobs.runners.JobsRunnerAsyncio import JobsRunnerAsyncio
14
- from edsl.utilities.decorators import remove_edsl_version
15
-
16
- from edsl.data.RemoteCacheSync import RemoteCacheSync
17
- from edsl.exceptions.coop import CoopServerResponseError
18
-
19
- if TYPE_CHECKING:
20
- from edsl.agents.Agent import Agent
21
- from edsl.agents.AgentList import AgentList
22
- from edsl.language_models.LanguageModel import LanguageModel
23
- from edsl.scenarios.Scenario import Scenario
24
- from edsl.surveys.Survey import Survey
25
- from edsl.results.Results import Results
26
- from edsl.results.Dataset import Dataset
27
-
28
-
29
- class Jobs(Base):
30
- """
31
- A collection of agents, scenarios and models and one survey.
32
- The actual running of a job is done by a `JobsRunner`, which is a subclass of `JobsRunner`.
33
- The `JobsRunner` is chosen by the user, and is stored in the `jobs_runner_name` attribute.
34
- """
35
-
36
- __documentation__ = "https://docs.expectedparrot.com/en/latest/jobs.html"
37
-
38
- def __init__(
39
- self,
40
- survey: "Survey",
41
- agents: Optional[list["Agent"]] = None,
42
- models: Optional[list["LanguageModel"]] = None,
43
- scenarios: Optional[list["Scenario"]] = None,
44
- ):
45
- """Initialize a Jobs instance.
46
-
47
- :param survey: the survey to be used in the job
48
- :param agents: a list of agents
49
- :param models: a list of models
50
- :param scenarios: a list of scenarios
51
- """
52
- self.survey = survey
53
- self.agents: "AgentList" = agents
54
- self.scenarios: "ScenarioList" = scenarios
55
- self.models = models
56
-
57
- self.__bucket_collection = None
58
-
59
- # these setters and getters are used to ensure that the agents, models, and scenarios are stored as AgentList, ModelList, and ScenarioList objects
60
-
61
- @property
62
- def models(self):
63
- return self._models
64
-
65
- @models.setter
66
- def models(self, value):
67
- from edsl import ModelList
68
-
69
- if value:
70
- if not isinstance(value, ModelList):
71
- self._models = ModelList(value)
72
- else:
73
- self._models = value
74
- else:
75
- self._models = ModelList([])
76
-
77
- @property
78
- def agents(self):
79
- return self._agents
80
-
81
- @agents.setter
82
- def agents(self, value):
83
- from edsl import AgentList
84
-
85
- if value:
86
- if not isinstance(value, AgentList):
87
- self._agents = AgentList(value)
88
- else:
89
- self._agents = value
90
- else:
91
- self._agents = AgentList([])
92
-
93
- @property
94
- def scenarios(self):
95
- return self._scenarios
96
-
97
- @scenarios.setter
98
- def scenarios(self, value):
99
- from edsl import ScenarioList
100
- from edsl.results.Dataset import Dataset
101
-
102
- if value:
103
- if isinstance(
104
- value, Dataset
105
- ): # if the user passes in a Dataset, convert it to a ScenarioList
106
- value = value.to_scenario_list()
107
-
108
- if not isinstance(value, ScenarioList):
109
- self._scenarios = ScenarioList(value)
110
- else:
111
- self._scenarios = value
112
- else:
113
- self._scenarios = ScenarioList([])
114
-
115
- def by(
116
- self,
117
- *args: Union[
118
- "Agent",
119
- "Scenario",
120
- "LanguageModel",
121
- Sequence[Union["Agent", "Scenario", "LanguageModel"]],
122
- ],
123
- ) -> Jobs:
124
- """
125
- Add Agents, Scenarios and LanguageModels to a job. If no objects of this type exist in the Jobs instance, it stores the new objects as a list in the corresponding attribute. Otherwise, it combines the new objects with existing objects using the object's `__add__` method.
126
-
127
- This 'by' is intended to create a fluent interface.
128
-
129
- >>> from edsl import Survey
130
- >>> from edsl import QuestionFreeText
131
- >>> q = QuestionFreeText(question_name="name", question_text="What is your name?")
132
- >>> j = Jobs(survey = Survey(questions=[q]))
133
- >>> j
134
- Jobs(survey=Survey(...), agents=AgentList([]), models=ModelList([]), scenarios=ScenarioList([]))
135
- >>> from edsl import Agent; a = Agent(traits = {"status": "Sad"})
136
- >>> j.by(a).agents
137
- AgentList([Agent(traits = {'status': 'Sad'})])
138
-
139
- :param args: objects or a sequence (list, tuple, ...) of objects of the same type
140
-
141
- Notes:
142
- - all objects must implement the 'get_value', 'set_value', and `__add__` methods
143
- - agents: traits of new agents are combined with traits of existing agents. New and existing agents should not have overlapping traits, and do not increase the # agents in the instance
144
- - scenarios: traits of new scenarios are combined with traits of old existing. New scenarios will overwrite overlapping traits, and do not increase the number of scenarios in the instance
145
- - models: new models overwrite old models.
146
- """
147
- from edsl.results.Dataset import Dataset
148
-
149
- if isinstance(
150
- args[0], Dataset
151
- ): # let the user user a Dataset as if it were a ScenarioList
152
- args = args[0].to_scenario_list()
153
-
154
- passed_objects = self._turn_args_to_list(
155
- args
156
- ) # objects can also be passed comma-separated
157
-
158
- current_objects, objects_key = self._get_current_objects_of_this_type(
159
- passed_objects[0]
160
- )
161
-
162
- if not current_objects:
163
- new_objects = passed_objects
164
- else:
165
- new_objects = self._merge_objects(passed_objects, current_objects)
166
-
167
- setattr(self, objects_key, new_objects) # update the job
168
- return self
169
-
170
- def prompts(self) -> "Dataset":
171
- """Return a Dataset of prompts that will be used.
172
-
173
-
174
- >>> from edsl.jobs import Jobs
175
- >>> Jobs.example().prompts()
176
- Dataset(...)
177
- """
178
- from edsl.jobs.JobsPrompts import JobsPrompts
179
-
180
- j = JobsPrompts(self)
181
- return j.prompts()
182
-
183
- def show_prompts(self, all=False) -> None:
184
- """Print the prompts."""
185
- if all:
186
- return self.prompts().to_scenario_list().table()
187
- else:
188
- return (
189
- self.prompts().to_scenario_list().table("user_prompt", "system_prompt")
190
- )
191
-
192
- @staticmethod
193
- def estimate_prompt_cost(
194
- system_prompt: str,
195
- user_prompt: str,
196
- price_lookup: dict,
197
- inference_service: str,
198
- model: str,
199
- ) -> dict:
200
- """
201
- Estimate the cost of running the prompts.
202
- :param iterations: the number of iterations to run
203
- """
204
- from edsl.jobs.JobsPrompts import JobsPrompts
205
-
206
- return JobsPrompts.estimate_prompt_cost(
207
- system_prompt, user_prompt, price_lookup, inference_service, model
208
- )
209
-
210
- def estimate_job_cost(self, iterations: int = 1) -> dict:
211
- """
212
- Estimate the cost of running the job.
213
-
214
- :param iterations: the number of iterations to run
215
- """
216
- from edsl.jobs.JobsPrompts import JobsPrompts
217
-
218
- j = JobsPrompts(self)
219
- return j.estimate_job_cost(iterations)
220
-
221
- def estimate_job_cost_from_external_prices(
222
- self, price_lookup: dict, iterations: int = 1
223
- ) -> dict:
224
- from edsl.jobs.JobsPrompts import JobsPrompts
225
-
226
- j = JobsPrompts(self)
227
- return j.estimate_job_cost_from_external_prices(price_lookup, iterations)
228
-
229
- @staticmethod
230
- def compute_job_cost(job_results: Results) -> float:
231
- """
232
- Computes the cost of a completed job in USD.
233
- """
234
- return job_results.compute_job_cost()
235
-
236
- @staticmethod
237
- def _get_container_class(object):
238
- from edsl.agents.AgentList import AgentList
239
- from edsl.agents.Agent import Agent
240
- from edsl.scenarios.Scenario import Scenario
241
- from edsl.scenarios.ScenarioList import ScenarioList
242
- from edsl.language_models.ModelList import ModelList
243
-
244
- if isinstance(object, Agent):
245
- return AgentList
246
- elif isinstance(object, Scenario):
247
- return ScenarioList
248
- elif isinstance(object, ModelList):
249
- return ModelList
250
- else:
251
- return list
252
-
253
- @staticmethod
254
- def _turn_args_to_list(args):
255
- """Return a list of the first argument if it is a sequence, otherwise returns a list of all the arguments.
256
-
257
- Example:
258
-
259
- >>> Jobs._turn_args_to_list([1,2,3])
260
- [1, 2, 3]
261
-
262
- """
263
-
264
- def did_user_pass_a_sequence(args):
265
- """Return True if the user passed a sequence, False otherwise.
266
-
267
- Example:
268
-
269
- >>> did_user_pass_a_sequence([1,2,3])
270
- True
271
-
272
- >>> did_user_pass_a_sequence(1)
273
- False
274
- """
275
- return len(args) == 1 and isinstance(args[0], Sequence)
276
-
277
- if did_user_pass_a_sequence(args):
278
- container_class = Jobs._get_container_class(args[0][0])
279
- return container_class(args[0])
280
- else:
281
- container_class = Jobs._get_container_class(args[0])
282
- return container_class(args)
283
-
284
- def _get_current_objects_of_this_type(
285
- self, object: Union["Agent", "Scenario", "LanguageModel"]
286
- ) -> tuple[list, str]:
287
- from edsl.agents.Agent import Agent
288
- from edsl.scenarios.Scenario import Scenario
289
- from edsl.language_models.LanguageModel import LanguageModel
290
-
291
- """Return the current objects of the same type as the first argument.
292
-
293
- >>> from edsl.jobs import Jobs
294
- >>> j = Jobs.example()
295
- >>> j._get_current_objects_of_this_type(j.agents[0])
296
- (AgentList([Agent(traits = {'status': 'Joyful'}), Agent(traits = {'status': 'Sad'})]), 'agents')
297
- """
298
- class_to_key = {
299
- Agent: "agents",
300
- Scenario: "scenarios",
301
- LanguageModel: "models",
302
- }
303
- for class_type in class_to_key:
304
- if isinstance(object, class_type) or issubclass(
305
- object.__class__, class_type
306
- ):
307
- key = class_to_key[class_type]
308
- break
309
- else:
310
- raise ValueError(
311
- f"First argument must be an Agent, Scenario, or LanguageModel, not {object}"
312
- )
313
- current_objects = getattr(self, key, None)
314
- return current_objects, key
315
-
316
- @staticmethod
317
- def _get_empty_container_object(object):
318
- from edsl.agents.AgentList import AgentList
319
- from edsl.scenarios.ScenarioList import ScenarioList
320
-
321
- return {"Agent": AgentList([]), "Scenario": ScenarioList([])}.get(
322
- object.__class__.__name__, []
323
- )
324
-
325
- @staticmethod
326
- def _merge_objects(passed_objects, current_objects) -> list:
327
- """
328
- Combine all the existing objects with the new objects.
329
-
330
- For example, if the user passes in 3 agents,
331
- and there are 2 existing agents, this will create 6 new agents
332
-
333
- >>> Jobs(survey = [])._merge_objects([1,2,3], [4,5,6])
334
- [5, 6, 7, 6, 7, 8, 7, 8, 9]
335
- """
336
- new_objects = Jobs._get_empty_container_object(passed_objects[0])
337
- for current_object in current_objects:
338
- for new_object in passed_objects:
339
- new_objects.append(current_object + new_object)
340
- return new_objects
341
-
342
- def interviews(self) -> list[Interview]:
343
- """
344
- Return a list of :class:`edsl.jobs.interviews.Interview` objects.
345
-
346
- It returns one Interview for each combination of Agent, Scenario, and LanguageModel.
347
- If any of Agents, Scenarios, or LanguageModels are missing, it fills in with defaults.
348
-
349
- >>> from edsl.jobs import Jobs
350
- >>> j = Jobs.example()
351
- >>> len(j.interviews())
352
- 4
353
- >>> j.interviews()[0]
354
- Interview(agent = Agent(traits = {'status': 'Joyful'}), survey = Survey(...), scenario = Scenario({'period': 'morning'}), model = Model(...))
355
- """
356
- if hasattr(self, "_interviews"):
357
- return self._interviews
358
- else:
359
- return list(self._create_interviews())
360
-
361
- @classmethod
362
- def from_interviews(cls, interview_list):
363
- """Return a Jobs instance from a list of interviews.
364
-
365
- This is useful when you have, say, a list of failed interviews and you want to create
366
- a new job with only those interviews.
367
- """
368
- survey = interview_list[0].survey
369
- # get all the models
370
- models = list(set([interview.model for interview in interview_list]))
371
- jobs = cls(survey)
372
- jobs.models = models
373
- jobs._interviews = interview_list
374
- return jobs
375
-
376
- def _create_interviews(self) -> Generator[Interview, None, None]:
377
- """
378
- Generate interviews.
379
-
380
- Note that this sets the agents, model and scenarios if they have not been set. This is a side effect of the method.
381
- This is useful because a user can create a job without setting the agents, models, or scenarios, and the job will still run,
382
- with us filling in defaults.
383
-
384
-
385
- """
386
- # if no agents, models, or scenarios are set, set them to defaults
387
- from edsl.agents.Agent import Agent
388
- from edsl.language_models.registry import Model
389
- from edsl.scenarios.Scenario import Scenario
390
-
391
- self.agents = self.agents or [Agent()]
392
- self.models = self.models or [Model()]
393
- self.scenarios = self.scenarios or [Scenario()]
394
- for agent, scenario, model in product(self.agents, self.scenarios, self.models):
395
- yield Interview(
396
- survey=self.survey,
397
- agent=agent,
398
- scenario=scenario,
399
- model=model,
400
- skip_retry=self.skip_retry,
401
- raise_validation_errors=self.raise_validation_errors,
402
- )
403
-
404
- def create_bucket_collection(self) -> BucketCollection:
405
- """
406
- Create a collection of buckets for each model.
407
-
408
- These buckets are used to track API calls and token usage.
409
-
410
- >>> from edsl.jobs import Jobs
411
- >>> from edsl import Model
412
- >>> j = Jobs.example().by(Model(temperature = 1), Model(temperature = 0.5))
413
- >>> bc = j.create_bucket_collection()
414
- >>> bc
415
- BucketCollection(...)
416
- """
417
- bucket_collection = BucketCollection()
418
- for model in self.models:
419
- bucket_collection.add_model(model)
420
- return bucket_collection
421
-
422
- @property
423
- def bucket_collection(self) -> BucketCollection:
424
- """Return the bucket collection. If it does not exist, create it."""
425
- if self.__bucket_collection is None:
426
- self.__bucket_collection = self.create_bucket_collection()
427
- return self.__bucket_collection
428
-
429
- def html(self):
430
- """Return the HTML representations for each scenario"""
431
- links = []
432
- for index, scenario in enumerate(self.scenarios):
433
- links.append(
434
- self.survey.html(
435
- scenario=scenario, return_link=True, cta=f"Scenario {index}"
436
- )
437
- )
438
- return links
439
-
440
- def __hash__(self):
441
- """Allow the model to be used as a key in a dictionary.
442
-
443
- >>> from edsl.jobs import Jobs
444
- >>> hash(Jobs.example())
445
- 846655441787442972
446
-
447
- """
448
- from edsl.utilities.utilities import dict_hash
449
-
450
- return dict_hash(self.to_dict(add_edsl_version=False))
451
-
452
- def _output(self, message) -> None:
453
- """Check if a Job is verbose. If so, print the message."""
454
- if hasattr(self, "verbose") and self.verbose:
455
- print(message)
456
-
457
- def _check_parameters(self, strict=False, warn=False) -> None:
458
- """Check if the parameters in the survey and scenarios are consistent.
459
-
460
- >>> from edsl import QuestionFreeText
461
- >>> from edsl import Survey
462
- >>> from edsl import Scenario
463
- >>> q = QuestionFreeText(question_text = "{{poo}}", question_name = "ugly_question")
464
- >>> j = Jobs(survey = Survey(questions=[q]))
465
- >>> with warnings.catch_warnings(record=True) as w:
466
- ... j._check_parameters(warn = True)
467
- ... assert len(w) == 1
468
- ... assert issubclass(w[-1].category, UserWarning)
469
- ... assert "The following parameters are in the survey but not in the scenarios" in str(w[-1].message)
470
-
471
- >>> q = QuestionFreeText(question_text = "{{poo}}", question_name = "ugly_question")
472
- >>> s = Scenario({'plop': "A", 'poo': "B"})
473
- >>> j = Jobs(survey = Survey(questions=[q])).by(s)
474
- >>> j._check_parameters(strict = True)
475
- Traceback (most recent call last):
476
- ...
477
- ValueError: The following parameters are in the scenarios but not in the survey: {'plop'}
478
-
479
- >>> q = QuestionFreeText(question_text = "Hello", question_name = "ugly_question")
480
- >>> s = Scenario({'ugly_question': "B"})
481
- >>> j = Jobs(survey = Survey(questions=[q])).by(s)
482
- >>> j._check_parameters()
483
- Traceback (most recent call last):
484
- ...
485
- ValueError: The following names are in both the survey question_names and the scenario keys: {'ugly_question'}. This will create issues.
486
- """
487
- survey_parameters: set = self.survey.parameters
488
- scenario_parameters: set = self.scenarios.parameters
489
-
490
- msg0, msg1, msg2 = None, None, None
491
-
492
- # look for key issues
493
- if intersection := set(self.scenarios.parameters) & set(
494
- self.survey.question_names
495
- ):
496
- msg0 = f"The following names are in both the survey question_names and the scenario keys: {intersection}. This will create issues."
497
-
498
- raise ValueError(msg0)
499
-
500
- if in_survey_but_not_in_scenarios := survey_parameters - scenario_parameters:
501
- msg1 = f"The following parameters are in the survey but not in the scenarios: {in_survey_but_not_in_scenarios}"
502
- if in_scenarios_but_not_in_survey := scenario_parameters - survey_parameters:
503
- msg2 = f"The following parameters are in the scenarios but not in the survey: {in_scenarios_but_not_in_survey}"
504
-
505
- if msg1 or msg2:
506
- message = "\n".join(filter(None, [msg1, msg2]))
507
- if strict:
508
- raise ValueError(message)
509
- else:
510
- if warn:
511
- warnings.warn(message)
512
-
513
- if self.scenarios.has_jinja_braces:
514
- warnings.warn(
515
- "The scenarios have Jinja braces ({{ and }}). Converting to '<<' and '>>'. If you want a different conversion, use the convert_jinja_braces method first to modify the scenario."
516
- )
517
- self.scenarios = self.scenarios.convert_jinja_braces()
518
-
519
- @property
520
- def skip_retry(self):
521
- if not hasattr(self, "_skip_retry"):
522
- return False
523
- return self._skip_retry
524
-
525
- @property
526
- def raise_validation_errors(self):
527
- if not hasattr(self, "_raise_validation_errors"):
528
- return False
529
- return self._raise_validation_errors
530
-
531
- def use_remote_cache(self, disable_remote_cache: bool) -> bool:
532
- if disable_remote_cache:
533
- return False
534
- if not disable_remote_cache:
535
- try:
536
- from edsl import Coop
537
-
538
- user_edsl_settings = Coop().edsl_settings
539
- return user_edsl_settings.get("remote_caching", False)
540
- except requests.ConnectionError:
541
- pass
542
- except CoopServerResponseError as e:
543
- pass
544
-
545
- return False
546
-
547
- def run(
548
- self,
549
- n: int = 1,
550
- progress_bar: bool = False,
551
- stop_on_exception: bool = False,
552
- cache: Union[Cache, bool] = None,
553
- check_api_keys: bool = False,
554
- sidecar_model: Optional[LanguageModel] = None,
555
- verbose: bool = True,
556
- print_exceptions=True,
557
- remote_cache_description: Optional[str] = None,
558
- remote_inference_description: Optional[str] = None,
559
- remote_inference_results_visibility: Optional[
560
- Literal["private", "public", "unlisted"]
561
- ] = "unlisted",
562
- skip_retry: bool = False,
563
- raise_validation_errors: bool = False,
564
- disable_remote_cache: bool = False,
565
- disable_remote_inference: bool = False,
566
- ) -> Results:
567
- """
568
- Runs the Job: conducts Interviews and returns their results.
569
-
570
- :param n: How many times to run each interview
571
- :param progress_bar: Whether to show a progress bar
572
- :param stop_on_exception: Stops the job if an exception is raised
573
- :param cache: A Cache object to store results
574
- :param check_api_keys: Raises an error if API keys are invalid
575
- :param verbose: Prints extra messages
576
- :param remote_cache_description: Specifies a description for this group of entries in the remote cache
577
- :param remote_inference_description: Specifies a description for the remote inference job
578
- :param remote_inference_results_visibility: The initial visibility of the Results object on Coop. This will only be used for remote jobs!
579
- :param disable_remote_cache: If True, the job will not use remote cache. This only works for local jobs!
580
- :param disable_remote_inference: If True, the job will not use remote inference
581
- """
582
- from edsl.coop.coop import Coop
583
-
584
- self._check_parameters()
585
- self._skip_retry = skip_retry
586
- self._raise_validation_errors = raise_validation_errors
587
-
588
- self.verbose = verbose
589
-
590
- from edsl.jobs.JobsChecks import JobsChecks
591
-
592
- jc = JobsChecks(self)
593
-
594
- # check if the user has all the keys they need
595
- if jc.needs_key_process():
596
- jc.key_process()
597
-
598
- from edsl.jobs.JobsRemoteInferenceHandler import JobsRemoteInferenceHandler
599
-
600
- jh = JobsRemoteInferenceHandler(self, verbose=verbose)
601
- if jh.use_remote_inference(disable_remote_inference):
602
- jh.create_remote_inference_job(
603
- iterations=n,
604
- remote_inference_description=remote_inference_description,
605
- remote_inference_results_visibility=remote_inference_results_visibility,
606
- )
607
- results = jh.poll_remote_inference_job()
608
- return results
609
-
610
- if check_api_keys:
611
- jc.check_api_keys()
612
-
613
- # handle cache
614
- if cache is None or cache is True:
615
- from edsl.data.CacheHandler import CacheHandler
616
-
617
- cache = CacheHandler().get_cache()
618
- if cache is False:
619
- from edsl.data.Cache import Cache
620
-
621
- cache = Cache()
622
-
623
- remote_cache = self.use_remote_cache(disable_remote_cache)
624
- with RemoteCacheSync(
625
- coop=Coop(),
626
- cache=cache,
627
- output_func=self._output,
628
- remote_cache=remote_cache,
629
- remote_cache_description=remote_cache_description,
630
- ) as r:
631
- results = self._run_local(
632
- n=n,
633
- progress_bar=progress_bar,
634
- cache=cache,
635
- stop_on_exception=stop_on_exception,
636
- sidecar_model=sidecar_model,
637
- print_exceptions=print_exceptions,
638
- raise_validation_errors=raise_validation_errors,
639
- )
640
-
641
- # results.cache = cache.new_entries_cache()
642
- return results
643
-
644
- async def run_async(
645
- self,
646
- cache=None,
647
- n=1,
648
- disable_remote_inference: bool = False,
649
- remote_inference_description: Optional[str] = None,
650
- remote_inference_results_visibility: Optional[
651
- Literal["private", "public", "unlisted"]
652
- ] = "unlisted",
653
- **kwargs,
654
- ):
655
- """Run the job asynchronously, either locally or remotely.
656
-
657
- :param cache: Cache object or boolean
658
- :param n: Number of iterations
659
- :param disable_remote_inference: If True, forces local execution
660
- :param remote_inference_description: Description for remote jobs
661
- :param remote_inference_results_visibility: Visibility setting for remote results
662
- :param kwargs: Additional arguments passed to local execution
663
- :return: Results object
664
- """
665
- # Check if we should use remote inference
666
- from edsl.jobs.JobsRemoteInferenceHandler import JobsRemoteInferenceHandler
667
-
668
- jh = JobsRemoteInferenceHandler(self, verbose=False)
669
- if jh.use_remote_inference(disable_remote_inference):
670
- results = await jh.create_and_poll_remote_job(
671
- iterations=n,
672
- remote_inference_description=remote_inference_description,
673
- remote_inference_results_visibility=remote_inference_results_visibility,
674
- )
675
- return results
676
-
677
- # If not using remote inference, run locally with async
678
- return await JobsRunnerAsyncio(self).run_async(cache=cache, n=n, **kwargs)
679
-
680
- def _run_local(self, *args, **kwargs):
681
- """Run the job locally."""
682
-
683
- results = JobsRunnerAsyncio(self).run(*args, **kwargs)
684
- return results
685
-
686
- def all_question_parameters(self):
687
- """Return all the fields in the questions in the survey.
688
- >>> from edsl.jobs import Jobs
689
- >>> Jobs.example().all_question_parameters()
690
- {'period'}
691
- """
692
- return set.union(*[question.parameters for question in self.survey.questions])
693
-
694
- def __repr__(self) -> str:
695
- """Return an eval-able string representation of the Jobs instance."""
696
- return f"Jobs(survey={repr(self.survey)}, agents={repr(self.agents)}, models={repr(self.models)}, scenarios={repr(self.scenarios)})"
697
-
698
- def _summary(self):
699
- return {
700
- "EDSL Class": "Jobs",
701
- "Number of questions": len(self.survey),
702
- "Number of agents": len(self.agents),
703
- "Number of models": len(self.models),
704
- "Number of scenarios": len(self.scenarios),
705
- }
706
-
707
- def _repr_html_(self) -> str:
708
- footer = f"<a href={self.__documentation__}>(docs)</a>"
709
- return str(self.summary(format="html")) + footer
710
-
711
- def __len__(self) -> int:
712
- """Return the maximum number of questions that will be asked while running this job.
713
- Note that this is the maximum number of questions, not the actual number of questions that will be asked, as some questions may be skipped.
714
-
715
- >>> from edsl.jobs import Jobs
716
- >>> len(Jobs.example())
717
- 8
718
- """
719
- number_of_questions = (
720
- len(self.agents or [1])
721
- * len(self.scenarios or [1])
722
- * len(self.models or [1])
723
- * len(self.survey)
724
- )
725
- return number_of_questions
726
-
727
- #######################
728
- # Serialization methods
729
- #######################
730
-
731
- def to_dict(self, add_edsl_version=True):
732
- d = {
733
- "survey": self.survey.to_dict(add_edsl_version=add_edsl_version),
734
- "agents": [
735
- agent.to_dict(add_edsl_version=add_edsl_version)
736
- for agent in self.agents
737
- ],
738
- "models": [
739
- model.to_dict(add_edsl_version=add_edsl_version)
740
- for model in self.models
741
- ],
742
- "scenarios": [
743
- scenario.to_dict(add_edsl_version=add_edsl_version)
744
- for scenario in self.scenarios
745
- ],
746
- }
747
- if add_edsl_version:
748
- from edsl import __version__
749
-
750
- d["edsl_version"] = __version__
751
- d["edsl_class_name"] = "Jobs"
752
-
753
- return d
754
-
755
- @classmethod
756
- @remove_edsl_version
757
- def from_dict(cls, data: dict) -> Jobs:
758
- """Creates a Jobs instance from a dictionary."""
759
- from edsl import Survey
760
- from edsl.agents.Agent import Agent
761
- from edsl.language_models.LanguageModel import LanguageModel
762
- from edsl.scenarios.Scenario import Scenario
763
-
764
- return cls(
765
- survey=Survey.from_dict(data["survey"]),
766
- agents=[Agent.from_dict(agent) for agent in data["agents"]],
767
- models=[LanguageModel.from_dict(model) for model in data["models"]],
768
- scenarios=[Scenario.from_dict(scenario) for scenario in data["scenarios"]],
769
- )
770
-
771
- def __eq__(self, other: Jobs) -> bool:
772
- """Return True if the Jobs instance is equal to another Jobs instance.
773
-
774
- >>> from edsl.jobs import Jobs
775
- >>> Jobs.example() == Jobs.example()
776
- True
777
-
778
- """
779
- return hash(self) == hash(other)
780
-
781
- #######################
782
- # Example methods
783
- #######################
784
- @classmethod
785
- def example(
786
- cls,
787
- throw_exception_probability: float = 0.0,
788
- randomize: bool = False,
789
- test_model=False,
790
- ) -> Jobs:
791
- """Return an example Jobs instance.
792
-
793
- :param throw_exception_probability: the probability that an exception will be thrown when answering a question. This is useful for testing error handling.
794
- :param randomize: whether to randomize the job by adding a random string to the period
795
- :param test_model: whether to use a test model
796
-
797
- >>> Jobs.example()
798
- Jobs(...)
799
-
800
- """
801
- import random
802
- from uuid import uuid4
803
- from edsl.questions import QuestionMultipleChoice
804
- from edsl.agents.Agent import Agent
805
- from edsl.scenarios.Scenario import Scenario
806
-
807
- addition = "" if not randomize else str(uuid4())
808
-
809
- if test_model:
810
- from edsl.language_models import LanguageModel
811
-
812
- m = LanguageModel.example(test_model=True)
813
-
814
- # (status, question, period)
815
- agent_answers = {
816
- ("Joyful", "how_feeling", "morning"): "OK",
817
- ("Joyful", "how_feeling", "afternoon"): "Great",
818
- ("Joyful", "how_feeling_yesterday", "morning"): "Great",
819
- ("Joyful", "how_feeling_yesterday", "afternoon"): "Good",
820
- ("Sad", "how_feeling", "morning"): "Terrible",
821
- ("Sad", "how_feeling", "afternoon"): "OK",
822
- ("Sad", "how_feeling_yesterday", "morning"): "OK",
823
- ("Sad", "how_feeling_yesterday", "afternoon"): "Terrible",
824
- }
825
-
826
- def answer_question_directly(self, question, scenario):
827
- """Return the answer to a question. This is a method that can be added to an agent."""
828
-
829
- if random.random() < throw_exception_probability:
830
- raise Exception("Error!")
831
- return agent_answers[
832
- (self.traits["status"], question.question_name, scenario["period"])
833
- ]
834
-
835
- sad_agent = Agent(traits={"status": "Sad"})
836
- joy_agent = Agent(traits={"status": "Joyful"})
837
-
838
- sad_agent.add_direct_question_answering_method(answer_question_directly)
839
- joy_agent.add_direct_question_answering_method(answer_question_directly)
840
-
841
- q1 = QuestionMultipleChoice(
842
- question_text="How are you this {{ period }}?",
843
- question_options=["Good", "Great", "OK", "Terrible"],
844
- question_name="how_feeling",
845
- )
846
- q2 = QuestionMultipleChoice(
847
- question_text="How were you feeling yesterday {{ period }}?",
848
- question_options=["Good", "Great", "OK", "Terrible"],
849
- question_name="how_feeling_yesterday",
850
- )
851
- from edsl import Survey, ScenarioList
852
-
853
- base_survey = Survey(questions=[q1, q2])
854
-
855
- scenario_list = ScenarioList(
856
- [
857
- Scenario({"period": f"morning{addition}"}),
858
- Scenario({"period": "afternoon"}),
859
- ]
860
- )
861
- if test_model:
862
- job = base_survey.by(m).by(scenario_list).by(joy_agent, sad_agent)
863
- else:
864
- job = base_survey.by(scenario_list).by(joy_agent, sad_agent)
865
-
866
- return job
867
-
868
- def rich_print(self):
869
- """Print a rich representation of the Jobs instance."""
870
- from rich.table import Table
871
-
872
- table = Table(title="Jobs")
873
- table.add_column("Jobs")
874
- table.add_row(self.survey.rich_print())
875
- return table
876
-
877
- def code(self):
878
- """Return the code to create this instance."""
879
- raise NotImplementedError
880
-
881
-
882
- def main():
883
- """Run the module's doctests."""
884
- from edsl.jobs import Jobs
885
- from edsl.data.Cache import Cache
886
-
887
- job = Jobs.example()
888
- len(job) == 8
889
- results = job.run(cache=Cache())
890
- len(results) == 8
891
- results
892
-
893
-
894
- if __name__ == "__main__":
895
- """Run the module's doctests."""
896
- import doctest
897
-
898
- doctest.testmod(optionflags=doctest.ELLIPSIS)
1
+ # """The Jobs class is a collection of agents, scenarios and models and one survey."""
2
+ from __future__ import annotations
3
+ import asyncio
4
+ from inspect import signature
5
+ from typing import (
6
+ Literal,
7
+ Optional,
8
+ Union,
9
+ Sequence,
10
+ Generator,
11
+ TYPE_CHECKING,
12
+ Callable,
13
+ Tuple,
14
+ )
15
+
16
+ from edsl.Base import Base
17
+
18
+ from edsl.jobs.buckets.BucketCollection import BucketCollection
19
+ from edsl.jobs.JobsPrompts import JobsPrompts
20
+ from edsl.jobs.interviews.Interview import Interview
21
+ from edsl.utilities.remove_edsl_version import remove_edsl_version
22
+ from edsl.jobs.runners.JobsRunnerAsyncio import JobsRunnerAsyncio
23
+ from edsl.data.RemoteCacheSync import RemoteCacheSync
24
+ from edsl.exceptions.coop import CoopServerResponseError
25
+
26
+ from edsl.jobs.JobsChecks import JobsChecks
27
+ from edsl.jobs.data_structures import RunEnvironment, RunParameters, RunConfig
28
+
29
+ if TYPE_CHECKING:
30
+ from edsl.agents.Agent import Agent
31
+ from edsl.agents.AgentList import AgentList
32
+ from edsl.language_models.LanguageModel import LanguageModel
33
+ from edsl.scenarios.Scenario import Scenario
34
+ from edsl.scenarios.ScenarioList import ScenarioList
35
+ from edsl.surveys.Survey import Survey
36
+ from edsl.results.Results import Results
37
+ from edsl.results.Dataset import Dataset
38
+ from edsl.language_models.ModelList import ModelList
39
+ from edsl.data.Cache import Cache
40
+ from edsl.language_models.key_management.KeyLookup import KeyLookup
41
+
42
+ VisibilityType = Literal["private", "public", "unlisted"]
43
+
44
+ from dataclasses import dataclass
45
+ from typing import Optional, Union, TypeVar, Callable, cast
46
+ from functools import wraps
47
+
48
+ try:
49
+ from typing import ParamSpec
50
+ except ImportError:
51
+ from typing_extensions import ParamSpec
52
+
53
+
54
+ P = ParamSpec("P")
55
+ T = TypeVar("T")
56
+
57
+
58
+ from edsl.jobs.check_survey_scenario_compatibility import (
59
+ CheckSurveyScenarioCompatibility,
60
+ )
61
+
62
+
63
+ def with_config(f: Callable[P, T]) -> Callable[P, T]:
64
+ "This decorator make it so that the run function parameters match the RunConfig dataclass."
65
+ parameter_fields = {
66
+ name: field.default
67
+ for name, field in RunParameters.__dataclass_fields__.items()
68
+ }
69
+ environment_fields = {
70
+ name: field.default
71
+ for name, field in RunEnvironment.__dataclass_fields__.items()
72
+ }
73
+ combined = {**parameter_fields, **environment_fields}
74
+
75
+ @wraps(f)
76
+ def wrapper(*args: P.args, **kwargs: P.kwargs) -> T:
77
+ environment = RunEnvironment(
78
+ **{k: v for k, v in kwargs.items() if k in environment_fields}
79
+ )
80
+ parameters = RunParameters(
81
+ **{k: v for k, v in kwargs.items() if k in parameter_fields}
82
+ )
83
+ config = RunConfig(environment=environment, parameters=parameters)
84
+ return f(*args, config=config)
85
+
86
+ # Update the wrapper's signature to include all RunConfig parameters
87
+ # old_sig = signature(f)
88
+ # wrapper.__signature__ = old_sig.replace(
89
+ # parameters=list(old_sig.parameters.values())[:-1]
90
+ # + [
91
+ # old_sig.parameters["config"].replace(
92
+ # default=parameter_fields[name], name=name
93
+ # )
94
+ # for name in combined
95
+ # ]
96
+ # )
97
+
98
+ return cast(Callable[P, T], wrapper)
99
+
100
+
101
+ class Jobs(Base):
102
+ """
103
+ A collection of agents, scenarios and models and one survey that creates 'interviews'
104
+ """
105
+
106
+ __documentation__ = "https://docs.expectedparrot.com/en/latest/jobs.html"
107
+
108
+ def __init__(
109
+ self,
110
+ survey: "Survey",
111
+ agents: Optional[Union[list[Agent], AgentList]] = None,
112
+ models: Optional[Union[ModelList, list[LanguageModel]]] = None,
113
+ scenarios: Optional[Union[ScenarioList, list[Scenario]]] = None,
114
+ ):
115
+ """Initialize a Jobs instance.
116
+
117
+ :param survey: the survey to be used in the job
118
+ :param agents: a list of agents
119
+ :param models: a list of models
120
+ :param scenarios: a list of scenarios
121
+ """
122
+ self.run_config = RunConfig(
123
+ environment=RunEnvironment(), parameters=RunParameters()
124
+ )
125
+
126
+ self.survey = survey
127
+ self.agents: AgentList = agents
128
+ self.scenarios: ScenarioList = scenarios
129
+ self.models: ModelList = models
130
+
131
+ def add_running_env(self, running_env: RunEnvironment):
132
+ self.run_config.add_environment(running_env)
133
+ return self
134
+
135
+ def using_cache(self, cache: "Cache") -> Jobs:
136
+ """
137
+ Add a Cache to the job.
138
+
139
+ :param cache: the cache to add
140
+ """
141
+ self.run_config.add_cache(cache)
142
+ return self
143
+
144
+ def using_bucket_collection(self, bucket_collection: BucketCollection) -> Jobs:
145
+ """
146
+ Add a BucketCollection to the job.
147
+
148
+ :param bucket_collection: the bucket collection to add
149
+ """
150
+ self.run_config.add_bucket_collection(bucket_collection)
151
+ return self
152
+
153
+ def using_key_lookup(self, key_lookup: KeyLookup) -> Jobs:
154
+ """
155
+ Add a KeyLookup to the job.
156
+
157
+ :param key_lookup: the key lookup to add
158
+ """
159
+ self.run_config.add_key_lookup(key_lookup)
160
+ return self
161
+
162
+ def using(self, obj: Union[Cache, BucketCollection, KeyLookup]) -> Jobs:
163
+ """
164
+ Add a Cache, BucketCollection, or KeyLookup to the job.
165
+
166
+ :param obj: the object to add
167
+ """
168
+ from edsl.data.Cache import Cache
169
+ from edsl.language_models.key_management.KeyLookup import KeyLookup
170
+
171
+ if isinstance(obj, Cache):
172
+ self.using_cache(obj)
173
+ elif isinstance(obj, BucketCollection):
174
+ self.using_bucket_collection(obj)
175
+ elif isinstance(obj, KeyLookup):
176
+ self.using_key_lookup(obj)
177
+ return self
178
+
179
+ @property
180
+ def models(self):
181
+ return self._models
182
+
183
+ @models.setter
184
+ def models(self, value):
185
+ from edsl.language_models.ModelList import ModelList
186
+
187
+ if value:
188
+ if not isinstance(value, ModelList):
189
+ self._models = ModelList(value)
190
+ else:
191
+ self._models = value
192
+ else:
193
+ self._models = ModelList([])
194
+
195
+ # update the bucket collection if it exists
196
+ if self.run_config.environment.bucket_collection is None:
197
+ self.run_config.environment.bucket_collection = (
198
+ self.create_bucket_collection()
199
+ )
200
+
201
+ @property
202
+ def agents(self):
203
+ return self._agents
204
+
205
+ @agents.setter
206
+ def agents(self, value):
207
+ from edsl.agents.AgentList import AgentList
208
+
209
+ if value:
210
+ if not isinstance(value, AgentList):
211
+ self._agents = AgentList(value)
212
+ else:
213
+ self._agents = value
214
+ else:
215
+ self._agents = AgentList([])
216
+
217
+ @property
218
+ def scenarios(self):
219
+ return self._scenarios
220
+
221
+ @scenarios.setter
222
+ def scenarios(self, value):
223
+ from edsl.scenarios.ScenarioList import ScenarioList
224
+ from edsl.results.Dataset import Dataset
225
+
226
+ if value:
227
+ if isinstance(
228
+ value, Dataset
229
+ ): # if the user passes in a Dataset, convert it to a ScenarioList
230
+ value = value.to_scenario_list()
231
+
232
+ if not isinstance(value, ScenarioList):
233
+ self._scenarios = ScenarioList(value)
234
+ else:
235
+ self._scenarios = value
236
+ else:
237
+ self._scenarios = ScenarioList([])
238
+
239
+ def by(
240
+ self,
241
+ *args: Union[
242
+ Agent,
243
+ Scenario,
244
+ LanguageModel,
245
+ Sequence[Union["Agent", "Scenario", "LanguageModel"]],
246
+ ],
247
+ ) -> Jobs:
248
+ """
249
+ Add Agents, Scenarios and LanguageModels to a job.
250
+
251
+ :param args: objects or a sequence (list, tuple, ...) of objects of the same type
252
+
253
+ If no objects of this type exist in the Jobs instance, it stores the new objects as a list in the corresponding attribute.
254
+ Otherwise, it combines the new objects with existing objects using the object's `__add__` method.
255
+
256
+ This 'by' is intended to create a fluent interface.
257
+
258
+ >>> from edsl.surveys.Survey import Survey
259
+ >>> from edsl.questions.QuestionFreeText import QuestionFreeText
260
+ >>> q = QuestionFreeText(question_name="name", question_text="What is your name?")
261
+ >>> j = Jobs(survey = Survey(questions=[q]))
262
+ >>> j
263
+ Jobs(survey=Survey(...), agents=AgentList([]), models=ModelList([]), scenarios=ScenarioList([]))
264
+ >>> from edsl.agents.Agent import Agent; a = Agent(traits = {"status": "Sad"})
265
+ >>> j.by(a).agents
266
+ AgentList([Agent(traits = {'status': 'Sad'})])
267
+
268
+
269
+ Notes:
270
+ - all objects must implement the 'get_value', 'set_value', and `__add__` methods
271
+ - agents: traits of new agents are combined with traits of existing agents. New and existing agents should not have overlapping traits, and do not increase the # agents in the instance
272
+ - scenarios: traits of new scenarios are combined with traits of old existing. New scenarios will overwrite overlapping traits, and do not increase the number of scenarios in the instance
273
+ - models: new models overwrite old models.
274
+ """
275
+ from edsl.jobs.JobsComponentConstructor import JobsComponentConstructor
276
+
277
+ return JobsComponentConstructor(self).by(*args)
278
+
279
+ def prompts(self) -> "Dataset":
280
+ """Return a Dataset of prompts that will be used.
281
+
282
+
283
+ >>> from edsl.jobs import Jobs
284
+ >>> Jobs.example().prompts()
285
+ Dataset(...)
286
+ """
287
+ return JobsPrompts(self).prompts()
288
+
289
+ def show_prompts(self, all: bool = False) -> None:
290
+ """Print the prompts."""
291
+ if all:
292
+ return self.prompts().to_scenario_list().table()
293
+ else:
294
+ return (
295
+ self.prompts().to_scenario_list().table("user_prompt", "system_prompt")
296
+ )
297
+
298
+ @staticmethod
299
+ def estimate_prompt_cost(
300
+ system_prompt: str,
301
+ user_prompt: str,
302
+ price_lookup: dict,
303
+ inference_service: str,
304
+ model: str,
305
+ ) -> dict:
306
+ """
307
+ Estimate the cost of running the prompts.
308
+ :param iterations: the number of iterations to run
309
+ :param system_prompt: the system prompt
310
+ :param user_prompt: the user prompt
311
+ :param price_lookup: the price lookup
312
+ :param inference_service: the inference service
313
+ :param model: the model name
314
+ """
315
+ return JobsPrompts.estimate_prompt_cost(
316
+ system_prompt, user_prompt, price_lookup, inference_service, model
317
+ )
318
+
319
+ def estimate_job_cost(self, iterations: int = 1) -> dict:
320
+ """
321
+ Estimate the cost of running the job.
322
+
323
+ :param iterations: the number of iterations to run
324
+ """
325
+ return JobsPrompts(self).estimate_job_cost(iterations)
326
+
327
+ def estimate_job_cost_from_external_prices(
328
+ self, price_lookup: dict, iterations: int = 1
329
+ ) -> dict:
330
+ return JobsPrompts(self).estimate_job_cost_from_external_prices(
331
+ price_lookup, iterations
332
+ )
333
+
334
+ @staticmethod
335
+ def compute_job_cost(job_results: Results) -> float:
336
+ """
337
+ Computes the cost of a completed job in USD.
338
+ """
339
+ return job_results.compute_job_cost()
340
+
341
+ def replace_missing_objects(self) -> None:
342
+ from edsl.agents.Agent import Agent
343
+ from edsl.language_models.model import Model
344
+ from edsl.scenarios.Scenario import Scenario
345
+
346
+ self.agents = self.agents or [Agent()]
347
+ self.models = self.models or [Model()]
348
+ self.scenarios = self.scenarios or [Scenario()]
349
+
350
+ def generate_interviews(self) -> Generator[Interview, None, None]:
351
+ """
352
+ Generate interviews.
353
+
354
+ Note that this sets the agents, model and scenarios if they have not been set. This is a side effect of the method.
355
+ This is useful because a user can create a job without setting the agents, models, or scenarios, and the job will still run,
356
+ with us filling in defaults.
357
+
358
+ """
359
+ from edsl.jobs.InterviewsConstructor import InterviewsConstructor
360
+
361
+ self.replace_missing_objects()
362
+ yield from InterviewsConstructor(
363
+ self, cache=self.run_config.environment.cache
364
+ ).create_interviews()
365
+
366
+ def interviews(self) -> list[Interview]:
367
+ """
368
+ Return a list of :class:`edsl.jobs.interviews.Interview` objects.
369
+
370
+ It returns one Interview for each combination of Agent, Scenario, and LanguageModel.
371
+ If any of Agents, Scenarios, or LanguageModels are missing, it fills in with defaults.
372
+
373
+ >>> from edsl.jobs import Jobs
374
+ >>> j = Jobs.example()
375
+ >>> len(j.interviews())
376
+ 4
377
+ >>> j.interviews()[0]
378
+ Interview(agent = Agent(traits = {'status': 'Joyful'}), survey = Survey(...), scenario = Scenario({'period': 'morning'}), model = Model(...))
379
+ """
380
+ return list(self.generate_interviews())
381
+
382
+ @classmethod
383
+ def from_interviews(cls, interview_list) -> "Jobs":
384
+ """Return a Jobs instance from a list of interviews.
385
+
386
+ This is useful when you have, say, a list of failed interviews and you want to create
387
+ a new job with only those interviews.
388
+ """
389
+ survey = interview_list[0].survey
390
+ # get all the models
391
+ models = list(set([interview.model for interview in interview_list]))
392
+ jobs = cls(survey)
393
+ jobs.models = models
394
+ jobs._interviews = interview_list
395
+ return jobs
396
+
397
+ def create_bucket_collection(self) -> BucketCollection:
398
+ """
399
+ Create a collection of buckets for each model.
400
+
401
+ These buckets are used to track API calls and token usage.
402
+
403
+ >>> from edsl.jobs import Jobs
404
+ >>> from edsl import Model
405
+ >>> j = Jobs.example().by(Model(temperature = 1), Model(temperature = 0.5))
406
+ >>> bc = j.create_bucket_collection()
407
+ >>> bc
408
+ BucketCollection(...)
409
+ """
410
+ return BucketCollection.from_models(self.models)
411
+
412
+ def html(self):
413
+ """Return the HTML representations for each scenario"""
414
+ links = []
415
+ for index, scenario in enumerate(self.scenarios):
416
+ links.append(
417
+ self.survey.html(
418
+ scenario=scenario, return_link=True, cta=f"Scenario {index}"
419
+ )
420
+ )
421
+ return links
422
+
423
+ def __hash__(self):
424
+ """Allow the model to be used as a key in a dictionary.
425
+
426
+ >>> from edsl.jobs import Jobs
427
+ >>> hash(Jobs.example())
428
+ 846655441787442972
429
+
430
+ """
431
+ from edsl.utilities.utilities import dict_hash
432
+
433
+ return dict_hash(self.to_dict(add_edsl_version=False))
434
+
435
+ def _output(self, message) -> None:
436
+ """Check if a Job is verbose. If so, print the message."""
437
+ if self.run_config.parameters.verbose:
438
+ print(message)
439
+ # if hasattr(self, "verbose") and self.verbose:
440
+ # print(message)
441
+
442
+ def all_question_parameters(self) -> set:
443
+ """Return all the fields in the questions in the survey.
444
+ >>> from edsl.jobs import Jobs
445
+ >>> Jobs.example().all_question_parameters()
446
+ {'period'}
447
+ """
448
+ return set.union(*[question.parameters for question in self.survey.questions])
449
+
450
+ def use_remote_cache(self) -> bool:
451
+ import requests
452
+
453
+ if self.run_config.parameters.disable_remote_cache:
454
+ return False
455
+ if not self.run_config.parameters.disable_remote_cache:
456
+ try:
457
+ from edsl.coop.coop import Coop
458
+
459
+ user_edsl_settings = Coop().edsl_settings
460
+ return user_edsl_settings.get("remote_caching", False)
461
+ except requests.ConnectionError:
462
+ pass
463
+ except CoopServerResponseError as e:
464
+ pass
465
+
466
+ return False
467
+
468
+ def _remote_results(
469
+ self,
470
+ ) -> Union["Results", None]:
471
+ from edsl.jobs.JobsRemoteInferenceHandler import JobsRemoteInferenceHandler
472
+
473
+ jh = JobsRemoteInferenceHandler(
474
+ self, verbose=self.run_config.parameters.verbose
475
+ )
476
+ if jh.use_remote_inference(self.run_config.parameters.disable_remote_inference):
477
+ job_info = jh.create_remote_inference_job(
478
+ iterations=self.run_config.parameters.n,
479
+ remote_inference_description=self.run_config.parameters.remote_inference_description,
480
+ remote_inference_results_visibility=self.run_config.parameters.remote_inference_results_visibility,
481
+ )
482
+ results = jh.poll_remote_inference_job(job_info)
483
+ return results
484
+ else:
485
+ return None
486
+
487
+ def _prepare_to_run(self) -> None:
488
+ "This makes sure that the job is ready to run and that keys are in place for a remote job."
489
+ CheckSurveyScenarioCompatibility(self.survey, self.scenarios).check()
490
+
491
+ def _check_if_remote_keys_ok(self):
492
+ jc = JobsChecks(self)
493
+ if jc.needs_key_process():
494
+ jc.key_process()
495
+
496
+ def _check_if_local_keys_ok(self):
497
+ jc = JobsChecks(self)
498
+ if self.run_config.parameters.check_api_keys:
499
+ jc.check_api_keys()
500
+
501
+ async def _execute_with_remote_cache(self, run_job_async: bool) -> Results:
502
+
503
+ use_remote_cache = self.use_remote_cache()
504
+
505
+ from edsl.coop.coop import Coop
506
+ from edsl.jobs.runners.JobsRunnerAsyncio import JobsRunnerAsyncio
507
+ from edsl.data.Cache import Cache
508
+
509
+ assert isinstance(self.run_config.environment.cache, Cache)
510
+
511
+ with RemoteCacheSync(
512
+ coop=Coop(),
513
+ cache=self.run_config.environment.cache,
514
+ output_func=self._output,
515
+ remote_cache=use_remote_cache,
516
+ remote_cache_description=self.run_config.parameters.remote_cache_description,
517
+ ):
518
+ runner = JobsRunnerAsyncio(self, environment=self.run_config.environment)
519
+ if run_job_async:
520
+ results = await runner.run_async(self.run_config.parameters)
521
+ else:
522
+ results = runner.run(self.run_config.parameters)
523
+ return results
524
+
525
+ def _setup_and_check(self) -> Tuple[RunConfig, Optional[Results]]:
526
+
527
+ self._prepare_to_run()
528
+ self._check_if_remote_keys_ok()
529
+
530
+ # first try to run the job remotely
531
+ if results := self._remote_results():
532
+ return results
533
+
534
+ self._check_if_local_keys_ok()
535
+ return None
536
+
537
+ @property
538
+ def num_interviews(self):
539
+ if self.run_config.parameters.n is None:
540
+ return len(self)
541
+ else:
542
+ len(self) * self.run_config.parameters.n
543
+
544
+ def _run(self, config: RunConfig):
545
+ "Shared code for run and run_async"
546
+ if config.environment.cache is not None:
547
+ self.run_config.environment.cache = config.environment.cache
548
+
549
+ if config.environment.bucket_collection is not None:
550
+ self.run_config.environment.bucket_collection = (
551
+ config.environment.bucket_collection
552
+ )
553
+
554
+ if config.environment.key_lookup is not None:
555
+ self.run_config.environment.key_lookup = config.environment.key_lookup
556
+
557
+ # replace the parameters with the ones from the config
558
+ self.run_config.parameters = config.parameters
559
+
560
+ self.replace_missing_objects()
561
+
562
+ # try to run remotely first
563
+ self._prepare_to_run()
564
+ self._check_if_remote_keys_ok()
565
+
566
+ if (
567
+ self.run_config.environment.cache is None
568
+ or self.run_config.environment.cache is True
569
+ ):
570
+ from edsl.data.CacheHandler import CacheHandler
571
+
572
+ self.run_config.environment.cache = CacheHandler().get_cache()
573
+
574
+ if self.run_config.environment.cache is False:
575
+ from edsl.data.Cache import Cache
576
+
577
+ self.run_config.environment.cache = Cache(immediate_write=False)
578
+
579
+ # first try to run the job remotely
580
+ if results := self._remote_results():
581
+ return results
582
+
583
+ self._check_if_local_keys_ok()
584
+
585
+ if config.environment.bucket_collection is None:
586
+ self.run_config.environment.bucket_collection = (
587
+ self.create_bucket_collection()
588
+ )
589
+
590
+ @with_config
591
+ def run(self, *, config: RunConfig) -> "Results":
592
+ """
593
+ Runs the Job: conducts Interviews and returns their results.
594
+
595
+ :param n: How many times to run each interview
596
+ :param progress_bar: Whether to show a progress bar
597
+ :param stop_on_exception: Stops the job if an exception is raised
598
+ :param check_api_keys: Raises an error if API keys are invalid
599
+ :param verbose: Prints extra messages
600
+ :param remote_cache_description: Specifies a description for this group of entries in the remote cache
601
+ :param remote_inference_description: Specifies a description for the remote inference job
602
+ :param remote_inference_results_visibility: The initial visibility of the Results object on Coop. This will only be used for remote jobs!
603
+ :param disable_remote_cache: If True, the job will not use remote cache. This only works for local jobs!
604
+ :param disable_remote_inference: If True, the job will not use remote inference
605
+ :param cache: A Cache object to store results
606
+ :param bucket_collection: A BucketCollection object to track API calls
607
+ :param key_lookup: A KeyLookup object to manage API keys
608
+ """
609
+ self._run(config)
610
+
611
+ return asyncio.run(self._execute_with_remote_cache(run_job_async=False))
612
+
613
+ @with_config
614
+ async def run_async(self, *, config: RunConfig) -> "Results":
615
+ """
616
+ Runs the Job: conducts Interviews and returns their results.
617
+
618
+ :param n: How many times to run each interview
619
+ :param progress_bar: Whether to show a progress bar
620
+ :param stop_on_exception: Stops the job if an exception is raised
621
+ :param check_api_keys: Raises an error if API keys are invalid
622
+ :param verbose: Prints extra messages
623
+ :param remote_cache_description: Specifies a description for this group of entries in the remote cache
624
+ :param remote_inference_description: Specifies a description for the remote inference job
625
+ :param remote_inference_results_visibility: The initial visibility of the Results object on Coop. This will only be used for remote jobs!
626
+ :param disable_remote_cache: If True, the job will not use remote cache. This only works for local jobs!
627
+ :param disable_remote_inference: If True, the job will not use remote inference
628
+ :param cache: A Cache object to store results
629
+ :param bucket_collection: A BucketCollection object to track API calls
630
+ :param key_lookup: A KeyLookup object to manage API keys
631
+ """
632
+ self._run(config)
633
+
634
+ return await self._execute_with_remote_cache(run_job_async=True)
635
+
636
+ def __repr__(self) -> str:
637
+ """Return an eval-able string representation of the Jobs instance."""
638
+ return f"Jobs(survey={repr(self.survey)}, agents={repr(self.agents)}, models={repr(self.models)}, scenarios={repr(self.scenarios)})"
639
+
640
+ def _summary(self):
641
+ return {
642
+ "questions": len(self.survey),
643
+ "agents": len(self.agents or [1]),
644
+ "models": len(self.models or [1]),
645
+ "scenarios": len(self.scenarios or [1]),
646
+ }
647
+
648
+ def __len__(self) -> int:
649
+ """Return the maximum number of questions that will be asked while running this job.
650
+ Note that this is the maximum number of questions, not the actual number of questions that will be asked, as some questions may be skipped.
651
+
652
+ >>> from edsl.jobs import Jobs
653
+ >>> len(Jobs.example())
654
+ 8
655
+ """
656
+ number_of_questions = (
657
+ len(self.agents or [1])
658
+ * len(self.scenarios or [1])
659
+ * len(self.models or [1])
660
+ * len(self.survey)
661
+ )
662
+ return number_of_questions
663
+
664
+ def to_dict(self, add_edsl_version=True):
665
+ d = {
666
+ "survey": self.survey.to_dict(add_edsl_version=add_edsl_version),
667
+ "agents": [
668
+ agent.to_dict(add_edsl_version=add_edsl_version)
669
+ for agent in self.agents
670
+ ],
671
+ "models": [
672
+ model.to_dict(add_edsl_version=add_edsl_version)
673
+ for model in self.models
674
+ ],
675
+ "scenarios": [
676
+ scenario.to_dict(add_edsl_version=add_edsl_version)
677
+ for scenario in self.scenarios
678
+ ],
679
+ }
680
+ if add_edsl_version:
681
+ from edsl import __version__
682
+
683
+ d["edsl_version"] = __version__
684
+ d["edsl_class_name"] = "Jobs"
685
+
686
+ return d
687
+
688
+ def table(self):
689
+ return self.prompts().to_scenario_list().table()
690
+
691
+ @classmethod
692
+ @remove_edsl_version
693
+ def from_dict(cls, data: dict) -> Jobs:
694
+ """Creates a Jobs instance from a dictionary."""
695
+ from edsl.surveys.Survey import Survey
696
+ from edsl.agents.Agent import Agent
697
+ from edsl.language_models.LanguageModel import LanguageModel
698
+ from edsl.scenarios.Scenario import Scenario
699
+
700
+ return cls(
701
+ survey=Survey.from_dict(data["survey"]),
702
+ agents=[Agent.from_dict(agent) for agent in data["agents"]],
703
+ models=[LanguageModel.from_dict(model) for model in data["models"]],
704
+ scenarios=[Scenario.from_dict(scenario) for scenario in data["scenarios"]],
705
+ )
706
+
707
+ def __eq__(self, other: Jobs) -> bool:
708
+ """Return True if the Jobs instance is equal to another Jobs instance.
709
+
710
+ >>> from edsl.jobs import Jobs
711
+ >>> Jobs.example() == Jobs.example()
712
+ True
713
+
714
+ """
715
+ return hash(self) == hash(other)
716
+
717
+ @classmethod
718
+ def example(
719
+ cls,
720
+ throw_exception_probability: float = 0.0,
721
+ randomize: bool = False,
722
+ test_model=False,
723
+ ) -> Jobs:
724
+ """Return an example Jobs instance.
725
+
726
+ :param throw_exception_probability: the probability that an exception will be thrown when answering a question. This is useful for testing error handling.
727
+ :param randomize: whether to randomize the job by adding a random string to the period
728
+ :param test_model: whether to use a test model
729
+
730
+ >>> Jobs.example()
731
+ Jobs(...)
732
+
733
+ """
734
+ import random
735
+ from uuid import uuid4
736
+ from edsl.questions.QuestionMultipleChoice import QuestionMultipleChoice
737
+ from edsl.agents.Agent import Agent
738
+ from edsl.scenarios.Scenario import Scenario
739
+
740
+ addition = "" if not randomize else str(uuid4())
741
+
742
+ if test_model:
743
+ from edsl.language_models.LanguageModel import LanguageModel
744
+
745
+ m = LanguageModel.example(test_model=True)
746
+
747
+ # (status, question, period)
748
+ agent_answers = {
749
+ ("Joyful", "how_feeling", "morning"): "OK",
750
+ ("Joyful", "how_feeling", "afternoon"): "Great",
751
+ ("Joyful", "how_feeling_yesterday", "morning"): "Great",
752
+ ("Joyful", "how_feeling_yesterday", "afternoon"): "Good",
753
+ ("Sad", "how_feeling", "morning"): "Terrible",
754
+ ("Sad", "how_feeling", "afternoon"): "OK",
755
+ ("Sad", "how_feeling_yesterday", "morning"): "OK",
756
+ ("Sad", "how_feeling_yesterday", "afternoon"): "Terrible",
757
+ }
758
+
759
+ def answer_question_directly(self, question, scenario):
760
+ """Return the answer to a question. This is a method that can be added to an agent."""
761
+
762
+ if random.random() < throw_exception_probability:
763
+ raise Exception("Error!")
764
+ return agent_answers[
765
+ (self.traits["status"], question.question_name, scenario["period"])
766
+ ]
767
+
768
+ sad_agent = Agent(traits={"status": "Sad"})
769
+ joy_agent = Agent(traits={"status": "Joyful"})
770
+
771
+ sad_agent.add_direct_question_answering_method(answer_question_directly)
772
+ joy_agent.add_direct_question_answering_method(answer_question_directly)
773
+
774
+ q1 = QuestionMultipleChoice(
775
+ question_text="How are you this {{ period }}?",
776
+ question_options=["Good", "Great", "OK", "Terrible"],
777
+ question_name="how_feeling",
778
+ )
779
+ q2 = QuestionMultipleChoice(
780
+ question_text="How were you feeling yesterday {{ period }}?",
781
+ question_options=["Good", "Great", "OK", "Terrible"],
782
+ question_name="how_feeling_yesterday",
783
+ )
784
+ from edsl.surveys.Survey import Survey
785
+ from edsl.scenarios.ScenarioList import ScenarioList
786
+
787
+ base_survey = Survey(questions=[q1, q2])
788
+
789
+ scenario_list = ScenarioList(
790
+ [
791
+ Scenario({"period": f"morning{addition}"}),
792
+ Scenario({"period": "afternoon"}),
793
+ ]
794
+ )
795
+ if test_model:
796
+ job = base_survey.by(m).by(scenario_list).by(joy_agent, sad_agent)
797
+ else:
798
+ job = base_survey.by(scenario_list).by(joy_agent, sad_agent)
799
+
800
+ return job
801
+
802
+ def code(self):
803
+ """Return the code to create this instance."""
804
+ raise NotImplementedError
805
+
806
+
807
+ def main():
808
+ """Run the module's doctests."""
809
+ from edsl.jobs.Jobs import Jobs
810
+ from edsl.data.Cache import Cache
811
+
812
+ job = Jobs.example()
813
+ len(job) == 8
814
+ results = job.run(cache=Cache())
815
+ len(results) == 8
816
+ results
817
+
818
+
819
+ if __name__ == "__main__":
820
+ """Run the module's doctests."""
821
+ import doctest
822
+
823
+ doctest.testmod(optionflags=doctest.ELLIPSIS)