edsl 0.1.36.dev6__py3-none-any.whl → 0.1.37__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (261) hide show
  1. edsl/Base.py +303 -303
  2. edsl/BaseDiff.py +260 -260
  3. edsl/TemplateLoader.py +24 -24
  4. edsl/__init__.py +48 -47
  5. edsl/__version__.py +1 -1
  6. edsl/agents/Agent.py +855 -804
  7. edsl/agents/AgentList.py +350 -337
  8. edsl/agents/Invigilator.py +222 -222
  9. edsl/agents/InvigilatorBase.py +284 -294
  10. edsl/agents/PromptConstructor.py +353 -312
  11. edsl/agents/__init__.py +3 -3
  12. edsl/agents/descriptors.py +99 -86
  13. edsl/agents/prompt_helpers.py +129 -129
  14. edsl/auto/AutoStudy.py +117 -117
  15. edsl/auto/StageBase.py +230 -230
  16. edsl/auto/StageGenerateSurvey.py +178 -178
  17. edsl/auto/StageLabelQuestions.py +125 -125
  18. edsl/auto/StagePersona.py +61 -61
  19. edsl/auto/StagePersonaDimensionValueRanges.py +88 -88
  20. edsl/auto/StagePersonaDimensionValues.py +74 -74
  21. edsl/auto/StagePersonaDimensions.py +69 -69
  22. edsl/auto/StageQuestions.py +73 -73
  23. edsl/auto/SurveyCreatorPipeline.py +21 -21
  24. edsl/auto/utilities.py +224 -224
  25. edsl/base/Base.py +289 -289
  26. edsl/config.py +149 -149
  27. edsl/conjure/AgentConstructionMixin.py +160 -152
  28. edsl/conjure/Conjure.py +62 -62
  29. edsl/conjure/InputData.py +659 -659
  30. edsl/conjure/InputDataCSV.py +48 -48
  31. edsl/conjure/InputDataMixinQuestionStats.py +182 -182
  32. edsl/conjure/InputDataPyRead.py +91 -91
  33. edsl/conjure/InputDataSPSS.py +8 -8
  34. edsl/conjure/InputDataStata.py +8 -8
  35. edsl/conjure/QuestionOptionMixin.py +76 -76
  36. edsl/conjure/QuestionTypeMixin.py +23 -23
  37. edsl/conjure/RawQuestion.py +65 -65
  38. edsl/conjure/SurveyResponses.py +7 -7
  39. edsl/conjure/__init__.py +9 -9
  40. edsl/conjure/naming_utilities.py +263 -263
  41. edsl/conjure/utilities.py +201 -201
  42. edsl/conversation/Conversation.py +290 -238
  43. edsl/conversation/car_buying.py +58 -58
  44. edsl/conversation/chips.py +95 -0
  45. edsl/conversation/mug_negotiation.py +81 -81
  46. edsl/conversation/next_speaker_utilities.py +93 -93
  47. edsl/coop/PriceFetcher.py +54 -54
  48. edsl/coop/__init__.py +2 -2
  49. edsl/coop/coop.py +958 -849
  50. edsl/coop/utils.py +131 -131
  51. edsl/data/Cache.py +527 -527
  52. edsl/data/CacheEntry.py +228 -228
  53. edsl/data/CacheHandler.py +149 -149
  54. edsl/data/RemoteCacheSync.py +97 -84
  55. edsl/data/SQLiteDict.py +292 -292
  56. edsl/data/__init__.py +4 -4
  57. edsl/data/orm.py +10 -10
  58. edsl/data_transfer_models.py +73 -73
  59. edsl/enums.py +173 -173
  60. edsl/exceptions/BaseException.py +21 -0
  61. edsl/exceptions/__init__.py +54 -50
  62. edsl/exceptions/agents.py +38 -40
  63. edsl/exceptions/configuration.py +16 -16
  64. edsl/exceptions/coop.py +10 -10
  65. edsl/exceptions/data.py +14 -14
  66. edsl/exceptions/general.py +34 -34
  67. edsl/exceptions/jobs.py +33 -33
  68. edsl/exceptions/language_models.py +63 -63
  69. edsl/exceptions/prompts.py +15 -15
  70. edsl/exceptions/questions.py +91 -91
  71. edsl/exceptions/results.py +29 -26
  72. edsl/exceptions/scenarios.py +22 -0
  73. edsl/exceptions/surveys.py +37 -34
  74. edsl/inference_services/AnthropicService.py +87 -87
  75. edsl/inference_services/AwsBedrock.py +120 -115
  76. edsl/inference_services/AzureAI.py +217 -217
  77. edsl/inference_services/DeepInfraService.py +18 -18
  78. edsl/inference_services/GoogleService.py +156 -156
  79. edsl/inference_services/GroqService.py +20 -20
  80. edsl/inference_services/InferenceServiceABC.py +147 -147
  81. edsl/inference_services/InferenceServicesCollection.py +97 -72
  82. edsl/inference_services/MistralAIService.py +123 -123
  83. edsl/inference_services/OllamaService.py +18 -18
  84. edsl/inference_services/OpenAIService.py +224 -224
  85. edsl/inference_services/TestService.py +89 -89
  86. edsl/inference_services/TogetherAIService.py +170 -170
  87. edsl/inference_services/models_available_cache.py +118 -118
  88. edsl/inference_services/rate_limits_cache.py +25 -25
  89. edsl/inference_services/registry.py +39 -39
  90. edsl/inference_services/write_available.py +10 -10
  91. edsl/jobs/Answers.py +56 -56
  92. edsl/jobs/Jobs.py +1347 -1112
  93. edsl/jobs/__init__.py +1 -1
  94. edsl/jobs/buckets/BucketCollection.py +63 -63
  95. edsl/jobs/buckets/ModelBuckets.py +65 -65
  96. edsl/jobs/buckets/TokenBucket.py +248 -248
  97. edsl/jobs/interviews/Interview.py +661 -651
  98. edsl/jobs/interviews/InterviewExceptionCollection.py +99 -99
  99. edsl/jobs/interviews/InterviewExceptionEntry.py +186 -182
  100. edsl/jobs/interviews/InterviewStatistic.py +63 -63
  101. edsl/jobs/interviews/InterviewStatisticsCollection.py +25 -25
  102. edsl/jobs/interviews/InterviewStatusDictionary.py +78 -78
  103. edsl/jobs/interviews/InterviewStatusLog.py +92 -92
  104. edsl/jobs/interviews/ReportErrors.py +66 -66
  105. edsl/jobs/interviews/interview_status_enum.py +9 -9
  106. edsl/jobs/runners/JobsRunnerAsyncio.py +338 -337
  107. edsl/jobs/runners/JobsRunnerStatus.py +332 -332
  108. edsl/jobs/tasks/QuestionTaskCreator.py +242 -242
  109. edsl/jobs/tasks/TaskCreators.py +64 -64
  110. edsl/jobs/tasks/TaskHistory.py +442 -441
  111. edsl/jobs/tasks/TaskStatusLog.py +23 -23
  112. edsl/jobs/tasks/task_status_enum.py +163 -163
  113. edsl/jobs/tokens/InterviewTokenUsage.py +27 -27
  114. edsl/jobs/tokens/TokenUsage.py +34 -34
  115. edsl/language_models/KeyLookup.py +30 -0
  116. edsl/language_models/LanguageModel.py +706 -718
  117. edsl/language_models/ModelList.py +102 -102
  118. edsl/language_models/RegisterLanguageModelsMeta.py +184 -184
  119. edsl/language_models/__init__.py +3 -2
  120. edsl/language_models/fake_openai_call.py +15 -15
  121. edsl/language_models/fake_openai_service.py +61 -61
  122. edsl/language_models/registry.py +137 -137
  123. edsl/language_models/repair.py +156 -156
  124. edsl/language_models/unused/ReplicateBase.py +83 -83
  125. edsl/language_models/utilities.py +64 -64
  126. edsl/notebooks/Notebook.py +259 -259
  127. edsl/notebooks/__init__.py +1 -1
  128. edsl/prompts/Prompt.py +357 -358
  129. edsl/prompts/__init__.py +2 -2
  130. edsl/questions/AnswerValidatorMixin.py +289 -289
  131. edsl/questions/QuestionBase.py +656 -616
  132. edsl/questions/QuestionBaseGenMixin.py +161 -161
  133. edsl/questions/QuestionBasePromptsMixin.py +234 -266
  134. edsl/questions/QuestionBudget.py +227 -227
  135. edsl/questions/QuestionCheckBox.py +359 -359
  136. edsl/questions/QuestionExtract.py +183 -183
  137. edsl/questions/QuestionFreeText.py +114 -113
  138. edsl/questions/QuestionFunctional.py +159 -159
  139. edsl/questions/QuestionList.py +231 -231
  140. edsl/questions/QuestionMultipleChoice.py +286 -286
  141. edsl/questions/QuestionNumerical.py +153 -153
  142. edsl/questions/QuestionRank.py +324 -324
  143. edsl/questions/Quick.py +41 -41
  144. edsl/questions/RegisterQuestionsMeta.py +71 -71
  145. edsl/questions/ResponseValidatorABC.py +174 -174
  146. edsl/questions/SimpleAskMixin.py +73 -73
  147. edsl/questions/__init__.py +26 -26
  148. edsl/questions/compose_questions.py +98 -98
  149. edsl/questions/decorators.py +21 -21
  150. edsl/questions/derived/QuestionLikertFive.py +76 -76
  151. edsl/questions/derived/QuestionLinearScale.py +87 -87
  152. edsl/questions/derived/QuestionTopK.py +91 -91
  153. edsl/questions/derived/QuestionYesNo.py +82 -82
  154. edsl/questions/descriptors.py +413 -418
  155. edsl/questions/prompt_templates/question_budget.jinja +13 -13
  156. edsl/questions/prompt_templates/question_checkbox.jinja +32 -32
  157. edsl/questions/prompt_templates/question_extract.jinja +11 -11
  158. edsl/questions/prompt_templates/question_free_text.jinja +3 -3
  159. edsl/questions/prompt_templates/question_linear_scale.jinja +11 -11
  160. edsl/questions/prompt_templates/question_list.jinja +17 -17
  161. edsl/questions/prompt_templates/question_multiple_choice.jinja +33 -33
  162. edsl/questions/prompt_templates/question_numerical.jinja +36 -36
  163. edsl/questions/question_registry.py +147 -147
  164. edsl/questions/settings.py +12 -12
  165. edsl/questions/templates/budget/answering_instructions.jinja +7 -7
  166. edsl/questions/templates/budget/question_presentation.jinja +7 -7
  167. edsl/questions/templates/checkbox/answering_instructions.jinja +10 -10
  168. edsl/questions/templates/checkbox/question_presentation.jinja +22 -22
  169. edsl/questions/templates/extract/answering_instructions.jinja +7 -7
  170. edsl/questions/templates/likert_five/answering_instructions.jinja +10 -10
  171. edsl/questions/templates/likert_five/question_presentation.jinja +11 -11
  172. edsl/questions/templates/linear_scale/answering_instructions.jinja +5 -5
  173. edsl/questions/templates/linear_scale/question_presentation.jinja +5 -5
  174. edsl/questions/templates/list/answering_instructions.jinja +3 -3
  175. edsl/questions/templates/list/question_presentation.jinja +5 -5
  176. edsl/questions/templates/multiple_choice/answering_instructions.jinja +9 -9
  177. edsl/questions/templates/multiple_choice/question_presentation.jinja +11 -11
  178. edsl/questions/templates/numerical/answering_instructions.jinja +6 -6
  179. edsl/questions/templates/numerical/question_presentation.jinja +6 -6
  180. edsl/questions/templates/rank/answering_instructions.jinja +11 -11
  181. edsl/questions/templates/rank/question_presentation.jinja +15 -15
  182. edsl/questions/templates/top_k/answering_instructions.jinja +8 -8
  183. edsl/questions/templates/top_k/question_presentation.jinja +22 -22
  184. edsl/questions/templates/yes_no/answering_instructions.jinja +6 -6
  185. edsl/questions/templates/yes_no/question_presentation.jinja +11 -11
  186. edsl/results/Dataset.py +293 -293
  187. edsl/results/DatasetExportMixin.py +717 -693
  188. edsl/results/DatasetTree.py +145 -145
  189. edsl/results/Result.py +450 -433
  190. edsl/results/Results.py +1071 -1158
  191. edsl/results/ResultsDBMixin.py +238 -238
  192. edsl/results/ResultsExportMixin.py +43 -43
  193. edsl/results/ResultsFetchMixin.py +33 -33
  194. edsl/results/ResultsGGMixin.py +121 -121
  195. edsl/results/ResultsToolsMixin.py +98 -98
  196. edsl/results/Selector.py +135 -118
  197. edsl/results/__init__.py +2 -2
  198. edsl/results/tree_explore.py +115 -115
  199. edsl/scenarios/FileStore.py +458 -443
  200. edsl/scenarios/Scenario.py +546 -507
  201. edsl/scenarios/ScenarioHtmlMixin.py +64 -59
  202. edsl/scenarios/ScenarioList.py +1112 -1101
  203. edsl/scenarios/ScenarioListExportMixin.py +52 -52
  204. edsl/scenarios/ScenarioListPdfMixin.py +261 -261
  205. edsl/scenarios/__init__.py +4 -2
  206. edsl/shared.py +1 -1
  207. edsl/study/ObjectEntry.py +173 -173
  208. edsl/study/ProofOfWork.py +113 -113
  209. edsl/study/SnapShot.py +80 -80
  210. edsl/study/Study.py +528 -528
  211. edsl/study/__init__.py +4 -4
  212. edsl/surveys/DAG.py +148 -148
  213. edsl/surveys/Memory.py +31 -31
  214. edsl/surveys/MemoryPlan.py +244 -244
  215. edsl/surveys/Rule.py +330 -324
  216. edsl/surveys/RuleCollection.py +387 -387
  217. edsl/surveys/Survey.py +1795 -1772
  218. edsl/surveys/SurveyCSS.py +261 -261
  219. edsl/surveys/SurveyExportMixin.py +259 -259
  220. edsl/surveys/SurveyFlowVisualizationMixin.py +121 -121
  221. edsl/surveys/SurveyQualtricsImport.py +284 -284
  222. edsl/surveys/__init__.py +3 -3
  223. edsl/surveys/base.py +53 -53
  224. edsl/surveys/descriptors.py +56 -56
  225. edsl/surveys/instructions/ChangeInstruction.py +47 -47
  226. edsl/surveys/instructions/Instruction.py +51 -51
  227. edsl/surveys/instructions/InstructionCollection.py +77 -77
  228. edsl/templates/error_reporting/base.html +23 -23
  229. edsl/templates/error_reporting/exceptions_by_model.html +34 -34
  230. edsl/templates/error_reporting/exceptions_by_question_name.html +16 -16
  231. edsl/templates/error_reporting/exceptions_by_type.html +16 -16
  232. edsl/templates/error_reporting/interview_details.html +115 -115
  233. edsl/templates/error_reporting/interviews.html +9 -9
  234. edsl/templates/error_reporting/overview.html +4 -4
  235. edsl/templates/error_reporting/performance_plot.html +1 -1
  236. edsl/templates/error_reporting/report.css +73 -73
  237. edsl/templates/error_reporting/report.html +117 -117
  238. edsl/templates/error_reporting/report.js +25 -25
  239. edsl/tools/__init__.py +1 -1
  240. edsl/tools/clusters.py +192 -192
  241. edsl/tools/embeddings.py +27 -27
  242. edsl/tools/embeddings_plotting.py +118 -118
  243. edsl/tools/plotting.py +112 -112
  244. edsl/tools/summarize.py +18 -18
  245. edsl/utilities/SystemInfo.py +28 -28
  246. edsl/utilities/__init__.py +22 -22
  247. edsl/utilities/ast_utilities.py +25 -25
  248. edsl/utilities/data/Registry.py +6 -6
  249. edsl/utilities/data/__init__.py +1 -1
  250. edsl/utilities/data/scooter_results.json +1 -1
  251. edsl/utilities/decorators.py +77 -77
  252. edsl/utilities/gcp_bucket/cloud_storage.py +96 -96
  253. edsl/utilities/interface.py +627 -627
  254. edsl/utilities/repair_functions.py +28 -28
  255. edsl/utilities/restricted_python.py +70 -70
  256. edsl/utilities/utilities.py +409 -391
  257. {edsl-0.1.36.dev6.dist-info → edsl-0.1.37.dist-info}/LICENSE +21 -21
  258. {edsl-0.1.36.dev6.dist-info → edsl-0.1.37.dist-info}/METADATA +1 -1
  259. edsl-0.1.37.dist-info/RECORD +283 -0
  260. edsl-0.1.36.dev6.dist-info/RECORD +0 -279
  261. {edsl-0.1.36.dev6.dist-info → edsl-0.1.37.dist-info}/WHEEL +0 -0
@@ -1,72 +1,97 @@
1
- from edsl.inference_services.InferenceServiceABC import InferenceServiceABC
2
- import warnings
3
-
4
-
5
- class InferenceServicesCollection:
6
- added_models = {}
7
-
8
- def __init__(self, services: list[InferenceServiceABC] = None):
9
- self.services = services or []
10
-
11
- @classmethod
12
- def add_model(cls, service_name, model_name):
13
- if service_name not in cls.added_models:
14
- cls.added_models[service_name] = []
15
- cls.added_models[service_name].append(model_name)
16
-
17
- @staticmethod
18
- def _get_service_available(service, warn: bool = False) -> list[str]:
19
- from_api = True
20
- try:
21
- service_models = service.available()
22
- except Exception as e:
23
- if warn:
24
- warnings.warn(
25
- f"""Error getting models for {service._inference_service_}.
26
- Check that you have properly stored your Expected Parrot API key and activated remote inference, or stored your own API keys for the language models that you want to use.
27
- See https://docs.expectedparrot.com/en/latest/api_keys.html for instructions on storing API keys.
28
- Relying on cache.""",
29
- UserWarning,
30
- )
31
- from edsl.inference_services.models_available_cache import models_available
32
-
33
- service_models = models_available.get(service._inference_service_, [])
34
- # cache results
35
- service._models_list_cache = service_models
36
- from_api = False
37
- return service_models # , from_api
38
-
39
- def available(self):
40
- total_models = []
41
- for service in self.services:
42
- service_models = self._get_service_available(service)
43
- for model in service_models:
44
- total_models.append([model, service._inference_service_, -1])
45
-
46
- for model in self.added_models.get(service._inference_service_, []):
47
- total_models.append([model, service._inference_service_, -1])
48
-
49
- sorted_models = sorted(total_models)
50
- for i, model in enumerate(sorted_models):
51
- model[2] = i
52
- model = tuple(model)
53
- return sorted_models
54
-
55
- def register(self, service):
56
- self.services.append(service)
57
-
58
- def create_model_factory(self, model_name: str, service_name=None, index=None):
59
- from edsl.inference_services.TestService import TestService
60
-
61
- if model_name == "test":
62
- return TestService.create_model(model_name)
63
- for service in self.services:
64
- if model_name in self._get_service_available(service):
65
- if service_name is None or service_name == service._inference_service_:
66
- return service.create_model(model_name)
67
-
68
- # if model_name == "test":
69
- # from edsl.language_models import LanguageModel
70
- # return LanguageModel(test = True)
71
-
72
- raise Exception(f"Model {model_name} not found in any of the services")
1
+ from edsl.inference_services.InferenceServiceABC import InferenceServiceABC
2
+ import warnings
3
+
4
+
5
+ class InferenceServicesCollection:
6
+ added_models = {}
7
+
8
+ def __init__(self, services: list[InferenceServiceABC] = None):
9
+ self.services = services or []
10
+
11
+ @classmethod
12
+ def add_model(cls, service_name, model_name):
13
+ if service_name not in cls.added_models:
14
+ cls.added_models[service_name] = []
15
+ cls.added_models[service_name].append(model_name)
16
+
17
+ @staticmethod
18
+ def _get_service_available(service, warn: bool = False) -> list[str]:
19
+ try:
20
+ service_models = service.available()
21
+ except Exception:
22
+ if warn:
23
+ warnings.warn(
24
+ f"""Error getting models for {service._inference_service_}.
25
+ Check that you have properly stored your Expected Parrot API key and activated remote inference, or stored your own API keys for the language models that you want to use.
26
+ See https://docs.expectedparrot.com/en/latest/api_keys.html for instructions on storing API keys.
27
+ Relying on Coop.""",
28
+ UserWarning,
29
+ )
30
+
31
+ # Use the list of models on Coop as a fallback
32
+ try:
33
+ from edsl import Coop
34
+
35
+ c = Coop()
36
+ models_from_coop = c.fetch_models()
37
+ service_models = models_from_coop.get(service._inference_service_, [])
38
+
39
+ # cache results
40
+ service._models_list_cache = service_models
41
+
42
+ # Finally, use the available models cache from the Python file
43
+ except Exception:
44
+ if warn:
45
+ warnings.warn(
46
+ f"""Error getting models for {service._inference_service_}.
47
+ Relying on EDSL cache.""",
48
+ UserWarning,
49
+ )
50
+
51
+ from edsl.inference_services.models_available_cache import (
52
+ models_available,
53
+ )
54
+
55
+ service_models = models_available.get(service._inference_service_, [])
56
+
57
+ # cache results
58
+ service._models_list_cache = service_models
59
+
60
+ return service_models
61
+
62
+ def available(self):
63
+ total_models = []
64
+ for service in self.services:
65
+ service_models = self._get_service_available(service)
66
+ for model in service_models:
67
+ total_models.append([model, service._inference_service_, -1])
68
+
69
+ for model in self.added_models.get(service._inference_service_, []):
70
+ total_models.append([model, service._inference_service_, -1])
71
+
72
+ sorted_models = sorted(total_models)
73
+ for i, model in enumerate(sorted_models):
74
+ model[2] = i
75
+ model = tuple(model)
76
+ return sorted_models
77
+
78
+ def register(self, service):
79
+ self.services.append(service)
80
+
81
+ def create_model_factory(self, model_name: str, service_name=None, index=None):
82
+ from edsl.inference_services.TestService import TestService
83
+
84
+ if model_name == "test":
85
+ return TestService.create_model(model_name)
86
+
87
+ if service_name:
88
+ for service in self.services:
89
+ if service_name == service._inference_service_:
90
+ return service.create_model(model_name)
91
+
92
+ for service in self.services:
93
+ if model_name in self._get_service_available(service):
94
+ if service_name is None or service_name == service._inference_service_:
95
+ return service.create_model(model_name)
96
+
97
+ raise Exception(f"Model {model_name} not found in any of the services")
@@ -1,123 +1,123 @@
1
- import os
2
- from typing import Any, List, Optional
3
- from edsl.inference_services.InferenceServiceABC import InferenceServiceABC
4
- from edsl.language_models.LanguageModel import LanguageModel
5
- import asyncio
6
- from mistralai import Mistral
7
-
8
- from edsl.exceptions.language_models import LanguageModelBadResponseError
9
-
10
-
11
- class MistralAIService(InferenceServiceABC):
12
- """Mistral AI service class."""
13
-
14
- key_sequence = ["choices", 0, "message", "content"]
15
- usage_sequence = ["usage"]
16
-
17
- _inference_service_ = "mistral"
18
- _env_key_name_ = "MISTRAL_API_KEY" # Environment variable for Mistral API key
19
- input_token_name = "prompt_tokens"
20
- output_token_name = "completion_tokens"
21
-
22
- _sync_client_instance = None
23
- _async_client_instance = None
24
-
25
- _sync_client = Mistral
26
- _async_client = Mistral
27
-
28
- _models_list_cache: List[str] = []
29
- model_exclude_list = []
30
-
31
- def __init_subclass__(cls, **kwargs):
32
- super().__init_subclass__(**kwargs)
33
- # so subclasses have to create their own instances of the clients
34
- cls._sync_client_instance = None
35
- cls._async_client_instance = None
36
-
37
- @classmethod
38
- def sync_client(cls):
39
- if cls._sync_client_instance is None:
40
- cls._sync_client_instance = cls._sync_client(
41
- api_key=os.getenv(cls._env_key_name_)
42
- )
43
- return cls._sync_client_instance
44
-
45
- @classmethod
46
- def async_client(cls):
47
- if cls._async_client_instance is None:
48
- cls._async_client_instance = cls._async_client(
49
- api_key=os.getenv(cls._env_key_name_)
50
- )
51
- return cls._async_client_instance
52
-
53
- @classmethod
54
- def available(cls) -> list[str]:
55
- if not cls._models_list_cache:
56
- cls._models_list_cache = [
57
- m.id for m in cls.sync_client().models.list().data
58
- ]
59
-
60
- return cls._models_list_cache
61
-
62
- @classmethod
63
- def create_model(
64
- cls, model_name: str = "mistral", model_class_name=None
65
- ) -> LanguageModel:
66
- if model_class_name is None:
67
- model_class_name = cls.to_class_name(model_name)
68
-
69
- class LLM(LanguageModel):
70
- """
71
- Child class of LanguageModel for interacting with Mistral models.
72
- """
73
-
74
- key_sequence = cls.key_sequence
75
- usage_sequence = cls.usage_sequence
76
-
77
- input_token_name = cls.input_token_name
78
- output_token_name = cls.output_token_name
79
-
80
- _inference_service_ = cls._inference_service_
81
- _model_ = model_name
82
- _parameters_ = {
83
- "temperature": 0.5,
84
- "max_tokens": 512,
85
- "top_p": 0.9,
86
- }
87
-
88
- _tpm = cls.get_tpm(cls)
89
- _rpm = cls.get_rpm(cls)
90
-
91
- def sync_client(self):
92
- return cls.sync_client()
93
-
94
- def async_client(self):
95
- return cls.async_client()
96
-
97
- async def async_execute_model_call(
98
- self,
99
- user_prompt: str,
100
- system_prompt: str = "",
101
- files_list: Optional[List["FileStore"]] = None,
102
- ) -> dict[str, Any]:
103
- """Calls the Mistral API and returns the API response."""
104
- s = self.async_client()
105
-
106
- try:
107
- res = await s.chat.complete_async(
108
- model=model_name,
109
- messages=[
110
- {
111
- "content": user_prompt,
112
- "role": "user",
113
- },
114
- ],
115
- )
116
- except Exception as e:
117
- raise LanguageModelBadResponseError(f"Error with Mistral API: {e}")
118
-
119
- return res.model_dump()
120
-
121
- LLM.__name__ = model_class_name
122
-
123
- return LLM
1
+ import os
2
+ from typing import Any, List, Optional
3
+ from edsl.inference_services.InferenceServiceABC import InferenceServiceABC
4
+ from edsl.language_models.LanguageModel import LanguageModel
5
+ import asyncio
6
+ from mistralai import Mistral
7
+
8
+ from edsl.exceptions.language_models import LanguageModelBadResponseError
9
+
10
+
11
+ class MistralAIService(InferenceServiceABC):
12
+ """Mistral AI service class."""
13
+
14
+ key_sequence = ["choices", 0, "message", "content"]
15
+ usage_sequence = ["usage"]
16
+
17
+ _inference_service_ = "mistral"
18
+ _env_key_name_ = "MISTRAL_API_KEY" # Environment variable for Mistral API key
19
+ input_token_name = "prompt_tokens"
20
+ output_token_name = "completion_tokens"
21
+
22
+ _sync_client_instance = None
23
+ _async_client_instance = None
24
+
25
+ _sync_client = Mistral
26
+ _async_client = Mistral
27
+
28
+ _models_list_cache: List[str] = []
29
+ model_exclude_list = []
30
+
31
+ def __init_subclass__(cls, **kwargs):
32
+ super().__init_subclass__(**kwargs)
33
+ # so subclasses have to create their own instances of the clients
34
+ cls._sync_client_instance = None
35
+ cls._async_client_instance = None
36
+
37
+ @classmethod
38
+ def sync_client(cls):
39
+ if cls._sync_client_instance is None:
40
+ cls._sync_client_instance = cls._sync_client(
41
+ api_key=os.getenv(cls._env_key_name_)
42
+ )
43
+ return cls._sync_client_instance
44
+
45
+ @classmethod
46
+ def async_client(cls):
47
+ if cls._async_client_instance is None:
48
+ cls._async_client_instance = cls._async_client(
49
+ api_key=os.getenv(cls._env_key_name_)
50
+ )
51
+ return cls._async_client_instance
52
+
53
+ @classmethod
54
+ def available(cls) -> list[str]:
55
+ if not cls._models_list_cache:
56
+ cls._models_list_cache = [
57
+ m.id for m in cls.sync_client().models.list().data
58
+ ]
59
+
60
+ return cls._models_list_cache
61
+
62
+ @classmethod
63
+ def create_model(
64
+ cls, model_name: str = "mistral", model_class_name=None
65
+ ) -> LanguageModel:
66
+ if model_class_name is None:
67
+ model_class_name = cls.to_class_name(model_name)
68
+
69
+ class LLM(LanguageModel):
70
+ """
71
+ Child class of LanguageModel for interacting with Mistral models.
72
+ """
73
+
74
+ key_sequence = cls.key_sequence
75
+ usage_sequence = cls.usage_sequence
76
+
77
+ input_token_name = cls.input_token_name
78
+ output_token_name = cls.output_token_name
79
+
80
+ _inference_service_ = cls._inference_service_
81
+ _model_ = model_name
82
+ _parameters_ = {
83
+ "temperature": 0.5,
84
+ "max_tokens": 512,
85
+ "top_p": 0.9,
86
+ }
87
+
88
+ _tpm = cls.get_tpm(cls)
89
+ _rpm = cls.get_rpm(cls)
90
+
91
+ def sync_client(self):
92
+ return cls.sync_client()
93
+
94
+ def async_client(self):
95
+ return cls.async_client()
96
+
97
+ async def async_execute_model_call(
98
+ self,
99
+ user_prompt: str,
100
+ system_prompt: str = "",
101
+ files_list: Optional[List["FileStore"]] = None,
102
+ ) -> dict[str, Any]:
103
+ """Calls the Mistral API and returns the API response."""
104
+ s = self.async_client()
105
+
106
+ try:
107
+ res = await s.chat.complete_async(
108
+ model=model_name,
109
+ messages=[
110
+ {
111
+ "content": user_prompt,
112
+ "role": "user",
113
+ },
114
+ ],
115
+ )
116
+ except Exception as e:
117
+ raise LanguageModelBadResponseError(f"Error with Mistral API: {e}")
118
+
119
+ return res.model_dump()
120
+
121
+ LLM.__name__ = model_class_name
122
+
123
+ return LLM
@@ -1,18 +1,18 @@
1
- import aiohttp
2
- import json
3
- import requests
4
- from typing import Any, List
5
-
6
- # from edsl.inference_services.InferenceServiceABC import InferenceServiceABC
7
- from edsl.language_models import LanguageModel
8
-
9
- from edsl.inference_services.OpenAIService import OpenAIService
10
-
11
-
12
- class OllamaService(OpenAIService):
13
- """DeepInfra service class."""
14
-
15
- _inference_service_ = "ollama"
16
- _env_key_name_ = "DEEP_INFRA_API_KEY"
17
- _base_url_ = "http://localhost:11434/v1"
18
- _models_list_cache: List[str] = []
1
+ import aiohttp
2
+ import json
3
+ import requests
4
+ from typing import Any, List
5
+
6
+ # from edsl.inference_services.InferenceServiceABC import InferenceServiceABC
7
+ from edsl.language_models import LanguageModel
8
+
9
+ from edsl.inference_services.OpenAIService import OpenAIService
10
+
11
+
12
+ class OllamaService(OpenAIService):
13
+ """DeepInfra service class."""
14
+
15
+ _inference_service_ = "ollama"
16
+ _env_key_name_ = "DEEP_INFRA_API_KEY"
17
+ _base_url_ = "http://localhost:11434/v1"
18
+ _models_list_cache: List[str] = []