edsl 0.1.39.dev1__py3-none-any.whl → 0.1.39.dev3__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (256) hide show
  1. edsl/Base.py +332 -332
  2. edsl/BaseDiff.py +260 -260
  3. edsl/TemplateLoader.py +24 -24
  4. edsl/__init__.py +49 -49
  5. edsl/__version__.py +1 -1
  6. edsl/agents/Agent.py +867 -867
  7. edsl/agents/AgentList.py +413 -413
  8. edsl/agents/Invigilator.py +233 -233
  9. edsl/agents/InvigilatorBase.py +270 -265
  10. edsl/agents/PromptConstructor.py +354 -354
  11. edsl/agents/__init__.py +3 -3
  12. edsl/agents/descriptors.py +99 -99
  13. edsl/agents/prompt_helpers.py +129 -129
  14. edsl/auto/AutoStudy.py +117 -117
  15. edsl/auto/StageBase.py +230 -230
  16. edsl/auto/StageGenerateSurvey.py +178 -178
  17. edsl/auto/StageLabelQuestions.py +125 -125
  18. edsl/auto/StagePersona.py +61 -61
  19. edsl/auto/StagePersonaDimensionValueRanges.py +88 -88
  20. edsl/auto/StagePersonaDimensionValues.py +74 -74
  21. edsl/auto/StagePersonaDimensions.py +69 -69
  22. edsl/auto/StageQuestions.py +73 -73
  23. edsl/auto/SurveyCreatorPipeline.py +21 -21
  24. edsl/auto/utilities.py +224 -224
  25. edsl/base/Base.py +279 -279
  26. edsl/config.py +157 -157
  27. edsl/conversation/Conversation.py +290 -290
  28. edsl/conversation/car_buying.py +58 -58
  29. edsl/conversation/chips.py +95 -95
  30. edsl/conversation/mug_negotiation.py +81 -81
  31. edsl/conversation/next_speaker_utilities.py +93 -93
  32. edsl/coop/PriceFetcher.py +54 -54
  33. edsl/coop/__init__.py +2 -2
  34. edsl/coop/coop.py +1028 -1028
  35. edsl/coop/utils.py +131 -131
  36. edsl/data/Cache.py +555 -555
  37. edsl/data/CacheEntry.py +233 -233
  38. edsl/data/CacheHandler.py +149 -149
  39. edsl/data/RemoteCacheSync.py +78 -78
  40. edsl/data/SQLiteDict.py +292 -292
  41. edsl/data/__init__.py +4 -4
  42. edsl/data/orm.py +10 -10
  43. edsl/data_transfer_models.py +73 -73
  44. edsl/enums.py +175 -175
  45. edsl/exceptions/BaseException.py +21 -21
  46. edsl/exceptions/__init__.py +54 -54
  47. edsl/exceptions/agents.py +42 -42
  48. edsl/exceptions/cache.py +5 -5
  49. edsl/exceptions/configuration.py +16 -16
  50. edsl/exceptions/coop.py +10 -10
  51. edsl/exceptions/data.py +14 -14
  52. edsl/exceptions/general.py +34 -34
  53. edsl/exceptions/jobs.py +33 -33
  54. edsl/exceptions/language_models.py +63 -63
  55. edsl/exceptions/prompts.py +15 -15
  56. edsl/exceptions/questions.py +91 -91
  57. edsl/exceptions/results.py +29 -29
  58. edsl/exceptions/scenarios.py +22 -22
  59. edsl/exceptions/surveys.py +37 -37
  60. edsl/inference_services/AnthropicService.py +87 -87
  61. edsl/inference_services/AwsBedrock.py +120 -120
  62. edsl/inference_services/AzureAI.py +217 -217
  63. edsl/inference_services/DeepInfraService.py +18 -18
  64. edsl/inference_services/GoogleService.py +148 -148
  65. edsl/inference_services/GroqService.py +20 -20
  66. edsl/inference_services/InferenceServiceABC.py +147 -147
  67. edsl/inference_services/InferenceServicesCollection.py +97 -97
  68. edsl/inference_services/MistralAIService.py +123 -123
  69. edsl/inference_services/OllamaService.py +18 -18
  70. edsl/inference_services/OpenAIService.py +224 -224
  71. edsl/inference_services/PerplexityService.py +163 -163
  72. edsl/inference_services/TestService.py +89 -89
  73. edsl/inference_services/TogetherAIService.py +170 -170
  74. edsl/inference_services/models_available_cache.py +118 -118
  75. edsl/inference_services/rate_limits_cache.py +25 -25
  76. edsl/inference_services/registry.py +41 -41
  77. edsl/inference_services/write_available.py +10 -10
  78. edsl/jobs/Answers.py +56 -56
  79. edsl/jobs/Jobs.py +898 -898
  80. edsl/jobs/JobsChecks.py +147 -147
  81. edsl/jobs/JobsPrompts.py +268 -268
  82. edsl/jobs/JobsRemoteInferenceHandler.py +239 -239
  83. edsl/jobs/__init__.py +1 -1
  84. edsl/jobs/buckets/BucketCollection.py +63 -63
  85. edsl/jobs/buckets/ModelBuckets.py +65 -65
  86. edsl/jobs/buckets/TokenBucket.py +251 -251
  87. edsl/jobs/interviews/Interview.py +661 -661
  88. edsl/jobs/interviews/InterviewExceptionCollection.py +99 -99
  89. edsl/jobs/interviews/InterviewExceptionEntry.py +186 -186
  90. edsl/jobs/interviews/InterviewStatistic.py +63 -63
  91. edsl/jobs/interviews/InterviewStatisticsCollection.py +25 -25
  92. edsl/jobs/interviews/InterviewStatusDictionary.py +78 -78
  93. edsl/jobs/interviews/InterviewStatusLog.py +92 -92
  94. edsl/jobs/interviews/ReportErrors.py +66 -66
  95. edsl/jobs/interviews/interview_status_enum.py +9 -9
  96. edsl/jobs/runners/JobsRunnerAsyncio.py +466 -466
  97. edsl/jobs/runners/JobsRunnerStatus.py +330 -330
  98. edsl/jobs/tasks/QuestionTaskCreator.py +242 -242
  99. edsl/jobs/tasks/TaskCreators.py +64 -64
  100. edsl/jobs/tasks/TaskHistory.py +450 -450
  101. edsl/jobs/tasks/TaskStatusLog.py +23 -23
  102. edsl/jobs/tasks/task_status_enum.py +163 -163
  103. edsl/jobs/tokens/InterviewTokenUsage.py +27 -27
  104. edsl/jobs/tokens/TokenUsage.py +34 -34
  105. edsl/language_models/KeyLookup.py +30 -30
  106. edsl/language_models/LanguageModel.py +668 -668
  107. edsl/language_models/ModelList.py +155 -155
  108. edsl/language_models/RegisterLanguageModelsMeta.py +184 -184
  109. edsl/language_models/__init__.py +3 -3
  110. edsl/language_models/fake_openai_call.py +15 -15
  111. edsl/language_models/fake_openai_service.py +61 -61
  112. edsl/language_models/registry.py +190 -190
  113. edsl/language_models/repair.py +156 -156
  114. edsl/language_models/unused/ReplicateBase.py +83 -83
  115. edsl/language_models/utilities.py +64 -64
  116. edsl/notebooks/Notebook.py +258 -258
  117. edsl/notebooks/__init__.py +1 -1
  118. edsl/prompts/Prompt.py +362 -362
  119. edsl/prompts/__init__.py +2 -2
  120. edsl/questions/AnswerValidatorMixin.py +289 -289
  121. edsl/questions/QuestionBase.py +664 -664
  122. edsl/questions/QuestionBaseGenMixin.py +161 -161
  123. edsl/questions/QuestionBasePromptsMixin.py +217 -217
  124. edsl/questions/QuestionBudget.py +227 -227
  125. edsl/questions/QuestionCheckBox.py +359 -359
  126. edsl/questions/QuestionExtract.py +182 -182
  127. edsl/questions/QuestionFreeText.py +114 -114
  128. edsl/questions/QuestionFunctional.py +166 -166
  129. edsl/questions/QuestionList.py +231 -231
  130. edsl/questions/QuestionMultipleChoice.py +286 -286
  131. edsl/questions/QuestionNumerical.py +153 -153
  132. edsl/questions/QuestionRank.py +324 -324
  133. edsl/questions/Quick.py +41 -41
  134. edsl/questions/RegisterQuestionsMeta.py +71 -71
  135. edsl/questions/ResponseValidatorABC.py +174 -174
  136. edsl/questions/SimpleAskMixin.py +73 -73
  137. edsl/questions/__init__.py +26 -26
  138. edsl/questions/compose_questions.py +98 -98
  139. edsl/questions/decorators.py +21 -21
  140. edsl/questions/derived/QuestionLikertFive.py +76 -76
  141. edsl/questions/derived/QuestionLinearScale.py +87 -87
  142. edsl/questions/derived/QuestionTopK.py +93 -93
  143. edsl/questions/derived/QuestionYesNo.py +82 -82
  144. edsl/questions/descriptors.py +413 -413
  145. edsl/questions/prompt_templates/question_budget.jinja +13 -13
  146. edsl/questions/prompt_templates/question_checkbox.jinja +32 -32
  147. edsl/questions/prompt_templates/question_extract.jinja +11 -11
  148. edsl/questions/prompt_templates/question_free_text.jinja +3 -3
  149. edsl/questions/prompt_templates/question_linear_scale.jinja +11 -11
  150. edsl/questions/prompt_templates/question_list.jinja +17 -17
  151. edsl/questions/prompt_templates/question_multiple_choice.jinja +33 -33
  152. edsl/questions/prompt_templates/question_numerical.jinja +36 -36
  153. edsl/questions/question_registry.py +177 -177
  154. edsl/questions/settings.py +12 -12
  155. edsl/questions/templates/budget/answering_instructions.jinja +7 -7
  156. edsl/questions/templates/budget/question_presentation.jinja +7 -7
  157. edsl/questions/templates/checkbox/answering_instructions.jinja +10 -10
  158. edsl/questions/templates/checkbox/question_presentation.jinja +22 -22
  159. edsl/questions/templates/extract/answering_instructions.jinja +7 -7
  160. edsl/questions/templates/likert_five/answering_instructions.jinja +10 -10
  161. edsl/questions/templates/likert_five/question_presentation.jinja +11 -11
  162. edsl/questions/templates/linear_scale/answering_instructions.jinja +5 -5
  163. edsl/questions/templates/linear_scale/question_presentation.jinja +5 -5
  164. edsl/questions/templates/list/answering_instructions.jinja +3 -3
  165. edsl/questions/templates/list/question_presentation.jinja +5 -5
  166. edsl/questions/templates/multiple_choice/answering_instructions.jinja +9 -9
  167. edsl/questions/templates/multiple_choice/question_presentation.jinja +11 -11
  168. edsl/questions/templates/numerical/answering_instructions.jinja +6 -6
  169. edsl/questions/templates/numerical/question_presentation.jinja +6 -6
  170. edsl/questions/templates/rank/answering_instructions.jinja +11 -11
  171. edsl/questions/templates/rank/question_presentation.jinja +15 -15
  172. edsl/questions/templates/top_k/answering_instructions.jinja +8 -8
  173. edsl/questions/templates/top_k/question_presentation.jinja +22 -22
  174. edsl/questions/templates/yes_no/answering_instructions.jinja +6 -6
  175. edsl/questions/templates/yes_no/question_presentation.jinja +11 -11
  176. edsl/results/CSSParameterizer.py +108 -108
  177. edsl/results/Dataset.py +424 -424
  178. edsl/results/DatasetExportMixin.py +731 -731
  179. edsl/results/DatasetTree.py +275 -275
  180. edsl/results/Result.py +465 -465
  181. edsl/results/Results.py +1165 -1165
  182. edsl/results/ResultsDBMixin.py +238 -238
  183. edsl/results/ResultsExportMixin.py +43 -43
  184. edsl/results/ResultsFetchMixin.py +33 -33
  185. edsl/results/ResultsGGMixin.py +121 -121
  186. edsl/results/ResultsToolsMixin.py +98 -98
  187. edsl/results/Selector.py +135 -135
  188. edsl/results/TableDisplay.py +198 -198
  189. edsl/results/__init__.py +2 -2
  190. edsl/results/table_display.css +77 -77
  191. edsl/results/tree_explore.py +115 -115
  192. edsl/scenarios/FileStore.py +632 -632
  193. edsl/scenarios/Scenario.py +601 -601
  194. edsl/scenarios/ScenarioHtmlMixin.py +64 -64
  195. edsl/scenarios/ScenarioJoin.py +127 -127
  196. edsl/scenarios/ScenarioList.py +1287 -1287
  197. edsl/scenarios/ScenarioListExportMixin.py +52 -52
  198. edsl/scenarios/ScenarioListPdfMixin.py +261 -261
  199. edsl/scenarios/__init__.py +4 -4
  200. edsl/shared.py +1 -1
  201. edsl/study/ObjectEntry.py +173 -173
  202. edsl/study/ProofOfWork.py +113 -113
  203. edsl/study/SnapShot.py +80 -80
  204. edsl/study/Study.py +528 -528
  205. edsl/study/__init__.py +4 -4
  206. edsl/surveys/DAG.py +148 -148
  207. edsl/surveys/Memory.py +31 -31
  208. edsl/surveys/MemoryPlan.py +244 -244
  209. edsl/surveys/Rule.py +326 -326
  210. edsl/surveys/RuleCollection.py +387 -387
  211. edsl/surveys/Survey.py +1801 -1801
  212. edsl/surveys/SurveyCSS.py +261 -261
  213. edsl/surveys/SurveyExportMixin.py +259 -259
  214. edsl/surveys/SurveyFlowVisualizationMixin.py +179 -179
  215. edsl/surveys/SurveyQualtricsImport.py +284 -284
  216. edsl/surveys/__init__.py +3 -3
  217. edsl/surveys/base.py +53 -53
  218. edsl/surveys/descriptors.py +56 -56
  219. edsl/surveys/instructions/ChangeInstruction.py +49 -49
  220. edsl/surveys/instructions/Instruction.py +65 -65
  221. edsl/surveys/instructions/InstructionCollection.py +77 -77
  222. edsl/templates/error_reporting/base.html +23 -23
  223. edsl/templates/error_reporting/exceptions_by_model.html +34 -34
  224. edsl/templates/error_reporting/exceptions_by_question_name.html +16 -16
  225. edsl/templates/error_reporting/exceptions_by_type.html +16 -16
  226. edsl/templates/error_reporting/interview_details.html +115 -115
  227. edsl/templates/error_reporting/interviews.html +19 -19
  228. edsl/templates/error_reporting/overview.html +4 -4
  229. edsl/templates/error_reporting/performance_plot.html +1 -1
  230. edsl/templates/error_reporting/report.css +73 -73
  231. edsl/templates/error_reporting/report.html +117 -117
  232. edsl/templates/error_reporting/report.js +25 -25
  233. edsl/tools/__init__.py +1 -1
  234. edsl/tools/clusters.py +192 -192
  235. edsl/tools/embeddings.py +27 -27
  236. edsl/tools/embeddings_plotting.py +118 -118
  237. edsl/tools/plotting.py +112 -112
  238. edsl/tools/summarize.py +18 -18
  239. edsl/utilities/SystemInfo.py +28 -28
  240. edsl/utilities/__init__.py +22 -22
  241. edsl/utilities/ast_utilities.py +25 -25
  242. edsl/utilities/data/Registry.py +6 -6
  243. edsl/utilities/data/__init__.py +1 -1
  244. edsl/utilities/data/scooter_results.json +1 -1
  245. edsl/utilities/decorators.py +77 -77
  246. edsl/utilities/gcp_bucket/cloud_storage.py +96 -96
  247. edsl/utilities/interface.py +627 -627
  248. edsl/utilities/naming_utilities.py +263 -263
  249. edsl/utilities/repair_functions.py +28 -28
  250. edsl/utilities/restricted_python.py +70 -70
  251. edsl/utilities/utilities.py +424 -424
  252. {edsl-0.1.39.dev1.dist-info → edsl-0.1.39.dev3.dist-info}/LICENSE +21 -21
  253. {edsl-0.1.39.dev1.dist-info → edsl-0.1.39.dev3.dist-info}/METADATA +1 -1
  254. edsl-0.1.39.dev3.dist-info/RECORD +277 -0
  255. edsl-0.1.39.dev1.dist-info/RECORD +0 -277
  256. {edsl-0.1.39.dev1.dist-info → edsl-0.1.39.dev3.dist-info}/WHEEL +0 -0
@@ -1,148 +1,148 @@
1
- import os
2
- from typing import Any, Dict, List, Optional
3
- import google
4
- import google.generativeai as genai
5
- from google.generativeai.types import GenerationConfig
6
- from google.api_core.exceptions import InvalidArgument
7
-
8
- from edsl.exceptions import MissingAPIKeyError
9
- from edsl.language_models.LanguageModel import LanguageModel
10
- from edsl.inference_services.InferenceServiceABC import InferenceServiceABC
11
- from edsl.coop import Coop
12
-
13
- safety_settings = [
14
- {
15
- "category": "HARM_CATEGORY_HARASSMENT",
16
- "threshold": "BLOCK_NONE",
17
- },
18
- {
19
- "category": "HARM_CATEGORY_HATE_SPEECH",
20
- "threshold": "BLOCK_NONE",
21
- },
22
- {
23
- "category": "HARM_CATEGORY_SEXUALLY_EXPLICIT",
24
- "threshold": "BLOCK_NONE",
25
- },
26
- {
27
- "category": "HARM_CATEGORY_DANGEROUS_CONTENT",
28
- "threshold": "BLOCK_NONE",
29
- },
30
- ]
31
-
32
-
33
- class GoogleService(InferenceServiceABC):
34
- _inference_service_ = "google"
35
- key_sequence = ["candidates", 0, "content", "parts", 0, "text"]
36
- usage_sequence = ["usage_metadata"]
37
- input_token_name = "prompt_token_count"
38
- output_token_name = "candidates_token_count"
39
-
40
- model_exclude_list = []
41
-
42
- # @classmethod
43
- # def available(cls) -> List[str]:
44
- # return ["gemini-pro", "gemini-1.5-pro", "gemini-1.5-flash", "gemini-1.0-pro"]
45
-
46
- @classmethod
47
- def available(cls) -> List[str]:
48
- model_list = []
49
- for m in genai.list_models():
50
- if "generateContent" in m.supported_generation_methods:
51
- model_list.append(m.name.split("/")[-1])
52
- return model_list
53
-
54
- @classmethod
55
- def create_model(
56
- cls, model_name: str = "gemini-pro", model_class_name=None
57
- ) -> LanguageModel:
58
- if model_class_name is None:
59
- model_class_name = cls.to_class_name(model_name)
60
-
61
- class LLM(LanguageModel):
62
- _model_ = model_name
63
- key_sequence = cls.key_sequence
64
- usage_sequence = cls.usage_sequence
65
- input_token_name = cls.input_token_name
66
- output_token_name = cls.output_token_name
67
- _inference_service_ = cls._inference_service_
68
-
69
- _tpm = cls.get_tpm(cls)
70
- _rpm = cls.get_rpm(cls)
71
-
72
- _parameters_ = {
73
- "temperature": 0.5,
74
- "topP": 1,
75
- "topK": 1,
76
- "maxOutputTokens": 2048,
77
- "stopSequences": [],
78
- }
79
-
80
- api_token = None
81
- model = None
82
-
83
- def __init__(self, *args, **kwargs):
84
- super().__init__(*args, **kwargs)
85
-
86
- def get_generation_config(self) -> GenerationConfig:
87
- return GenerationConfig(
88
- temperature=self.temperature,
89
- top_p=self.topP,
90
- top_k=self.topK,
91
- max_output_tokens=self.maxOutputTokens,
92
- stop_sequences=self.stopSequences,
93
- )
94
-
95
- async def async_execute_model_call(
96
- self,
97
- user_prompt: str,
98
- system_prompt: str = "",
99
- files_list: Optional["Files"] = None,
100
- ) -> Dict[str, Any]:
101
- generation_config = self.get_generation_config()
102
-
103
- if files_list is None:
104
- files_list = []
105
-
106
- genai.configure(api_key=self.api_token)
107
- if (
108
- system_prompt is not None
109
- and system_prompt != ""
110
- and self._model_ != "gemini-pro"
111
- ):
112
- try:
113
- self.generative_model = genai.GenerativeModel(
114
- self._model_,
115
- safety_settings=safety_settings,
116
- system_instruction=system_prompt,
117
- )
118
- except InvalidArgument as e:
119
- print(
120
- f"This model, {self._model_}, does not support system_instruction"
121
- )
122
- print("Will add system_prompt to user_prompt")
123
- user_prompt = f"{system_prompt}\n{user_prompt}"
124
- else:
125
- self.generative_model = genai.GenerativeModel(
126
- self._model_,
127
- safety_settings=safety_settings,
128
- )
129
- combined_prompt = [user_prompt]
130
- for file in files_list:
131
- if "google" not in file.external_locations:
132
- _ = file.upload_google()
133
- gen_ai_file = google.generativeai.types.file_types.File(
134
- file.external_locations["google"]
135
- )
136
- combined_prompt.append(gen_ai_file)
137
-
138
- response = await self.generative_model.generate_content_async(
139
- combined_prompt, generation_config=generation_config
140
- )
141
- return response.to_dict()
142
-
143
- LLM.__name__ = model_name
144
- return LLM
145
-
146
-
147
- if __name__ == "__main__":
148
- pass
1
+ import os
2
+ from typing import Any, Dict, List, Optional
3
+ import google
4
+ import google.generativeai as genai
5
+ from google.generativeai.types import GenerationConfig
6
+ from google.api_core.exceptions import InvalidArgument
7
+
8
+ from edsl.exceptions import MissingAPIKeyError
9
+ from edsl.language_models.LanguageModel import LanguageModel
10
+ from edsl.inference_services.InferenceServiceABC import InferenceServiceABC
11
+ from edsl.coop import Coop
12
+
13
+ safety_settings = [
14
+ {
15
+ "category": "HARM_CATEGORY_HARASSMENT",
16
+ "threshold": "BLOCK_NONE",
17
+ },
18
+ {
19
+ "category": "HARM_CATEGORY_HATE_SPEECH",
20
+ "threshold": "BLOCK_NONE",
21
+ },
22
+ {
23
+ "category": "HARM_CATEGORY_SEXUALLY_EXPLICIT",
24
+ "threshold": "BLOCK_NONE",
25
+ },
26
+ {
27
+ "category": "HARM_CATEGORY_DANGEROUS_CONTENT",
28
+ "threshold": "BLOCK_NONE",
29
+ },
30
+ ]
31
+
32
+
33
+ class GoogleService(InferenceServiceABC):
34
+ _inference_service_ = "google"
35
+ key_sequence = ["candidates", 0, "content", "parts", 0, "text"]
36
+ usage_sequence = ["usage_metadata"]
37
+ input_token_name = "prompt_token_count"
38
+ output_token_name = "candidates_token_count"
39
+
40
+ model_exclude_list = []
41
+
42
+ # @classmethod
43
+ # def available(cls) -> List[str]:
44
+ # return ["gemini-pro", "gemini-1.5-pro", "gemini-1.5-flash", "gemini-1.0-pro"]
45
+
46
+ @classmethod
47
+ def available(cls) -> List[str]:
48
+ model_list = []
49
+ for m in genai.list_models():
50
+ if "generateContent" in m.supported_generation_methods:
51
+ model_list.append(m.name.split("/")[-1])
52
+ return model_list
53
+
54
+ @classmethod
55
+ def create_model(
56
+ cls, model_name: str = "gemini-pro", model_class_name=None
57
+ ) -> LanguageModel:
58
+ if model_class_name is None:
59
+ model_class_name = cls.to_class_name(model_name)
60
+
61
+ class LLM(LanguageModel):
62
+ _model_ = model_name
63
+ key_sequence = cls.key_sequence
64
+ usage_sequence = cls.usage_sequence
65
+ input_token_name = cls.input_token_name
66
+ output_token_name = cls.output_token_name
67
+ _inference_service_ = cls._inference_service_
68
+
69
+ _tpm = cls.get_tpm(cls)
70
+ _rpm = cls.get_rpm(cls)
71
+
72
+ _parameters_ = {
73
+ "temperature": 0.5,
74
+ "topP": 1,
75
+ "topK": 1,
76
+ "maxOutputTokens": 2048,
77
+ "stopSequences": [],
78
+ }
79
+
80
+ api_token = None
81
+ model = None
82
+
83
+ def __init__(self, *args, **kwargs):
84
+ super().__init__(*args, **kwargs)
85
+
86
+ def get_generation_config(self) -> GenerationConfig:
87
+ return GenerationConfig(
88
+ temperature=self.temperature,
89
+ top_p=self.topP,
90
+ top_k=self.topK,
91
+ max_output_tokens=self.maxOutputTokens,
92
+ stop_sequences=self.stopSequences,
93
+ )
94
+
95
+ async def async_execute_model_call(
96
+ self,
97
+ user_prompt: str,
98
+ system_prompt: str = "",
99
+ files_list: Optional["Files"] = None,
100
+ ) -> Dict[str, Any]:
101
+ generation_config = self.get_generation_config()
102
+
103
+ if files_list is None:
104
+ files_list = []
105
+
106
+ genai.configure(api_key=self.api_token)
107
+ if (
108
+ system_prompt is not None
109
+ and system_prompt != ""
110
+ and self._model_ != "gemini-pro"
111
+ ):
112
+ try:
113
+ self.generative_model = genai.GenerativeModel(
114
+ self._model_,
115
+ safety_settings=safety_settings,
116
+ system_instruction=system_prompt,
117
+ )
118
+ except InvalidArgument as e:
119
+ print(
120
+ f"This model, {self._model_}, does not support system_instruction"
121
+ )
122
+ print("Will add system_prompt to user_prompt")
123
+ user_prompt = f"{system_prompt}\n{user_prompt}"
124
+ else:
125
+ self.generative_model = genai.GenerativeModel(
126
+ self._model_,
127
+ safety_settings=safety_settings,
128
+ )
129
+ combined_prompt = [user_prompt]
130
+ for file in files_list:
131
+ if "google" not in file.external_locations:
132
+ _ = file.upload_google()
133
+ gen_ai_file = google.generativeai.types.file_types.File(
134
+ file.external_locations["google"]
135
+ )
136
+ combined_prompt.append(gen_ai_file)
137
+
138
+ response = await self.generative_model.generate_content_async(
139
+ combined_prompt, generation_config=generation_config
140
+ )
141
+ return response.to_dict()
142
+
143
+ LLM.__name__ = model_name
144
+ return LLM
145
+
146
+
147
+ if __name__ == "__main__":
148
+ pass
@@ -1,20 +1,20 @@
1
- from typing import Any, List
2
- from edsl.inference_services.OpenAIService import OpenAIService
3
-
4
- import groq
5
-
6
-
7
- class GroqService(OpenAIService):
8
- """DeepInfra service class."""
9
-
10
- _inference_service_ = "groq"
11
- _env_key_name_ = "GROQ_API_KEY"
12
-
13
- _sync_client_ = groq.Groq
14
- _async_client_ = groq.AsyncGroq
15
-
16
- model_exclude_list = ["whisper-large-v3", "distil-whisper-large-v3-en"]
17
-
18
- # _base_url_ = "https://api.deepinfra.com/v1/openai"
19
- _base_url_ = None
20
- _models_list_cache: List[str] = []
1
+ from typing import Any, List
2
+ from edsl.inference_services.OpenAIService import OpenAIService
3
+
4
+ import groq
5
+
6
+
7
+ class GroqService(OpenAIService):
8
+ """DeepInfra service class."""
9
+
10
+ _inference_service_ = "groq"
11
+ _env_key_name_ = "GROQ_API_KEY"
12
+
13
+ _sync_client_ = groq.Groq
14
+ _async_client_ = groq.AsyncGroq
15
+
16
+ model_exclude_list = ["whisper-large-v3", "distil-whisper-large-v3-en"]
17
+
18
+ # _base_url_ = "https://api.deepinfra.com/v1/openai"
19
+ _base_url_ = None
20
+ _models_list_cache: List[str] = []
@@ -1,147 +1,147 @@
1
- from abc import abstractmethod, ABC
2
- import os
3
- import re
4
- from datetime import datetime, timedelta
5
- from edsl.config import CONFIG
6
-
7
-
8
- class InferenceServiceABC(ABC):
9
- """
10
- Abstract class for inference services.
11
- Anthropic: https://docs.anthropic.com/en/api/rate-limits
12
- """
13
-
14
- _coop_config_vars = None
15
-
16
- default_levels = {
17
- "google": {"tpm": 2_000_000, "rpm": 15},
18
- "openai": {"tpm": 2_000_000, "rpm": 10_000},
19
- "anthropic": {"tpm": 2_000_000, "rpm": 500},
20
- }
21
-
22
- def __init_subclass__(cls):
23
- """
24
- Check that the subclass has the required attributes.
25
- - `key_sequence` attribute determines...
26
- - `model_exclude_list` attribute determines...
27
- """
28
- if not hasattr(cls, "key_sequence"):
29
- raise NotImplementedError(
30
- f"Class {cls.__name__} must have a 'key_sequence' attribute."
31
- )
32
- if not hasattr(cls, "model_exclude_list"):
33
- raise NotImplementedError(
34
- f"Class {cls.__name__} must have a 'model_exclude_list' attribute."
35
- )
36
-
37
- @classmethod
38
- def _should_refresh_coop_config_vars(cls):
39
- """
40
- Returns True if config vars have been fetched over 24 hours ago, and False otherwise.
41
- """
42
-
43
- if cls._last_config_fetch is None:
44
- return True
45
- return (datetime.now() - cls._last_config_fetch) > timedelta(hours=24)
46
-
47
- @classmethod
48
- def _get_limt(cls, limit_type: str) -> int:
49
- key = f"EDSL_SERVICE_{limit_type.upper()}_{cls._inference_service_.upper()}"
50
- if key in os.environ:
51
- return int(os.getenv(key))
52
-
53
- if cls._coop_config_vars is None or cls._should_refresh_coop_config_vars():
54
- try:
55
- from edsl import Coop
56
-
57
- c = Coop()
58
- cls._coop_config_vars = c.fetch_rate_limit_config_vars()
59
- cls._last_config_fetch = datetime.now()
60
- if key in cls._coop_config_vars:
61
- return cls._coop_config_vars[key]
62
- except Exception:
63
- cls._coop_config_vars = None
64
- else:
65
- if key in cls._coop_config_vars:
66
- return cls._coop_config_vars[key]
67
-
68
- if cls._inference_service_ in cls.default_levels:
69
- return int(cls.default_levels[cls._inference_service_][limit_type])
70
-
71
- return int(CONFIG.get(f"EDSL_SERVICE_{limit_type.upper()}_BASELINE"))
72
-
73
- def get_tpm(cls) -> int:
74
- """
75
- Returns the TPM for the service. If the service is not defined in the environment variables, it will return the baseline TPM.
76
- """
77
- return cls._get_limt(limit_type="tpm")
78
-
79
- def get_rpm(cls):
80
- """
81
- Returns the RPM for the service. If the service is not defined in the environment variables, it will return the baseline RPM.
82
- """
83
- return cls._get_limt(limit_type="rpm")
84
-
85
- @abstractmethod
86
- def available() -> list[str]:
87
- """
88
- Returns a list of available models for the service.
89
- """
90
- pass
91
-
92
- @abstractmethod
93
- def create_model():
94
- """
95
- Returns a LanguageModel object.
96
- """
97
- pass
98
-
99
- @staticmethod
100
- def to_class_name(s):
101
- """
102
- Converts a string to a valid class name.
103
-
104
- >>> InferenceServiceABC.to_class_name("hello world")
105
- 'HelloWorld'
106
- """
107
-
108
- s = re.sub(r"[^a-zA-Z0-9 ]", "", s)
109
- s = "".join(word.title() for word in s.split())
110
- if s and s[0].isdigit():
111
- s = "Class" + s
112
- return s
113
-
114
-
115
- if __name__ == "__main__":
116
- pass
117
- # deep_infra_service = DeepInfraService("deep_infra", "DEEP_INFRA_API_KEY")
118
- # deep_infra_service.available()
119
- # m = deep_infra_service.create_model("microsoft/WizardLM-2-7B")
120
- # response = m().hello()
121
- # print(response)
122
-
123
- # anthropic_service = AnthropicService("anthropic", "ANTHROPIC_API_KEY")
124
- # anthropic_service.available()
125
- # m = anthropic_service.create_model("claude-3-opus-20240229")
126
- # response = m().hello()
127
- # print(response)
128
- # factory = OpenAIService("openai", "OPENAI_API")
129
- # factory.available()
130
- # m = factory.create_model("gpt-3.5-turbo")
131
- # response = m().hello()
132
-
133
- # from edsl import QuestionFreeText
134
- # results = QuestionFreeText.example().by(m()).run()
135
-
136
- # collection = InferenceServicesCollection([
137
- # OpenAIService,
138
- # AnthropicService,
139
- # DeepInfraService
140
- # ])
141
-
142
- # available = collection.available()
143
- # factory = collection.create_model_factory(*available[0])
144
- # m = factory()
145
- # from edsl import QuestionFreeText
146
- # results = QuestionFreeText.example().by(m).run()
147
- # print(results)
1
+ from abc import abstractmethod, ABC
2
+ import os
3
+ import re
4
+ from datetime import datetime, timedelta
5
+ from edsl.config import CONFIG
6
+
7
+
8
+ class InferenceServiceABC(ABC):
9
+ """
10
+ Abstract class for inference services.
11
+ Anthropic: https://docs.anthropic.com/en/api/rate-limits
12
+ """
13
+
14
+ _coop_config_vars = None
15
+
16
+ default_levels = {
17
+ "google": {"tpm": 2_000_000, "rpm": 15},
18
+ "openai": {"tpm": 2_000_000, "rpm": 10_000},
19
+ "anthropic": {"tpm": 2_000_000, "rpm": 500},
20
+ }
21
+
22
+ def __init_subclass__(cls):
23
+ """
24
+ Check that the subclass has the required attributes.
25
+ - `key_sequence` attribute determines...
26
+ - `model_exclude_list` attribute determines...
27
+ """
28
+ if not hasattr(cls, "key_sequence"):
29
+ raise NotImplementedError(
30
+ f"Class {cls.__name__} must have a 'key_sequence' attribute."
31
+ )
32
+ if not hasattr(cls, "model_exclude_list"):
33
+ raise NotImplementedError(
34
+ f"Class {cls.__name__} must have a 'model_exclude_list' attribute."
35
+ )
36
+
37
+ @classmethod
38
+ def _should_refresh_coop_config_vars(cls):
39
+ """
40
+ Returns True if config vars have been fetched over 24 hours ago, and False otherwise.
41
+ """
42
+
43
+ if cls._last_config_fetch is None:
44
+ return True
45
+ return (datetime.now() - cls._last_config_fetch) > timedelta(hours=24)
46
+
47
+ @classmethod
48
+ def _get_limt(cls, limit_type: str) -> int:
49
+ key = f"EDSL_SERVICE_{limit_type.upper()}_{cls._inference_service_.upper()}"
50
+ if key in os.environ:
51
+ return int(os.getenv(key))
52
+
53
+ if cls._coop_config_vars is None or cls._should_refresh_coop_config_vars():
54
+ try:
55
+ from edsl import Coop
56
+
57
+ c = Coop()
58
+ cls._coop_config_vars = c.fetch_rate_limit_config_vars()
59
+ cls._last_config_fetch = datetime.now()
60
+ if key in cls._coop_config_vars:
61
+ return cls._coop_config_vars[key]
62
+ except Exception:
63
+ cls._coop_config_vars = None
64
+ else:
65
+ if key in cls._coop_config_vars:
66
+ return cls._coop_config_vars[key]
67
+
68
+ if cls._inference_service_ in cls.default_levels:
69
+ return int(cls.default_levels[cls._inference_service_][limit_type])
70
+
71
+ return int(CONFIG.get(f"EDSL_SERVICE_{limit_type.upper()}_BASELINE"))
72
+
73
+ def get_tpm(cls) -> int:
74
+ """
75
+ Returns the TPM for the service. If the service is not defined in the environment variables, it will return the baseline TPM.
76
+ """
77
+ return cls._get_limt(limit_type="tpm")
78
+
79
+ def get_rpm(cls):
80
+ """
81
+ Returns the RPM for the service. If the service is not defined in the environment variables, it will return the baseline RPM.
82
+ """
83
+ return cls._get_limt(limit_type="rpm")
84
+
85
+ @abstractmethod
86
+ def available() -> list[str]:
87
+ """
88
+ Returns a list of available models for the service.
89
+ """
90
+ pass
91
+
92
+ @abstractmethod
93
+ def create_model():
94
+ """
95
+ Returns a LanguageModel object.
96
+ """
97
+ pass
98
+
99
+ @staticmethod
100
+ def to_class_name(s):
101
+ """
102
+ Converts a string to a valid class name.
103
+
104
+ >>> InferenceServiceABC.to_class_name("hello world")
105
+ 'HelloWorld'
106
+ """
107
+
108
+ s = re.sub(r"[^a-zA-Z0-9 ]", "", s)
109
+ s = "".join(word.title() for word in s.split())
110
+ if s and s[0].isdigit():
111
+ s = "Class" + s
112
+ return s
113
+
114
+
115
+ if __name__ == "__main__":
116
+ pass
117
+ # deep_infra_service = DeepInfraService("deep_infra", "DEEP_INFRA_API_KEY")
118
+ # deep_infra_service.available()
119
+ # m = deep_infra_service.create_model("microsoft/WizardLM-2-7B")
120
+ # response = m().hello()
121
+ # print(response)
122
+
123
+ # anthropic_service = AnthropicService("anthropic", "ANTHROPIC_API_KEY")
124
+ # anthropic_service.available()
125
+ # m = anthropic_service.create_model("claude-3-opus-20240229")
126
+ # response = m().hello()
127
+ # print(response)
128
+ # factory = OpenAIService("openai", "OPENAI_API")
129
+ # factory.available()
130
+ # m = factory.create_model("gpt-3.5-turbo")
131
+ # response = m().hello()
132
+
133
+ # from edsl import QuestionFreeText
134
+ # results = QuestionFreeText.example().by(m()).run()
135
+
136
+ # collection = InferenceServicesCollection([
137
+ # OpenAIService,
138
+ # AnthropicService,
139
+ # DeepInfraService
140
+ # ])
141
+
142
+ # available = collection.available()
143
+ # factory = collection.create_model_factory(*available[0])
144
+ # m = factory()
145
+ # from edsl import QuestionFreeText
146
+ # results = QuestionFreeText.example().by(m).run()
147
+ # print(results)