edsl 0.1.39.dev3__py3-none-any.whl → 0.1.39.dev4__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (344) hide show
  1. edsl/Base.py +413 -332
  2. edsl/BaseDiff.py +260 -260
  3. edsl/TemplateLoader.py +24 -24
  4. edsl/__init__.py +57 -49
  5. edsl/__version__.py +1 -1
  6. edsl/agents/Agent.py +1071 -867
  7. edsl/agents/AgentList.py +551 -413
  8. edsl/agents/Invigilator.py +284 -233
  9. edsl/agents/InvigilatorBase.py +257 -270
  10. edsl/agents/PromptConstructor.py +272 -354
  11. edsl/agents/QuestionInstructionPromptBuilder.py +128 -0
  12. edsl/agents/QuestionTemplateReplacementsBuilder.py +137 -0
  13. edsl/agents/__init__.py +2 -3
  14. edsl/agents/descriptors.py +99 -99
  15. edsl/agents/prompt_helpers.py +129 -129
  16. edsl/agents/question_option_processor.py +172 -0
  17. edsl/auto/AutoStudy.py +130 -117
  18. edsl/auto/StageBase.py +243 -230
  19. edsl/auto/StageGenerateSurvey.py +178 -178
  20. edsl/auto/StageLabelQuestions.py +125 -125
  21. edsl/auto/StagePersona.py +61 -61
  22. edsl/auto/StagePersonaDimensionValueRanges.py +88 -88
  23. edsl/auto/StagePersonaDimensionValues.py +74 -74
  24. edsl/auto/StagePersonaDimensions.py +69 -69
  25. edsl/auto/StageQuestions.py +74 -73
  26. edsl/auto/SurveyCreatorPipeline.py +21 -21
  27. edsl/auto/utilities.py +218 -224
  28. edsl/base/Base.py +279 -279
  29. edsl/config.py +177 -157
  30. edsl/conversation/Conversation.py +290 -290
  31. edsl/conversation/car_buying.py +59 -58
  32. edsl/conversation/chips.py +95 -95
  33. edsl/conversation/mug_negotiation.py +81 -81
  34. edsl/conversation/next_speaker_utilities.py +93 -93
  35. edsl/coop/CoopFunctionsMixin.py +15 -0
  36. edsl/coop/ExpectedParrotKeyHandler.py +125 -0
  37. edsl/coop/PriceFetcher.py +54 -54
  38. edsl/coop/__init__.py +2 -2
  39. edsl/coop/coop.py +1106 -1028
  40. edsl/coop/utils.py +131 -131
  41. edsl/data/Cache.py +573 -555
  42. edsl/data/CacheEntry.py +230 -233
  43. edsl/data/CacheHandler.py +168 -149
  44. edsl/data/RemoteCacheSync.py +186 -78
  45. edsl/data/SQLiteDict.py +292 -292
  46. edsl/data/__init__.py +5 -4
  47. edsl/data/hack.py +10 -0
  48. edsl/data/orm.py +10 -10
  49. edsl/data_transfer_models.py +74 -73
  50. edsl/enums.py +202 -175
  51. edsl/exceptions/BaseException.py +21 -21
  52. edsl/exceptions/__init__.py +54 -54
  53. edsl/exceptions/agents.py +54 -42
  54. edsl/exceptions/cache.py +5 -5
  55. edsl/exceptions/configuration.py +16 -16
  56. edsl/exceptions/coop.py +10 -10
  57. edsl/exceptions/data.py +14 -14
  58. edsl/exceptions/general.py +34 -34
  59. edsl/exceptions/inference_services.py +5 -0
  60. edsl/exceptions/jobs.py +33 -33
  61. edsl/exceptions/language_models.py +63 -63
  62. edsl/exceptions/prompts.py +15 -15
  63. edsl/exceptions/questions.py +109 -91
  64. edsl/exceptions/results.py +29 -29
  65. edsl/exceptions/scenarios.py +29 -22
  66. edsl/exceptions/surveys.py +37 -37
  67. edsl/inference_services/AnthropicService.py +106 -87
  68. edsl/inference_services/AvailableModelCacheHandler.py +184 -0
  69. edsl/inference_services/AvailableModelFetcher.py +215 -0
  70. edsl/inference_services/AwsBedrock.py +118 -120
  71. edsl/inference_services/AzureAI.py +215 -217
  72. edsl/inference_services/DeepInfraService.py +18 -18
  73. edsl/inference_services/GoogleService.py +143 -148
  74. edsl/inference_services/GroqService.py +20 -20
  75. edsl/inference_services/InferenceServiceABC.py +80 -147
  76. edsl/inference_services/InferenceServicesCollection.py +138 -97
  77. edsl/inference_services/MistralAIService.py +120 -123
  78. edsl/inference_services/OllamaService.py +18 -18
  79. edsl/inference_services/OpenAIService.py +236 -224
  80. edsl/inference_services/PerplexityService.py +160 -163
  81. edsl/inference_services/ServiceAvailability.py +135 -0
  82. edsl/inference_services/TestService.py +90 -89
  83. edsl/inference_services/TogetherAIService.py +172 -170
  84. edsl/inference_services/data_structures.py +134 -0
  85. edsl/inference_services/models_available_cache.py +118 -118
  86. edsl/inference_services/rate_limits_cache.py +25 -25
  87. edsl/inference_services/registry.py +41 -41
  88. edsl/inference_services/write_available.py +10 -10
  89. edsl/jobs/AnswerQuestionFunctionConstructor.py +223 -0
  90. edsl/jobs/Answers.py +43 -56
  91. edsl/jobs/FetchInvigilator.py +47 -0
  92. edsl/jobs/InterviewTaskManager.py +98 -0
  93. edsl/jobs/InterviewsConstructor.py +50 -0
  94. edsl/jobs/Jobs.py +823 -898
  95. edsl/jobs/JobsChecks.py +172 -147
  96. edsl/jobs/JobsComponentConstructor.py +189 -0
  97. edsl/jobs/JobsPrompts.py +270 -268
  98. edsl/jobs/JobsRemoteInferenceHandler.py +311 -239
  99. edsl/jobs/JobsRemoteInferenceLogger.py +239 -0
  100. edsl/jobs/RequestTokenEstimator.py +30 -0
  101. edsl/jobs/__init__.py +1 -1
  102. edsl/jobs/async_interview_runner.py +138 -0
  103. edsl/jobs/buckets/BucketCollection.py +104 -63
  104. edsl/jobs/buckets/ModelBuckets.py +65 -65
  105. edsl/jobs/buckets/TokenBucket.py +283 -251
  106. edsl/jobs/buckets/TokenBucketAPI.py +211 -0
  107. edsl/jobs/buckets/TokenBucketClient.py +191 -0
  108. edsl/jobs/check_survey_scenario_compatibility.py +85 -0
  109. edsl/jobs/data_structures.py +120 -0
  110. edsl/jobs/decorators.py +35 -0
  111. edsl/jobs/interviews/Interview.py +396 -661
  112. edsl/jobs/interviews/InterviewExceptionCollection.py +99 -99
  113. edsl/jobs/interviews/InterviewExceptionEntry.py +186 -186
  114. edsl/jobs/interviews/InterviewStatistic.py +63 -63
  115. edsl/jobs/interviews/InterviewStatisticsCollection.py +25 -25
  116. edsl/jobs/interviews/InterviewStatusDictionary.py +78 -78
  117. edsl/jobs/interviews/InterviewStatusLog.py +92 -92
  118. edsl/jobs/interviews/ReportErrors.py +66 -66
  119. edsl/jobs/interviews/interview_status_enum.py +9 -9
  120. edsl/jobs/jobs_status_enums.py +9 -0
  121. edsl/jobs/loggers/HTMLTableJobLogger.py +304 -0
  122. edsl/jobs/results_exceptions_handler.py +98 -0
  123. edsl/jobs/runners/JobsRunnerAsyncio.py +151 -466
  124. edsl/jobs/runners/JobsRunnerStatus.py +297 -330
  125. edsl/jobs/tasks/QuestionTaskCreator.py +244 -242
  126. edsl/jobs/tasks/TaskCreators.py +64 -64
  127. edsl/jobs/tasks/TaskHistory.py +470 -450
  128. edsl/jobs/tasks/TaskStatusLog.py +23 -23
  129. edsl/jobs/tasks/task_status_enum.py +161 -163
  130. edsl/jobs/tokens/InterviewTokenUsage.py +27 -27
  131. edsl/jobs/tokens/TokenUsage.py +34 -34
  132. edsl/language_models/ComputeCost.py +63 -0
  133. edsl/language_models/LanguageModel.py +626 -668
  134. edsl/language_models/ModelList.py +164 -155
  135. edsl/language_models/PriceManager.py +127 -0
  136. edsl/language_models/RawResponseHandler.py +106 -0
  137. edsl/language_models/RegisterLanguageModelsMeta.py +184 -184
  138. edsl/language_models/ServiceDataSources.py +0 -0
  139. edsl/language_models/__init__.py +2 -3
  140. edsl/language_models/fake_openai_call.py +15 -15
  141. edsl/language_models/fake_openai_service.py +61 -61
  142. edsl/language_models/key_management/KeyLookup.py +63 -0
  143. edsl/language_models/key_management/KeyLookupBuilder.py +273 -0
  144. edsl/language_models/key_management/KeyLookupCollection.py +38 -0
  145. edsl/language_models/key_management/__init__.py +0 -0
  146. edsl/language_models/key_management/models.py +131 -0
  147. edsl/language_models/model.py +256 -0
  148. edsl/language_models/repair.py +156 -156
  149. edsl/language_models/utilities.py +65 -64
  150. edsl/notebooks/Notebook.py +263 -258
  151. edsl/notebooks/NotebookToLaTeX.py +142 -0
  152. edsl/notebooks/__init__.py +1 -1
  153. edsl/prompts/Prompt.py +352 -362
  154. edsl/prompts/__init__.py +2 -2
  155. edsl/questions/ExceptionExplainer.py +77 -0
  156. edsl/questions/HTMLQuestion.py +103 -0
  157. edsl/questions/QuestionBase.py +518 -664
  158. edsl/questions/QuestionBasePromptsMixin.py +221 -217
  159. edsl/questions/QuestionBudget.py +227 -227
  160. edsl/questions/QuestionCheckBox.py +359 -359
  161. edsl/questions/QuestionExtract.py +180 -182
  162. edsl/questions/QuestionFreeText.py +113 -114
  163. edsl/questions/QuestionFunctional.py +166 -166
  164. edsl/questions/QuestionList.py +223 -231
  165. edsl/questions/QuestionMatrix.py +265 -0
  166. edsl/questions/QuestionMultipleChoice.py +330 -286
  167. edsl/questions/QuestionNumerical.py +151 -153
  168. edsl/questions/QuestionRank.py +314 -324
  169. edsl/questions/Quick.py +41 -41
  170. edsl/questions/SimpleAskMixin.py +74 -73
  171. edsl/questions/__init__.py +27 -26
  172. edsl/questions/{AnswerValidatorMixin.py → answer_validator_mixin.py} +334 -289
  173. edsl/questions/compose_questions.py +98 -98
  174. edsl/questions/data_structures.py +20 -0
  175. edsl/questions/decorators.py +21 -21
  176. edsl/questions/derived/QuestionLikertFive.py +76 -76
  177. edsl/questions/derived/QuestionLinearScale.py +90 -87
  178. edsl/questions/derived/QuestionTopK.py +93 -93
  179. edsl/questions/derived/QuestionYesNo.py +82 -82
  180. edsl/questions/descriptors.py +427 -413
  181. edsl/questions/loop_processor.py +149 -0
  182. edsl/questions/prompt_templates/question_budget.jinja +13 -13
  183. edsl/questions/prompt_templates/question_checkbox.jinja +32 -32
  184. edsl/questions/prompt_templates/question_extract.jinja +11 -11
  185. edsl/questions/prompt_templates/question_free_text.jinja +3 -3
  186. edsl/questions/prompt_templates/question_linear_scale.jinja +11 -11
  187. edsl/questions/prompt_templates/question_list.jinja +17 -17
  188. edsl/questions/prompt_templates/question_multiple_choice.jinja +33 -33
  189. edsl/questions/prompt_templates/question_numerical.jinja +36 -36
  190. edsl/questions/{QuestionBaseGenMixin.py → question_base_gen_mixin.py} +168 -161
  191. edsl/questions/question_registry.py +177 -177
  192. edsl/questions/{RegisterQuestionsMeta.py → register_questions_meta.py} +71 -71
  193. edsl/questions/{ResponseValidatorABC.py → response_validator_abc.py} +188 -174
  194. edsl/questions/response_validator_factory.py +34 -0
  195. edsl/questions/settings.py +12 -12
  196. edsl/questions/templates/budget/answering_instructions.jinja +7 -7
  197. edsl/questions/templates/budget/question_presentation.jinja +7 -7
  198. edsl/questions/templates/checkbox/answering_instructions.jinja +10 -10
  199. edsl/questions/templates/checkbox/question_presentation.jinja +22 -22
  200. edsl/questions/templates/extract/answering_instructions.jinja +7 -7
  201. edsl/questions/templates/likert_five/answering_instructions.jinja +10 -10
  202. edsl/questions/templates/likert_five/question_presentation.jinja +11 -11
  203. edsl/questions/templates/linear_scale/answering_instructions.jinja +5 -5
  204. edsl/questions/templates/linear_scale/question_presentation.jinja +5 -5
  205. edsl/questions/templates/list/answering_instructions.jinja +3 -3
  206. edsl/questions/templates/list/question_presentation.jinja +5 -5
  207. edsl/questions/templates/matrix/__init__.py +1 -0
  208. edsl/questions/templates/matrix/answering_instructions.jinja +5 -0
  209. edsl/questions/templates/matrix/question_presentation.jinja +20 -0
  210. edsl/questions/templates/multiple_choice/answering_instructions.jinja +9 -9
  211. edsl/questions/templates/multiple_choice/question_presentation.jinja +11 -11
  212. edsl/questions/templates/numerical/answering_instructions.jinja +6 -6
  213. edsl/questions/templates/numerical/question_presentation.jinja +6 -6
  214. edsl/questions/templates/rank/answering_instructions.jinja +11 -11
  215. edsl/questions/templates/rank/question_presentation.jinja +15 -15
  216. edsl/questions/templates/top_k/answering_instructions.jinja +8 -8
  217. edsl/questions/templates/top_k/question_presentation.jinja +22 -22
  218. edsl/questions/templates/yes_no/answering_instructions.jinja +6 -6
  219. edsl/questions/templates/yes_no/question_presentation.jinja +11 -11
  220. edsl/results/CSSParameterizer.py +108 -108
  221. edsl/results/Dataset.py +587 -424
  222. edsl/results/DatasetExportMixin.py +594 -731
  223. edsl/results/DatasetTree.py +295 -275
  224. edsl/results/MarkdownToDocx.py +122 -0
  225. edsl/results/MarkdownToPDF.py +111 -0
  226. edsl/results/Result.py +557 -465
  227. edsl/results/Results.py +1183 -1165
  228. edsl/results/ResultsExportMixin.py +45 -43
  229. edsl/results/ResultsGGMixin.py +121 -121
  230. edsl/results/TableDisplay.py +125 -198
  231. edsl/results/TextEditor.py +50 -0
  232. edsl/results/__init__.py +2 -2
  233. edsl/results/file_exports.py +252 -0
  234. edsl/results/{ResultsFetchMixin.py → results_fetch_mixin.py} +33 -33
  235. edsl/results/{Selector.py → results_selector.py} +145 -135
  236. edsl/results/{ResultsToolsMixin.py → results_tools_mixin.py} +98 -98
  237. edsl/results/smart_objects.py +96 -0
  238. edsl/results/table_data_class.py +12 -0
  239. edsl/results/table_display.css +77 -77
  240. edsl/results/table_renderers.py +118 -0
  241. edsl/results/tree_explore.py +115 -115
  242. edsl/scenarios/ConstructDownloadLink.py +109 -0
  243. edsl/scenarios/DocumentChunker.py +102 -0
  244. edsl/scenarios/DocxScenario.py +16 -0
  245. edsl/scenarios/FileStore.py +511 -632
  246. edsl/scenarios/PdfExtractor.py +40 -0
  247. edsl/scenarios/Scenario.py +498 -601
  248. edsl/scenarios/ScenarioHtmlMixin.py +65 -64
  249. edsl/scenarios/ScenarioList.py +1458 -1287
  250. edsl/scenarios/ScenarioListExportMixin.py +45 -52
  251. edsl/scenarios/ScenarioListPdfMixin.py +239 -261
  252. edsl/scenarios/__init__.py +3 -4
  253. edsl/scenarios/directory_scanner.py +96 -0
  254. edsl/scenarios/file_methods.py +85 -0
  255. edsl/scenarios/handlers/__init__.py +13 -0
  256. edsl/scenarios/handlers/csv.py +38 -0
  257. edsl/scenarios/handlers/docx.py +76 -0
  258. edsl/scenarios/handlers/html.py +37 -0
  259. edsl/scenarios/handlers/json.py +111 -0
  260. edsl/scenarios/handlers/latex.py +5 -0
  261. edsl/scenarios/handlers/md.py +51 -0
  262. edsl/scenarios/handlers/pdf.py +68 -0
  263. edsl/scenarios/handlers/png.py +39 -0
  264. edsl/scenarios/handlers/pptx.py +105 -0
  265. edsl/scenarios/handlers/py.py +294 -0
  266. edsl/scenarios/handlers/sql.py +313 -0
  267. edsl/scenarios/handlers/sqlite.py +149 -0
  268. edsl/scenarios/handlers/txt.py +33 -0
  269. edsl/scenarios/{ScenarioJoin.py → scenario_join.py} +131 -127
  270. edsl/scenarios/scenario_selector.py +156 -0
  271. edsl/shared.py +1 -1
  272. edsl/study/ObjectEntry.py +173 -173
  273. edsl/study/ProofOfWork.py +113 -113
  274. edsl/study/SnapShot.py +80 -80
  275. edsl/study/Study.py +521 -528
  276. edsl/study/__init__.py +4 -4
  277. edsl/surveys/ConstructDAG.py +92 -0
  278. edsl/surveys/DAG.py +148 -148
  279. edsl/surveys/EditSurvey.py +221 -0
  280. edsl/surveys/InstructionHandler.py +100 -0
  281. edsl/surveys/Memory.py +31 -31
  282. edsl/surveys/MemoryManagement.py +72 -0
  283. edsl/surveys/MemoryPlan.py +244 -244
  284. edsl/surveys/Rule.py +327 -326
  285. edsl/surveys/RuleCollection.py +385 -387
  286. edsl/surveys/RuleManager.py +172 -0
  287. edsl/surveys/Simulator.py +75 -0
  288. edsl/surveys/Survey.py +1280 -1801
  289. edsl/surveys/SurveyCSS.py +273 -261
  290. edsl/surveys/SurveyExportMixin.py +259 -259
  291. edsl/surveys/{SurveyFlowVisualizationMixin.py → SurveyFlowVisualization.py} +181 -179
  292. edsl/surveys/SurveyQualtricsImport.py +284 -284
  293. edsl/surveys/SurveyToApp.py +141 -0
  294. edsl/surveys/__init__.py +5 -3
  295. edsl/surveys/base.py +53 -53
  296. edsl/surveys/descriptors.py +60 -56
  297. edsl/surveys/instructions/ChangeInstruction.py +48 -49
  298. edsl/surveys/instructions/Instruction.py +56 -65
  299. edsl/surveys/instructions/InstructionCollection.py +82 -77
  300. edsl/templates/error_reporting/base.html +23 -23
  301. edsl/templates/error_reporting/exceptions_by_model.html +34 -34
  302. edsl/templates/error_reporting/exceptions_by_question_name.html +16 -16
  303. edsl/templates/error_reporting/exceptions_by_type.html +16 -16
  304. edsl/templates/error_reporting/interview_details.html +115 -115
  305. edsl/templates/error_reporting/interviews.html +19 -19
  306. edsl/templates/error_reporting/overview.html +4 -4
  307. edsl/templates/error_reporting/performance_plot.html +1 -1
  308. edsl/templates/error_reporting/report.css +73 -73
  309. edsl/templates/error_reporting/report.html +117 -117
  310. edsl/templates/error_reporting/report.js +25 -25
  311. edsl/test_h +1 -0
  312. edsl/tools/__init__.py +1 -1
  313. edsl/tools/clusters.py +192 -192
  314. edsl/tools/embeddings.py +27 -27
  315. edsl/tools/embeddings_plotting.py +118 -118
  316. edsl/tools/plotting.py +112 -112
  317. edsl/tools/summarize.py +18 -18
  318. edsl/utilities/PrettyList.py +56 -0
  319. edsl/utilities/SystemInfo.py +28 -28
  320. edsl/utilities/__init__.py +22 -22
  321. edsl/utilities/ast_utilities.py +25 -25
  322. edsl/utilities/data/Registry.py +6 -6
  323. edsl/utilities/data/__init__.py +1 -1
  324. edsl/utilities/data/scooter_results.json +1 -1
  325. edsl/utilities/decorators.py +77 -77
  326. edsl/utilities/gcp_bucket/cloud_storage.py +96 -96
  327. edsl/utilities/gcp_bucket/example.py +50 -0
  328. edsl/utilities/interface.py +627 -627
  329. edsl/utilities/is_notebook.py +18 -0
  330. edsl/utilities/is_valid_variable_name.py +11 -0
  331. edsl/utilities/naming_utilities.py +263 -263
  332. edsl/utilities/remove_edsl_version.py +24 -0
  333. edsl/utilities/repair_functions.py +28 -28
  334. edsl/utilities/restricted_python.py +70 -70
  335. edsl/utilities/utilities.py +436 -424
  336. {edsl-0.1.39.dev3.dist-info → edsl-0.1.39.dev4.dist-info}/LICENSE +21 -21
  337. {edsl-0.1.39.dev3.dist-info → edsl-0.1.39.dev4.dist-info}/METADATA +13 -11
  338. edsl-0.1.39.dev4.dist-info/RECORD +361 -0
  339. edsl/language_models/KeyLookup.py +0 -30
  340. edsl/language_models/registry.py +0 -190
  341. edsl/language_models/unused/ReplicateBase.py +0 -83
  342. edsl/results/ResultsDBMixin.py +0 -238
  343. edsl-0.1.39.dev3.dist-info/RECORD +0 -277
  344. {edsl-0.1.39.dev3.dist-info → edsl-0.1.39.dev4.dist-info}/WHEEL +0 -0
@@ -1,74 +1,74 @@
1
- body {
2
- font-family: Arial, sans-serif;
3
- line-height: 1.6;
4
- background-color: #f9f9f9;
5
- color: #333;
6
- margin: 20px;
7
- }
8
-
9
- .interview {
10
- font-size: 1.5em;
11
- margin-bottom: 10px;
12
- padding: 10px;
13
- background-color: #e3f2fd;
14
- border-left: 5px solid #2196f3;
15
- }
16
-
17
- .question {
18
- font-size: 1.2em;
19
- margin-bottom: 10px;
20
- padding: 10px;
21
- background-color: #fff9c4;
22
- border-left: 5px solid #ffeb3b;
23
- }
24
-
25
- .exception-detail {
26
- margin-bottom: 10px;
27
- background-color: #ffebee;
28
- border-left: 5px solid #f44336;
29
- }
30
-
31
- .exception-header {
32
- padding: 10px;
33
- cursor: pointer;
34
- display: flex;
35
- justify-content: space-between;
36
- align-items: center;
37
- }
38
-
39
- .exception-content {
40
- padding: 10px;
41
- display: none;
42
- }
43
-
44
- .exception-content.show {
45
- display: block;
46
- }
47
-
48
- .question-detail {
49
- border: 3px solid black;
50
- padding: 10px;
51
- }
52
-
53
- .exception-exception {
54
- font-weight: bold;
55
- color: #d32f2f;
56
- }
57
-
58
- .exception-time,
59
- .exception-traceback {
60
- font-style: italic;
61
- color: #555;
62
- }
63
-
64
- .toggle-btn {
65
- background: none;
66
- border: none;
67
- font-size: 1.2em;
68
- cursor: pointer;
69
- transition: transform 0.3s;
70
- }
71
-
72
- .toggle-btn.rotated {
73
- transform: rotate(180deg);
1
+ body {
2
+ font-family: Arial, sans-serif;
3
+ line-height: 1.6;
4
+ background-color: #f9f9f9;
5
+ color: #333;
6
+ margin: 20px;
7
+ }
8
+
9
+ .interview {
10
+ font-size: 1.5em;
11
+ margin-bottom: 10px;
12
+ padding: 10px;
13
+ background-color: #e3f2fd;
14
+ border-left: 5px solid #2196f3;
15
+ }
16
+
17
+ .question {
18
+ font-size: 1.2em;
19
+ margin-bottom: 10px;
20
+ padding: 10px;
21
+ background-color: #fff9c4;
22
+ border-left: 5px solid #ffeb3b;
23
+ }
24
+
25
+ .exception-detail {
26
+ margin-bottom: 10px;
27
+ background-color: #ffebee;
28
+ border-left: 5px solid #f44336;
29
+ }
30
+
31
+ .exception-header {
32
+ padding: 10px;
33
+ cursor: pointer;
34
+ display: flex;
35
+ justify-content: space-between;
36
+ align-items: center;
37
+ }
38
+
39
+ .exception-content {
40
+ padding: 10px;
41
+ display: none;
42
+ }
43
+
44
+ .exception-content.show {
45
+ display: block;
46
+ }
47
+
48
+ .question-detail {
49
+ border: 3px solid black;
50
+ padding: 10px;
51
+ }
52
+
53
+ .exception-exception {
54
+ font-weight: bold;
55
+ color: #d32f2f;
56
+ }
57
+
58
+ .exception-time,
59
+ .exception-traceback {
60
+ font-style: italic;
61
+ color: #555;
62
+ }
63
+
64
+ .toggle-btn {
65
+ background: none;
66
+ border: none;
67
+ font-size: 1.2em;
68
+ cursor: pointer;
69
+ transition: transform 0.3s;
70
+ }
71
+
72
+ .toggle-btn.rotated {
73
+ transform: rotate(180deg);
74
74
  }
@@ -1,118 +1,118 @@
1
- <!DOCTYPE html>
2
- <html lang="en">
3
- <head>
4
- <meta charset="UTF-8">
5
- <meta name="viewport" content="width=device-width, initial-scale=1.0">
6
- <title>Exception Details</title>
7
- <style>
8
- {{ css }}
9
- </style>
10
- </head>
11
- <body>
12
- <h1>Overview</h1>
13
- <p>There were {{ interviews|length }} total interviews. The number of interviews with exceptions was {{ num_exceptions }}.</p>
14
- <p>The models used were: {{ models_used }}.</p>
15
- <p>For documentation on dealing with exceptions on Expected Parrot,
16
- see <a href="https://docs.expectedparrot.com/en/latest/exceptions.html">here</a>.</p>
17
-
18
- <h2>Exceptions by Type</h2>
19
- <table>
20
- <thead>
21
- <tr>
22
- <th>Exception Type</th>
23
- <th>Number</th>
24
- </tr>
25
- </thead>
26
- <tbody>
27
- {% for exception_type, exceptions in exceptions_by_type.items() %}
28
- <tr>
29
- <td>{{ exception_type }}</td>
30
- <td>{{ exceptions }}</td>
31
- </tr>
32
- {% endfor %}
33
- </tbody>
34
- </table>
35
-
36
-
37
- <h2>Exceptions by Model</h2>
38
- <table>
39
- <thead>
40
- <tr>
41
- <th>Model</th>
42
- <th>Number</th>
43
- </tr>
44
- </thead>
45
- <tbody>
46
- {% for model, exceptions in exceptions_by_model.items() %}
47
- <tr>
48
- <td>{{ model }}</td>
49
- <td>{{ exceptions }}</td>
50
- </tr>
51
- {% endfor %}
52
- </tbody>
53
- </table>
54
-
55
-
56
- <h2>Exceptions by Question Name</h2>
57
- <table>
58
- <thead>
59
- <tr>
60
- <th>Question Name</th>
61
- <th>Number of Exceptions</th>
62
- </tr>
63
- </thead>
64
- <tbody>
65
- {% for question_name, exception_count in exceptions_by_question_name.items() %}
66
- <tr>
67
- <td>{{ question_name }}</td>
68
- <td>{{ exception_count }}</td>
69
- </tr>
70
- {% endfor %}
71
- </tbody>
72
- </table>
73
-
74
-
75
- {% for index, interview in interviews.items() %}
76
- {% if interview.exceptions != {} %}
77
- <div class="interview">Interview: {{ index }} </div>
78
- <h1>Failing questions</h1>
79
- {% endif %}
80
- {% for question, exceptions in interview.exceptions.items() %}
81
- <div class="question">question_name: {{ question }}</div>
82
-
83
- <h2>Question</h2>
84
- <div class="question-detail">
85
- {{ interview.survey.get_question(question).html() }}
86
- </div>
87
-
88
- <h2>Scenario</h2>
89
- <div class="scenario">
90
- {{ interview.scenario._repr_html_() }}
91
- </div>
92
-
93
- <h2>Agent</h2>
94
- <div class="agent">
95
- {{ interview.agent._repr_html_() }}
96
- </div>
97
-
98
- <h2>Model</h2>
99
- <div class="model">
100
- {{ interview.model._repr_html_() }}
101
- </div>
102
-
103
- <h2>Exception details</h2>
104
-
105
- {% for exception_message in exceptions %}
106
- <div class="exception-detail">
107
- <div class="exception-exception">Exception: {{ exception_message.exception }}</div>
108
- <div class="exception-time">Time: {{ exception_message.time }}</div>
109
- <div class="exception-traceback">Traceback: <pre>{{ exception_message.traceback }} </pre></div>
110
- </div>
111
- {% endfor %}
112
- {% endfor %}
113
- {% endfor %}
114
-
115
- <h1>Performance Plot</h1>
116
- {{ performance_plot_html }}
117
- </body>
1
+ <!DOCTYPE html>
2
+ <html lang="en">
3
+ <head>
4
+ <meta charset="UTF-8">
5
+ <meta name="viewport" content="width=device-width, initial-scale=1.0">
6
+ <title>Exception Details</title>
7
+ <style>
8
+ {{ css }}
9
+ </style>
10
+ </head>
11
+ <body>
12
+ <h1>Overview</h1>
13
+ <p>There were {{ interviews|length }} total interviews. The number of interviews with exceptions was {{ num_exceptions }}.</p>
14
+ <p>The models used were: {{ models_used }}.</p>
15
+ <p>For documentation on dealing with exceptions on Expected Parrot,
16
+ see <a href="https://docs.expectedparrot.com/en/latest/exceptions.html">here</a>.</p>
17
+
18
+ <h2>Exceptions by Type</h2>
19
+ <table>
20
+ <thead>
21
+ <tr>
22
+ <th>Exception Type</th>
23
+ <th>Number</th>
24
+ </tr>
25
+ </thead>
26
+ <tbody>
27
+ {% for exception_type, exceptions in exceptions_by_type.items() %}
28
+ <tr>
29
+ <td>{{ exception_type }}</td>
30
+ <td>{{ exceptions }}</td>
31
+ </tr>
32
+ {% endfor %}
33
+ </tbody>
34
+ </table>
35
+
36
+
37
+ <h2>Exceptions by Model</h2>
38
+ <table>
39
+ <thead>
40
+ <tr>
41
+ <th>Model</th>
42
+ <th>Number</th>
43
+ </tr>
44
+ </thead>
45
+ <tbody>
46
+ {% for model, exceptions in exceptions_by_model.items() %}
47
+ <tr>
48
+ <td>{{ model }}</td>
49
+ <td>{{ exceptions }}</td>
50
+ </tr>
51
+ {% endfor %}
52
+ </tbody>
53
+ </table>
54
+
55
+
56
+ <h2>Exceptions by Question Name</h2>
57
+ <table>
58
+ <thead>
59
+ <tr>
60
+ <th>Question Name</th>
61
+ <th>Number of Exceptions</th>
62
+ </tr>
63
+ </thead>
64
+ <tbody>
65
+ {% for question_name, exception_count in exceptions_by_question_name.items() %}
66
+ <tr>
67
+ <td>{{ question_name }}</td>
68
+ <td>{{ exception_count }}</td>
69
+ </tr>
70
+ {% endfor %}
71
+ </tbody>
72
+ </table>
73
+
74
+
75
+ {% for index, interview in interviews.items() %}
76
+ {% if interview.exceptions != {} %}
77
+ <div class="interview">Interview: {{ index }} </div>
78
+ <h1>Failing questions</h1>
79
+ {% endif %}
80
+ {% for question, exceptions in interview.exceptions.items() %}
81
+ <div class="question">question_name: {{ question }}</div>
82
+
83
+ <h2>Question</h2>
84
+ <div class="question-detail">
85
+ {{ interview.survey._get_question_by_name(question).html() }}
86
+ </div>
87
+
88
+ <h2>Scenario</h2>
89
+ <div class="scenario">
90
+ {{ interview.scenario._repr_html_() }}
91
+ </div>
92
+
93
+ <h2>Agent</h2>
94
+ <div class="agent">
95
+ {{ interview.agent._repr_html_() }}
96
+ </div>
97
+
98
+ <h2>Model</h2>
99
+ <div class="model">
100
+ {{ interview.model._repr_html_() }}
101
+ </div>
102
+
103
+ <h2>Exception details</h2>
104
+
105
+ {% for exception_message in exceptions %}
106
+ <div class="exception-detail">
107
+ <div class="exception-exception">Exception: {{ exception_message.exception }}</div>
108
+ <div class="exception-time">Time: {{ exception_message.time }}</div>
109
+ <div class="exception-traceback">Traceback: <pre>{{ exception_message.traceback }} </pre></div>
110
+ </div>
111
+ {% endfor %}
112
+ {% endfor %}
113
+ {% endfor %}
114
+
115
+ <h1>Performance Plot</h1>
116
+ {{ performance_plot_html }}
117
+ </body>
118
118
  </html>
@@ -1,25 +1,25 @@
1
- document.addEventListener('DOMContentLoaded', function() {
2
- const collapsibleSections = document.querySelectorAll('.exception-detail, .raw-model-response');
3
-
4
- collapsibleSections.forEach(section => {
5
- const header = section.querySelector('.exception-header, .response-header');
6
- const content = section.querySelector('.exception-content, .response-content');
7
- const toggleBtn = section.querySelector('.toggle-btn');
8
-
9
- header.addEventListener('click', function() {
10
- content.classList.toggle('show');
11
- toggleBtn.classList.toggle('rotated');
12
- });
13
- });
14
-
15
- });
16
-
17
- function copyCode() {
18
- const textarea = document.getElementById('codeToCopy');
19
- textarea.select();
20
- textarea.setSelectionRange(0, 99999); // For mobile devices
21
- document.execCommand("copy");
22
-
23
- // Optionally, you can display an alert or change the button text to indicate success
24
- alert("Code copied to clipboard!");
25
- }
1
+ document.addEventListener('DOMContentLoaded', function() {
2
+ const collapsibleSections = document.querySelectorAll('.exception-detail, .raw-model-response');
3
+
4
+ collapsibleSections.forEach(section => {
5
+ const header = section.querySelector('.exception-header, .response-header');
6
+ const content = section.querySelector('.exception-content, .response-content');
7
+ const toggleBtn = section.querySelector('.toggle-btn');
8
+
9
+ header.addEventListener('click', function() {
10
+ content.classList.toggle('show');
11
+ toggleBtn.classList.toggle('rotated');
12
+ });
13
+ });
14
+
15
+ });
16
+
17
+ function copyCode() {
18
+ const textarea = document.getElementById('codeToCopy');
19
+ textarea.select();
20
+ textarea.setSelectionRange(0, 99999); // For mobile devices
21
+ document.execCommand("copy");
22
+
23
+ // Optionally, you can display an alert or change the button text to indicate success
24
+ alert("Code copied to clipboard!");
25
+ }
edsl/test_h ADDED
@@ -0,0 +1 @@
1
+ [{'model_name': 'nvidia/Nemotron-4-340B-Instruct', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': 'Nemotron-4-340B-Instruct is a chat model intended for use for the English language, designed for Synthetic Data Generation', 'cover_img_url': 'https://shared.deepinfra.com/models/nvidia/Nemotron-4-340B-Instruct/cover_image.6223332b8dfbf8dba008d030548970e7dbee4cfb571412430cc5ff8c007b7072.webp', 'tags': [...], 'pricing': {...}, 'max_tokens': 4096, 'replaced_by': None, 'deprecated': None, 'quantization': 'bfloat16'}, {'model_name': 'meta-llama/Llama-2-7b-chat-hf', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': 'Llama 2 is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. This is the repository for the 7B fine-tuned model, optimized for dialogue use cases and converted for the Hugging Face Transformers format. ', 'cover_img_url': 'https://shared.deepinfra.com/models/meta-llama/Llama-2-7b-chat-hf/cover_image.10373e7a429dd725e0eb9e57cd20aeb815426c077217b27d9aedce37bd5c2173.jpg', 'tags': [...], 'pricing': {...}, 'max_tokens': 4096, 'replaced_by': 'meta-llama/Meta-Llama-3-8B-Instruct', 'deprecated': 1718309480, 'quantization': 'fp16'}, {'model_name': 'meta-llama/Llama-2-13b-chat-hf', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': 'Llama 2 is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. This is the repository for the 7B fine-tuned model, optimized for dialogue use cases and converted for the Hugging Face Transformers format. ', 'cover_img_url': 'https://shared.deepinfra.com/models/meta-llama/Llama-2-13b-chat-hf/cover_image.6cbca6a3445de3ac6e75983f8828fe6b6323ae8a2d3c025ae1561ac5e56e6326.jpg', 'tags': [...], 'pricing': {...}, 'max_tokens': 4096, 'replaced_by': 'meta-llama/Meta-Llama-3-8B-Instruct', 'deprecated': 1718228019, 'quantization': 'fp16'}, {'model_name': 'deepinfra/airoboros-70b', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': 'Latest version of the Airoboros model fine-tunned version of llama-2-70b using the Airoboros dataset. This model is currently running jondurbin/airoboros-l2-70b-2.2.1 ', 'cover_img_url': 'https://shared.deepinfra.com/models/deepinfra/airoboros-70b/cover_image.4df6a78233488497ec8cad9a032e070a1c2e0c510c7d3b2ebd5c159e15b69793.webp', 'tags': [...], 'pricing': {...}, 'max_tokens': 4096, 'replaced_by': 'lizpreciatior/lzlv_70b_fp16_hf', 'deprecated': 1718067221, 'quantization': 'fp16'}, {'model_name': '01-ai/Yi-34B-Chat', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': '', 'cover_img_url': 'https://shared.deepinfra.com/models/01-ai/Yi-34B-Chat/cover_image.ba50c187d0414ba799d008eef1756d102f2a016b33109023a5d05ac7370f21c8.webp', 'tags': [...], 'pricing': {...}, 'max_tokens': 4096, 'replaced_by': 'cognitivecomputations/dolphin-2.6-mixtral-8x7b', 'deprecated': None, 'quantization': 'fp16'}, {'model_name': 'mistralai/Mistral-7B-Instruct-v0.2', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': 'The Mistral-7B-Instruct-v0.2 Large Language Model (LLM) is a instruct fine-tuned version of the Mistral-7B-v0.2 generative text model using a variety of publicly available conversation datasets.', 'cover_img_url': 'https://shared.deepinfra.com/models/mistralai/Mistral-7B-Instruct-v0.2/cover_image.429fef8a2a09e5c4104ede511db12beaea9d3917c4754d709b05a65d5d6f6c1f.webp', 'tags': [...], 'pricing': {...}, 'max_tokens': 32768, 'replaced_by': None, 'deprecated': None, 'quantization': 'fp16'}, {'model_name': 'Qwen/Qwen2-7B-Instruct', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': 'The 7 billion parameter Qwen2 excels in language understanding, multilingual capabilities, coding, mathematics, and reasoning.', 'cover_img_url': 'https://shared.deepinfra.com/models/Qwen/Qwen2-7B-Instruct/cover_image.e8c2257ba46edbd20ef17c81343ad40fdf27f7bc3838837d079557accd2879e2.webp', 'tags': [...], 'pricing': {...}, 'max_tokens': 32768, 'replaced_by': None, 'deprecated': None, 'quantization': 'bfloat16'}, {'model_name': 'Qwen/Qwen2-72B-Instruct', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': 'The 72 billion parameter Qwen2 excels in language understanding, multilingual capabilities, coding, mathematics, and reasoning.', 'cover_img_url': 'https://shared.deepinfra.com/models/Qwen/Qwen2-72B-Instruct/cover_image.6771b7ff122c6ffaa9d5ae6b6ab54db0d46079a8f90441fa7770d60cfd4c6f4f.webp', 'tags': [...], 'pricing': {...}, 'max_tokens': 32768, 'replaced_by': None, 'deprecated': None, 'quantization': 'bfloat16'}, {'model_name': 'google/codegemma-7b-it', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': 'CodeGemma is a collection of lightweight open code models built on top of Gemma. CodeGemma models are text-to-text and text-to-code decoder-only models and are available as a 7 billion pretrained variant that specializes in code completion and code generation tasks, a 7 billion parameter instruction-tuned variant for code chat and instruction following and a 2 billion parameter pretrained variant for fast code completion.', 'cover_img_url': 'https://shared.deepinfra.com/models/google/codegemma-7b-it/cover_image.ecf88bf320526e3a113a1e51057501ffe65b8e264912e0c920ae7f95553f74b2.webp', 'tags': [...], 'pricing': {...}, 'max_tokens': 8192, 'replaced_by': 'google/gemma-1.1-7b-it', 'deprecated': 1718830428, 'quantization': 'fp16'}, {'model_name': 'mistralai/Mistral-7B-Instruct-v0.3', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': 'Mistral-7B-Instruct-v0.3 is an instruction-tuned model, next iteration of of Mistral 7B that has larger vocabulary, newer tokenizer and supports function calling.', 'cover_img_url': 'https://shared.deepinfra.com/models/mistralai/Mistral-7B-Instruct-v0.3/cover_image.716d64cdc98717436953bde0b80dede06d7071f66cb46f1df85caaa270e5cdd6.webp', 'tags': [...], 'pricing': {...}, 'max_tokens': 32768, 'replaced_by': None, 'deprecated': None, 'quantization': 'bfloat16'}, {'model_name': 'mistralai/Mixtral-8x7B-Instruct-v0.1', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': 'Mixtral is mixture of expert large language model (LLM) from Mistral AI. This is state of the art machine learning model using a mixture 8 of experts (MoE) 7b models. During inference 2 expers are selected. This architecture allows large models to be fast and cheap at inference. The Mixtral-8x7B outperforms Llama 2 70B on most benchmarks.', 'cover_img_url': 'https://shared.deepinfra.com/models/mistralai/Mixtral-8x7B-Instruct-v0.1/cover_image.a3146cc88bb3c77e6eae14b35d8db03d7952a597633a53378ef8182186c5a9d7.webp', 'tags': [...], 'pricing': {...}, 'max_tokens': 32768, 'replaced_by': None, 'deprecated': None, 'quantization': 'bfloat16'}, {'model_name': 'openchat/openchat_3.5', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': 'OpenChat is a library of open-source language models that have been fine-tuned with C-RLFT, a strategy inspired by offline reinforcement learning. These models can learn from mixed-quality data without preference labels and have achieved exceptional performance comparable to ChatGPT. The developers of OpenChat are dedicated to creating a high-performance, commercially viable, open-source large language model and are continuously making progress towards this goal.', 'cover_img_url': 'https://shared.deepinfra.com/models/openchat/openchat_3.5/cover_image.6112a8e07a704c30bd7c354351fa79c13904d9df7667a0064fb6b30bc80e728b.webp', 'tags': [...], 'pricing': {...}, 'max_tokens': 8192, 'replaced_by': None, 'deprecated': None, 'quantization': 'fp16'}, {'model_name': 'bigcode/starcoder2-15b-instruct-v0.1', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': 'We introduce StarCoder2-15B-Instruct-v0.1, the very first entirely self-aligned code Large Language Model (LLM) trained with a fully permissive and transparent pipeline. Our open-source pipeline uses StarCoder2-15B to generate thousands of instruction-response pairs, which are then used to fine-tune StarCoder-15B itself without any human annotations or distilled data from huge and proprietary LLMs.', 'cover_img_url': 'https://shared.deepinfra.com/models/bigcode/starcoder2-15b-instruct-v0.1/cover_image.f20836d5d430a93c8aa95eefc92154d14e93dc69b17967410ecce3187a07c601.webp', 'tags': [...], 'pricing': {...}, 'max_tokens': None, 'replaced_by': 'Phind/Phind-CodeLlama-34B-v2', 'deprecated': 1718830388, 'quantization': 'fp16'}, {'model_name': 'mistralai/Mistral-7B-Instruct-v0.1', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': 'The Mistral-7B-Instruct-v0.1 Large Language Model (LLM) is a instruct fine-tuned version of the Mistral-7B-v0.1 generative text model using a variety of publicly available conversation datasets.', 'cover_img_url': 'https://shared.deepinfra.com/models/mistralai/Mistral-7B-Instruct-v0.1/cover_image.cdb70679749ff93eb56b8480215bb8cd1382cbeffcec00a06bdb0145be9ef511.webp', 'tags': [...], 'pricing': {...}, 'max_tokens': 32768, 'replaced_by': None, 'deprecated': None, 'quantization': 'fp16'}, {'model_name': 'meta-llama/Meta-Llama-3-8B-Instruct', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': 'Meta developed and released the Meta Llama 3 family of large language models (LLMs), a collection of pretrained and instruction tuned generative text models in 8 and 70B sizes.', 'cover_img_url': 'https://shared.deepinfra.com/models/meta-llama/Meta-Llama-3-8B-Instruct/cover_image.9ea753fd36aabfbca4939ee488b859e08e95c4626ffff71ec3a385be66b1d3ba.webp', 'tags': [...], 'pricing': {...}, 'max_tokens': 8192, 'replaced_by': None, 'deprecated': None, 'quantization': 'bfloat16'}, {'model_name': 'codellama/CodeLlama-34b-Instruct-hf', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': 'Code Llama is a state-of-the-art LLM capable of generating code, and natural language about code, from both code and natural language prompts. This particular instance is the 34b instruct variant', 'cover_img_url': 'https://shared.deepinfra.com/models/codellama/CodeLlama-34b-Instruct-hf/cover_image.6b78b2c46fdafdd29a6eb2bcc388412b0afea3f4286d5ac90ab3c65e2da61be7.webp', 'tags': [...], 'pricing': {...}, 'max_tokens': 4096, 'replaced_by': 'Phind/Phind-CodeLlama-34B-v2', 'deprecated': None, 'quantization': 'fp16'}, {'model_name': 'HuggingFaceH4/zephyr-orpo-141b-A35b-v0.1', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': 'Zephyr 141B-A35B is an instruction-tuned (assistant) version of Mixtral-8x22B. It was fine-tuned on a mix of publicly available, synthetic datasets. It achieves strong performance on chat benchmarks.', 'cover_img_url': 'https://shared.deepinfra.com/models/HuggingFaceH4/zephyr-orpo-141b-A35b-v0.1/cover_image.06f8e4df28f71df5ed7d0f5c995f296ca5e259f6c9b3a3e8914ba58805fb83a6.webp', 'tags': [...], 'pricing': {...}, 'max_tokens': 65536, 'replaced_by': 'mistralai/Mixtral-8x22B-Instruct-v0.1', 'deprecated': None, 'quantization': 'fp8'}, {'model_name': 'Gryphe/MythoMax-L2-13b-turbo', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': 'Faster version of Gryphe/MythoMax-L2-13b running on multiple H100 cards in fp8 precision. Up to 160 tps. ', 'cover_img_url': 'https://shared.deepinfra.com/models/Gryphe/MythoMax-L2-13b-turbo/cover_image.1975a8dba4bcf6809a083fe29aff31cb2895a749171f89e0db650bc039da051a.webp', 'tags': [...], 'pricing': {...}, 'max_tokens': 4096, 'replaced_by': 'Gryphe/MythoMax-L2-13b', 'deprecated': 1718830497, 'quantization': 'fp8'}, {'model_name': 'codellama/CodeLlama-70b-Instruct-hf', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': 'CodeLlama-70b is the largest and latest code generation from the Code Llama collection. ', 'cover_img_url': 'https://shared.deepinfra.com/models/codellama/CodeLlama-70b-Instruct-hf/cover_image.6fc1f8b121f95b8075d0bff6f5d6fe39fdac93454078d88467697abe8ecd416b.webp', 'tags': [...], 'pricing': {...}, 'max_tokens': 4096, 'replaced_by': 'Phind/Phind-CodeLlama-34B-v2', 'deprecated': None, 'quantization': 'fp16'}, {'model_name': 'microsoft/WizardLM-2-8x22B', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': "WizardLM-2 8x22B is Microsoft AI's most advanced Wizard model. It demonstrates highly competitive performance compared to those leading proprietary models.", 'cover_img_url': 'https://shared.deepinfra.com/models/microsoft/WizardLM-2-8x22B/cover_image.395b63e0d661def89bf43c88976a699b066f69208b3b58ae5cc2663693033ee8.webp', 'tags': [...], 'pricing': {...}, 'max_tokens': 65536, 'replaced_by': None, 'deprecated': None, 'quantization': 'bfloat16'}, {'model_name': 'databricks/dbrx-instruct', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': 'DBRX is an open source LLM created by Databricks. It uses mixture-of-experts (MoE) architecture with 132B total parameters of which 36B parameters are active on any input. It outperforms existing open source LLMs like Llama 2 70B and Mixtral-8x7B on standard industry benchmarks for language understanding, programming, math, and logic.', 'cover_img_url': 'https://shared.deepinfra.com/models/databricks/dbrx-instruct/cover_image.061dbf63fed77872a3f1b97053081189ceefa2d6fafdc912d2d74111cfbd75c0.webp', 'tags': [...], 'pricing': {...}, 'max_tokens': 32768, 'replaced_by': 'cognitivecomputations/dolphin-2.6-mixtral-8x7b', 'deprecated': None, 'quantization': 'bfloat16'}, {'model_name': 'microsoft/WizardLM-2-7B', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': "WizardLM-2 7B is the smaller variant of Microsoft AI's latest Wizard model. It is the fastest and achieves comparable performance with existing 10x larger open-source leading models", 'cover_img_url': 'https://shared.deepinfra.com/models/microsoft/WizardLM-2-7B/cover_image.305876b2901f4833a14d17e09a3876b59ca7561b73017518970a75643fe9ff69.webp', 'tags': [...], 'pricing': {...}, 'max_tokens': 32768, 'replaced_by': None, 'deprecated': None, 'quantization': 'fp16'}, {'model_name': 'mistralai/Mixtral-8x22B-v0.1', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': 'Mixtral-8x22B is the latest and largest mixture of expert large language model (LLM) from Mistral AI. This is state of the art machine learning model using a mixture 8 of experts (MoE) 22b models. During inference 2 expers are selected. This architecture allows large models to be fast and cheap at inference. This model is not instruction tuned. ', 'cover_img_url': 'https://shared.deepinfra.com/models/mistralai/Mixtral-8x22B-v0.1/cover_image.eb92d1199149a5d7fa5e7b2dc17dc991f7398301747b92bd60032c3b7fc77a0f.webp', 'tags': [...], 'pricing': {...}, 'max_tokens': 65536, 'replaced_by': 'mistralai/Mixtral-8x22B-Instruct-v0.1', 'deprecated': None, 'quantization': 'fp16'}, {'model_name': 'mistralai/Mixtral-8x22B-Instruct-v0.1', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': 'This is the instruction fine-tuned version of Mixtral-8x22B - the latest and largest mixture of experts large language model (LLM) from Mistral AI. This state of the art machine learning model uses a mixture 8 of experts (MoE) 22b models. During inference 2 experts are selected. This architecture allows large models to be fast and cheap at inference.', 'cover_img_url': 'https://shared.deepinfra.com/models/mistralai/Mixtral-8x22B-Instruct-v0.1/cover_image.8bb1b015367a1537fd23c69d5b8117675a86b207c9bd3cce326b750ef877bcb6.webp', 'tags': [...], 'pricing': {...}, 'max_tokens': 65536, 'replaced_by': None, 'deprecated': None, 'quantization': 'bfloat16'}, {'model_name': 'meta-llama/Llama-2-70b-chat-hf', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': 'LLaMa 2 is a collections of LLMs trained by Meta. This is the 70B chat optimized version. This endpoint has per token pricing.', 'cover_img_url': 'https://shared.deepinfra.com/models/meta-llama/Llama-2-70b-chat-hf/cover_image.7b3407408b20bd422edfb75da90ee92d0a05649e94b59bf409c827e845fc3c46.webp', 'tags': [...], 'pricing': {...}, 'max_tokens': 4096, 'replaced_by': 'meta-llama/Meta-Llama-3-70B-Instruct', 'deprecated': 1718309527, 'quantization': 'fp16'}, {'model_name': 'google/gemma-1.1-7b-it', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': 'Gemma is an open-source model designed by Google. This is Gemma 1.1 7B (IT), an update over the original instruction-tuned Gemma release. Gemma 1.1 was trained using a novel RLHF method, leading to substantial gains on quality, coding capabilities, factuality, instruction following and multi-turn conversation quality.', 'cover_img_url': 'https://shared.deepinfra.com/models/google/gemma-1.1-7b-it/cover_image.18ce701e05fe6377f27c31c2e0d05649b7962f184f8dc894e31aad3ffb468f70.webp', 'tags': [...], 'pricing': {...}, 'max_tokens': 8192, 'replaced_by': None, 'deprecated': None, 'quantization': 'bfloat16'}, {'model_name': 'Gryphe/MythoMax-L2-13b', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': '', 'cover_img_url': 'https://shared.deepinfra.com/models/Gryphe/MythoMax-L2-13b/cover_image.7567613d62f797fa930227a88202f7aee5ef30da38e6c9c7b775979cc71220bc.webp', 'tags': [...], 'pricing': {...}, 'max_tokens': 4096, 'replaced_by': None, 'deprecated': None, 'quantization': 'fp16'}, {'model_name': 'openchat/openchat-3.6-8b', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': 'Openchat 3.6 is a LLama-3-8b fine tune that outperforms it on multiple benchmarks.', 'cover_img_url': 'https://shared.deepinfra.com/models/openchat/openchat-3.6-8b/cover_image.c5792073e4034a0847ff5112e00356adb411c4a5c900ed22c32ad65c5d97e8d1.webp', 'tags': [...], 'pricing': {...}, 'max_tokens': 8192, 'replaced_by': None, 'deprecated': None, 'quantization': 'bfloat16'}, {'model_name': 'Austism/chronos-hermes-13b-v2', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': 'This offers the imaginative writing style of chronos while still retaining coherency and being capable. Outputs are long and utilize exceptional prose. Supports a maxium context length of 4096. The model follows the Alpaca prompt format.', 'cover_img_url': '', 'tags': [...], 'pricing': {...}, 'max_tokens': 4096, 'replaced_by': None, 'deprecated': None, 'quantization': 'fp16'}, {'model_name': 'Phind/Phind-CodeLlama-34B-v2', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': 'Phind-CodeLlama-34B-v2 is an open-source language model that has been fine-tuned on 1.5B tokens of high-quality programming-related data and achieved a pass@1 rate of 73.8% on HumanEval. It is multi-lingual and proficient in Python, C/C++, TypeScript, Java, and more. It has been trained on a proprietary dataset of instruction-answer pairs instead of code completion examples. The model is instruction-tuned on the Alpaca/Vicuna format to be steerable and easy-to-use. It accepts the Alpaca/Vicuna instruction format and can generate one completion for each prompt.', 'cover_img_url': 'https://shared.deepinfra.com/models/Phind/Phind-CodeLlama-34B-v2/cover_image.0d7cb500b84d00e46b7bf490b75cf8eda73a3ad775fa4360c8deba541c3349b3.webp', 'tags': [...], 'pricing': {...}, 'max_tokens': 4096, 'replaced_by': None, 'deprecated': None, 'quantization': 'fp16'}, {'model_name': 'lizpreciatior/lzlv_70b_fp16_hf', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': 'A Mythomax/MLewd_13B-style merge of selected 70B models A multi-model merge of several LLaMA2 70B finetunes for roleplaying and creative work. The goal was to create a model that combines creativity with intelligence for an enhanced experience.', 'cover_img_url': 'https://shared.deepinfra.com/models/lizpreciatior/lzlv_70b_fp16_hf/cover_image.2bb893141f7dce176afce500c4ec8ca22cfe5e2b00253d997fea31a7f60adc1b.webp', 'tags': [...], 'pricing': {...}, 'max_tokens': 4096, 'replaced_by': None, 'deprecated': None, 'quantization': 'fp16'}, {'model_name': 'cognitivecomputations/dolphin-2.6-mixtral-8x7b', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': 'The Dolphin 2.6 Mixtral 8x7b model is a finetuned version of the Mixtral-8x7b model, trained on a variety of data including coding data, for 3 days on 4 A100 GPUs. It is uncensored and requires trust_remote_code. The model is very obedient and good at coding, but not DPO tuned. The dataset has been filtered for alignment and bias. The model is compliant with user requests and can be used for various purposes such as generating code or engaging in general chat.', 'cover_img_url': 'https://shared.deepinfra.com/models/cognitivecomputations/dolphin-2.6-mixtral-8x7b/cover_image.b265207e1a422c62c06f23a86e6ef6e8ee326de40a24bb1c5d9f102c1f2acd6b.webp', 'tags': [...], 'pricing': {...}, 'max_tokens': 32768, 'replaced_by': None, 'deprecated': None, 'quantization': 'bfloat16'}, {'model_name': 'bigcode/starcoder2-15b', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': 'StarCoder2-15B model is a 15B parameter model trained on 600+ programming languages. It specializes in code completion.', 'cover_img_url': 'https://shared.deepinfra.com/models/bigcode/starcoder2-15b/cover_image.8981f6e7c85d72bf816c2abcb90d811c905db812ce0560ba5216df8cbc314464.webp', 'tags': [...], 'pricing': {...}, 'max_tokens': 16384, 'replaced_by': 'Phind/Phind-CodeLlama-34B-v2', 'deprecated': None, 'quantization': 'fp16'}, {'model_name': 'llava-hf/llava-1.5-7b-hf', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': 'LLaVa is a multimodal model that supports vision and language models combined.', 'cover_img_url': 'https://shared.deepinfra.com/models/llava-hf/llava-1.5-7b-hf/cover_image.ed4fba7a25b147e7fe6675e9f760585e11274e8ee72596e6412447260493cd4f.webp', 'tags': [...], 'pricing': {...}, 'max_tokens': 4096, 'replaced_by': None, 'deprecated': None, 'quantization': 'fp16'}, {'model_name': 'meta-llama/Meta-Llama-3-70B-Instruct', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': 'Model Details Meta developed and released the Meta Llama 3 family of large language models (LLMs), a collection of pretrained and instruction tuned generative text models in 8 and 70B sizes.', 'cover_img_url': 'https://shared.deepinfra.com/models/meta-llama/Meta-Llama-3-70B-Instruct/cover_image.bcffae761540e7dd36aea32e2a576690d43592a0fc39b9edbe83a5420758aabf.webp', 'tags': [...], 'pricing': {...}, 'max_tokens': 8192, 'replaced_by': None, 'deprecated': None, 'quantization': 'bfloat16'}, {'model_name': 'microsoft/Phi-3-medium-4k-instruct', 'type': 'text-generation', 'reported_type': 'text-generation', 'description': 'The Phi-3-Medium-4K-Instruct is a powerful and lightweight language model with 14 billion parameters, trained on high-quality data to excel in instruction following and safety measures. It demonstrates exceptional performance across benchmarks, including common sense, language understanding, and logical reasoning, outperforming models of similar size.', 'cover_img_url': 'https://shared.deepinfra.com/models/microsoft/Phi-3-medium-4k-instruct/cover_image.6c7f810d3584719025ba43de13448d318bda84043a08af1b9718c61d9498b18c.webp', 'tags': [...], 'pricing': {...}, 'max_tokens': 4096, 'replaced_by': None, 'deprecated': None, 'quantization': 'bfloat16'}]
edsl/tools/__init__.py CHANGED
@@ -1 +1 @@
1
- from edsl.tools.plotting import barchart
1
+ from edsl.tools.plotting import barchart