crfm-helm 0.5.2__py3-none-any.whl → 0.5.4__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of crfm-helm might be problematic. Click here for more details.

Files changed (209) hide show
  1. {crfm_helm-0.5.2.dist-info → crfm_helm-0.5.4.dist-info}/METADATA +81 -112
  2. {crfm_helm-0.5.2.dist-info → crfm_helm-0.5.4.dist-info}/RECORD +165 -155
  3. {crfm_helm-0.5.2.dist-info → crfm_helm-0.5.4.dist-info}/WHEEL +1 -1
  4. helm/benchmark/adaptation/adapters/multiple_choice_joint_adapter.py +12 -5
  5. helm/benchmark/adaptation/adapters/test_generation_adapter.py +12 -12
  6. helm/benchmark/adaptation/adapters/test_language_modeling_adapter.py +8 -8
  7. helm/benchmark/adaptation/adapters/test_multiple_choice_joint_adapter.py +77 -9
  8. helm/benchmark/adaptation/common_adapter_specs.py +2 -0
  9. helm/benchmark/annotation/anthropic_red_team_annotator.py +57 -0
  10. helm/benchmark/annotation/call_center_annotator.py +258 -0
  11. helm/benchmark/annotation/financebench_annotator.py +79 -0
  12. helm/benchmark/annotation/harm_bench_annotator.py +55 -0
  13. helm/benchmark/annotation/{image2structure → image2struct}/latex_compiler_annotator.py +2 -2
  14. helm/benchmark/annotation/{image2structure → image2struct}/lilypond_compiler_annotator.py +5 -3
  15. helm/benchmark/annotation/{image2structure → image2struct}/webpage_compiler_annotator.py +5 -5
  16. helm/benchmark/annotation/live_qa_annotator.py +37 -45
  17. helm/benchmark/annotation/medication_qa_annotator.py +36 -44
  18. helm/benchmark/annotation/model_as_judge.py +96 -0
  19. helm/benchmark/annotation/simple_safety_tests_annotator.py +50 -0
  20. helm/benchmark/annotation/xstest_annotator.py +100 -0
  21. helm/benchmark/metrics/annotation_metrics.py +108 -0
  22. helm/benchmark/metrics/bhasa_metrics.py +188 -0
  23. helm/benchmark/metrics/bhasa_metrics_specs.py +10 -0
  24. helm/benchmark/metrics/code_metrics_helper.py +11 -1
  25. helm/benchmark/metrics/safety_metrics.py +79 -0
  26. helm/benchmark/metrics/summac/model_summac.py +3 -3
  27. helm/benchmark/metrics/tokens/test_ai21_token_cost_estimator.py +2 -2
  28. helm/benchmark/metrics/tokens/test_openai_token_cost_estimator.py +4 -4
  29. helm/benchmark/metrics/unitxt_metrics.py +17 -3
  30. helm/benchmark/metrics/vision_language/image_metrics.py +7 -3
  31. helm/benchmark/metrics/vision_language/image_utils.py +1 -1
  32. helm/benchmark/model_metadata_registry.py +3 -3
  33. helm/benchmark/presentation/create_plots.py +1 -1
  34. helm/benchmark/presentation/schema.py +3 -0
  35. helm/benchmark/presentation/summarize.py +106 -256
  36. helm/benchmark/presentation/test_run_entry.py +1 -0
  37. helm/benchmark/presentation/test_summarize.py +145 -3
  38. helm/benchmark/run.py +15 -0
  39. helm/benchmark/run_expander.py +83 -30
  40. helm/benchmark/run_specs/bhasa_run_specs.py +652 -0
  41. helm/benchmark/run_specs/call_center_run_specs.py +152 -0
  42. helm/benchmark/run_specs/decodingtrust_run_specs.py +8 -8
  43. helm/benchmark/run_specs/experimental_run_specs.py +52 -0
  44. helm/benchmark/run_specs/finance_run_specs.py +82 -1
  45. helm/benchmark/run_specs/safety_run_specs.py +154 -0
  46. helm/benchmark/run_specs/vlm_run_specs.py +100 -24
  47. helm/benchmark/scenarios/anthropic_red_team_scenario.py +71 -0
  48. helm/benchmark/scenarios/banking77_scenario.py +51 -0
  49. helm/benchmark/scenarios/bhasa_scenario.py +1942 -0
  50. helm/benchmark/scenarios/call_center_scenario.py +84 -0
  51. helm/benchmark/scenarios/decodingtrust_stereotype_bias_scenario.py +2 -1
  52. helm/benchmark/scenarios/ewok_scenario.py +116 -0
  53. helm/benchmark/scenarios/fin_qa_scenario.py +2 -0
  54. helm/benchmark/scenarios/financebench_scenario.py +53 -0
  55. helm/benchmark/scenarios/harm_bench_scenario.py +59 -0
  56. helm/benchmark/scenarios/raft_scenario.py +1 -1
  57. helm/benchmark/scenarios/scenario.py +1 -1
  58. helm/benchmark/scenarios/simple_safety_tests_scenario.py +33 -0
  59. helm/benchmark/scenarios/test_commonsense_scenario.py +21 -0
  60. helm/benchmark/scenarios/test_ewok_scenario.py +25 -0
  61. helm/benchmark/scenarios/test_financebench_scenario.py +26 -0
  62. helm/benchmark/scenarios/test_gsm_scenario.py +31 -0
  63. helm/benchmark/scenarios/test_legalbench_scenario.py +30 -0
  64. helm/benchmark/scenarios/test_math_scenario.py +2 -8
  65. helm/benchmark/scenarios/test_med_qa_scenario.py +30 -0
  66. helm/benchmark/scenarios/test_mmlu_scenario.py +33 -0
  67. helm/benchmark/scenarios/test_narrativeqa_scenario.py +73 -0
  68. helm/benchmark/scenarios/thai_exam_scenario.py +4 -4
  69. helm/benchmark/scenarios/vision_language/a_okvqa_scenario.py +1 -1
  70. helm/benchmark/scenarios/vision_language/bingo_scenario.py +2 -2
  71. helm/benchmark/scenarios/vision_language/crossmodal_3600_scenario.py +2 -1
  72. helm/benchmark/scenarios/vision_language/exams_v_scenario.py +104 -0
  73. helm/benchmark/scenarios/vision_language/fair_face_scenario.py +136 -0
  74. helm/benchmark/scenarios/vision_language/flickr30k_scenario.py +1 -1
  75. helm/benchmark/scenarios/vision_language/gqa_scenario.py +2 -2
  76. helm/benchmark/scenarios/vision_language/hateful_memes_scenario.py +1 -1
  77. helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/chart2csv_scenario.py +1 -1
  78. helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/latex_scenario.py +3 -3
  79. helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/musicsheet_scenario.py +1 -1
  80. helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/utils_latex.py +31 -39
  81. helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/webpage/driver.py +1 -1
  82. helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/webpage/utils.py +1 -1
  83. helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/webpage_scenario.py +41 -12
  84. helm/benchmark/scenarios/vision_language/math_vista_scenario.py +1 -1
  85. helm/benchmark/scenarios/vision_language/mementos_scenario.py +3 -3
  86. helm/benchmark/scenarios/vision_language/mm_safety_bench_scenario.py +2 -2
  87. helm/benchmark/scenarios/vision_language/mme_scenario.py +21 -18
  88. helm/benchmark/scenarios/vision_language/mmmu_scenario.py +1 -1
  89. helm/benchmark/scenarios/vision_language/pairs_scenario.py +1 -1
  90. helm/benchmark/scenarios/vision_language/pope_scenario.py +2 -1
  91. helm/benchmark/scenarios/vision_language/real_world_qa_scenario.py +57 -0
  92. helm/benchmark/scenarios/vision_language/seed_bench_scenario.py +7 -5
  93. helm/benchmark/scenarios/vision_language/unicorn_scenario.py +2 -2
  94. helm/benchmark/scenarios/vision_language/vibe_eval_scenario.py +6 -3
  95. helm/benchmark/scenarios/vision_language/viz_wiz_scenario.py +1 -1
  96. helm/benchmark/scenarios/vision_language/vqa_scenario.py +3 -1
  97. helm/benchmark/scenarios/xstest_scenario.py +35 -0
  98. helm/benchmark/server.py +1 -6
  99. helm/benchmark/static/schema_air_bench.yaml +750 -750
  100. helm/benchmark/static/schema_bhasa.yaml +709 -0
  101. helm/benchmark/static/schema_call_center.yaml +232 -0
  102. helm/benchmark/static/schema_cleva.yaml +768 -0
  103. helm/benchmark/static/schema_decodingtrust.yaml +444 -0
  104. helm/benchmark/static/schema_ewok.yaml +367 -0
  105. helm/benchmark/static/schema_finance.yaml +55 -9
  106. helm/benchmark/static/{schema_image2structure.yaml → schema_image2struct.yaml} +231 -90
  107. helm/benchmark/static/schema_legal.yaml +566 -0
  108. helm/benchmark/static/schema_safety.yaml +266 -0
  109. helm/benchmark/static/schema_tables.yaml +149 -8
  110. helm/benchmark/static/schema_thai.yaml +21 -0
  111. helm/benchmark/static/schema_vhelm.yaml +137 -101
  112. helm/benchmark/static_build/assets/accenture-6f97eeda.png +0 -0
  113. helm/benchmark/static_build/assets/aisingapore-6dfc9acf.png +0 -0
  114. helm/benchmark/static_build/assets/cresta-9e22b983.png +0 -0
  115. helm/benchmark/static_build/assets/cuhk-8c5631e9.png +0 -0
  116. helm/benchmark/static_build/assets/index-05c76bb1.css +1 -0
  117. helm/benchmark/static_build/assets/index-3ee38b3d.js +10 -0
  118. helm/benchmark/static_build/assets/scb10x-204bd786.png +0 -0
  119. helm/benchmark/static_build/assets/vhelm-aspects-1437d673.png +0 -0
  120. helm/benchmark/static_build/assets/vhelm-framework-a1ca3f3f.png +0 -0
  121. helm/benchmark/static_build/assets/vhelm-model-8afb7616.png +0 -0
  122. helm/benchmark/static_build/assets/wellsfargo-a86a6c4a.png +0 -0
  123. helm/benchmark/static_build/index.html +2 -2
  124. helm/benchmark/window_services/test_openai_window_service.py +8 -8
  125. helm/benchmark/window_services/tokenizer_service.py +0 -5
  126. helm/clients/ai21_client.py +71 -1
  127. helm/clients/anthropic_client.py +7 -19
  128. helm/clients/huggingface_client.py +38 -37
  129. helm/clients/nvidia_nim_client.py +35 -0
  130. helm/clients/openai_client.py +18 -4
  131. helm/clients/palmyra_client.py +24 -0
  132. helm/clients/perspective_api_client.py +11 -6
  133. helm/clients/test_client.py +4 -6
  134. helm/clients/together_client.py +22 -0
  135. helm/clients/vision_language/open_flamingo_client.py +1 -2
  136. helm/clients/vision_language/palmyra_vision_client.py +28 -13
  137. helm/common/cache.py +8 -30
  138. helm/common/images_utils.py +6 -0
  139. helm/common/key_value_store.py +9 -9
  140. helm/common/mongo_key_value_store.py +5 -4
  141. helm/common/request.py +16 -0
  142. helm/common/test_cache.py +1 -48
  143. helm/common/tokenization_request.py +0 -9
  144. helm/config/model_deployments.yaml +444 -329
  145. helm/config/model_metadata.yaml +513 -111
  146. helm/config/tokenizer_configs.yaml +140 -11
  147. helm/proxy/example_queries.py +14 -21
  148. helm/proxy/server.py +0 -9
  149. helm/proxy/services/remote_service.py +0 -6
  150. helm/proxy/services/server_service.py +6 -20
  151. helm/proxy/services/service.py +0 -6
  152. helm/proxy/token_counters/test_auto_token_counter.py +2 -2
  153. helm/tokenizers/ai21_tokenizer.py +51 -59
  154. helm/tokenizers/cohere_tokenizer.py +0 -75
  155. helm/tokenizers/huggingface_tokenizer.py +0 -1
  156. helm/tokenizers/test_ai21_tokenizer.py +48 -0
  157. helm/benchmark/data_overlap/data_overlap_spec.py +0 -86
  158. helm/benchmark/data_overlap/export_scenario_text.py +0 -119
  159. helm/benchmark/data_overlap/light_scenario.py +0 -60
  160. helm/benchmark/scenarios/vision_language/image2structure/webpage/__init__.py +0 -0
  161. helm/benchmark/static/benchmarking.css +0 -156
  162. helm/benchmark/static/benchmarking.js +0 -1705
  163. helm/benchmark/static/config.js +0 -3
  164. helm/benchmark/static/general.js +0 -122
  165. helm/benchmark/static/images/crfm-logo.png +0 -0
  166. helm/benchmark/static/images/helm-logo-simple.png +0 -0
  167. helm/benchmark/static/images/helm-logo.png +0 -0
  168. helm/benchmark/static/images/language-model-helm.png +0 -0
  169. helm/benchmark/static/images/organizations/ai21.png +0 -0
  170. helm/benchmark/static/images/organizations/anthropic.png +0 -0
  171. helm/benchmark/static/images/organizations/bigscience.png +0 -0
  172. helm/benchmark/static/images/organizations/cohere.png +0 -0
  173. helm/benchmark/static/images/organizations/eleutherai.png +0 -0
  174. helm/benchmark/static/images/organizations/google.png +0 -0
  175. helm/benchmark/static/images/organizations/meta.png +0 -0
  176. helm/benchmark/static/images/organizations/microsoft.png +0 -0
  177. helm/benchmark/static/images/organizations/nvidia.png +0 -0
  178. helm/benchmark/static/images/organizations/openai.png +0 -0
  179. helm/benchmark/static/images/organizations/together.png +0 -0
  180. helm/benchmark/static/images/organizations/tsinghua-keg.png +0 -0
  181. helm/benchmark/static/images/organizations/yandex.png +0 -0
  182. helm/benchmark/static/images/scenarios-by-metrics.png +0 -0
  183. helm/benchmark/static/images/taxonomy-scenarios.png +0 -0
  184. helm/benchmark/static/index.html +0 -68
  185. helm/benchmark/static/info-icon.png +0 -0
  186. helm/benchmark/static/json-urls.js +0 -69
  187. helm/benchmark/static/plot-captions.js +0 -27
  188. helm/benchmark/static/utils.js +0 -285
  189. helm/benchmark/static_build/assets/index-30dbceba.js +0 -10
  190. helm/benchmark/static_build/assets/index-66b02d40.css +0 -1
  191. helm/benchmark/static_build/assets/vhelm-framework-cde7618a.png +0 -0
  192. helm/benchmark/static_build/assets/vhelm-model-6d812526.png +0 -0
  193. helm/benchmark/window_services/ai21_window_service.py +0 -247
  194. helm/benchmark/window_services/cohere_window_service.py +0 -101
  195. helm/benchmark/window_services/test_ai21_window_service.py +0 -163
  196. helm/benchmark/window_services/test_cohere_window_service.py +0 -75
  197. helm/benchmark/window_services/test_cohere_window_service_utils.py +0 -8328
  198. helm/benchmark/window_services/test_ice_window_service.py +0 -327
  199. helm/tokenizers/ice_tokenizer.py +0 -30
  200. helm/tokenizers/test_ice_tokenizer.py +0 -57
  201. {crfm_helm-0.5.2.dist-info → crfm_helm-0.5.4.dist-info}/LICENSE +0 -0
  202. {crfm_helm-0.5.2.dist-info → crfm_helm-0.5.4.dist-info}/entry_points.txt +0 -0
  203. {crfm_helm-0.5.2.dist-info → crfm_helm-0.5.4.dist-info}/top_level.txt +0 -0
  204. /helm/benchmark/annotation/{image2structure → image2struct}/__init__.py +0 -0
  205. /helm/benchmark/annotation/{image2structure → image2struct}/image_compiler_annotator.py +0 -0
  206. /helm/benchmark/{data_overlap → scenarios/vision_language/image2struct}/__init__.py +0 -0
  207. /helm/benchmark/scenarios/vision_language/{image2structure/image2structure_scenario.py → image2struct/image2struct_scenario.py} +0 -0
  208. /helm/benchmark/scenarios/vision_language/{image2structure → image2struct/webpage}/__init__.py +0 -0
  209. /helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/webpage/jekyll_server.py +0 -0
@@ -28,74 +28,47 @@ model_deployments:
28
28
 
29
29
  # AI21 Labs
30
30
 
31
- # J1 models are Deprecated by AI21 Labs
32
- # API returns: Detail: Jurassic J1 models are deprecated
33
- - name: ai21/j1-jumbo
34
- deprecated: true
35
- model_name: ai21/j1-jumbo
36
- tokenizer_name: ai21/j1
37
- max_sequence_length: 2047
38
- client_spec:
39
- class_name: "helm.clients.ai21_client.AI21Client"
40
- window_service_spec:
41
- class_name: "helm.benchmark.window_services.ai21_window_service.AI21WindowService"
42
-
43
- - name: ai21/j1-large
44
- deprecated: true
45
- model_name: ai21/j1-large
46
- tokenizer_name: ai21/j1
47
- max_sequence_length: 2047
48
- client_spec:
49
- class_name: "helm.clients.ai21_client.AI21Client"
50
- window_service_spec:
51
- class_name: "helm.benchmark.window_services.ai21_window_service.AI21WindowService"
52
-
53
- - name: ai21/j1-grande
54
- deprecated: true
55
- model_name: ai21/j1-grande
56
- tokenizer_name: ai21/j1
31
+ - name: ai21/j2-large
32
+ model_name: ai21/j2-large
33
+ tokenizer_name: ai21/j2-tokenizer
57
34
  max_sequence_length: 2047
58
35
  client_spec:
59
36
  class_name: "helm.clients.ai21_client.AI21Client"
60
- window_service_spec:
61
- class_name: "helm.benchmark.window_services.ai21_window_service.AI21WindowService"
62
37
 
63
- - name: ai21/j1-grande-v2-beta
64
- deprecated: true
65
- model_name: ai21/j1-grande-v2-beta
66
- tokenizer_name: ai21/j1
38
+ - name: ai21/j2-grande
39
+ model_name: ai21/j2-grande
40
+ tokenizer_name: ai21/j2-tokenizer
67
41
  max_sequence_length: 2047
68
42
  client_spec:
69
43
  class_name: "helm.clients.ai21_client.AI21Client"
70
- window_service_spec:
71
- class_name: "helm.benchmark.window_services.ai21_window_service.AI21WindowService"
72
44
 
73
45
  - name: ai21/j2-jumbo
74
46
  model_name: ai21/j2-jumbo
75
- tokenizer_name: ai21/j1
47
+ tokenizer_name: ai21/j2-tokenizer
76
48
  max_sequence_length: 6000
77
49
  client_spec:
78
50
  class_name: "helm.clients.ai21_client.AI21Client"
79
- window_service_spec:
80
- class_name: "helm.benchmark.window_services.ai21_window_service.AI21WindowService"
81
51
 
82
- - name: ai21/j2-large
83
- model_name: ai21/j2-large
84
- tokenizer_name: ai21/j1
85
- max_sequence_length: 2047
52
+ - name: ai21/jamba-instruct
53
+ model_name: ai21/jamba-instruct
54
+ tokenizer_name: ai21/jamba-instruct-tokenizer
55
+ max_sequence_length: 256000
86
56
  client_spec:
87
- class_name: "helm.clients.ai21_client.AI21Client"
88
- window_service_spec:
89
- class_name: "helm.benchmark.window_services.ai21_window_service.AI21WindowService"
57
+ class_name: "helm.clients.ai21_client.AI21ChatClient"
90
58
 
91
- - name: ai21/j2-grande
92
- model_name: ai21/j2-grande
93
- tokenizer_name: ai21/j1
94
- max_sequence_length: 2047
59
+ - name: ai21/jamba-1.5-mini
60
+ model_name: ai21/jamba-1.5-mini
61
+ tokenizer_name: ai21/jamba-1.5-mini-tokenizer
62
+ max_sequence_length: 256000
95
63
  client_spec:
96
- class_name: "helm.clients.ai21_client.AI21Client"
97
- window_service_spec:
98
- class_name: "helm.benchmark.window_services.ai21_window_service.AI21WindowService"
64
+ class_name: "helm.clients.ai21_client.AI21ChatClient"
65
+
66
+ - name: ai21/jamba-1.5-large
67
+ model_name: ai21/jamba-1.5-large
68
+ tokenizer_name: ai21/jamba-1.5-large-tokenizer
69
+ max_sequence_length: 256000
70
+ client_spec:
71
+ class_name: "helm.clients.ai21_client.AI21ChatClient"
99
72
 
100
73
  # Aleph Alpha
101
74
  - name: AlephAlpha/luminous-base
@@ -216,6 +189,13 @@ model_deployments:
216
189
  client_spec:
217
190
  class_name: "helm.clients.anthropic_client.AnthropicMessagesClient"
218
191
 
192
+ - name: anthropic/claude-3-5-sonnet-20240620
193
+ model_name: anthropic/claude-3-5-sonnet-20240620
194
+ tokenizer_name: anthropic/claude
195
+ max_sequence_length: 200000
196
+ client_spec:
197
+ class_name: "helm.clients.anthropic_client.AnthropicMessagesClient"
198
+
219
199
  - name: anthropic/stanford-online-all-v4-s3
220
200
  deprecated: true # Closed model, not accessible via API
221
201
  model_name: anthropic/stanford-online-all-v4-s3
@@ -225,86 +205,6 @@ model_deployments:
225
205
  class_name: "helm.clients.anthropic_client.AnthropicLegacyClient"
226
206
 
227
207
  # Cohere
228
- - name: cohere/xlarge-20220609
229
- model_name: cohere/xlarge-20220609
230
- tokenizer_name: cohere/cohere
231
- max_sequence_length: 2047
232
- max_request_length: 2048
233
- client_spec:
234
- class_name: "helm.clients.cohere_client.CohereClient"
235
- window_service_spec:
236
- class_name: "helm.benchmark.window_services.cohere_window_service.CohereWindowService"
237
-
238
- - name: cohere/large-20220720
239
- model_name: cohere/large-20220720
240
- tokenizer_name: cohere/cohere
241
- max_sequence_length: 2047
242
- max_request_length: 2048
243
- client_spec:
244
- class_name: "helm.clients.cohere_client.CohereClient"
245
- window_service_spec:
246
- class_name: "helm.benchmark.window_services.cohere_window_service.CohereWindowService"
247
-
248
- - name: cohere/medium-20220720
249
- model_name: cohere/medium-20220720
250
- tokenizer_name: cohere/cohere
251
- max_sequence_length: 2047
252
- max_request_length: 2048
253
- client_spec:
254
- class_name: "helm.clients.cohere_client.CohereClient"
255
- window_service_spec:
256
- class_name: "helm.benchmark.window_services.cohere_window_service.CohereWindowService"
257
-
258
- - name: cohere/small-20220720
259
- model_name: cohere/small-20220720
260
- tokenizer_name: cohere/cohere
261
- max_sequence_length: 2047
262
- max_request_length: 2048
263
- client_spec:
264
- class_name: "helm.clients.cohere_client.CohereClient"
265
- window_service_spec:
266
- class_name: "helm.benchmark.window_services.cohere_window_service.CohereWindowService"
267
-
268
- - name: cohere/xlarge-20221108
269
- model_name: cohere/xlarge-20221108
270
- tokenizer_name: cohere/cohere
271
- max_sequence_length: 2047
272
- max_request_length: 2048
273
- client_spec:
274
- class_name: "helm.clients.cohere_client.CohereClient"
275
- window_service_spec:
276
- class_name: "helm.benchmark.window_services.cohere_window_service.CohereWindowService"
277
-
278
- - name: cohere/medium-20221108
279
- model_name: cohere/medium-20221108
280
- tokenizer_name: cohere/cohere
281
- max_sequence_length: 2047
282
- max_request_length: 2048
283
- client_spec:
284
- class_name: "helm.clients.cohere_client.CohereClient"
285
- window_service_spec:
286
- class_name: "helm.benchmark.window_services.cohere_window_service.CohereWindowService"
287
-
288
- - name: cohere/command-medium-beta
289
- model_name: cohere/command-medium-beta
290
- tokenizer_name: cohere/cohere
291
- max_sequence_length: 2019
292
- max_request_length: 2020
293
- client_spec:
294
- class_name: "helm.clients.cohere_client.CohereClient"
295
- window_service_spec:
296
- class_name: "helm.benchmark.window_services.cohere_window_service.CohereWindowService"
297
-
298
- - name: cohere/command-xlarge-beta
299
- model_name: cohere/command-xlarge-beta
300
- tokenizer_name: cohere/cohere
301
- max_sequence_length: 2019
302
- max_request_length: 2020
303
- client_spec:
304
- class_name: "helm.clients.cohere_client.CohereClient"
305
- window_service_spec:
306
- class_name: "helm.benchmark.window_services.cohere_window_service.CohereWindowService"
307
-
308
208
  - name: cohere/command
309
209
  model_name: cohere/command
310
210
  tokenizer_name: cohere/command
@@ -312,8 +212,6 @@ model_deployments:
312
212
  max_request_length: 2020 # TODO: verify this
313
213
  client_spec:
314
214
  class_name: "helm.clients.cohere_client.CohereClient"
315
- window_service_spec:
316
- class_name: "helm.benchmark.window_services.cohere_window_service.CohereWindowService"
317
215
 
318
216
  - name: cohere/command-light
319
217
  model_name: cohere/command-light
@@ -322,8 +220,6 @@ model_deployments:
322
220
  max_request_length: 2020 # TODO: verify this
323
221
  client_spec:
324
222
  class_name: "helm.clients.cohere_client.CohereClient"
325
- window_service_spec:
326
- class_name: "helm.benchmark.window_services.cohere_window_service.CohereWindowService"
327
223
 
328
224
  - name: cohere/command-r
329
225
  model_name: cohere/command-r
@@ -587,6 +483,20 @@ model_deployments:
587
483
  client_spec:
588
484
  class_name: "helm.clients.together_client.TogetherClient"
589
485
 
486
+ - name: together/gemma-2-9b-it
487
+ model_name: google/gemma-2-9b-it
488
+ tokenizer_name: google/gemma-2-9b
489
+ max_sequence_length: 8191
490
+ client_spec:
491
+ class_name: "helm.clients.together_client.TogetherClient"
492
+
493
+ - name: together/gemma-2-27b-it
494
+ model_name: google/gemma-2-27b-it
495
+ tokenizer_name: google/gemma-2-9b
496
+ max_sequence_length: 8191
497
+ client_spec:
498
+ class_name: "helm.clients.together_client.TogetherClient"
499
+
590
500
  ## MedLM
591
501
  - name: google/medlm-medium
592
502
  model_name: google/medlm-medium
@@ -716,6 +626,26 @@ model_deployments:
716
626
  args:
717
627
  trust_remote_code: true
718
628
 
629
+ - name: huggingface/llama3-8b-cpt-sea-lionv2-base
630
+ model_name: aisingapore/llama3-8b-cpt-sea-lionv2-base
631
+ tokenizer_name: meta/llama-3-8b-instruct
632
+ max_sequence_length: 8192
633
+ client_spec:
634
+ class_name: "helm.clients.huggingface_client.HuggingFaceClient"
635
+ args:
636
+ device_map: auto
637
+ torch_dtype: torch.bfloat16
638
+
639
+ - name: huggingface/llama3-8b-cpt-sea-lionv2.1-instruct
640
+ model_name: aisingapore/llama3-8b-cpt-sea-lionv2.1-instruct
641
+ tokenizer_name: meta/llama-3-8b-instruct
642
+ max_sequence_length: 8192
643
+ client_spec:
644
+ class_name: "helm.clients.huggingface_client.HuggingFaceClient"
645
+ args:
646
+ device_map: auto
647
+ torch_dtype: torch.bfloat16
648
+
719
649
  ## Bigcode
720
650
  - name: huggingface/santacoder
721
651
  model_name: bigcode/santacoder
@@ -807,6 +737,47 @@ model_deployments:
807
737
  client_spec:
808
738
  class_name: "helm.clients.huggingface_client.HuggingFaceClient"
809
739
 
740
+ ## Google
741
+ - name: huggingface/gemma-2-9b
742
+ model_name: google/gemma-2-9b
743
+ tokenizer_name: google/gemma-2-9b
744
+ max_sequence_length: 8192
745
+ client_spec:
746
+ class_name: "helm.clients.huggingface_client.HuggingFaceClient"
747
+ args:
748
+ device_map: auto
749
+ torch_dtype: torch.bfloat16
750
+
751
+ - name: huggingface/gemma-2-9b-it
752
+ model_name: google/gemma-2-9b-it
753
+ tokenizer_name: google/gemma-2-9b
754
+ max_sequence_length: 8192
755
+ client_spec:
756
+ class_name: "helm.clients.huggingface_client.HuggingFaceClient"
757
+ args:
758
+ device_map: auto
759
+ torch_dtype: torch.bfloat16
760
+
761
+ - name: huggingface/gemma-2-27b
762
+ model_name: google/gemma-2-27b
763
+ tokenizer_name: google/gemma-2-9b
764
+ max_sequence_length: 8192
765
+ client_spec:
766
+ class_name: "helm.clients.huggingface_client.HuggingFaceClient"
767
+ args:
768
+ device_map: auto
769
+ torch_dtype: torch.bfloat16
770
+
771
+ - name: huggingface/gemma-2-27b-it
772
+ model_name: google/gemma-2-27b-it
773
+ tokenizer_name: google/gemma-2-9b
774
+ max_sequence_length: 8192
775
+ client_spec:
776
+ class_name: "helm.clients.huggingface_client.HuggingFaceClient"
777
+ args:
778
+ device_map: auto
779
+ torch_dtype: torch.bfloat16
780
+
810
781
  ## LMSYS
811
782
  - name: huggingface/vicuna-7b-v1.3
812
783
  model_name: lmsys/vicuna-7b-v1.3
@@ -931,6 +902,7 @@ model_deployments:
931
902
  tokenizer_name: "anas-awadalla-2/mpt-7b"
932
903
  cross_attn_every_n_layers: 4
933
904
 
905
+ ## Microsoft
934
906
  - name: together/phi-2
935
907
  model_name: microsoft/phi-2
936
908
  tokenizer_name: microsoft/phi-2
@@ -938,6 +910,26 @@ model_deployments:
938
910
  client_spec:
939
911
  class_name: "helm.clients.together_client.TogetherClient"
940
912
 
913
+ - name: huggingface/phi-3-small-8k-instruct
914
+ model_name: microsoft/phi-3-small-8k-instruct
915
+ tokenizer_name: microsoft/phi-3-small-8k-instruct
916
+ max_sequence_length: 8192
917
+ client_spec:
918
+ class_name: "helm.clients.huggingface_client.HuggingFaceClient"
919
+ args:
920
+ torch_dtype: auto
921
+ trust_remote_code: true
922
+
923
+ - name: huggingface/phi-3-medium-4k-instruct
924
+ model_name: microsoft/phi-3-medium-4k-instruct
925
+ tokenizer_name: microsoft/phi-3-medium-4k-instruct
926
+ max_sequence_length: 4096
927
+ client_spec:
928
+ class_name: "helm.clients.huggingface_client.HuggingFaceClient"
929
+ args:
930
+ device_map: auto
931
+ torch_dtype: auto
932
+
941
933
  ## Mistral AI
942
934
  - name: huggingface/bakLlava-v1-hf
943
935
  model_name: mistralai/bakLlava-v1-hf
@@ -992,42 +984,134 @@ model_deployments:
992
984
  args:
993
985
  pretrained_model_name_or_path: openai-community/gpt2
994
986
 
987
+ ## OpenThaiGPT
988
+ - name: huggingface/openthaigpt-1.0.0-7b-chat
989
+ model_name: openthaigpt/openthaigpt-1.0.0-7b-chat
990
+ tokenizer_name: openthaigpt/openthaigpt-1.0.0-7b-chat
991
+ max_sequence_length: 4096
992
+ client_spec:
993
+ class_name: "helm.clients.huggingface_client.HuggingFaceClient"
994
+
995
+ - name: huggingface/openthaigpt-1.0.0-13b-chat
996
+ model_name: openthaigpt/openthaigpt-1.0.0-13b-chat
997
+ tokenizer_name: openthaigpt/openthaigpt-1.0.0-7b-chat
998
+ max_sequence_length: 4096
999
+ client_spec:
1000
+ class_name: "helm.clients.huggingface_client.HuggingFaceClient"
1001
+ args:
1002
+ device_map: auto
1003
+
1004
+ - name: huggingface/openthaigpt-1.0.0-70b-chat
1005
+ model_name: openthaigpt/openthaigpt-1.0.0-70b-chat
1006
+ tokenizer_name: huggingface/openthaigpt-1.0.0-7b-chat
1007
+ max_sequence_length: 4096
1008
+ client_spec:
1009
+ class_name: "helm.clients.huggingface_client.HuggingFaceClient"
1010
+ args:
1011
+ device_map: auto
1012
+
995
1013
  ## SAIL (SEA AI Lab)
996
- - name: sail/sailor-7b
1014
+ - name: huggingface/sailor-7b
997
1015
  model_name: sail/sailor-7b
998
1016
  tokenizer_name: qwen/qwen1.5-7b
999
1017
  max_sequence_length: 32768
1000
1018
  client_spec:
1001
1019
  class_name: "helm.clients.huggingface_client.HuggingFaceClient"
1002
1020
 
1003
- - name: sail/sailor-7b-chat
1021
+ - name: huggingface/sailor-7b-chat
1004
1022
  model_name: sail/sailor-7b-chat
1005
1023
  tokenizer_name: qwen/qwen1.5-7b
1006
1024
  max_sequence_length: 32768
1007
1025
  client_spec:
1008
1026
  class_name: "helm.clients.huggingface_client.HuggingFaceClient"
1009
1027
 
1010
- - name: sail/sailor-14b
1028
+ - name: huggingface/sailor-14b
1011
1029
  model_name: sail/sailor-14b
1012
1030
  tokenizer_name: qwen/qwen1.5-7b
1013
1031
  max_sequence_length: 32768
1014
1032
  client_spec:
1015
1033
  class_name: "helm.clients.huggingface_client.HuggingFaceClient"
1034
+ args:
1035
+ device_map: auto
1016
1036
 
1017
- - name: sail/sailor-14b-chat
1037
+ - name: huggingface/sailor-14b-chat
1018
1038
  model_name: sail/sailor-14b-chat
1019
1039
  tokenizer_name: qwen/qwen1.5-7b
1020
1040
  max_sequence_length: 32768
1021
1041
  client_spec:
1022
1042
  class_name: "helm.clients.huggingface_client.HuggingFaceClient"
1043
+ args:
1044
+ device_map: auto
1045
+
1046
+ # SambaNova
1047
+ - name: huggingface/sambalingo-thai-base
1048
+ model_name: sambanova/sambalingo-thai-base
1049
+ tokenizer_name: sambanova/sambalingo-thai-base
1050
+ max_sequence_length: 4096
1051
+ client_spec:
1052
+ class_name: "helm.clients.huggingface_client.HuggingFaceClient"
1053
+ args:
1054
+ pretrained_model_name_or_path: sambanovasystems/SambaLingo-Thai-Base
1055
+
1056
+ - name: huggingface/sambalingo-thai-chat
1057
+ model_name: sambanova/sambalingo-thai-chat
1058
+ tokenizer_name: sambanova/sambalingo-thai-base
1059
+ max_sequence_length: 4096
1060
+ client_spec:
1061
+ class_name: "helm.clients.huggingface_client.HuggingFaceClient"
1062
+ args:
1063
+ pretrained_model_name_or_path: sambanovasystems/SambaLingo-Thai-Base
1064
+
1065
+ - name: huggingface/sambalingo-thai-base-70b
1066
+ model_name: sambanova/sambalingo-thai-base-70b
1067
+ tokenizer_name: sambanova/sambalingo-thai-base
1068
+ max_sequence_length: 4096
1069
+ client_spec:
1070
+ class_name: "helm.clients.huggingface_client.HuggingFaceClient"
1071
+ args:
1072
+ pretrained_model_name_or_path: sambanovasystems/SambaLingo-Thai-Base-70B
1073
+ device_map: auto
1074
+
1075
+ - name: huggingface/sambalingo-thai-chat-70b
1076
+ model_name: sambanova/sambalingo-thai-chat-70b
1077
+ tokenizer_name: sambanova/sambalingo-thai-base
1078
+ max_sequence_length: 4096
1079
+ client_spec:
1080
+ class_name: "helm.clients.huggingface_client.HuggingFaceClient"
1081
+ args:
1082
+ pretrained_model_name_or_path: sambanovasystems/SambaLingo-Thai-Base-70B
1083
+ device_map: auto
1023
1084
 
1024
1085
  ## SCB10X
1086
+ - name: huggingface/typhoon-7b
1087
+ model_name: scb10x/typhoon-7b
1088
+ tokenizer_name: scb10x/typhoon-7b
1089
+ max_sequence_length: 4096
1090
+ client_spec:
1091
+ class_name: "helm.clients.huggingface_client.HuggingFaceClient"
1092
+
1093
+ - name: huggingface/typhoon-v1.5-8b
1094
+ model_name: scb10x/typhoon-v1.5-8b
1095
+ tokenizer_name: meta/llama-3-8b
1096
+ max_sequence_length: 8192
1097
+ client_spec:
1098
+ class_name: "helm.clients.huggingface_client.HuggingFaceClient"
1099
+
1100
+ - name: huggingface/typhoon-v1.5-8b-instruct
1101
+ model_name: scb10x/typhoon-v1.5-8b-instruct
1102
+ tokenizer_name: meta/llama-3-8b
1103
+ max_sequence_length: 8192
1104
+ client_spec:
1105
+ class_name: "helm.clients.huggingface_client.HuggingFaceClient"
1106
+
1025
1107
  - name: huggingface/typhoon-v1.5-72b
1026
1108
  model_name: scb10x/typhoon-v1.5-72b
1027
1109
  tokenizer_name: qwen/qwen1.5-7b
1028
1110
  max_sequence_length: 32768
1029
1111
  client_spec:
1030
1112
  class_name: "helm.clients.huggingface_client.HuggingFaceClient"
1113
+ args:
1114
+ device_map: auto
1031
1115
 
1032
1116
  - name: huggingface/typhoon-v1.5-72b-instruct
1033
1117
  model_name: scb10x/typhoon-v1.5-72b-instruct
@@ -1035,6 +1119,43 @@ model_deployments:
1035
1119
  max_sequence_length: 32768
1036
1120
  client_spec:
1037
1121
  class_name: "helm.clients.huggingface_client.HuggingFaceClient"
1122
+ args:
1123
+ device_map: auto
1124
+
1125
+ - name: huggingface/llama-3-typhoon-v1.5x-8b-instruct
1126
+ model_name: scb10x/llama-3-typhoon-v1.5x-8b-instruct
1127
+ tokenizer_name: meta/llama-3-8b
1128
+ max_sequence_length: 8192
1129
+ client_spec:
1130
+ class_name: "helm.clients.huggingface_client.HuggingFaceClient"
1131
+
1132
+ - name: huggingface/llama-3-typhoon-v1.5x-70b-instruct
1133
+ model_name: scb10x/llama-3-typhoon-v1.5x-70b-instruct
1134
+ tokenizer_name: meta/llama-3-8b
1135
+ max_sequence_length: 8192
1136
+ client_spec:
1137
+ class_name: "helm.clients.huggingface_client.HuggingFaceClient"
1138
+ args:
1139
+ device_map: auto
1140
+
1141
+ # Alibaba DAMO Academy
1142
+ - name: huggingface/seallm-7b-v2
1143
+ model_name: damo/seallm-7b-v2
1144
+ tokenizer_name: damo/seallm-7b-v2
1145
+ max_sequence_length: 4096
1146
+ client_spec:
1147
+ class_name: "helm.clients.huggingface_client.HuggingFaceClient"
1148
+ args:
1149
+ pretrained_model_name_or_path: SeaLLMs/SeaLLM-7B-v2
1150
+
1151
+ - name: huggingface/seallm-7b-v2.5
1152
+ model_name: damo/seallm-7b-v2.5
1153
+ tokenizer_name: damo/seallm-7b-v2.5
1154
+ max_sequence_length: 4096
1155
+ client_spec:
1156
+ class_name: "helm.clients.huggingface_client.HuggingFaceClient"
1157
+ args:
1158
+ pretrained_model_name_or_path: SeaLLMs/SeaLLM-7B-v2.5
1038
1159
 
1039
1160
  ## StabilityAI
1040
1161
  - name: huggingface/stablelm-base-alpha-3b
@@ -1315,6 +1436,20 @@ model_deployments:
1315
1436
  client_spec:
1316
1437
  class_name: "helm.clients.mistral_client.MistralAIClient"
1317
1438
 
1439
+ - name: mistralai/mistral-large-2407
1440
+ model_name: mistralai/mistral-large-2407
1441
+ tokenizer_name: mistralai/Mistral-Large-Instruct-2407
1442
+ max_sequence_length: 128000
1443
+ client_spec:
1444
+ class_name: "helm.clients.mistral_client.MistralAIClient"
1445
+
1446
+ - name: mistralai/open-mistral-nemo-2407
1447
+ model_name: mistralai/open-mistral-nemo-2407
1448
+ tokenizer_name: mistralai/Mistral-Nemo-Base-2407
1449
+ max_sequence_length: 128000
1450
+ client_spec:
1451
+ class_name: "helm.clients.mistral_client.MistralAIClient"
1452
+
1318
1453
  # Neurips
1319
1454
  - name: neurips/local
1320
1455
  model_name: neurips/local
@@ -1331,6 +1466,13 @@ model_deployments:
1331
1466
  client_spec:
1332
1467
  class_name: "helm.clients.megatron_client.MegatronClient"
1333
1468
 
1469
+ - name: nvidia/nemotron-4-340b-instruct
1470
+ model_name: nvidia/nemotron-4-340b-instruct
1471
+ tokenizer_name: nvidia/nemotron-4-340b-instruct
1472
+ max_sequence_length: 4085
1473
+ client_spec:
1474
+ class_name: "helm.clients.nvidia_nim_client.NvidiaNimClient"
1475
+
1334
1476
  # OpenAI
1335
1477
 
1336
1478
  ## GPT 3 Models
@@ -1353,99 +1495,6 @@ model_deployments:
1353
1495
  client_spec:
1354
1496
  class_name: "helm.clients.openai_client.OpenAIClient"
1355
1497
 
1356
- # The list of models can be found here: https://beta.openai.com/docs/engines/gpt-3
1357
- # DEPRECATED: Announced on July 06 2023 that these models will be shut down on January 04 2024.
1358
-
1359
- - name: openai/davinci
1360
- deprecated: true
1361
- model_name: openai/davinci
1362
- tokenizer_name: huggingface/gpt2
1363
- max_sequence_length: 2048
1364
- max_request_length: 2049
1365
- client_spec:
1366
- class_name: "helm.clients.openai_client.OpenAIClient"
1367
-
1368
- - name: openai/curie
1369
- deprecated: true
1370
- model_name: openai/curie
1371
- tokenizer_name: huggingface/gpt2
1372
- max_sequence_length: 2048
1373
- max_request_length: 2049
1374
- client_spec:
1375
- class_name: "helm.clients.openai_client.OpenAIClient"
1376
-
1377
- - name: openai/babbage
1378
- deprecated: true
1379
- model_name: openai/babbage
1380
- tokenizer_name: huggingface/gpt2
1381
- max_sequence_length: 2048
1382
- max_request_length: 2049
1383
- client_spec:
1384
- class_name: "helm.clients.openai_client.OpenAIClient"
1385
-
1386
- - name: openai/ada
1387
- deprecated: true
1388
- model_name: openai/ada
1389
- tokenizer_name: huggingface/gpt2
1390
- max_sequence_length: 2048
1391
- max_request_length: 2049
1392
- client_spec:
1393
- class_name: "helm.clients.openai_client.OpenAIClient"
1394
-
1395
- - name: openai/text-davinci-003
1396
- deprecated: true
1397
- model_name: openai/text-davinci-003
1398
- tokenizer_name: huggingface/gpt2
1399
- max_sequence_length: 4000
1400
- max_request_length: 4001
1401
- client_spec:
1402
- class_name: "helm.clients.openai_client.OpenAIClient"
1403
-
1404
- - name: openai/text-davinci-002
1405
- deprecated: true
1406
- model_name: openai/text-davinci-002
1407
- tokenizer_name: huggingface/gpt2
1408
- max_sequence_length: 4000
1409
- max_request_length: 4001
1410
- client_spec:
1411
- class_name: "helm.clients.openai_client.OpenAIClient"
1412
-
1413
- - name: openai/text-davinci-001
1414
- deprecated: true
1415
- model_name: openai/text-davinci-001
1416
- tokenizer_name: huggingface/gpt2
1417
- max_sequence_length: 2048
1418
- max_request_length: 2049
1419
- client_spec:
1420
- class_name: "helm.clients.openai_client.OpenAIClient"
1421
-
1422
- - name: openai/text-curie-001
1423
- deprecated: true
1424
- model_name: openai/text-curie-001
1425
- tokenizer_name: huggingface/gpt2
1426
- max_sequence_length: 2048
1427
- max_request_length: 2049
1428
- client_spec:
1429
- class_name: "helm.clients.openai_client.OpenAIClient"
1430
-
1431
- - name: openai/text-babbage-001
1432
- deprecated: true
1433
- model_name: openai/text-babbage-001
1434
- tokenizer_name: huggingface/gpt2
1435
- max_sequence_length: 2048
1436
- max_request_length: 2049
1437
- client_spec:
1438
- class_name: "helm.clients.openai_client.OpenAIClient"
1439
-
1440
- - name: openai/text-ada-001
1441
- deprecated: true
1442
- model_name: openai/text-ada-001
1443
- tokenizer_name: huggingface/gpt2
1444
- max_sequence_length: 2048
1445
- max_request_length: 2049
1446
- client_spec:
1447
- class_name: "helm.clients.openai_client.OpenAIClient"
1448
-
1449
1498
  ## GPT 3.5 Turbo Models
1450
1499
  # ChatGPT: https://openai.com/blog/chatgpt
1451
1500
 
@@ -1580,6 +1629,20 @@ model_deployments:
1580
1629
  client_spec:
1581
1630
  class_name: "helm.clients.openai_client.OpenAIClient"
1582
1631
 
1632
+ - name: openai/gpt-4o-2024-08-06
1633
+ model_name: openai/gpt-4o-2024-08-06
1634
+ tokenizer_name: openai/o200k_base
1635
+ max_sequence_length: 128000
1636
+ client_spec:
1637
+ class_name: "helm.clients.openai_client.OpenAIClient"
1638
+
1639
+ - name: openai/gpt-4o-mini-2024-07-18
1640
+ model_name: openai/gpt-4o-mini-2024-07-18
1641
+ tokenizer_name: openai/o200k_base
1642
+ max_sequence_length: 128000
1643
+ client_spec:
1644
+ class_name: "helm.clients.openai_client.OpenAIClient"
1645
+
1583
1646
  - name: openai/gpt-4-vision-preview
1584
1647
  model_name: openai/gpt-4-vision-preview
1585
1648
  tokenizer_name: openai/cl100k_base
@@ -1598,33 +1661,18 @@ model_deployments:
1598
1661
  client_spec:
1599
1662
  class_name: "helm.clients.openai_client.OpenAIClient"
1600
1663
 
1601
- ## Codex Models
1602
- # DEPRECATED: Codex models have been shut down on March 23 2023.
1603
-
1604
- - name: openai/code-davinci-002
1605
- deprecated: true
1606
- model_name: openai/code-davinci-002
1607
- tokenizer_name: huggingface/gpt2
1608
- max_sequence_length: 4000
1609
- max_request_length: 4001
1610
- client_spec:
1611
- class_name: "helm.clients.openai_client.OpenAIClient"
1612
-
1613
- - name: openai/code-davinci-001
1614
- deprecated: true
1615
- model_name: openai/code-davinci-001
1616
- tokenizer_name: huggingface/gpt2
1617
- max_sequence_length: 2048
1618
- max_request_length: 2049
1664
+ ## o1 Models
1665
+ - name: openai/o1-preview-2024-09-12
1666
+ model_name: openai/o1-preview-2024-09-12
1667
+ tokenizer_name: openai/cl100k_base
1668
+ max_sequence_length: 128000
1619
1669
  client_spec:
1620
1670
  class_name: "helm.clients.openai_client.OpenAIClient"
1621
1671
 
1622
- - name: openai/code-cushman-001
1623
- deprecated: true
1624
- model_name: openai/code-cushman-001
1625
- tokenizer_name: huggingface/gpt2
1626
- max_sequence_length: 2048
1627
- max_request_length: 2049
1672
+ - name: openai/o1-mini-2024-09-12
1673
+ model_name: openai/o1-mini-2024-09-12
1674
+ tokenizer_name: openai/cl100k_base
1675
+ max_sequence_length: 128000
1628
1676
  client_spec:
1629
1677
  class_name: "helm.clients.openai_client.OpenAIClient"
1630
1678
 
@@ -1632,44 +1680,6 @@ model_deployments:
1632
1680
  # OpenAI similarity embedding models: https://beta.openai.com/docs/guides/embeddings
1633
1681
  # The number of parameters is guessed based on the number of parameters of the
1634
1682
  # corresponding GPT-3 model.
1635
- # DEPRECATED: Announced on July 06 2023 that first generation embeddings models
1636
- # will be shut down on January 04 2024.
1637
-
1638
- - name: openai/text-similarity-davinci-001
1639
- deprecated: true
1640
- model_name: openai/text-similarity-davinci-001
1641
- tokenizer_name: huggingface/gpt2
1642
- max_sequence_length: 2048
1643
- max_request_length: 2049
1644
- client_spec:
1645
- class_name: "helm.clients.openai_client.OpenAIClient"
1646
-
1647
- - name: openai/text-similarity-curie-001
1648
- deprecated: true
1649
- model_name: openai/text-similarity-curie-001
1650
- tokenizer_name: huggingface/gpt2
1651
- max_sequence_length: 2048
1652
- max_request_length: 2049
1653
- client_spec:
1654
- class_name: "helm.clients.openai_client.OpenAIClient"
1655
-
1656
- - name: openai/text-similarity-babbage-001
1657
- deprecated: true
1658
- model_name: openai/text-similarity-babbage-001
1659
- tokenizer_name: huggingface/gpt2
1660
- max_sequence_length: 2048
1661
- max_request_length: 2049
1662
- client_spec:
1663
- class_name: "helm.clients.openai_client.OpenAIClient"
1664
-
1665
- - name: openai/text-similarity-ada-001
1666
- deprecated: true
1667
- model_name: openai/text-similarity-ada-001
1668
- tokenizer_name: huggingface/gpt2
1669
- max_sequence_length: 2048
1670
- max_request_length: 2049
1671
- client_spec:
1672
- class_name: "helm.clients.openai_client.OpenAIClient"
1673
1683
 
1674
1684
  # As of 2023-11-07, text-embedding-ada-002 is not deprecated:
1675
1685
  # "We recommend using text-embedding-ada-002 for nearly all use cases."
@@ -1856,7 +1866,25 @@ model_deployments:
1856
1866
  client_spec:
1857
1867
  class_name: "helm.clients.together_client.TogetherClient"
1858
1868
  args:
1859
- together_model: meta-llama/Meta-Llama-3-8B
1869
+ together_model: meta-llama/Llama-3-8b-hf
1870
+
1871
+ - name: together/llama-3-8b-instruct-turbo
1872
+ model_name: meta/llama-3-8b-instruct-turbo
1873
+ tokenizer_name: meta/llama-3-8b
1874
+ max_sequence_length: 8191
1875
+ client_spec:
1876
+ class_name: "helm.clients.together_client.TogetherClient"
1877
+ args:
1878
+ together_model: meta-llama/Meta-Llama-3-8B-Instruct-Turbo
1879
+
1880
+ - name: together/llama-3-8b-instruct-lite
1881
+ model_name: meta/llama-3-8b-instruct-lite
1882
+ tokenizer_name: meta/llama-3-8b
1883
+ max_sequence_length: 8191
1884
+ client_spec:
1885
+ class_name: "helm.clients.together_client.TogetherClient"
1886
+ args:
1887
+ together_model: meta-llama/Meta-Llama-3-8B-Instruct-Lite
1860
1888
 
1861
1889
  - name: together/llama-3-70b
1862
1890
  model_name: meta/llama-3-70b
@@ -1867,10 +1895,55 @@ model_deployments:
1867
1895
  args:
1868
1896
  together_model: meta-llama/Meta-Llama-3-70B
1869
1897
 
1870
- - name: together/llama-3-8b-chat
1871
- model_name: meta/llama-3-8b-chat
1898
+ - name: together/llama-3-70b-instruct-turbo
1899
+ model_name: meta/llama-3-70b-instruct-turbo
1872
1900
  tokenizer_name: meta/llama-3-8b
1873
1901
  max_sequence_length: 8191
1902
+ client_spec:
1903
+ class_name: "helm.clients.together_client.TogetherClient"
1904
+ args:
1905
+ together_model: meta-llama/Meta-Llama-3-70B-Instruct-Turbo
1906
+
1907
+ - name: together/llama-3-70b-instruct-lite
1908
+ model_name: meta/llama-3-70b-instruct-lite
1909
+ tokenizer_name: meta/llama-3-8b
1910
+ max_sequence_length: 8191
1911
+ client_spec:
1912
+ class_name: "helm.clients.together_client.TogetherClient"
1913
+ args:
1914
+ together_model: meta-llama/Meta-Llama-3-70B-Instruct-Lite
1915
+
1916
+ - name: together/llama-3.1-8b-instruct-turbo
1917
+ model_name: meta/llama-3.1-8b-instruct-turbo
1918
+ tokenizer_name: meta/llama-3.1-8b
1919
+ max_sequence_length: 128000
1920
+ client_spec:
1921
+ class_name: "helm.clients.together_client.TogetherChatClient"
1922
+ args:
1923
+ together_model: meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo
1924
+
1925
+ - name: together/llama-3.1-70b-instruct-turbo
1926
+ model_name: meta/llama-3.1-70b-instruct-turbo
1927
+ tokenizer_name: meta/llama-3.1-8b
1928
+ max_sequence_length: 128000
1929
+ client_spec:
1930
+ class_name: "helm.clients.together_client.TogetherChatClient"
1931
+ args:
1932
+ together_model: meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo
1933
+
1934
+ - name: together/llama-3.1-405b-instruct-turbo
1935
+ model_name: meta/llama-3.1-405b-instruct-turbo
1936
+ tokenizer_name: meta/llama-3.1-8b
1937
+ max_sequence_length: 128000
1938
+ client_spec:
1939
+ class_name: "helm.clients.together_client.TogetherChatClient"
1940
+ args:
1941
+ together_model: meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo
1942
+
1943
+ - name: together/llama-3-8b-chat
1944
+ model_name: meta/llama-3-8b-chat
1945
+ tokenizer_name: meta/llama-3-8b-instruct
1946
+ max_sequence_length: 8182
1874
1947
  client_spec:
1875
1948
  class_name: "helm.clients.together_client.TogetherChatClient"
1876
1949
  args:
@@ -1878,13 +1951,40 @@ model_deployments:
1878
1951
 
1879
1952
  - name: together/llama-3-70b-chat
1880
1953
  model_name: meta/llama-3-70b-chat
1881
- tokenizer_name: meta/llama-3-8b
1882
- max_sequence_length: 8191
1954
+ tokenizer_name: meta/llama-3-8b-instruct
1955
+ max_sequence_length: 8182
1883
1956
  client_spec:
1884
1957
  class_name: "helm.clients.together_client.TogetherChatClient"
1885
1958
  args:
1886
1959
  together_model: meta-llama/Llama-3-70b-chat-hf
1887
1960
 
1961
+ - name: together/llama-3.2-3b-instruct-turbo
1962
+ model_name: meta/llama-3.2-3b-instruct-turbo
1963
+ tokenizer_name: meta/llama-3.2-3b-instruct
1964
+ max_sequence_length: 128000
1965
+ client_spec:
1966
+ class_name: "helm.clients.together_client.TogetherChatClient"
1967
+ args:
1968
+ together_model: meta-llama/Llama-3.2-3B-Instruct-Turbo
1969
+
1970
+ - name: together/llama-3.2-11b-vision-instruct-turbo
1971
+ model_name: meta/llama-3.2-11b-vision-instruct-turbo
1972
+ tokenizer_name: meta/llama-3.2-11b-vision-instruct
1973
+ max_sequence_length: 128000
1974
+ client_spec:
1975
+ class_name: "helm.clients.together_client.TogetherChatClient"
1976
+ args:
1977
+ together_model: meta-llama/Llama-3.2-11B-Vision-Instruct-Turbo
1978
+
1979
+ - name: together/llama-3.2-90b-vision-instruct-turbo
1980
+ model_name: meta/llama-3.2-90b-vision-instruct-turbo
1981
+ tokenizer_name: meta/llama-3.2-11b-vision-instruct
1982
+ max_sequence_length: 128000
1983
+ client_spec:
1984
+ class_name: "helm.clients.together_client.TogetherChatClient"
1985
+ args:
1986
+ together_model: meta-llama/Llama-3.2-90B-Vision-Instruct-Turbo
1987
+
1888
1988
  - name: together/llama-guard-7b
1889
1989
  model_name: meta/llama-guard-7b
1890
1990
  tokenizer_name: meta-llama/Llama-2-7b-hf
@@ -1903,6 +2003,15 @@ model_deployments:
1903
2003
  args:
1904
2004
  together_model: meta-llama/llamaguard-2-8b
1905
2005
 
2006
+ - name: together/llama-guard-3-8b
2007
+ model_name: meta/llama-guard-3-8b
2008
+ tokenizer_name: meta/llama-3.1-8b
2009
+ max_sequence_length: 128000
2010
+ client_spec:
2011
+ class_name: "helm.clients.together_client.TogetherClient"
2012
+ args:
2013
+ together_model: meta-llama/Meta-Llama-Guard-3-8B
2014
+
1906
2015
  # 01.AI
1907
2016
  - name: together/yi-6b
1908
2017
  model_name: 01-ai/yi-6b
@@ -2029,7 +2138,7 @@ model_deployments:
2029
2138
  - name: together/mixtral-8x7b-instruct-v0.1
2030
2139
  model_name: mistralai/mixtral-8x7b-instruct-v0.1
2031
2140
  tokenizer_name: mistralai/Mistral-7B-v0.1
2032
- max_sequence_length: 4095 # Subtract 1 token to work around a off-by-one bug in Together's input validation token counting (#2080)
2141
+ max_sequence_length: 32767
2033
2142
  client_spec:
2034
2143
  class_name: "helm.clients.together_client.TogetherChatClient"
2035
2144
 
@@ -2165,18 +2274,6 @@ model_deployments:
2165
2274
  args:
2166
2275
  together_model: togethercomputer/RedPajama-INCITE-7B-Instruct
2167
2276
 
2168
- ## Tsinghua
2169
- - name: together/glm
2170
- deprecated: true # Removed from Together
2171
- model_name: tsinghua/glm
2172
- tokenizer_name: TsinghuaKEG/ice
2173
- max_sequence_length: 2048
2174
- max_request_length: 2049
2175
- client_spec:
2176
- class_name: "helm.clients.together_client.TogetherClient"
2177
- window_service_spec:
2178
- class_name: "helm.benchmark.window_services.ice_window_service.ICEWindowService"
2179
-
2180
2277
  - name: thudm/cogview2
2181
2278
  model_name: thudm/cogview2
2182
2279
  tokenizer_name: openai/clip-vit-large-patch14
@@ -2215,22 +2312,6 @@ model_deployments:
2215
2312
  client_spec:
2216
2313
  class_name: "helm.clients.palmyra_client.PalmyraClient"
2217
2314
 
2218
- - name: writer/palmyra-instruct-30
2219
- model_name: writer/palmyra-instruct-30
2220
- tokenizer_name: writer/gpt2
2221
- max_sequence_length: 2048
2222
- max_sequence_and_generated_tokens_length: 2048
2223
- client_spec:
2224
- class_name: "helm.clients.palmyra_client.PalmyraClient"
2225
-
2226
- - name: writer/palmyra-e
2227
- model_name: writer/palmyra-e
2228
- tokenizer_name: writer/gpt2
2229
- max_sequence_length: 2048
2230
- max_sequence_and_generated_tokens_length: 2048
2231
- client_spec:
2232
- class_name: "helm.clients.palmyra_client.PalmyraClient"
2233
-
2234
2315
  - name: writer/silk-road
2235
2316
  model_name: writer/silk-road
2236
2317
  tokenizer_name: writer/gpt2
@@ -2279,6 +2360,40 @@ model_deployments:
2279
2360
  client_spec:
2280
2361
  class_name: "helm.clients.vision_language.palmyra_vision_client.PalmyraVisionClient"
2281
2362
 
2363
+ - name: writer/palmyra-x-004
2364
+ model_name: writer/palmyra-x-004
2365
+ # Actual tokenizer is Llama 2, but it cannot be used in HELM due to this issue:
2366
+ # https://github.com/stanford-crfm/helm/issues/2467
2367
+ # Work around by using Llama 3 tokenizer for now.
2368
+ tokenizer_name: meta/llama-3-8b
2369
+ max_sequence_length: 8192
2370
+ client_spec:
2371
+ class_name: "helm.clients.palmyra_client.PalmyraChatClient"
2372
+
2373
+ - name: writer/palmyra-med-32k
2374
+ model_name: writer/palmyra-med-32k
2375
+ # Palmyra-Med uses the "<|end_of_text|>" as the end of text token, which is used by meta/llama-3-8b,
2376
+ # rather than "<|eot_id|>", which is used by meta/llama-3-8b-instruct
2377
+ tokenizer_name: meta/llama-3-8b
2378
+ max_sequence_length: 32000
2379
+ client_spec:
2380
+ class_name: "helm.clients.palmyra_client.PalmyraChatClient"
2381
+
2382
+ - name: writer/palmyra-med
2383
+ model_name: writer/palmyra-med
2384
+ # Palmyra-Med uses the "<|end_of_text|>" as the end of text token, which is used by meta/llama-3-8b,
2385
+ # rather than "<|eot_id|>", which is used by meta/llama-3-8b-instruct
2386
+ tokenizer_name: meta/llama-3-8b
2387
+ max_sequence_length: 4096
2388
+ client_spec:
2389
+ class_name: "helm.clients.palmyra_client.PalmyraChatClient"
2390
+
2391
+ - name: writer/palmyra-fin-32k
2392
+ model_name: writer/palmyra-fin-32k
2393
+ tokenizer_name: meta/llama-3-8b-instruct
2394
+ max_sequence_length: 32000
2395
+ client_spec:
2396
+ class_name: "helm.clients.palmyra_client.PalmyraChatClient"
2282
2397
 
2283
2398
  # Qwen
2284
2399
 
@@ -2431,4 +2546,4 @@ model_deployments:
2431
2546
  tokenizer_name: openai/cl100k_base
2432
2547
  max_sequence_length: 64000
2433
2548
  client_spec:
2434
- class_name: "helm.clients.reka_client.RekaClient"
2549
+ class_name: "helm.clients.reka_client.RekaClient"