crfm-helm 0.5.2__py3-none-any.whl → 0.5.4__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of crfm-helm might be problematic. Click here for more details.
- {crfm_helm-0.5.2.dist-info → crfm_helm-0.5.4.dist-info}/METADATA +81 -112
- {crfm_helm-0.5.2.dist-info → crfm_helm-0.5.4.dist-info}/RECORD +165 -155
- {crfm_helm-0.5.2.dist-info → crfm_helm-0.5.4.dist-info}/WHEEL +1 -1
- helm/benchmark/adaptation/adapters/multiple_choice_joint_adapter.py +12 -5
- helm/benchmark/adaptation/adapters/test_generation_adapter.py +12 -12
- helm/benchmark/adaptation/adapters/test_language_modeling_adapter.py +8 -8
- helm/benchmark/adaptation/adapters/test_multiple_choice_joint_adapter.py +77 -9
- helm/benchmark/adaptation/common_adapter_specs.py +2 -0
- helm/benchmark/annotation/anthropic_red_team_annotator.py +57 -0
- helm/benchmark/annotation/call_center_annotator.py +258 -0
- helm/benchmark/annotation/financebench_annotator.py +79 -0
- helm/benchmark/annotation/harm_bench_annotator.py +55 -0
- helm/benchmark/annotation/{image2structure → image2struct}/latex_compiler_annotator.py +2 -2
- helm/benchmark/annotation/{image2structure → image2struct}/lilypond_compiler_annotator.py +5 -3
- helm/benchmark/annotation/{image2structure → image2struct}/webpage_compiler_annotator.py +5 -5
- helm/benchmark/annotation/live_qa_annotator.py +37 -45
- helm/benchmark/annotation/medication_qa_annotator.py +36 -44
- helm/benchmark/annotation/model_as_judge.py +96 -0
- helm/benchmark/annotation/simple_safety_tests_annotator.py +50 -0
- helm/benchmark/annotation/xstest_annotator.py +100 -0
- helm/benchmark/metrics/annotation_metrics.py +108 -0
- helm/benchmark/metrics/bhasa_metrics.py +188 -0
- helm/benchmark/metrics/bhasa_metrics_specs.py +10 -0
- helm/benchmark/metrics/code_metrics_helper.py +11 -1
- helm/benchmark/metrics/safety_metrics.py +79 -0
- helm/benchmark/metrics/summac/model_summac.py +3 -3
- helm/benchmark/metrics/tokens/test_ai21_token_cost_estimator.py +2 -2
- helm/benchmark/metrics/tokens/test_openai_token_cost_estimator.py +4 -4
- helm/benchmark/metrics/unitxt_metrics.py +17 -3
- helm/benchmark/metrics/vision_language/image_metrics.py +7 -3
- helm/benchmark/metrics/vision_language/image_utils.py +1 -1
- helm/benchmark/model_metadata_registry.py +3 -3
- helm/benchmark/presentation/create_plots.py +1 -1
- helm/benchmark/presentation/schema.py +3 -0
- helm/benchmark/presentation/summarize.py +106 -256
- helm/benchmark/presentation/test_run_entry.py +1 -0
- helm/benchmark/presentation/test_summarize.py +145 -3
- helm/benchmark/run.py +15 -0
- helm/benchmark/run_expander.py +83 -30
- helm/benchmark/run_specs/bhasa_run_specs.py +652 -0
- helm/benchmark/run_specs/call_center_run_specs.py +152 -0
- helm/benchmark/run_specs/decodingtrust_run_specs.py +8 -8
- helm/benchmark/run_specs/experimental_run_specs.py +52 -0
- helm/benchmark/run_specs/finance_run_specs.py +82 -1
- helm/benchmark/run_specs/safety_run_specs.py +154 -0
- helm/benchmark/run_specs/vlm_run_specs.py +100 -24
- helm/benchmark/scenarios/anthropic_red_team_scenario.py +71 -0
- helm/benchmark/scenarios/banking77_scenario.py +51 -0
- helm/benchmark/scenarios/bhasa_scenario.py +1942 -0
- helm/benchmark/scenarios/call_center_scenario.py +84 -0
- helm/benchmark/scenarios/decodingtrust_stereotype_bias_scenario.py +2 -1
- helm/benchmark/scenarios/ewok_scenario.py +116 -0
- helm/benchmark/scenarios/fin_qa_scenario.py +2 -0
- helm/benchmark/scenarios/financebench_scenario.py +53 -0
- helm/benchmark/scenarios/harm_bench_scenario.py +59 -0
- helm/benchmark/scenarios/raft_scenario.py +1 -1
- helm/benchmark/scenarios/scenario.py +1 -1
- helm/benchmark/scenarios/simple_safety_tests_scenario.py +33 -0
- helm/benchmark/scenarios/test_commonsense_scenario.py +21 -0
- helm/benchmark/scenarios/test_ewok_scenario.py +25 -0
- helm/benchmark/scenarios/test_financebench_scenario.py +26 -0
- helm/benchmark/scenarios/test_gsm_scenario.py +31 -0
- helm/benchmark/scenarios/test_legalbench_scenario.py +30 -0
- helm/benchmark/scenarios/test_math_scenario.py +2 -8
- helm/benchmark/scenarios/test_med_qa_scenario.py +30 -0
- helm/benchmark/scenarios/test_mmlu_scenario.py +33 -0
- helm/benchmark/scenarios/test_narrativeqa_scenario.py +73 -0
- helm/benchmark/scenarios/thai_exam_scenario.py +4 -4
- helm/benchmark/scenarios/vision_language/a_okvqa_scenario.py +1 -1
- helm/benchmark/scenarios/vision_language/bingo_scenario.py +2 -2
- helm/benchmark/scenarios/vision_language/crossmodal_3600_scenario.py +2 -1
- helm/benchmark/scenarios/vision_language/exams_v_scenario.py +104 -0
- helm/benchmark/scenarios/vision_language/fair_face_scenario.py +136 -0
- helm/benchmark/scenarios/vision_language/flickr30k_scenario.py +1 -1
- helm/benchmark/scenarios/vision_language/gqa_scenario.py +2 -2
- helm/benchmark/scenarios/vision_language/hateful_memes_scenario.py +1 -1
- helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/chart2csv_scenario.py +1 -1
- helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/latex_scenario.py +3 -3
- helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/musicsheet_scenario.py +1 -1
- helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/utils_latex.py +31 -39
- helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/webpage/driver.py +1 -1
- helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/webpage/utils.py +1 -1
- helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/webpage_scenario.py +41 -12
- helm/benchmark/scenarios/vision_language/math_vista_scenario.py +1 -1
- helm/benchmark/scenarios/vision_language/mementos_scenario.py +3 -3
- helm/benchmark/scenarios/vision_language/mm_safety_bench_scenario.py +2 -2
- helm/benchmark/scenarios/vision_language/mme_scenario.py +21 -18
- helm/benchmark/scenarios/vision_language/mmmu_scenario.py +1 -1
- helm/benchmark/scenarios/vision_language/pairs_scenario.py +1 -1
- helm/benchmark/scenarios/vision_language/pope_scenario.py +2 -1
- helm/benchmark/scenarios/vision_language/real_world_qa_scenario.py +57 -0
- helm/benchmark/scenarios/vision_language/seed_bench_scenario.py +7 -5
- helm/benchmark/scenarios/vision_language/unicorn_scenario.py +2 -2
- helm/benchmark/scenarios/vision_language/vibe_eval_scenario.py +6 -3
- helm/benchmark/scenarios/vision_language/viz_wiz_scenario.py +1 -1
- helm/benchmark/scenarios/vision_language/vqa_scenario.py +3 -1
- helm/benchmark/scenarios/xstest_scenario.py +35 -0
- helm/benchmark/server.py +1 -6
- helm/benchmark/static/schema_air_bench.yaml +750 -750
- helm/benchmark/static/schema_bhasa.yaml +709 -0
- helm/benchmark/static/schema_call_center.yaml +232 -0
- helm/benchmark/static/schema_cleva.yaml +768 -0
- helm/benchmark/static/schema_decodingtrust.yaml +444 -0
- helm/benchmark/static/schema_ewok.yaml +367 -0
- helm/benchmark/static/schema_finance.yaml +55 -9
- helm/benchmark/static/{schema_image2structure.yaml → schema_image2struct.yaml} +231 -90
- helm/benchmark/static/schema_legal.yaml +566 -0
- helm/benchmark/static/schema_safety.yaml +266 -0
- helm/benchmark/static/schema_tables.yaml +149 -8
- helm/benchmark/static/schema_thai.yaml +21 -0
- helm/benchmark/static/schema_vhelm.yaml +137 -101
- helm/benchmark/static_build/assets/accenture-6f97eeda.png +0 -0
- helm/benchmark/static_build/assets/aisingapore-6dfc9acf.png +0 -0
- helm/benchmark/static_build/assets/cresta-9e22b983.png +0 -0
- helm/benchmark/static_build/assets/cuhk-8c5631e9.png +0 -0
- helm/benchmark/static_build/assets/index-05c76bb1.css +1 -0
- helm/benchmark/static_build/assets/index-3ee38b3d.js +10 -0
- helm/benchmark/static_build/assets/scb10x-204bd786.png +0 -0
- helm/benchmark/static_build/assets/vhelm-aspects-1437d673.png +0 -0
- helm/benchmark/static_build/assets/vhelm-framework-a1ca3f3f.png +0 -0
- helm/benchmark/static_build/assets/vhelm-model-8afb7616.png +0 -0
- helm/benchmark/static_build/assets/wellsfargo-a86a6c4a.png +0 -0
- helm/benchmark/static_build/index.html +2 -2
- helm/benchmark/window_services/test_openai_window_service.py +8 -8
- helm/benchmark/window_services/tokenizer_service.py +0 -5
- helm/clients/ai21_client.py +71 -1
- helm/clients/anthropic_client.py +7 -19
- helm/clients/huggingface_client.py +38 -37
- helm/clients/nvidia_nim_client.py +35 -0
- helm/clients/openai_client.py +18 -4
- helm/clients/palmyra_client.py +24 -0
- helm/clients/perspective_api_client.py +11 -6
- helm/clients/test_client.py +4 -6
- helm/clients/together_client.py +22 -0
- helm/clients/vision_language/open_flamingo_client.py +1 -2
- helm/clients/vision_language/palmyra_vision_client.py +28 -13
- helm/common/cache.py +8 -30
- helm/common/images_utils.py +6 -0
- helm/common/key_value_store.py +9 -9
- helm/common/mongo_key_value_store.py +5 -4
- helm/common/request.py +16 -0
- helm/common/test_cache.py +1 -48
- helm/common/tokenization_request.py +0 -9
- helm/config/model_deployments.yaml +444 -329
- helm/config/model_metadata.yaml +513 -111
- helm/config/tokenizer_configs.yaml +140 -11
- helm/proxy/example_queries.py +14 -21
- helm/proxy/server.py +0 -9
- helm/proxy/services/remote_service.py +0 -6
- helm/proxy/services/server_service.py +6 -20
- helm/proxy/services/service.py +0 -6
- helm/proxy/token_counters/test_auto_token_counter.py +2 -2
- helm/tokenizers/ai21_tokenizer.py +51 -59
- helm/tokenizers/cohere_tokenizer.py +0 -75
- helm/tokenizers/huggingface_tokenizer.py +0 -1
- helm/tokenizers/test_ai21_tokenizer.py +48 -0
- helm/benchmark/data_overlap/data_overlap_spec.py +0 -86
- helm/benchmark/data_overlap/export_scenario_text.py +0 -119
- helm/benchmark/data_overlap/light_scenario.py +0 -60
- helm/benchmark/scenarios/vision_language/image2structure/webpage/__init__.py +0 -0
- helm/benchmark/static/benchmarking.css +0 -156
- helm/benchmark/static/benchmarking.js +0 -1705
- helm/benchmark/static/config.js +0 -3
- helm/benchmark/static/general.js +0 -122
- helm/benchmark/static/images/crfm-logo.png +0 -0
- helm/benchmark/static/images/helm-logo-simple.png +0 -0
- helm/benchmark/static/images/helm-logo.png +0 -0
- helm/benchmark/static/images/language-model-helm.png +0 -0
- helm/benchmark/static/images/organizations/ai21.png +0 -0
- helm/benchmark/static/images/organizations/anthropic.png +0 -0
- helm/benchmark/static/images/organizations/bigscience.png +0 -0
- helm/benchmark/static/images/organizations/cohere.png +0 -0
- helm/benchmark/static/images/organizations/eleutherai.png +0 -0
- helm/benchmark/static/images/organizations/google.png +0 -0
- helm/benchmark/static/images/organizations/meta.png +0 -0
- helm/benchmark/static/images/organizations/microsoft.png +0 -0
- helm/benchmark/static/images/organizations/nvidia.png +0 -0
- helm/benchmark/static/images/organizations/openai.png +0 -0
- helm/benchmark/static/images/organizations/together.png +0 -0
- helm/benchmark/static/images/organizations/tsinghua-keg.png +0 -0
- helm/benchmark/static/images/organizations/yandex.png +0 -0
- helm/benchmark/static/images/scenarios-by-metrics.png +0 -0
- helm/benchmark/static/images/taxonomy-scenarios.png +0 -0
- helm/benchmark/static/index.html +0 -68
- helm/benchmark/static/info-icon.png +0 -0
- helm/benchmark/static/json-urls.js +0 -69
- helm/benchmark/static/plot-captions.js +0 -27
- helm/benchmark/static/utils.js +0 -285
- helm/benchmark/static_build/assets/index-30dbceba.js +0 -10
- helm/benchmark/static_build/assets/index-66b02d40.css +0 -1
- helm/benchmark/static_build/assets/vhelm-framework-cde7618a.png +0 -0
- helm/benchmark/static_build/assets/vhelm-model-6d812526.png +0 -0
- helm/benchmark/window_services/ai21_window_service.py +0 -247
- helm/benchmark/window_services/cohere_window_service.py +0 -101
- helm/benchmark/window_services/test_ai21_window_service.py +0 -163
- helm/benchmark/window_services/test_cohere_window_service.py +0 -75
- helm/benchmark/window_services/test_cohere_window_service_utils.py +0 -8328
- helm/benchmark/window_services/test_ice_window_service.py +0 -327
- helm/tokenizers/ice_tokenizer.py +0 -30
- helm/tokenizers/test_ice_tokenizer.py +0 -57
- {crfm_helm-0.5.2.dist-info → crfm_helm-0.5.4.dist-info}/LICENSE +0 -0
- {crfm_helm-0.5.2.dist-info → crfm_helm-0.5.4.dist-info}/entry_points.txt +0 -0
- {crfm_helm-0.5.2.dist-info → crfm_helm-0.5.4.dist-info}/top_level.txt +0 -0
- /helm/benchmark/annotation/{image2structure → image2struct}/__init__.py +0 -0
- /helm/benchmark/annotation/{image2structure → image2struct}/image_compiler_annotator.py +0 -0
- /helm/benchmark/{data_overlap → scenarios/vision_language/image2struct}/__init__.py +0 -0
- /helm/benchmark/scenarios/vision_language/{image2structure/image2structure_scenario.py → image2struct/image2struct_scenario.py} +0 -0
- /helm/benchmark/scenarios/vision_language/{image2structure → image2struct/webpage}/__init__.py +0 -0
- /helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/webpage/jekyll_server.py +0 -0
|
@@ -28,74 +28,47 @@ model_deployments:
|
|
|
28
28
|
|
|
29
29
|
# AI21 Labs
|
|
30
30
|
|
|
31
|
-
|
|
32
|
-
|
|
33
|
-
|
|
34
|
-
deprecated: true
|
|
35
|
-
model_name: ai21/j1-jumbo
|
|
36
|
-
tokenizer_name: ai21/j1
|
|
37
|
-
max_sequence_length: 2047
|
|
38
|
-
client_spec:
|
|
39
|
-
class_name: "helm.clients.ai21_client.AI21Client"
|
|
40
|
-
window_service_spec:
|
|
41
|
-
class_name: "helm.benchmark.window_services.ai21_window_service.AI21WindowService"
|
|
42
|
-
|
|
43
|
-
- name: ai21/j1-large
|
|
44
|
-
deprecated: true
|
|
45
|
-
model_name: ai21/j1-large
|
|
46
|
-
tokenizer_name: ai21/j1
|
|
47
|
-
max_sequence_length: 2047
|
|
48
|
-
client_spec:
|
|
49
|
-
class_name: "helm.clients.ai21_client.AI21Client"
|
|
50
|
-
window_service_spec:
|
|
51
|
-
class_name: "helm.benchmark.window_services.ai21_window_service.AI21WindowService"
|
|
52
|
-
|
|
53
|
-
- name: ai21/j1-grande
|
|
54
|
-
deprecated: true
|
|
55
|
-
model_name: ai21/j1-grande
|
|
56
|
-
tokenizer_name: ai21/j1
|
|
31
|
+
- name: ai21/j2-large
|
|
32
|
+
model_name: ai21/j2-large
|
|
33
|
+
tokenizer_name: ai21/j2-tokenizer
|
|
57
34
|
max_sequence_length: 2047
|
|
58
35
|
client_spec:
|
|
59
36
|
class_name: "helm.clients.ai21_client.AI21Client"
|
|
60
|
-
window_service_spec:
|
|
61
|
-
class_name: "helm.benchmark.window_services.ai21_window_service.AI21WindowService"
|
|
62
37
|
|
|
63
|
-
- name: ai21/
|
|
64
|
-
|
|
65
|
-
|
|
66
|
-
tokenizer_name: ai21/j1
|
|
38
|
+
- name: ai21/j2-grande
|
|
39
|
+
model_name: ai21/j2-grande
|
|
40
|
+
tokenizer_name: ai21/j2-tokenizer
|
|
67
41
|
max_sequence_length: 2047
|
|
68
42
|
client_spec:
|
|
69
43
|
class_name: "helm.clients.ai21_client.AI21Client"
|
|
70
|
-
window_service_spec:
|
|
71
|
-
class_name: "helm.benchmark.window_services.ai21_window_service.AI21WindowService"
|
|
72
44
|
|
|
73
45
|
- name: ai21/j2-jumbo
|
|
74
46
|
model_name: ai21/j2-jumbo
|
|
75
|
-
tokenizer_name: ai21/
|
|
47
|
+
tokenizer_name: ai21/j2-tokenizer
|
|
76
48
|
max_sequence_length: 6000
|
|
77
49
|
client_spec:
|
|
78
50
|
class_name: "helm.clients.ai21_client.AI21Client"
|
|
79
|
-
window_service_spec:
|
|
80
|
-
class_name: "helm.benchmark.window_services.ai21_window_service.AI21WindowService"
|
|
81
51
|
|
|
82
|
-
- name: ai21/
|
|
83
|
-
model_name: ai21/
|
|
84
|
-
tokenizer_name: ai21/
|
|
85
|
-
max_sequence_length:
|
|
52
|
+
- name: ai21/jamba-instruct
|
|
53
|
+
model_name: ai21/jamba-instruct
|
|
54
|
+
tokenizer_name: ai21/jamba-instruct-tokenizer
|
|
55
|
+
max_sequence_length: 256000
|
|
86
56
|
client_spec:
|
|
87
|
-
class_name: "helm.clients.ai21_client.
|
|
88
|
-
window_service_spec:
|
|
89
|
-
class_name: "helm.benchmark.window_services.ai21_window_service.AI21WindowService"
|
|
57
|
+
class_name: "helm.clients.ai21_client.AI21ChatClient"
|
|
90
58
|
|
|
91
|
-
- name: ai21/
|
|
92
|
-
model_name: ai21/
|
|
93
|
-
tokenizer_name: ai21/
|
|
94
|
-
max_sequence_length:
|
|
59
|
+
- name: ai21/jamba-1.5-mini
|
|
60
|
+
model_name: ai21/jamba-1.5-mini
|
|
61
|
+
tokenizer_name: ai21/jamba-1.5-mini-tokenizer
|
|
62
|
+
max_sequence_length: 256000
|
|
95
63
|
client_spec:
|
|
96
|
-
class_name: "helm.clients.ai21_client.
|
|
97
|
-
|
|
98
|
-
|
|
64
|
+
class_name: "helm.clients.ai21_client.AI21ChatClient"
|
|
65
|
+
|
|
66
|
+
- name: ai21/jamba-1.5-large
|
|
67
|
+
model_name: ai21/jamba-1.5-large
|
|
68
|
+
tokenizer_name: ai21/jamba-1.5-large-tokenizer
|
|
69
|
+
max_sequence_length: 256000
|
|
70
|
+
client_spec:
|
|
71
|
+
class_name: "helm.clients.ai21_client.AI21ChatClient"
|
|
99
72
|
|
|
100
73
|
# Aleph Alpha
|
|
101
74
|
- name: AlephAlpha/luminous-base
|
|
@@ -216,6 +189,13 @@ model_deployments:
|
|
|
216
189
|
client_spec:
|
|
217
190
|
class_name: "helm.clients.anthropic_client.AnthropicMessagesClient"
|
|
218
191
|
|
|
192
|
+
- name: anthropic/claude-3-5-sonnet-20240620
|
|
193
|
+
model_name: anthropic/claude-3-5-sonnet-20240620
|
|
194
|
+
tokenizer_name: anthropic/claude
|
|
195
|
+
max_sequence_length: 200000
|
|
196
|
+
client_spec:
|
|
197
|
+
class_name: "helm.clients.anthropic_client.AnthropicMessagesClient"
|
|
198
|
+
|
|
219
199
|
- name: anthropic/stanford-online-all-v4-s3
|
|
220
200
|
deprecated: true # Closed model, not accessible via API
|
|
221
201
|
model_name: anthropic/stanford-online-all-v4-s3
|
|
@@ -225,86 +205,6 @@ model_deployments:
|
|
|
225
205
|
class_name: "helm.clients.anthropic_client.AnthropicLegacyClient"
|
|
226
206
|
|
|
227
207
|
# Cohere
|
|
228
|
-
- name: cohere/xlarge-20220609
|
|
229
|
-
model_name: cohere/xlarge-20220609
|
|
230
|
-
tokenizer_name: cohere/cohere
|
|
231
|
-
max_sequence_length: 2047
|
|
232
|
-
max_request_length: 2048
|
|
233
|
-
client_spec:
|
|
234
|
-
class_name: "helm.clients.cohere_client.CohereClient"
|
|
235
|
-
window_service_spec:
|
|
236
|
-
class_name: "helm.benchmark.window_services.cohere_window_service.CohereWindowService"
|
|
237
|
-
|
|
238
|
-
- name: cohere/large-20220720
|
|
239
|
-
model_name: cohere/large-20220720
|
|
240
|
-
tokenizer_name: cohere/cohere
|
|
241
|
-
max_sequence_length: 2047
|
|
242
|
-
max_request_length: 2048
|
|
243
|
-
client_spec:
|
|
244
|
-
class_name: "helm.clients.cohere_client.CohereClient"
|
|
245
|
-
window_service_spec:
|
|
246
|
-
class_name: "helm.benchmark.window_services.cohere_window_service.CohereWindowService"
|
|
247
|
-
|
|
248
|
-
- name: cohere/medium-20220720
|
|
249
|
-
model_name: cohere/medium-20220720
|
|
250
|
-
tokenizer_name: cohere/cohere
|
|
251
|
-
max_sequence_length: 2047
|
|
252
|
-
max_request_length: 2048
|
|
253
|
-
client_spec:
|
|
254
|
-
class_name: "helm.clients.cohere_client.CohereClient"
|
|
255
|
-
window_service_spec:
|
|
256
|
-
class_name: "helm.benchmark.window_services.cohere_window_service.CohereWindowService"
|
|
257
|
-
|
|
258
|
-
- name: cohere/small-20220720
|
|
259
|
-
model_name: cohere/small-20220720
|
|
260
|
-
tokenizer_name: cohere/cohere
|
|
261
|
-
max_sequence_length: 2047
|
|
262
|
-
max_request_length: 2048
|
|
263
|
-
client_spec:
|
|
264
|
-
class_name: "helm.clients.cohere_client.CohereClient"
|
|
265
|
-
window_service_spec:
|
|
266
|
-
class_name: "helm.benchmark.window_services.cohere_window_service.CohereWindowService"
|
|
267
|
-
|
|
268
|
-
- name: cohere/xlarge-20221108
|
|
269
|
-
model_name: cohere/xlarge-20221108
|
|
270
|
-
tokenizer_name: cohere/cohere
|
|
271
|
-
max_sequence_length: 2047
|
|
272
|
-
max_request_length: 2048
|
|
273
|
-
client_spec:
|
|
274
|
-
class_name: "helm.clients.cohere_client.CohereClient"
|
|
275
|
-
window_service_spec:
|
|
276
|
-
class_name: "helm.benchmark.window_services.cohere_window_service.CohereWindowService"
|
|
277
|
-
|
|
278
|
-
- name: cohere/medium-20221108
|
|
279
|
-
model_name: cohere/medium-20221108
|
|
280
|
-
tokenizer_name: cohere/cohere
|
|
281
|
-
max_sequence_length: 2047
|
|
282
|
-
max_request_length: 2048
|
|
283
|
-
client_spec:
|
|
284
|
-
class_name: "helm.clients.cohere_client.CohereClient"
|
|
285
|
-
window_service_spec:
|
|
286
|
-
class_name: "helm.benchmark.window_services.cohere_window_service.CohereWindowService"
|
|
287
|
-
|
|
288
|
-
- name: cohere/command-medium-beta
|
|
289
|
-
model_name: cohere/command-medium-beta
|
|
290
|
-
tokenizer_name: cohere/cohere
|
|
291
|
-
max_sequence_length: 2019
|
|
292
|
-
max_request_length: 2020
|
|
293
|
-
client_spec:
|
|
294
|
-
class_name: "helm.clients.cohere_client.CohereClient"
|
|
295
|
-
window_service_spec:
|
|
296
|
-
class_name: "helm.benchmark.window_services.cohere_window_service.CohereWindowService"
|
|
297
|
-
|
|
298
|
-
- name: cohere/command-xlarge-beta
|
|
299
|
-
model_name: cohere/command-xlarge-beta
|
|
300
|
-
tokenizer_name: cohere/cohere
|
|
301
|
-
max_sequence_length: 2019
|
|
302
|
-
max_request_length: 2020
|
|
303
|
-
client_spec:
|
|
304
|
-
class_name: "helm.clients.cohere_client.CohereClient"
|
|
305
|
-
window_service_spec:
|
|
306
|
-
class_name: "helm.benchmark.window_services.cohere_window_service.CohereWindowService"
|
|
307
|
-
|
|
308
208
|
- name: cohere/command
|
|
309
209
|
model_name: cohere/command
|
|
310
210
|
tokenizer_name: cohere/command
|
|
@@ -312,8 +212,6 @@ model_deployments:
|
|
|
312
212
|
max_request_length: 2020 # TODO: verify this
|
|
313
213
|
client_spec:
|
|
314
214
|
class_name: "helm.clients.cohere_client.CohereClient"
|
|
315
|
-
window_service_spec:
|
|
316
|
-
class_name: "helm.benchmark.window_services.cohere_window_service.CohereWindowService"
|
|
317
215
|
|
|
318
216
|
- name: cohere/command-light
|
|
319
217
|
model_name: cohere/command-light
|
|
@@ -322,8 +220,6 @@ model_deployments:
|
|
|
322
220
|
max_request_length: 2020 # TODO: verify this
|
|
323
221
|
client_spec:
|
|
324
222
|
class_name: "helm.clients.cohere_client.CohereClient"
|
|
325
|
-
window_service_spec:
|
|
326
|
-
class_name: "helm.benchmark.window_services.cohere_window_service.CohereWindowService"
|
|
327
223
|
|
|
328
224
|
- name: cohere/command-r
|
|
329
225
|
model_name: cohere/command-r
|
|
@@ -587,6 +483,20 @@ model_deployments:
|
|
|
587
483
|
client_spec:
|
|
588
484
|
class_name: "helm.clients.together_client.TogetherClient"
|
|
589
485
|
|
|
486
|
+
- name: together/gemma-2-9b-it
|
|
487
|
+
model_name: google/gemma-2-9b-it
|
|
488
|
+
tokenizer_name: google/gemma-2-9b
|
|
489
|
+
max_sequence_length: 8191
|
|
490
|
+
client_spec:
|
|
491
|
+
class_name: "helm.clients.together_client.TogetherClient"
|
|
492
|
+
|
|
493
|
+
- name: together/gemma-2-27b-it
|
|
494
|
+
model_name: google/gemma-2-27b-it
|
|
495
|
+
tokenizer_name: google/gemma-2-9b
|
|
496
|
+
max_sequence_length: 8191
|
|
497
|
+
client_spec:
|
|
498
|
+
class_name: "helm.clients.together_client.TogetherClient"
|
|
499
|
+
|
|
590
500
|
## MedLM
|
|
591
501
|
- name: google/medlm-medium
|
|
592
502
|
model_name: google/medlm-medium
|
|
@@ -716,6 +626,26 @@ model_deployments:
|
|
|
716
626
|
args:
|
|
717
627
|
trust_remote_code: true
|
|
718
628
|
|
|
629
|
+
- name: huggingface/llama3-8b-cpt-sea-lionv2-base
|
|
630
|
+
model_name: aisingapore/llama3-8b-cpt-sea-lionv2-base
|
|
631
|
+
tokenizer_name: meta/llama-3-8b-instruct
|
|
632
|
+
max_sequence_length: 8192
|
|
633
|
+
client_spec:
|
|
634
|
+
class_name: "helm.clients.huggingface_client.HuggingFaceClient"
|
|
635
|
+
args:
|
|
636
|
+
device_map: auto
|
|
637
|
+
torch_dtype: torch.bfloat16
|
|
638
|
+
|
|
639
|
+
- name: huggingface/llama3-8b-cpt-sea-lionv2.1-instruct
|
|
640
|
+
model_name: aisingapore/llama3-8b-cpt-sea-lionv2.1-instruct
|
|
641
|
+
tokenizer_name: meta/llama-3-8b-instruct
|
|
642
|
+
max_sequence_length: 8192
|
|
643
|
+
client_spec:
|
|
644
|
+
class_name: "helm.clients.huggingface_client.HuggingFaceClient"
|
|
645
|
+
args:
|
|
646
|
+
device_map: auto
|
|
647
|
+
torch_dtype: torch.bfloat16
|
|
648
|
+
|
|
719
649
|
## Bigcode
|
|
720
650
|
- name: huggingface/santacoder
|
|
721
651
|
model_name: bigcode/santacoder
|
|
@@ -807,6 +737,47 @@ model_deployments:
|
|
|
807
737
|
client_spec:
|
|
808
738
|
class_name: "helm.clients.huggingface_client.HuggingFaceClient"
|
|
809
739
|
|
|
740
|
+
## Google
|
|
741
|
+
- name: huggingface/gemma-2-9b
|
|
742
|
+
model_name: google/gemma-2-9b
|
|
743
|
+
tokenizer_name: google/gemma-2-9b
|
|
744
|
+
max_sequence_length: 8192
|
|
745
|
+
client_spec:
|
|
746
|
+
class_name: "helm.clients.huggingface_client.HuggingFaceClient"
|
|
747
|
+
args:
|
|
748
|
+
device_map: auto
|
|
749
|
+
torch_dtype: torch.bfloat16
|
|
750
|
+
|
|
751
|
+
- name: huggingface/gemma-2-9b-it
|
|
752
|
+
model_name: google/gemma-2-9b-it
|
|
753
|
+
tokenizer_name: google/gemma-2-9b
|
|
754
|
+
max_sequence_length: 8192
|
|
755
|
+
client_spec:
|
|
756
|
+
class_name: "helm.clients.huggingface_client.HuggingFaceClient"
|
|
757
|
+
args:
|
|
758
|
+
device_map: auto
|
|
759
|
+
torch_dtype: torch.bfloat16
|
|
760
|
+
|
|
761
|
+
- name: huggingface/gemma-2-27b
|
|
762
|
+
model_name: google/gemma-2-27b
|
|
763
|
+
tokenizer_name: google/gemma-2-9b
|
|
764
|
+
max_sequence_length: 8192
|
|
765
|
+
client_spec:
|
|
766
|
+
class_name: "helm.clients.huggingface_client.HuggingFaceClient"
|
|
767
|
+
args:
|
|
768
|
+
device_map: auto
|
|
769
|
+
torch_dtype: torch.bfloat16
|
|
770
|
+
|
|
771
|
+
- name: huggingface/gemma-2-27b-it
|
|
772
|
+
model_name: google/gemma-2-27b-it
|
|
773
|
+
tokenizer_name: google/gemma-2-9b
|
|
774
|
+
max_sequence_length: 8192
|
|
775
|
+
client_spec:
|
|
776
|
+
class_name: "helm.clients.huggingface_client.HuggingFaceClient"
|
|
777
|
+
args:
|
|
778
|
+
device_map: auto
|
|
779
|
+
torch_dtype: torch.bfloat16
|
|
780
|
+
|
|
810
781
|
## LMSYS
|
|
811
782
|
- name: huggingface/vicuna-7b-v1.3
|
|
812
783
|
model_name: lmsys/vicuna-7b-v1.3
|
|
@@ -931,6 +902,7 @@ model_deployments:
|
|
|
931
902
|
tokenizer_name: "anas-awadalla-2/mpt-7b"
|
|
932
903
|
cross_attn_every_n_layers: 4
|
|
933
904
|
|
|
905
|
+
## Microsoft
|
|
934
906
|
- name: together/phi-2
|
|
935
907
|
model_name: microsoft/phi-2
|
|
936
908
|
tokenizer_name: microsoft/phi-2
|
|
@@ -938,6 +910,26 @@ model_deployments:
|
|
|
938
910
|
client_spec:
|
|
939
911
|
class_name: "helm.clients.together_client.TogetherClient"
|
|
940
912
|
|
|
913
|
+
- name: huggingface/phi-3-small-8k-instruct
|
|
914
|
+
model_name: microsoft/phi-3-small-8k-instruct
|
|
915
|
+
tokenizer_name: microsoft/phi-3-small-8k-instruct
|
|
916
|
+
max_sequence_length: 8192
|
|
917
|
+
client_spec:
|
|
918
|
+
class_name: "helm.clients.huggingface_client.HuggingFaceClient"
|
|
919
|
+
args:
|
|
920
|
+
torch_dtype: auto
|
|
921
|
+
trust_remote_code: true
|
|
922
|
+
|
|
923
|
+
- name: huggingface/phi-3-medium-4k-instruct
|
|
924
|
+
model_name: microsoft/phi-3-medium-4k-instruct
|
|
925
|
+
tokenizer_name: microsoft/phi-3-medium-4k-instruct
|
|
926
|
+
max_sequence_length: 4096
|
|
927
|
+
client_spec:
|
|
928
|
+
class_name: "helm.clients.huggingface_client.HuggingFaceClient"
|
|
929
|
+
args:
|
|
930
|
+
device_map: auto
|
|
931
|
+
torch_dtype: auto
|
|
932
|
+
|
|
941
933
|
## Mistral AI
|
|
942
934
|
- name: huggingface/bakLlava-v1-hf
|
|
943
935
|
model_name: mistralai/bakLlava-v1-hf
|
|
@@ -992,42 +984,134 @@ model_deployments:
|
|
|
992
984
|
args:
|
|
993
985
|
pretrained_model_name_or_path: openai-community/gpt2
|
|
994
986
|
|
|
987
|
+
## OpenThaiGPT
|
|
988
|
+
- name: huggingface/openthaigpt-1.0.0-7b-chat
|
|
989
|
+
model_name: openthaigpt/openthaigpt-1.0.0-7b-chat
|
|
990
|
+
tokenizer_name: openthaigpt/openthaigpt-1.0.0-7b-chat
|
|
991
|
+
max_sequence_length: 4096
|
|
992
|
+
client_spec:
|
|
993
|
+
class_name: "helm.clients.huggingface_client.HuggingFaceClient"
|
|
994
|
+
|
|
995
|
+
- name: huggingface/openthaigpt-1.0.0-13b-chat
|
|
996
|
+
model_name: openthaigpt/openthaigpt-1.0.0-13b-chat
|
|
997
|
+
tokenizer_name: openthaigpt/openthaigpt-1.0.0-7b-chat
|
|
998
|
+
max_sequence_length: 4096
|
|
999
|
+
client_spec:
|
|
1000
|
+
class_name: "helm.clients.huggingface_client.HuggingFaceClient"
|
|
1001
|
+
args:
|
|
1002
|
+
device_map: auto
|
|
1003
|
+
|
|
1004
|
+
- name: huggingface/openthaigpt-1.0.0-70b-chat
|
|
1005
|
+
model_name: openthaigpt/openthaigpt-1.0.0-70b-chat
|
|
1006
|
+
tokenizer_name: huggingface/openthaigpt-1.0.0-7b-chat
|
|
1007
|
+
max_sequence_length: 4096
|
|
1008
|
+
client_spec:
|
|
1009
|
+
class_name: "helm.clients.huggingface_client.HuggingFaceClient"
|
|
1010
|
+
args:
|
|
1011
|
+
device_map: auto
|
|
1012
|
+
|
|
995
1013
|
## SAIL (SEA AI Lab)
|
|
996
|
-
- name:
|
|
1014
|
+
- name: huggingface/sailor-7b
|
|
997
1015
|
model_name: sail/sailor-7b
|
|
998
1016
|
tokenizer_name: qwen/qwen1.5-7b
|
|
999
1017
|
max_sequence_length: 32768
|
|
1000
1018
|
client_spec:
|
|
1001
1019
|
class_name: "helm.clients.huggingface_client.HuggingFaceClient"
|
|
1002
1020
|
|
|
1003
|
-
- name:
|
|
1021
|
+
- name: huggingface/sailor-7b-chat
|
|
1004
1022
|
model_name: sail/sailor-7b-chat
|
|
1005
1023
|
tokenizer_name: qwen/qwen1.5-7b
|
|
1006
1024
|
max_sequence_length: 32768
|
|
1007
1025
|
client_spec:
|
|
1008
1026
|
class_name: "helm.clients.huggingface_client.HuggingFaceClient"
|
|
1009
1027
|
|
|
1010
|
-
- name:
|
|
1028
|
+
- name: huggingface/sailor-14b
|
|
1011
1029
|
model_name: sail/sailor-14b
|
|
1012
1030
|
tokenizer_name: qwen/qwen1.5-7b
|
|
1013
1031
|
max_sequence_length: 32768
|
|
1014
1032
|
client_spec:
|
|
1015
1033
|
class_name: "helm.clients.huggingface_client.HuggingFaceClient"
|
|
1034
|
+
args:
|
|
1035
|
+
device_map: auto
|
|
1016
1036
|
|
|
1017
|
-
- name:
|
|
1037
|
+
- name: huggingface/sailor-14b-chat
|
|
1018
1038
|
model_name: sail/sailor-14b-chat
|
|
1019
1039
|
tokenizer_name: qwen/qwen1.5-7b
|
|
1020
1040
|
max_sequence_length: 32768
|
|
1021
1041
|
client_spec:
|
|
1022
1042
|
class_name: "helm.clients.huggingface_client.HuggingFaceClient"
|
|
1043
|
+
args:
|
|
1044
|
+
device_map: auto
|
|
1045
|
+
|
|
1046
|
+
# SambaNova
|
|
1047
|
+
- name: huggingface/sambalingo-thai-base
|
|
1048
|
+
model_name: sambanova/sambalingo-thai-base
|
|
1049
|
+
tokenizer_name: sambanova/sambalingo-thai-base
|
|
1050
|
+
max_sequence_length: 4096
|
|
1051
|
+
client_spec:
|
|
1052
|
+
class_name: "helm.clients.huggingface_client.HuggingFaceClient"
|
|
1053
|
+
args:
|
|
1054
|
+
pretrained_model_name_or_path: sambanovasystems/SambaLingo-Thai-Base
|
|
1055
|
+
|
|
1056
|
+
- name: huggingface/sambalingo-thai-chat
|
|
1057
|
+
model_name: sambanova/sambalingo-thai-chat
|
|
1058
|
+
tokenizer_name: sambanova/sambalingo-thai-base
|
|
1059
|
+
max_sequence_length: 4096
|
|
1060
|
+
client_spec:
|
|
1061
|
+
class_name: "helm.clients.huggingface_client.HuggingFaceClient"
|
|
1062
|
+
args:
|
|
1063
|
+
pretrained_model_name_or_path: sambanovasystems/SambaLingo-Thai-Base
|
|
1064
|
+
|
|
1065
|
+
- name: huggingface/sambalingo-thai-base-70b
|
|
1066
|
+
model_name: sambanova/sambalingo-thai-base-70b
|
|
1067
|
+
tokenizer_name: sambanova/sambalingo-thai-base
|
|
1068
|
+
max_sequence_length: 4096
|
|
1069
|
+
client_spec:
|
|
1070
|
+
class_name: "helm.clients.huggingface_client.HuggingFaceClient"
|
|
1071
|
+
args:
|
|
1072
|
+
pretrained_model_name_or_path: sambanovasystems/SambaLingo-Thai-Base-70B
|
|
1073
|
+
device_map: auto
|
|
1074
|
+
|
|
1075
|
+
- name: huggingface/sambalingo-thai-chat-70b
|
|
1076
|
+
model_name: sambanova/sambalingo-thai-chat-70b
|
|
1077
|
+
tokenizer_name: sambanova/sambalingo-thai-base
|
|
1078
|
+
max_sequence_length: 4096
|
|
1079
|
+
client_spec:
|
|
1080
|
+
class_name: "helm.clients.huggingface_client.HuggingFaceClient"
|
|
1081
|
+
args:
|
|
1082
|
+
pretrained_model_name_or_path: sambanovasystems/SambaLingo-Thai-Base-70B
|
|
1083
|
+
device_map: auto
|
|
1023
1084
|
|
|
1024
1085
|
## SCB10X
|
|
1086
|
+
- name: huggingface/typhoon-7b
|
|
1087
|
+
model_name: scb10x/typhoon-7b
|
|
1088
|
+
tokenizer_name: scb10x/typhoon-7b
|
|
1089
|
+
max_sequence_length: 4096
|
|
1090
|
+
client_spec:
|
|
1091
|
+
class_name: "helm.clients.huggingface_client.HuggingFaceClient"
|
|
1092
|
+
|
|
1093
|
+
- name: huggingface/typhoon-v1.5-8b
|
|
1094
|
+
model_name: scb10x/typhoon-v1.5-8b
|
|
1095
|
+
tokenizer_name: meta/llama-3-8b
|
|
1096
|
+
max_sequence_length: 8192
|
|
1097
|
+
client_spec:
|
|
1098
|
+
class_name: "helm.clients.huggingface_client.HuggingFaceClient"
|
|
1099
|
+
|
|
1100
|
+
- name: huggingface/typhoon-v1.5-8b-instruct
|
|
1101
|
+
model_name: scb10x/typhoon-v1.5-8b-instruct
|
|
1102
|
+
tokenizer_name: meta/llama-3-8b
|
|
1103
|
+
max_sequence_length: 8192
|
|
1104
|
+
client_spec:
|
|
1105
|
+
class_name: "helm.clients.huggingface_client.HuggingFaceClient"
|
|
1106
|
+
|
|
1025
1107
|
- name: huggingface/typhoon-v1.5-72b
|
|
1026
1108
|
model_name: scb10x/typhoon-v1.5-72b
|
|
1027
1109
|
tokenizer_name: qwen/qwen1.5-7b
|
|
1028
1110
|
max_sequence_length: 32768
|
|
1029
1111
|
client_spec:
|
|
1030
1112
|
class_name: "helm.clients.huggingface_client.HuggingFaceClient"
|
|
1113
|
+
args:
|
|
1114
|
+
device_map: auto
|
|
1031
1115
|
|
|
1032
1116
|
- name: huggingface/typhoon-v1.5-72b-instruct
|
|
1033
1117
|
model_name: scb10x/typhoon-v1.5-72b-instruct
|
|
@@ -1035,6 +1119,43 @@ model_deployments:
|
|
|
1035
1119
|
max_sequence_length: 32768
|
|
1036
1120
|
client_spec:
|
|
1037
1121
|
class_name: "helm.clients.huggingface_client.HuggingFaceClient"
|
|
1122
|
+
args:
|
|
1123
|
+
device_map: auto
|
|
1124
|
+
|
|
1125
|
+
- name: huggingface/llama-3-typhoon-v1.5x-8b-instruct
|
|
1126
|
+
model_name: scb10x/llama-3-typhoon-v1.5x-8b-instruct
|
|
1127
|
+
tokenizer_name: meta/llama-3-8b
|
|
1128
|
+
max_sequence_length: 8192
|
|
1129
|
+
client_spec:
|
|
1130
|
+
class_name: "helm.clients.huggingface_client.HuggingFaceClient"
|
|
1131
|
+
|
|
1132
|
+
- name: huggingface/llama-3-typhoon-v1.5x-70b-instruct
|
|
1133
|
+
model_name: scb10x/llama-3-typhoon-v1.5x-70b-instruct
|
|
1134
|
+
tokenizer_name: meta/llama-3-8b
|
|
1135
|
+
max_sequence_length: 8192
|
|
1136
|
+
client_spec:
|
|
1137
|
+
class_name: "helm.clients.huggingface_client.HuggingFaceClient"
|
|
1138
|
+
args:
|
|
1139
|
+
device_map: auto
|
|
1140
|
+
|
|
1141
|
+
# Alibaba DAMO Academy
|
|
1142
|
+
- name: huggingface/seallm-7b-v2
|
|
1143
|
+
model_name: damo/seallm-7b-v2
|
|
1144
|
+
tokenizer_name: damo/seallm-7b-v2
|
|
1145
|
+
max_sequence_length: 4096
|
|
1146
|
+
client_spec:
|
|
1147
|
+
class_name: "helm.clients.huggingface_client.HuggingFaceClient"
|
|
1148
|
+
args:
|
|
1149
|
+
pretrained_model_name_or_path: SeaLLMs/SeaLLM-7B-v2
|
|
1150
|
+
|
|
1151
|
+
- name: huggingface/seallm-7b-v2.5
|
|
1152
|
+
model_name: damo/seallm-7b-v2.5
|
|
1153
|
+
tokenizer_name: damo/seallm-7b-v2.5
|
|
1154
|
+
max_sequence_length: 4096
|
|
1155
|
+
client_spec:
|
|
1156
|
+
class_name: "helm.clients.huggingface_client.HuggingFaceClient"
|
|
1157
|
+
args:
|
|
1158
|
+
pretrained_model_name_or_path: SeaLLMs/SeaLLM-7B-v2.5
|
|
1038
1159
|
|
|
1039
1160
|
## StabilityAI
|
|
1040
1161
|
- name: huggingface/stablelm-base-alpha-3b
|
|
@@ -1315,6 +1436,20 @@ model_deployments:
|
|
|
1315
1436
|
client_spec:
|
|
1316
1437
|
class_name: "helm.clients.mistral_client.MistralAIClient"
|
|
1317
1438
|
|
|
1439
|
+
- name: mistralai/mistral-large-2407
|
|
1440
|
+
model_name: mistralai/mistral-large-2407
|
|
1441
|
+
tokenizer_name: mistralai/Mistral-Large-Instruct-2407
|
|
1442
|
+
max_sequence_length: 128000
|
|
1443
|
+
client_spec:
|
|
1444
|
+
class_name: "helm.clients.mistral_client.MistralAIClient"
|
|
1445
|
+
|
|
1446
|
+
- name: mistralai/open-mistral-nemo-2407
|
|
1447
|
+
model_name: mistralai/open-mistral-nemo-2407
|
|
1448
|
+
tokenizer_name: mistralai/Mistral-Nemo-Base-2407
|
|
1449
|
+
max_sequence_length: 128000
|
|
1450
|
+
client_spec:
|
|
1451
|
+
class_name: "helm.clients.mistral_client.MistralAIClient"
|
|
1452
|
+
|
|
1318
1453
|
# Neurips
|
|
1319
1454
|
- name: neurips/local
|
|
1320
1455
|
model_name: neurips/local
|
|
@@ -1331,6 +1466,13 @@ model_deployments:
|
|
|
1331
1466
|
client_spec:
|
|
1332
1467
|
class_name: "helm.clients.megatron_client.MegatronClient"
|
|
1333
1468
|
|
|
1469
|
+
- name: nvidia/nemotron-4-340b-instruct
|
|
1470
|
+
model_name: nvidia/nemotron-4-340b-instruct
|
|
1471
|
+
tokenizer_name: nvidia/nemotron-4-340b-instruct
|
|
1472
|
+
max_sequence_length: 4085
|
|
1473
|
+
client_spec:
|
|
1474
|
+
class_name: "helm.clients.nvidia_nim_client.NvidiaNimClient"
|
|
1475
|
+
|
|
1334
1476
|
# OpenAI
|
|
1335
1477
|
|
|
1336
1478
|
## GPT 3 Models
|
|
@@ -1353,99 +1495,6 @@ model_deployments:
|
|
|
1353
1495
|
client_spec:
|
|
1354
1496
|
class_name: "helm.clients.openai_client.OpenAIClient"
|
|
1355
1497
|
|
|
1356
|
-
# The list of models can be found here: https://beta.openai.com/docs/engines/gpt-3
|
|
1357
|
-
# DEPRECATED: Announced on July 06 2023 that these models will be shut down on January 04 2024.
|
|
1358
|
-
|
|
1359
|
-
- name: openai/davinci
|
|
1360
|
-
deprecated: true
|
|
1361
|
-
model_name: openai/davinci
|
|
1362
|
-
tokenizer_name: huggingface/gpt2
|
|
1363
|
-
max_sequence_length: 2048
|
|
1364
|
-
max_request_length: 2049
|
|
1365
|
-
client_spec:
|
|
1366
|
-
class_name: "helm.clients.openai_client.OpenAIClient"
|
|
1367
|
-
|
|
1368
|
-
- name: openai/curie
|
|
1369
|
-
deprecated: true
|
|
1370
|
-
model_name: openai/curie
|
|
1371
|
-
tokenizer_name: huggingface/gpt2
|
|
1372
|
-
max_sequence_length: 2048
|
|
1373
|
-
max_request_length: 2049
|
|
1374
|
-
client_spec:
|
|
1375
|
-
class_name: "helm.clients.openai_client.OpenAIClient"
|
|
1376
|
-
|
|
1377
|
-
- name: openai/babbage
|
|
1378
|
-
deprecated: true
|
|
1379
|
-
model_name: openai/babbage
|
|
1380
|
-
tokenizer_name: huggingface/gpt2
|
|
1381
|
-
max_sequence_length: 2048
|
|
1382
|
-
max_request_length: 2049
|
|
1383
|
-
client_spec:
|
|
1384
|
-
class_name: "helm.clients.openai_client.OpenAIClient"
|
|
1385
|
-
|
|
1386
|
-
- name: openai/ada
|
|
1387
|
-
deprecated: true
|
|
1388
|
-
model_name: openai/ada
|
|
1389
|
-
tokenizer_name: huggingface/gpt2
|
|
1390
|
-
max_sequence_length: 2048
|
|
1391
|
-
max_request_length: 2049
|
|
1392
|
-
client_spec:
|
|
1393
|
-
class_name: "helm.clients.openai_client.OpenAIClient"
|
|
1394
|
-
|
|
1395
|
-
- name: openai/text-davinci-003
|
|
1396
|
-
deprecated: true
|
|
1397
|
-
model_name: openai/text-davinci-003
|
|
1398
|
-
tokenizer_name: huggingface/gpt2
|
|
1399
|
-
max_sequence_length: 4000
|
|
1400
|
-
max_request_length: 4001
|
|
1401
|
-
client_spec:
|
|
1402
|
-
class_name: "helm.clients.openai_client.OpenAIClient"
|
|
1403
|
-
|
|
1404
|
-
- name: openai/text-davinci-002
|
|
1405
|
-
deprecated: true
|
|
1406
|
-
model_name: openai/text-davinci-002
|
|
1407
|
-
tokenizer_name: huggingface/gpt2
|
|
1408
|
-
max_sequence_length: 4000
|
|
1409
|
-
max_request_length: 4001
|
|
1410
|
-
client_spec:
|
|
1411
|
-
class_name: "helm.clients.openai_client.OpenAIClient"
|
|
1412
|
-
|
|
1413
|
-
- name: openai/text-davinci-001
|
|
1414
|
-
deprecated: true
|
|
1415
|
-
model_name: openai/text-davinci-001
|
|
1416
|
-
tokenizer_name: huggingface/gpt2
|
|
1417
|
-
max_sequence_length: 2048
|
|
1418
|
-
max_request_length: 2049
|
|
1419
|
-
client_spec:
|
|
1420
|
-
class_name: "helm.clients.openai_client.OpenAIClient"
|
|
1421
|
-
|
|
1422
|
-
- name: openai/text-curie-001
|
|
1423
|
-
deprecated: true
|
|
1424
|
-
model_name: openai/text-curie-001
|
|
1425
|
-
tokenizer_name: huggingface/gpt2
|
|
1426
|
-
max_sequence_length: 2048
|
|
1427
|
-
max_request_length: 2049
|
|
1428
|
-
client_spec:
|
|
1429
|
-
class_name: "helm.clients.openai_client.OpenAIClient"
|
|
1430
|
-
|
|
1431
|
-
- name: openai/text-babbage-001
|
|
1432
|
-
deprecated: true
|
|
1433
|
-
model_name: openai/text-babbage-001
|
|
1434
|
-
tokenizer_name: huggingface/gpt2
|
|
1435
|
-
max_sequence_length: 2048
|
|
1436
|
-
max_request_length: 2049
|
|
1437
|
-
client_spec:
|
|
1438
|
-
class_name: "helm.clients.openai_client.OpenAIClient"
|
|
1439
|
-
|
|
1440
|
-
- name: openai/text-ada-001
|
|
1441
|
-
deprecated: true
|
|
1442
|
-
model_name: openai/text-ada-001
|
|
1443
|
-
tokenizer_name: huggingface/gpt2
|
|
1444
|
-
max_sequence_length: 2048
|
|
1445
|
-
max_request_length: 2049
|
|
1446
|
-
client_spec:
|
|
1447
|
-
class_name: "helm.clients.openai_client.OpenAIClient"
|
|
1448
|
-
|
|
1449
1498
|
## GPT 3.5 Turbo Models
|
|
1450
1499
|
# ChatGPT: https://openai.com/blog/chatgpt
|
|
1451
1500
|
|
|
@@ -1580,6 +1629,20 @@ model_deployments:
|
|
|
1580
1629
|
client_spec:
|
|
1581
1630
|
class_name: "helm.clients.openai_client.OpenAIClient"
|
|
1582
1631
|
|
|
1632
|
+
- name: openai/gpt-4o-2024-08-06
|
|
1633
|
+
model_name: openai/gpt-4o-2024-08-06
|
|
1634
|
+
tokenizer_name: openai/o200k_base
|
|
1635
|
+
max_sequence_length: 128000
|
|
1636
|
+
client_spec:
|
|
1637
|
+
class_name: "helm.clients.openai_client.OpenAIClient"
|
|
1638
|
+
|
|
1639
|
+
- name: openai/gpt-4o-mini-2024-07-18
|
|
1640
|
+
model_name: openai/gpt-4o-mini-2024-07-18
|
|
1641
|
+
tokenizer_name: openai/o200k_base
|
|
1642
|
+
max_sequence_length: 128000
|
|
1643
|
+
client_spec:
|
|
1644
|
+
class_name: "helm.clients.openai_client.OpenAIClient"
|
|
1645
|
+
|
|
1583
1646
|
- name: openai/gpt-4-vision-preview
|
|
1584
1647
|
model_name: openai/gpt-4-vision-preview
|
|
1585
1648
|
tokenizer_name: openai/cl100k_base
|
|
@@ -1598,33 +1661,18 @@ model_deployments:
|
|
|
1598
1661
|
client_spec:
|
|
1599
1662
|
class_name: "helm.clients.openai_client.OpenAIClient"
|
|
1600
1663
|
|
|
1601
|
-
##
|
|
1602
|
-
|
|
1603
|
-
|
|
1604
|
-
|
|
1605
|
-
|
|
1606
|
-
model_name: openai/code-davinci-002
|
|
1607
|
-
tokenizer_name: huggingface/gpt2
|
|
1608
|
-
max_sequence_length: 4000
|
|
1609
|
-
max_request_length: 4001
|
|
1610
|
-
client_spec:
|
|
1611
|
-
class_name: "helm.clients.openai_client.OpenAIClient"
|
|
1612
|
-
|
|
1613
|
-
- name: openai/code-davinci-001
|
|
1614
|
-
deprecated: true
|
|
1615
|
-
model_name: openai/code-davinci-001
|
|
1616
|
-
tokenizer_name: huggingface/gpt2
|
|
1617
|
-
max_sequence_length: 2048
|
|
1618
|
-
max_request_length: 2049
|
|
1664
|
+
## o1 Models
|
|
1665
|
+
- name: openai/o1-preview-2024-09-12
|
|
1666
|
+
model_name: openai/o1-preview-2024-09-12
|
|
1667
|
+
tokenizer_name: openai/cl100k_base
|
|
1668
|
+
max_sequence_length: 128000
|
|
1619
1669
|
client_spec:
|
|
1620
1670
|
class_name: "helm.clients.openai_client.OpenAIClient"
|
|
1621
1671
|
|
|
1622
|
-
- name: openai/
|
|
1623
|
-
|
|
1624
|
-
|
|
1625
|
-
|
|
1626
|
-
max_sequence_length: 2048
|
|
1627
|
-
max_request_length: 2049
|
|
1672
|
+
- name: openai/o1-mini-2024-09-12
|
|
1673
|
+
model_name: openai/o1-mini-2024-09-12
|
|
1674
|
+
tokenizer_name: openai/cl100k_base
|
|
1675
|
+
max_sequence_length: 128000
|
|
1628
1676
|
client_spec:
|
|
1629
1677
|
class_name: "helm.clients.openai_client.OpenAIClient"
|
|
1630
1678
|
|
|
@@ -1632,44 +1680,6 @@ model_deployments:
|
|
|
1632
1680
|
# OpenAI similarity embedding models: https://beta.openai.com/docs/guides/embeddings
|
|
1633
1681
|
# The number of parameters is guessed based on the number of parameters of the
|
|
1634
1682
|
# corresponding GPT-3 model.
|
|
1635
|
-
# DEPRECATED: Announced on July 06 2023 that first generation embeddings models
|
|
1636
|
-
# will be shut down on January 04 2024.
|
|
1637
|
-
|
|
1638
|
-
- name: openai/text-similarity-davinci-001
|
|
1639
|
-
deprecated: true
|
|
1640
|
-
model_name: openai/text-similarity-davinci-001
|
|
1641
|
-
tokenizer_name: huggingface/gpt2
|
|
1642
|
-
max_sequence_length: 2048
|
|
1643
|
-
max_request_length: 2049
|
|
1644
|
-
client_spec:
|
|
1645
|
-
class_name: "helm.clients.openai_client.OpenAIClient"
|
|
1646
|
-
|
|
1647
|
-
- name: openai/text-similarity-curie-001
|
|
1648
|
-
deprecated: true
|
|
1649
|
-
model_name: openai/text-similarity-curie-001
|
|
1650
|
-
tokenizer_name: huggingface/gpt2
|
|
1651
|
-
max_sequence_length: 2048
|
|
1652
|
-
max_request_length: 2049
|
|
1653
|
-
client_spec:
|
|
1654
|
-
class_name: "helm.clients.openai_client.OpenAIClient"
|
|
1655
|
-
|
|
1656
|
-
- name: openai/text-similarity-babbage-001
|
|
1657
|
-
deprecated: true
|
|
1658
|
-
model_name: openai/text-similarity-babbage-001
|
|
1659
|
-
tokenizer_name: huggingface/gpt2
|
|
1660
|
-
max_sequence_length: 2048
|
|
1661
|
-
max_request_length: 2049
|
|
1662
|
-
client_spec:
|
|
1663
|
-
class_name: "helm.clients.openai_client.OpenAIClient"
|
|
1664
|
-
|
|
1665
|
-
- name: openai/text-similarity-ada-001
|
|
1666
|
-
deprecated: true
|
|
1667
|
-
model_name: openai/text-similarity-ada-001
|
|
1668
|
-
tokenizer_name: huggingface/gpt2
|
|
1669
|
-
max_sequence_length: 2048
|
|
1670
|
-
max_request_length: 2049
|
|
1671
|
-
client_spec:
|
|
1672
|
-
class_name: "helm.clients.openai_client.OpenAIClient"
|
|
1673
1683
|
|
|
1674
1684
|
# As of 2023-11-07, text-embedding-ada-002 is not deprecated:
|
|
1675
1685
|
# "We recommend using text-embedding-ada-002 for nearly all use cases."
|
|
@@ -1856,7 +1866,25 @@ model_deployments:
|
|
|
1856
1866
|
client_spec:
|
|
1857
1867
|
class_name: "helm.clients.together_client.TogetherClient"
|
|
1858
1868
|
args:
|
|
1859
|
-
together_model: meta-llama/
|
|
1869
|
+
together_model: meta-llama/Llama-3-8b-hf
|
|
1870
|
+
|
|
1871
|
+
- name: together/llama-3-8b-instruct-turbo
|
|
1872
|
+
model_name: meta/llama-3-8b-instruct-turbo
|
|
1873
|
+
tokenizer_name: meta/llama-3-8b
|
|
1874
|
+
max_sequence_length: 8191
|
|
1875
|
+
client_spec:
|
|
1876
|
+
class_name: "helm.clients.together_client.TogetherClient"
|
|
1877
|
+
args:
|
|
1878
|
+
together_model: meta-llama/Meta-Llama-3-8B-Instruct-Turbo
|
|
1879
|
+
|
|
1880
|
+
- name: together/llama-3-8b-instruct-lite
|
|
1881
|
+
model_name: meta/llama-3-8b-instruct-lite
|
|
1882
|
+
tokenizer_name: meta/llama-3-8b
|
|
1883
|
+
max_sequence_length: 8191
|
|
1884
|
+
client_spec:
|
|
1885
|
+
class_name: "helm.clients.together_client.TogetherClient"
|
|
1886
|
+
args:
|
|
1887
|
+
together_model: meta-llama/Meta-Llama-3-8B-Instruct-Lite
|
|
1860
1888
|
|
|
1861
1889
|
- name: together/llama-3-70b
|
|
1862
1890
|
model_name: meta/llama-3-70b
|
|
@@ -1867,10 +1895,55 @@ model_deployments:
|
|
|
1867
1895
|
args:
|
|
1868
1896
|
together_model: meta-llama/Meta-Llama-3-70B
|
|
1869
1897
|
|
|
1870
|
-
- name: together/llama-3-
|
|
1871
|
-
model_name: meta/llama-3-
|
|
1898
|
+
- name: together/llama-3-70b-instruct-turbo
|
|
1899
|
+
model_name: meta/llama-3-70b-instruct-turbo
|
|
1872
1900
|
tokenizer_name: meta/llama-3-8b
|
|
1873
1901
|
max_sequence_length: 8191
|
|
1902
|
+
client_spec:
|
|
1903
|
+
class_name: "helm.clients.together_client.TogetherClient"
|
|
1904
|
+
args:
|
|
1905
|
+
together_model: meta-llama/Meta-Llama-3-70B-Instruct-Turbo
|
|
1906
|
+
|
|
1907
|
+
- name: together/llama-3-70b-instruct-lite
|
|
1908
|
+
model_name: meta/llama-3-70b-instruct-lite
|
|
1909
|
+
tokenizer_name: meta/llama-3-8b
|
|
1910
|
+
max_sequence_length: 8191
|
|
1911
|
+
client_spec:
|
|
1912
|
+
class_name: "helm.clients.together_client.TogetherClient"
|
|
1913
|
+
args:
|
|
1914
|
+
together_model: meta-llama/Meta-Llama-3-70B-Instruct-Lite
|
|
1915
|
+
|
|
1916
|
+
- name: together/llama-3.1-8b-instruct-turbo
|
|
1917
|
+
model_name: meta/llama-3.1-8b-instruct-turbo
|
|
1918
|
+
tokenizer_name: meta/llama-3.1-8b
|
|
1919
|
+
max_sequence_length: 128000
|
|
1920
|
+
client_spec:
|
|
1921
|
+
class_name: "helm.clients.together_client.TogetherChatClient"
|
|
1922
|
+
args:
|
|
1923
|
+
together_model: meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo
|
|
1924
|
+
|
|
1925
|
+
- name: together/llama-3.1-70b-instruct-turbo
|
|
1926
|
+
model_name: meta/llama-3.1-70b-instruct-turbo
|
|
1927
|
+
tokenizer_name: meta/llama-3.1-8b
|
|
1928
|
+
max_sequence_length: 128000
|
|
1929
|
+
client_spec:
|
|
1930
|
+
class_name: "helm.clients.together_client.TogetherChatClient"
|
|
1931
|
+
args:
|
|
1932
|
+
together_model: meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo
|
|
1933
|
+
|
|
1934
|
+
- name: together/llama-3.1-405b-instruct-turbo
|
|
1935
|
+
model_name: meta/llama-3.1-405b-instruct-turbo
|
|
1936
|
+
tokenizer_name: meta/llama-3.1-8b
|
|
1937
|
+
max_sequence_length: 128000
|
|
1938
|
+
client_spec:
|
|
1939
|
+
class_name: "helm.clients.together_client.TogetherChatClient"
|
|
1940
|
+
args:
|
|
1941
|
+
together_model: meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo
|
|
1942
|
+
|
|
1943
|
+
- name: together/llama-3-8b-chat
|
|
1944
|
+
model_name: meta/llama-3-8b-chat
|
|
1945
|
+
tokenizer_name: meta/llama-3-8b-instruct
|
|
1946
|
+
max_sequence_length: 8182
|
|
1874
1947
|
client_spec:
|
|
1875
1948
|
class_name: "helm.clients.together_client.TogetherChatClient"
|
|
1876
1949
|
args:
|
|
@@ -1878,13 +1951,40 @@ model_deployments:
|
|
|
1878
1951
|
|
|
1879
1952
|
- name: together/llama-3-70b-chat
|
|
1880
1953
|
model_name: meta/llama-3-70b-chat
|
|
1881
|
-
tokenizer_name: meta/llama-3-8b
|
|
1882
|
-
max_sequence_length:
|
|
1954
|
+
tokenizer_name: meta/llama-3-8b-instruct
|
|
1955
|
+
max_sequence_length: 8182
|
|
1883
1956
|
client_spec:
|
|
1884
1957
|
class_name: "helm.clients.together_client.TogetherChatClient"
|
|
1885
1958
|
args:
|
|
1886
1959
|
together_model: meta-llama/Llama-3-70b-chat-hf
|
|
1887
1960
|
|
|
1961
|
+
- name: together/llama-3.2-3b-instruct-turbo
|
|
1962
|
+
model_name: meta/llama-3.2-3b-instruct-turbo
|
|
1963
|
+
tokenizer_name: meta/llama-3.2-3b-instruct
|
|
1964
|
+
max_sequence_length: 128000
|
|
1965
|
+
client_spec:
|
|
1966
|
+
class_name: "helm.clients.together_client.TogetherChatClient"
|
|
1967
|
+
args:
|
|
1968
|
+
together_model: meta-llama/Llama-3.2-3B-Instruct-Turbo
|
|
1969
|
+
|
|
1970
|
+
- name: together/llama-3.2-11b-vision-instruct-turbo
|
|
1971
|
+
model_name: meta/llama-3.2-11b-vision-instruct-turbo
|
|
1972
|
+
tokenizer_name: meta/llama-3.2-11b-vision-instruct
|
|
1973
|
+
max_sequence_length: 128000
|
|
1974
|
+
client_spec:
|
|
1975
|
+
class_name: "helm.clients.together_client.TogetherChatClient"
|
|
1976
|
+
args:
|
|
1977
|
+
together_model: meta-llama/Llama-3.2-11B-Vision-Instruct-Turbo
|
|
1978
|
+
|
|
1979
|
+
- name: together/llama-3.2-90b-vision-instruct-turbo
|
|
1980
|
+
model_name: meta/llama-3.2-90b-vision-instruct-turbo
|
|
1981
|
+
tokenizer_name: meta/llama-3.2-11b-vision-instruct
|
|
1982
|
+
max_sequence_length: 128000
|
|
1983
|
+
client_spec:
|
|
1984
|
+
class_name: "helm.clients.together_client.TogetherChatClient"
|
|
1985
|
+
args:
|
|
1986
|
+
together_model: meta-llama/Llama-3.2-90B-Vision-Instruct-Turbo
|
|
1987
|
+
|
|
1888
1988
|
- name: together/llama-guard-7b
|
|
1889
1989
|
model_name: meta/llama-guard-7b
|
|
1890
1990
|
tokenizer_name: meta-llama/Llama-2-7b-hf
|
|
@@ -1903,6 +2003,15 @@ model_deployments:
|
|
|
1903
2003
|
args:
|
|
1904
2004
|
together_model: meta-llama/llamaguard-2-8b
|
|
1905
2005
|
|
|
2006
|
+
- name: together/llama-guard-3-8b
|
|
2007
|
+
model_name: meta/llama-guard-3-8b
|
|
2008
|
+
tokenizer_name: meta/llama-3.1-8b
|
|
2009
|
+
max_sequence_length: 128000
|
|
2010
|
+
client_spec:
|
|
2011
|
+
class_name: "helm.clients.together_client.TogetherClient"
|
|
2012
|
+
args:
|
|
2013
|
+
together_model: meta-llama/Meta-Llama-Guard-3-8B
|
|
2014
|
+
|
|
1906
2015
|
# 01.AI
|
|
1907
2016
|
- name: together/yi-6b
|
|
1908
2017
|
model_name: 01-ai/yi-6b
|
|
@@ -2029,7 +2138,7 @@ model_deployments:
|
|
|
2029
2138
|
- name: together/mixtral-8x7b-instruct-v0.1
|
|
2030
2139
|
model_name: mistralai/mixtral-8x7b-instruct-v0.1
|
|
2031
2140
|
tokenizer_name: mistralai/Mistral-7B-v0.1
|
|
2032
|
-
max_sequence_length:
|
|
2141
|
+
max_sequence_length: 32767
|
|
2033
2142
|
client_spec:
|
|
2034
2143
|
class_name: "helm.clients.together_client.TogetherChatClient"
|
|
2035
2144
|
|
|
@@ -2165,18 +2274,6 @@ model_deployments:
|
|
|
2165
2274
|
args:
|
|
2166
2275
|
together_model: togethercomputer/RedPajama-INCITE-7B-Instruct
|
|
2167
2276
|
|
|
2168
|
-
## Tsinghua
|
|
2169
|
-
- name: together/glm
|
|
2170
|
-
deprecated: true # Removed from Together
|
|
2171
|
-
model_name: tsinghua/glm
|
|
2172
|
-
tokenizer_name: TsinghuaKEG/ice
|
|
2173
|
-
max_sequence_length: 2048
|
|
2174
|
-
max_request_length: 2049
|
|
2175
|
-
client_spec:
|
|
2176
|
-
class_name: "helm.clients.together_client.TogetherClient"
|
|
2177
|
-
window_service_spec:
|
|
2178
|
-
class_name: "helm.benchmark.window_services.ice_window_service.ICEWindowService"
|
|
2179
|
-
|
|
2180
2277
|
- name: thudm/cogview2
|
|
2181
2278
|
model_name: thudm/cogview2
|
|
2182
2279
|
tokenizer_name: openai/clip-vit-large-patch14
|
|
@@ -2215,22 +2312,6 @@ model_deployments:
|
|
|
2215
2312
|
client_spec:
|
|
2216
2313
|
class_name: "helm.clients.palmyra_client.PalmyraClient"
|
|
2217
2314
|
|
|
2218
|
-
- name: writer/palmyra-instruct-30
|
|
2219
|
-
model_name: writer/palmyra-instruct-30
|
|
2220
|
-
tokenizer_name: writer/gpt2
|
|
2221
|
-
max_sequence_length: 2048
|
|
2222
|
-
max_sequence_and_generated_tokens_length: 2048
|
|
2223
|
-
client_spec:
|
|
2224
|
-
class_name: "helm.clients.palmyra_client.PalmyraClient"
|
|
2225
|
-
|
|
2226
|
-
- name: writer/palmyra-e
|
|
2227
|
-
model_name: writer/palmyra-e
|
|
2228
|
-
tokenizer_name: writer/gpt2
|
|
2229
|
-
max_sequence_length: 2048
|
|
2230
|
-
max_sequence_and_generated_tokens_length: 2048
|
|
2231
|
-
client_spec:
|
|
2232
|
-
class_name: "helm.clients.palmyra_client.PalmyraClient"
|
|
2233
|
-
|
|
2234
2315
|
- name: writer/silk-road
|
|
2235
2316
|
model_name: writer/silk-road
|
|
2236
2317
|
tokenizer_name: writer/gpt2
|
|
@@ -2279,6 +2360,40 @@ model_deployments:
|
|
|
2279
2360
|
client_spec:
|
|
2280
2361
|
class_name: "helm.clients.vision_language.palmyra_vision_client.PalmyraVisionClient"
|
|
2281
2362
|
|
|
2363
|
+
- name: writer/palmyra-x-004
|
|
2364
|
+
model_name: writer/palmyra-x-004
|
|
2365
|
+
# Actual tokenizer is Llama 2, but it cannot be used in HELM due to this issue:
|
|
2366
|
+
# https://github.com/stanford-crfm/helm/issues/2467
|
|
2367
|
+
# Work around by using Llama 3 tokenizer for now.
|
|
2368
|
+
tokenizer_name: meta/llama-3-8b
|
|
2369
|
+
max_sequence_length: 8192
|
|
2370
|
+
client_spec:
|
|
2371
|
+
class_name: "helm.clients.palmyra_client.PalmyraChatClient"
|
|
2372
|
+
|
|
2373
|
+
- name: writer/palmyra-med-32k
|
|
2374
|
+
model_name: writer/palmyra-med-32k
|
|
2375
|
+
# Palmyra-Med uses the "<|end_of_text|>" as the end of text token, which is used by meta/llama-3-8b,
|
|
2376
|
+
# rather than "<|eot_id|>", which is used by meta/llama-3-8b-instruct
|
|
2377
|
+
tokenizer_name: meta/llama-3-8b
|
|
2378
|
+
max_sequence_length: 32000
|
|
2379
|
+
client_spec:
|
|
2380
|
+
class_name: "helm.clients.palmyra_client.PalmyraChatClient"
|
|
2381
|
+
|
|
2382
|
+
- name: writer/palmyra-med
|
|
2383
|
+
model_name: writer/palmyra-med
|
|
2384
|
+
# Palmyra-Med uses the "<|end_of_text|>" as the end of text token, which is used by meta/llama-3-8b,
|
|
2385
|
+
# rather than "<|eot_id|>", which is used by meta/llama-3-8b-instruct
|
|
2386
|
+
tokenizer_name: meta/llama-3-8b
|
|
2387
|
+
max_sequence_length: 4096
|
|
2388
|
+
client_spec:
|
|
2389
|
+
class_name: "helm.clients.palmyra_client.PalmyraChatClient"
|
|
2390
|
+
|
|
2391
|
+
- name: writer/palmyra-fin-32k
|
|
2392
|
+
model_name: writer/palmyra-fin-32k
|
|
2393
|
+
tokenizer_name: meta/llama-3-8b-instruct
|
|
2394
|
+
max_sequence_length: 32000
|
|
2395
|
+
client_spec:
|
|
2396
|
+
class_name: "helm.clients.palmyra_client.PalmyraChatClient"
|
|
2282
2397
|
|
|
2283
2398
|
# Qwen
|
|
2284
2399
|
|
|
@@ -2431,4 +2546,4 @@ model_deployments:
|
|
|
2431
2546
|
tokenizer_name: openai/cl100k_base
|
|
2432
2547
|
max_sequence_length: 64000
|
|
2433
2548
|
client_spec:
|
|
2434
|
-
class_name: "helm.clients.reka_client.RekaClient"
|
|
2549
|
+
class_name: "helm.clients.reka_client.RekaClient"
|