crfm-helm 0.5.2__py3-none-any.whl → 0.5.3__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of crfm-helm might be problematic. Click here for more details.
- {crfm_helm-0.5.2.dist-info → crfm_helm-0.5.3.dist-info}/METADATA +29 -55
- {crfm_helm-0.5.2.dist-info → crfm_helm-0.5.3.dist-info}/RECORD +146 -134
- {crfm_helm-0.5.2.dist-info → crfm_helm-0.5.3.dist-info}/WHEEL +1 -1
- helm/benchmark/adaptation/adapters/multiple_choice_joint_adapter.py +12 -5
- helm/benchmark/adaptation/adapters/test_generation_adapter.py +12 -12
- helm/benchmark/adaptation/adapters/test_language_modeling_adapter.py +8 -8
- helm/benchmark/adaptation/adapters/test_multiple_choice_joint_adapter.py +77 -9
- helm/benchmark/adaptation/common_adapter_specs.py +2 -0
- helm/benchmark/annotation/anthropic_red_team_annotator.py +70 -0
- helm/benchmark/annotation/call_center_annotator.py +247 -0
- helm/benchmark/annotation/financebench_annotator.py +79 -0
- helm/benchmark/annotation/harm_bench_annotator.py +68 -0
- helm/benchmark/annotation/{image2structure → image2struct}/latex_compiler_annotator.py +2 -2
- helm/benchmark/annotation/{image2structure → image2struct}/lilypond_compiler_annotator.py +5 -3
- helm/benchmark/annotation/{image2structure → image2struct}/webpage_compiler_annotator.py +5 -5
- helm/benchmark/annotation/live_qa_annotator.py +32 -45
- helm/benchmark/annotation/medication_qa_annotator.py +31 -44
- helm/benchmark/annotation/model_as_judge.py +45 -0
- helm/benchmark/annotation/simple_safety_tests_annotator.py +64 -0
- helm/benchmark/annotation/xstest_annotator.py +110 -0
- helm/benchmark/metrics/annotation_metrics.py +108 -0
- helm/benchmark/metrics/bhasa_metrics.py +188 -0
- helm/benchmark/metrics/bhasa_metrics_specs.py +10 -0
- helm/benchmark/metrics/code_metrics_helper.py +11 -1
- helm/benchmark/metrics/safety_metrics.py +57 -0
- helm/benchmark/metrics/summac/model_summac.py +3 -3
- helm/benchmark/metrics/tokens/test_ai21_token_cost_estimator.py +2 -2
- helm/benchmark/metrics/tokens/test_openai_token_cost_estimator.py +4 -4
- helm/benchmark/metrics/vision_language/image_metrics.py +1 -1
- helm/benchmark/metrics/vision_language/image_utils.py +1 -1
- helm/benchmark/model_metadata_registry.py +3 -3
- helm/benchmark/presentation/test_run_entry.py +1 -0
- helm/benchmark/run.py +15 -0
- helm/benchmark/run_expander.py +56 -30
- helm/benchmark/run_specs/bhasa_run_specs.py +638 -0
- helm/benchmark/run_specs/call_center_run_specs.py +152 -0
- helm/benchmark/run_specs/decodingtrust_run_specs.py +8 -8
- helm/benchmark/run_specs/experimental_run_specs.py +52 -0
- helm/benchmark/run_specs/finance_run_specs.py +78 -1
- helm/benchmark/run_specs/safety_run_specs.py +154 -0
- helm/benchmark/run_specs/vlm_run_specs.py +92 -21
- helm/benchmark/scenarios/anthropic_red_team_scenario.py +71 -0
- helm/benchmark/scenarios/banking77_scenario.py +51 -0
- helm/benchmark/scenarios/bhasa_scenario.py +1798 -0
- helm/benchmark/scenarios/call_center_scenario.py +84 -0
- helm/benchmark/scenarios/decodingtrust_stereotype_bias_scenario.py +2 -1
- helm/benchmark/scenarios/ewok_scenario.py +116 -0
- helm/benchmark/scenarios/fin_qa_scenario.py +2 -0
- helm/benchmark/scenarios/financebench_scenario.py +53 -0
- helm/benchmark/scenarios/harm_bench_scenario.py +59 -0
- helm/benchmark/scenarios/scenario.py +1 -1
- helm/benchmark/scenarios/simple_safety_tests_scenario.py +33 -0
- helm/benchmark/scenarios/test_commonsense_scenario.py +21 -0
- helm/benchmark/scenarios/test_ewok_scenario.py +25 -0
- helm/benchmark/scenarios/test_financebench_scenario.py +26 -0
- helm/benchmark/scenarios/test_gsm_scenario.py +31 -0
- helm/benchmark/scenarios/test_legalbench_scenario.py +30 -0
- helm/benchmark/scenarios/test_math_scenario.py +2 -8
- helm/benchmark/scenarios/test_med_qa_scenario.py +30 -0
- helm/benchmark/scenarios/test_mmlu_scenario.py +33 -0
- helm/benchmark/scenarios/test_narrativeqa_scenario.py +73 -0
- helm/benchmark/scenarios/thai_exam_scenario.py +4 -4
- helm/benchmark/scenarios/vision_language/a_okvqa_scenario.py +1 -1
- helm/benchmark/scenarios/vision_language/bingo_scenario.py +2 -2
- helm/benchmark/scenarios/vision_language/crossmodal_3600_scenario.py +2 -1
- helm/benchmark/scenarios/vision_language/exams_v_scenario.py +104 -0
- helm/benchmark/scenarios/vision_language/fair_face_scenario.py +136 -0
- helm/benchmark/scenarios/vision_language/flickr30k_scenario.py +1 -1
- helm/benchmark/scenarios/vision_language/gqa_scenario.py +2 -2
- helm/benchmark/scenarios/vision_language/hateful_memes_scenario.py +1 -1
- helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/chart2csv_scenario.py +1 -1
- helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/latex_scenario.py +3 -3
- helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/musicsheet_scenario.py +1 -1
- helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/utils_latex.py +31 -39
- helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/webpage/driver.py +1 -1
- helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/webpage/utils.py +1 -1
- helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/webpage_scenario.py +41 -12
- helm/benchmark/scenarios/vision_language/math_vista_scenario.py +1 -1
- helm/benchmark/scenarios/vision_language/mementos_scenario.py +3 -3
- helm/benchmark/scenarios/vision_language/mm_safety_bench_scenario.py +2 -2
- helm/benchmark/scenarios/vision_language/mme_scenario.py +21 -18
- helm/benchmark/scenarios/vision_language/mmmu_scenario.py +1 -1
- helm/benchmark/scenarios/vision_language/pairs_scenario.py +1 -1
- helm/benchmark/scenarios/vision_language/pope_scenario.py +2 -1
- helm/benchmark/scenarios/vision_language/real_world_qa_scenario.py +57 -0
- helm/benchmark/scenarios/vision_language/seed_bench_scenario.py +7 -5
- helm/benchmark/scenarios/vision_language/unicorn_scenario.py +2 -2
- helm/benchmark/scenarios/vision_language/vibe_eval_scenario.py +6 -3
- helm/benchmark/scenarios/vision_language/viz_wiz_scenario.py +1 -1
- helm/benchmark/scenarios/vision_language/vqa_scenario.py +3 -1
- helm/benchmark/scenarios/xstest_scenario.py +35 -0
- helm/benchmark/server.py +1 -6
- helm/benchmark/static/schema_air_bench.yaml +750 -750
- helm/benchmark/static/schema_bhasa.yaml +709 -0
- helm/benchmark/static/schema_call_center.yaml +232 -0
- helm/benchmark/static/schema_cleva.yaml +768 -0
- helm/benchmark/static/schema_decodingtrust.yaml +444 -0
- helm/benchmark/static/schema_ewok.yaml +367 -0
- helm/benchmark/static/schema_finance.yaml +55 -9
- helm/benchmark/static/{schema_image2structure.yaml → schema_image2struct.yaml} +231 -90
- helm/benchmark/static/schema_safety.yaml +247 -0
- helm/benchmark/static/schema_tables.yaml +124 -7
- helm/benchmark/static/schema_thai.yaml +21 -0
- helm/benchmark/static/schema_vhelm.yaml +96 -91
- helm/benchmark/static_build/assets/accenture-6f97eeda.png +0 -0
- helm/benchmark/static_build/assets/aisingapore-6dfc9acf.png +0 -0
- helm/benchmark/static_build/assets/cresta-9e22b983.png +0 -0
- helm/benchmark/static_build/assets/cuhk-8c5631e9.png +0 -0
- helm/benchmark/static_build/assets/index-05c76bb1.css +1 -0
- helm/benchmark/static_build/assets/index-58f97dcd.js +10 -0
- helm/benchmark/static_build/assets/scb10x-204bd786.png +0 -0
- helm/benchmark/static_build/assets/wellsfargo-a86a6c4a.png +0 -0
- helm/benchmark/static_build/index.html +2 -2
- helm/benchmark/window_services/test_openai_window_service.py +8 -8
- helm/clients/ai21_client.py +71 -1
- helm/clients/anthropic_client.py +7 -19
- helm/clients/huggingface_client.py +38 -37
- helm/clients/nvidia_nim_client.py +35 -0
- helm/clients/openai_client.py +2 -3
- helm/clients/palmyra_client.py +25 -0
- helm/clients/perspective_api_client.py +11 -6
- helm/clients/test_client.py +4 -6
- helm/clients/vision_language/open_flamingo_client.py +1 -2
- helm/clients/vision_language/palmyra_vision_client.py +28 -13
- helm/common/images_utils.py +6 -0
- helm/common/mongo_key_value_store.py +2 -1
- helm/common/request.py +16 -0
- helm/config/model_deployments.yaml +315 -332
- helm/config/model_metadata.yaml +384 -110
- helm/config/tokenizer_configs.yaml +116 -11
- helm/proxy/example_queries.py +14 -21
- helm/proxy/services/server_service.py +1 -2
- helm/proxy/token_counters/test_auto_token_counter.py +2 -2
- helm/tokenizers/ai21_tokenizer.py +51 -59
- helm/tokenizers/cohere_tokenizer.py +0 -75
- helm/tokenizers/huggingface_tokenizer.py +0 -1
- helm/tokenizers/test_ai21_tokenizer.py +48 -0
- helm/benchmark/static/benchmarking.css +0 -156
- helm/benchmark/static/benchmarking.js +0 -1705
- helm/benchmark/static/config.js +0 -3
- helm/benchmark/static/general.js +0 -122
- helm/benchmark/static/images/crfm-logo.png +0 -0
- helm/benchmark/static/images/helm-logo-simple.png +0 -0
- helm/benchmark/static/images/helm-logo.png +0 -0
- helm/benchmark/static/images/language-model-helm.png +0 -0
- helm/benchmark/static/images/organizations/ai21.png +0 -0
- helm/benchmark/static/images/organizations/anthropic.png +0 -0
- helm/benchmark/static/images/organizations/bigscience.png +0 -0
- helm/benchmark/static/images/organizations/cohere.png +0 -0
- helm/benchmark/static/images/organizations/eleutherai.png +0 -0
- helm/benchmark/static/images/organizations/google.png +0 -0
- helm/benchmark/static/images/organizations/meta.png +0 -0
- helm/benchmark/static/images/organizations/microsoft.png +0 -0
- helm/benchmark/static/images/organizations/nvidia.png +0 -0
- helm/benchmark/static/images/organizations/openai.png +0 -0
- helm/benchmark/static/images/organizations/together.png +0 -0
- helm/benchmark/static/images/organizations/tsinghua-keg.png +0 -0
- helm/benchmark/static/images/organizations/yandex.png +0 -0
- helm/benchmark/static/images/scenarios-by-metrics.png +0 -0
- helm/benchmark/static/images/taxonomy-scenarios.png +0 -0
- helm/benchmark/static/index.html +0 -68
- helm/benchmark/static/info-icon.png +0 -0
- helm/benchmark/static/json-urls.js +0 -69
- helm/benchmark/static/plot-captions.js +0 -27
- helm/benchmark/static/utils.js +0 -285
- helm/benchmark/static_build/assets/index-30dbceba.js +0 -10
- helm/benchmark/static_build/assets/index-66b02d40.css +0 -1
- helm/benchmark/window_services/ai21_window_service.py +0 -247
- helm/benchmark/window_services/cohere_window_service.py +0 -101
- helm/benchmark/window_services/test_ai21_window_service.py +0 -163
- helm/benchmark/window_services/test_cohere_window_service.py +0 -75
- helm/benchmark/window_services/test_cohere_window_service_utils.py +0 -8328
- helm/benchmark/window_services/test_ice_window_service.py +0 -327
- helm/tokenizers/ice_tokenizer.py +0 -30
- helm/tokenizers/test_ice_tokenizer.py +0 -57
- {crfm_helm-0.5.2.dist-info → crfm_helm-0.5.3.dist-info}/LICENSE +0 -0
- {crfm_helm-0.5.2.dist-info → crfm_helm-0.5.3.dist-info}/entry_points.txt +0 -0
- {crfm_helm-0.5.2.dist-info → crfm_helm-0.5.3.dist-info}/top_level.txt +0 -0
- /helm/benchmark/annotation/{image2structure → image2struct}/__init__.py +0 -0
- /helm/benchmark/annotation/{image2structure → image2struct}/image_compiler_annotator.py +0 -0
- /helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/__init__.py +0 -0
- /helm/benchmark/scenarios/vision_language/{image2structure/image2structure_scenario.py → image2struct/image2struct_scenario.py} +0 -0
- /helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/webpage/__init__.py +0 -0
- /helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/webpage/jekyll_server.py +0 -0
helm/config/model_metadata.yaml
CHANGED
|
@@ -31,50 +31,41 @@ models:
|
|
|
31
31
|
|
|
32
32
|
|
|
33
33
|
# AI21 Labs
|
|
34
|
-
- name: ai21/j1-jumbo
|
|
34
|
+
- name: ai21/j1-jumbo
|
|
35
35
|
display_name: J1-Jumbo v1 (178B)
|
|
36
36
|
description: Jurassic-1 Jumbo (178B parameters) ([docs](https://studio.ai21.com/docs/jurassic1-language-models/), [tech report](https://uploads-ssl.webflow.com/60fd4503684b466578c0d307/61138924626a6981ee09caf6_jurassic_tech_paper.pdf)).
|
|
37
37
|
creator_organization_name: AI21 Labs
|
|
38
38
|
access: limited
|
|
39
39
|
num_parameters: 178000000000
|
|
40
40
|
release_date: 2021-08-11
|
|
41
|
-
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
41
|
+
tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
42
42
|
|
|
43
|
-
- name: ai21/j1-large
|
|
43
|
+
- name: ai21/j1-large
|
|
44
44
|
display_name: J1-Large v1 (7.5B)
|
|
45
45
|
description: Jurassic-1 Large (7.5B parameters) ([docs](https://studio.ai21.com/docs/jurassic1-language-models/), [tech report](https://uploads-ssl.webflow.com/60fd4503684b466578c0d307/61138924626a6981ee09caf6_jurassic_tech_paper.pdf)).
|
|
46
46
|
creator_organization_name: AI21 Labs
|
|
47
47
|
access: limited
|
|
48
48
|
num_parameters: 7500000000
|
|
49
49
|
release_date: 2021-08-11
|
|
50
|
-
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
50
|
+
tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
51
51
|
|
|
52
|
-
- name: ai21/j1-grande
|
|
52
|
+
- name: ai21/j1-grande
|
|
53
53
|
display_name: J1-Grande v1 (17B)
|
|
54
54
|
description: Jurassic-1 Grande (17B parameters) with a "few tweaks" to the training process ([docs](https://studio.ai21.com/docs/jurassic1-language-models/), [tech report](https://uploads-ssl.webflow.com/60fd4503684b466578c0d307/61138924626a6981ee09caf6_jurassic_tech_paper.pdf)).
|
|
55
55
|
creator_organization_name: AI21 Labs
|
|
56
56
|
access: limited
|
|
57
57
|
num_parameters: 17000000000
|
|
58
58
|
release_date: 2022-05-03
|
|
59
|
-
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
59
|
+
tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
60
60
|
|
|
61
|
-
- name: ai21/j1-grande-v2-beta
|
|
61
|
+
- name: ai21/j1-grande-v2-beta
|
|
62
62
|
display_name: J1-Grande v2 beta (17B)
|
|
63
63
|
description: Jurassic-1 Grande v2 beta (17B parameters)
|
|
64
64
|
creator_organization_name: AI21 Labs
|
|
65
65
|
access: limited
|
|
66
66
|
num_parameters: 17000000000
|
|
67
67
|
release_date: 2022-10-28
|
|
68
|
-
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
69
|
-
|
|
70
|
-
- name: ai21/j2-jumbo
|
|
71
|
-
display_name: Jurassic-2 Jumbo (178B)
|
|
72
|
-
description: Jurassic-2 Jumbo (178B parameters) ([docs](https://www.ai21.com/blog/introducing-j2))
|
|
73
|
-
creator_organization_name: AI21 Labs
|
|
74
|
-
access: limited
|
|
75
|
-
num_parameters: 178000000000
|
|
76
|
-
release_date: 2023-03-09
|
|
77
|
-
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
68
|
+
tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
78
69
|
|
|
79
70
|
- name: ai21/j2-large
|
|
80
71
|
display_name: Jurassic-2 Large (7.5B)
|
|
@@ -83,7 +74,7 @@ models:
|
|
|
83
74
|
access: limited
|
|
84
75
|
num_parameters: 7500000000
|
|
85
76
|
release_date: 2023-03-09
|
|
86
|
-
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
77
|
+
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
87
78
|
|
|
88
79
|
- name: ai21/j2-grande
|
|
89
80
|
display_name: Jurassic-2 Grande (17B)
|
|
@@ -92,13 +83,48 @@ models:
|
|
|
92
83
|
access: limited
|
|
93
84
|
num_parameters: 17000000000
|
|
94
85
|
release_date: 2023-03-09
|
|
95
|
-
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
86
|
+
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
87
|
+
|
|
88
|
+
- name: ai21/j2-jumbo
|
|
89
|
+
display_name: Jurassic-2 Jumbo (178B)
|
|
90
|
+
description: Jurassic-2 Jumbo (178B parameters) ([docs](https://www.ai21.com/blog/introducing-j2))
|
|
91
|
+
creator_organization_name: AI21 Labs
|
|
92
|
+
access: limited
|
|
93
|
+
num_parameters: 178000000000
|
|
94
|
+
release_date: 2023-03-09
|
|
95
|
+
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
96
96
|
|
|
97
97
|
# TODO(1524): Change AI21 model names
|
|
98
98
|
# - j2-jumbo -> j2-ultra
|
|
99
99
|
# - j2-grande -> j2-mid
|
|
100
100
|
# - j2-large -> j2-light
|
|
101
101
|
|
|
102
|
+
- name: ai21/jamba-instruct
|
|
103
|
+
display_name: Jamba Instruct
|
|
104
|
+
description: Jamba Instruct is an instruction tuned version of Jamba, which uses a hybrid Transformer-Mamba mixture-of-experts (MoE) architecture that interleaves blocks of Transformer and Mamba layers. ([blog](https://www.ai21.com/blog/announcing-jamba-instruct))
|
|
105
|
+
creator_organization_name: AI21 Labs
|
|
106
|
+
access: limited
|
|
107
|
+
num_parameters: 52000000000
|
|
108
|
+
release_date: 2024-05-02
|
|
109
|
+
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
110
|
+
|
|
111
|
+
- name: ai21/jamba-1.5-mini
|
|
112
|
+
display_name: Jamba 1.5 Mini
|
|
113
|
+
description: Jamba 1.5 Mini is a long-context, hybrid SSM-Transformer instruction following foundation model that is optimized for function calling, structured output, and grounded generation. ([blog](https://www.ai21.com/blog/announcing-jamba-model-family))
|
|
114
|
+
creator_organization_name: AI21 Labs
|
|
115
|
+
access: open
|
|
116
|
+
num_parameters: 51600000000
|
|
117
|
+
release_date: 2024-08-22
|
|
118
|
+
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
119
|
+
|
|
120
|
+
- name: ai21/jamba-1.5-large
|
|
121
|
+
display_name: Jamba 1.5 Large
|
|
122
|
+
description: Jamba 1.5 Large is a long-context, hybrid SSM-Transformer instruction following foundation model that is optimized for function calling, structured output, and grounded generation. ([blog](https://www.ai21.com/blog/announcing-jamba-model-family))
|
|
123
|
+
creator_organization_name: AI21 Labs
|
|
124
|
+
access: open
|
|
125
|
+
num_parameters: 399000000000
|
|
126
|
+
release_date: 2024-08-22
|
|
127
|
+
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
102
128
|
|
|
103
129
|
# AI Singapore
|
|
104
130
|
- name: aisingapore/sea-lion-7b
|
|
@@ -117,7 +143,7 @@ models:
|
|
|
117
143
|
access: open
|
|
118
144
|
num_parameters: 7000000000
|
|
119
145
|
release_date: 2023-02-24
|
|
120
|
-
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
146
|
+
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
121
147
|
|
|
122
148
|
|
|
123
149
|
# Aleph Alpha
|
|
@@ -272,7 +298,14 @@ models:
|
|
|
272
298
|
release_date: 2024-03-04 # https://www.anthropic.com/news/claude-3-family
|
|
273
299
|
tags: [ANTHROPIC_CLAUDE_3_MODEL_TAG, TEXT_MODEL_TAG, VISION_LANGUAGE_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
274
300
|
|
|
275
|
-
|
|
301
|
+
- name: anthropic/claude-3-5-sonnet-20240620
|
|
302
|
+
display_name: Claude 3.5 Sonnet (20240620)
|
|
303
|
+
description: Claude 3.5 Sonnet is a Claude 3 family model which outperforms Claude 3 Opus while operating faster and at a lower cost. ([blog](https://www.anthropic.com/news/claude-3-5-sonnet))
|
|
304
|
+
creator_organization_name: Anthropic
|
|
305
|
+
access: limited
|
|
306
|
+
release_date: 2024-06-20
|
|
307
|
+
tags: [ANTHROPIC_CLAUDE_3_MODEL_TAG, TEXT_MODEL_TAG, VISION_LANGUAGE_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
308
|
+
|
|
276
309
|
- name: anthropic/stanford-online-all-v4-s3
|
|
277
310
|
display_name: Anthropic-LM v4-s3 (52B)
|
|
278
311
|
description: A 52B parameter language model, trained using reinforcement learning from human feedback [paper](https://arxiv.org/pdf/2204.05862.pdf).
|
|
@@ -280,7 +313,7 @@ models:
|
|
|
280
313
|
access: closed
|
|
281
314
|
num_parameters: 52000000000
|
|
282
315
|
release_date: 2021-12-01
|
|
283
|
-
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG, ABLATION_MODEL_TAG]
|
|
316
|
+
tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG, ABLATION_MODEL_TAG]
|
|
284
317
|
|
|
285
318
|
|
|
286
319
|
|
|
@@ -401,16 +434,16 @@ models:
|
|
|
401
434
|
access: limited
|
|
402
435
|
num_parameters: 52400000000
|
|
403
436
|
release_date: 2022-06-09
|
|
404
|
-
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
437
|
+
tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
405
438
|
|
|
406
|
-
- name: cohere/large-20220720
|
|
439
|
+
- name: cohere/large-20220720
|
|
407
440
|
display_name: Cohere large v20220720 (13.1B)
|
|
408
441
|
description: Cohere large v20220720 (13.1B parameters), which is deprecated by Cohere as of December 2, 2022.
|
|
409
442
|
creator_organization_name: Cohere
|
|
410
443
|
access: limited
|
|
411
444
|
num_parameters: 13100000000
|
|
412
445
|
release_date: 2022-07-20
|
|
413
|
-
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
446
|
+
tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
414
447
|
|
|
415
448
|
- name: cohere/medium-20220720
|
|
416
449
|
display_name: Cohere medium v20220720 (6.1B)
|
|
@@ -419,16 +452,16 @@ models:
|
|
|
419
452
|
access: limited
|
|
420
453
|
num_parameters: 6100000000
|
|
421
454
|
release_date: 2022-07-20
|
|
422
|
-
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
455
|
+
tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
423
456
|
|
|
424
|
-
- name: cohere/small-20220720
|
|
457
|
+
- name: cohere/small-20220720
|
|
425
458
|
display_name: Cohere small v20220720 (410M)
|
|
426
459
|
description: Cohere small v20220720 (410M parameters), which is deprecated by Cohere as of December 2, 2022.
|
|
427
460
|
creator_organization_name: Cohere
|
|
428
461
|
access: limited
|
|
429
462
|
num_parameters: 410000000
|
|
430
463
|
release_date: 2022-07-20
|
|
431
|
-
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
464
|
+
tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
432
465
|
|
|
433
466
|
- name: cohere/xlarge-20221108
|
|
434
467
|
display_name: Cohere xlarge v20221108 (52.4B)
|
|
@@ -437,34 +470,34 @@ models:
|
|
|
437
470
|
access: limited
|
|
438
471
|
num_parameters: 52400000000
|
|
439
472
|
release_date: 2022-11-08
|
|
440
|
-
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
473
|
+
tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
441
474
|
|
|
442
|
-
- name: cohere/medium-20221108
|
|
475
|
+
- name: cohere/medium-20221108
|
|
443
476
|
display_name: Cohere medium v20221108 (6.1B)
|
|
444
477
|
description: Cohere medium v20221108 (6.1B parameters)
|
|
445
478
|
creator_organization_name: Cohere
|
|
446
479
|
access: limited
|
|
447
480
|
num_parameters: 6100000000
|
|
448
481
|
release_date: 2022-11-08
|
|
449
|
-
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
482
|
+
tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
450
483
|
|
|
451
|
-
- name: cohere/command-medium-beta
|
|
484
|
+
- name: cohere/command-medium-beta
|
|
452
485
|
display_name: Command beta (6.1B)
|
|
453
|
-
description:
|
|
486
|
+
description: Command beta (6.1B parameters) is fine-tuned from the medium model to respond well with instruction-like prompts ([details](https://docs.cohere.ai/docs/command-beta)).
|
|
454
487
|
creator_organization_name: Cohere
|
|
455
488
|
access: limited
|
|
456
489
|
num_parameters: 6100000000
|
|
457
490
|
release_date: 2022-11-08
|
|
458
|
-
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
491
|
+
tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
459
492
|
|
|
460
|
-
- name: cohere/command-xlarge-beta
|
|
493
|
+
- name: cohere/command-xlarge-beta
|
|
461
494
|
display_name: Command beta (52.4B)
|
|
462
|
-
description:
|
|
495
|
+
description: Command beta (52.4B parameters) is fine-tuned from the XL model to respond well with instruction-like prompts ([details](https://docs.cohere.ai/docs/command-beta)).
|
|
463
496
|
creator_organization_name: Cohere
|
|
464
497
|
access: limited
|
|
465
498
|
num_parameters: 52400000000
|
|
466
499
|
release_date: 2022-11-08
|
|
467
|
-
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
500
|
+
tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
468
501
|
|
|
469
502
|
- name: cohere/command
|
|
470
503
|
display_name: Command
|
|
@@ -837,8 +870,7 @@ models:
|
|
|
837
870
|
|
|
838
871
|
- name: google/gemma-2b
|
|
839
872
|
display_name: Gemma (2B)
|
|
840
|
-
|
|
841
|
-
description: TBD
|
|
873
|
+
description: Gemma is a family of lightweight, open models built from the research and technology that Google used to create the Gemini models. ([model card](https://www.kaggle.com/models/google/gemma), [blog post](https://blog.google/technology/developers/gemma-open-models/))
|
|
842
874
|
creator_organization_name: Google
|
|
843
875
|
access: open
|
|
844
876
|
release_date: 2024-02-21
|
|
@@ -846,8 +878,7 @@ models:
|
|
|
846
878
|
|
|
847
879
|
- name: google/gemma-2b-it
|
|
848
880
|
display_name: Gemma Instruct (2B)
|
|
849
|
-
|
|
850
|
-
description: TBD
|
|
881
|
+
description: Gemma is a family of lightweight, open models built from the research and technology that Google used to create the Gemini models. ([model card](https://www.kaggle.com/models/google/gemma), [blog post](https://blog.google/technology/developers/gemma-open-models/))
|
|
851
882
|
creator_organization_name: Google
|
|
852
883
|
access: open
|
|
853
884
|
release_date: 2024-02-21
|
|
@@ -855,8 +886,7 @@ models:
|
|
|
855
886
|
|
|
856
887
|
- name: google/gemma-7b
|
|
857
888
|
display_name: Gemma (7B)
|
|
858
|
-
|
|
859
|
-
description: TBD
|
|
889
|
+
description: Gemma is a family of lightweight, open models built from the research and technology that Google used to create the Gemini models. ([model card](https://www.kaggle.com/models/google/gemma), [blog post](https://blog.google/technology/developers/gemma-open-models/))
|
|
860
890
|
creator_organization_name: Google
|
|
861
891
|
access: open
|
|
862
892
|
release_date: 2024-02-21
|
|
@@ -864,12 +894,42 @@ models:
|
|
|
864
894
|
|
|
865
895
|
- name: google/gemma-7b-it
|
|
866
896
|
display_name: Gemma Instruct (7B)
|
|
867
|
-
|
|
868
|
-
description: TBD
|
|
897
|
+
description: Gemma is a family of lightweight, open models built from the research and technology that Google used to create the Gemini models. ([model card](https://www.kaggle.com/models/google/gemma), [blog post](https://blog.google/technology/developers/gemma-open-models/))
|
|
869
898
|
creator_organization_name: Google
|
|
870
899
|
access: open
|
|
871
900
|
release_date: 2024-02-21
|
|
872
|
-
|
|
901
|
+
tags: [TEXT_MODEL_TAG, GOOGLE_GEMMA_INSTRUCT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
902
|
+
|
|
903
|
+
- name: google/gemma-2-9b
|
|
904
|
+
display_name: Gemma 2 (9B)
|
|
905
|
+
description: Gemma is a family of lightweight, open models built from the research and technology that Google used to create the Gemini models. ([model card](https://www.kaggle.com/models/google/gemma), [blog post](https://blog.google/technology/developers/google-gemma-2/))
|
|
906
|
+
creator_organization_name: Google
|
|
907
|
+
access: open
|
|
908
|
+
release_date: 2024-06-27
|
|
909
|
+
tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
910
|
+
|
|
911
|
+
- name: google/gemma-2-9b-it
|
|
912
|
+
display_name: Gemma 2 Instruct (9B)
|
|
913
|
+
description: Gemma is a family of lightweight, open models built from the research and technology that Google used to create the Gemini models. ([model card](https://www.kaggle.com/models/google/gemma), [blog post](https://blog.google/technology/developers/google-gemma-2/))
|
|
914
|
+
creator_organization_name: Google
|
|
915
|
+
access: open
|
|
916
|
+
release_date: 2024-06-27
|
|
917
|
+
tags: [TEXT_MODEL_TAG, GOOGLE_GEMMA_INSTRUCT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
918
|
+
|
|
919
|
+
- name: google/gemma-2-27b
|
|
920
|
+
display_name: Gemma 2 (27B)
|
|
921
|
+
description: Gemma is a family of lightweight, open models built from the research and technology that Google used to create the Gemini models. ([model card](https://www.kaggle.com/models/google/gemma), [blog post](https://blog.google/technology/developers/google-gemma-2/))
|
|
922
|
+
creator_organization_name: Google
|
|
923
|
+
access: open
|
|
924
|
+
release_date: 2024-06-27
|
|
925
|
+
tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
926
|
+
|
|
927
|
+
- name: google/gemma-2-27b-it
|
|
928
|
+
display_name: Gemma 2 Instruct (27B)
|
|
929
|
+
description: Gemma is a family of lightweight, open models built from the research and technology that Google used to create the Gemini models. ([model card](https://www.kaggle.com/models/google/gemma), [blog post](https://blog.google/technology/developers/google-gemma-2/))
|
|
930
|
+
creator_organization_name: Google
|
|
931
|
+
access: open
|
|
932
|
+
release_date: 2024-06-27
|
|
873
933
|
tags: [TEXT_MODEL_TAG, GOOGLE_GEMMA_INSTRUCT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
874
934
|
|
|
875
935
|
- name: google/paligemma-3b-mix-224
|
|
@@ -1361,7 +1421,7 @@ models:
|
|
|
1361
1421
|
|
|
1362
1422
|
- name: meta/llama-3-8b
|
|
1363
1423
|
display_name: Llama 3 (8B)
|
|
1364
|
-
description: Llama 3 is a family of language models that have been trained on more than 15 trillion tokens, and use Grouped-Query Attention (GQA) for improved inference scalability.
|
|
1424
|
+
description: Llama 3 is a family of language models that have been trained on more than 15 trillion tokens, and use Grouped-Query Attention (GQA) for improved inference scalability. ([paper](https://ai.meta.com/research/publications/the-llama-3-herd-of-models/)
|
|
1365
1425
|
creator_organization_name: Meta
|
|
1366
1426
|
access: open
|
|
1367
1427
|
num_parameters: 8000000000
|
|
@@ -1370,16 +1430,43 @@ models:
|
|
|
1370
1430
|
|
|
1371
1431
|
- name: meta/llama-3-70b
|
|
1372
1432
|
display_name: Llama 3 (70B)
|
|
1373
|
-
description: Llama 3 is a family of language models that have been trained on more than 15 trillion tokens, and use Grouped-Query Attention (GQA) for improved inference scalability.
|
|
1433
|
+
description: Llama 3 is a family of language models that have been trained on more than 15 trillion tokens, and use Grouped-Query Attention (GQA) for improved inference scalability. ([paper](https://ai.meta.com/research/publications/the-llama-3-herd-of-models/)
|
|
1374
1434
|
creator_organization_name: Meta
|
|
1375
1435
|
access: open
|
|
1376
1436
|
num_parameters: 70000000000
|
|
1377
1437
|
release_date: 2024-04-18
|
|
1378
1438
|
tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
1379
1439
|
|
|
1440
|
+
- name: meta/llama-3.1-8b-instruct-turbo
|
|
1441
|
+
display_name: Llama 3.1 Instruct Turbo (8B)
|
|
1442
|
+
description: Llama 3.1 (8B) is part of the Llama 3 family of dense Transformer models that that natively support multilinguality, coding, reasoning, and tool usage. ([paper](https://ai.meta.com/research/publications/the-llama-3-herd-of-models/), [blog](https://ai.meta.com/blog/meta-llama-3-1/)) Turbo is Together's implementation, providing a near negligible difference in quality from the reference implementation with faster performance and lower cost, currently using FP8 quantization. ([blog](https://www.together.ai/blog/llama-31-quality))
|
|
1443
|
+
creator_organization_name: Meta
|
|
1444
|
+
access: open
|
|
1445
|
+
num_parameters: 8000000000
|
|
1446
|
+
release_date: 2024-07-23
|
|
1447
|
+
tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
1448
|
+
|
|
1449
|
+
- name: meta/llama-3.1-70b-instruct-turbo
|
|
1450
|
+
display_name: Llama 3.1 Instruct Turbo (70B)
|
|
1451
|
+
description: Llama 3.1 (70B) is part of the Llama 3 family of dense Transformer models that that natively support multilinguality, coding, reasoning, and tool usage. ([paper](https://ai.meta.com/research/publications/the-llama-3-herd-of-models/), [blog](https://ai.meta.com/blog/meta-llama-3-1/)) Turbo is Together's implementation, providing a near negligible difference in quality from the reference implementation with faster performance and lower cost, currently using FP8 quantization. ([blog](https://www.together.ai/blog/llama-31-quality))
|
|
1452
|
+
creator_organization_name: Meta
|
|
1453
|
+
access: open
|
|
1454
|
+
num_parameters: 70000000000
|
|
1455
|
+
release_date: 2024-07-23
|
|
1456
|
+
tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
1457
|
+
|
|
1458
|
+
- name: meta/llama-3.1-405b-instruct-turbo
|
|
1459
|
+
display_name: Llama 3.1 Instruct Turbo (405B)
|
|
1460
|
+
description: Llama 3.1 (405B) is part of the Llama 3 family of dense Transformer models that that natively support multilinguality, coding, reasoning, and tool usage. ([paper](https://ai.meta.com/research/publications/the-llama-3-herd-of-models/), [blog](https://ai.meta.com/blog/meta-llama-3-1/)) Turbo is Together's implementation, providing a near negligible difference in quality from the reference implementation with faster performance and lower cost, currently using FP8 quantization. ([blog](https://www.together.ai/blog/llama-31-quality))
|
|
1461
|
+
creator_organization_name: Meta
|
|
1462
|
+
access: open
|
|
1463
|
+
num_parameters: 405000000000
|
|
1464
|
+
release_date: 2024-07-23
|
|
1465
|
+
tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
1466
|
+
|
|
1380
1467
|
- name: meta/llama-3-8b-chat
|
|
1381
1468
|
display_name: Llama 3 Instruct (8B)
|
|
1382
|
-
description: Llama 3 is a family of language models that have been trained on more than 15 trillion tokens, and use Grouped-Query Attention (GQA) for improved inference scalability. It used SFT, rejection sampling, PPO and DPO for post-training.
|
|
1469
|
+
description: Llama 3 is a family of language models that have been trained on more than 15 trillion tokens, and use Grouped-Query Attention (GQA) for improved inference scalability. It used SFT, rejection sampling, PPO and DPO for post-training. ([paper](https://ai.meta.com/research/publications/the-llama-3-herd-of-models/)
|
|
1383
1470
|
creator_organization_name: Meta
|
|
1384
1471
|
access: open
|
|
1385
1472
|
num_parameters: 8000000000
|
|
@@ -1388,7 +1475,7 @@ models:
|
|
|
1388
1475
|
|
|
1389
1476
|
- name: meta/llama-3-70b-chat
|
|
1390
1477
|
display_name: Llama 3 Instruct (70B)
|
|
1391
|
-
description: Llama 3 is a family of language models that have been trained on more than 15 trillion tokens, and use Grouped-Query Attention (GQA) for improved inference scalability. It used SFT, rejection sampling, PPO and DPO for post-training.
|
|
1478
|
+
description: Llama 3 is a family of language models that have been trained on more than 15 trillion tokens, and use Grouped-Query Attention (GQA) for improved inference scalability. It used SFT, rejection sampling, PPO and DPO for post-training. ([paper](https://ai.meta.com/research/publications/the-llama-3-herd-of-models/)
|
|
1392
1479
|
creator_organization_name: Meta
|
|
1393
1480
|
access: open
|
|
1394
1481
|
num_parameters: 70000000000
|
|
@@ -1413,6 +1500,16 @@ models:
|
|
|
1413
1500
|
release_date: 2024-04-18
|
|
1414
1501
|
tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
1415
1502
|
|
|
1503
|
+
- name: meta/llama-guard-3-8b
|
|
1504
|
+
display_name: Llama Guard 3 (8B)
|
|
1505
|
+
description: Llama Guard 3 is an 8B parameter Llama 3.1-based LLM safeguard model. Similar to Llama Guard, it can be used for classifying content in both LLM inputs (prompt classification) and in LLM responses (response classification). It acts as an LLM – it generates text in its output that indicates whether a given prompt or response is safe or unsafe, and if unsafe, it also lists the content categories violated.
|
|
1506
|
+
creator_organization_name: Meta
|
|
1507
|
+
access: open
|
|
1508
|
+
num_parameters: 8000000000
|
|
1509
|
+
release_date: 2024-07-23
|
|
1510
|
+
tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
1511
|
+
|
|
1512
|
+
|
|
1416
1513
|
|
|
1417
1514
|
|
|
1418
1515
|
|
|
@@ -1424,7 +1521,7 @@ models:
|
|
|
1424
1521
|
access: closed
|
|
1425
1522
|
num_parameters: 530000000000
|
|
1426
1523
|
release_date: 2022-01-28
|
|
1427
|
-
tags: [
|
|
1524
|
+
tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
1428
1525
|
|
|
1429
1526
|
- name: microsoft/TNLGv2_7B
|
|
1430
1527
|
display_name: TNLG v2 (6.7B)
|
|
@@ -1433,7 +1530,7 @@ models:
|
|
|
1433
1530
|
access: closed
|
|
1434
1531
|
num_parameters: 6700000000
|
|
1435
1532
|
release_date: 2022-01-28
|
|
1436
|
-
tags: [
|
|
1533
|
+
tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
1437
1534
|
|
|
1438
1535
|
- name: microsoft/llava-1.5-7b-hf
|
|
1439
1536
|
display_name: LLaVA 1.5 (7B)
|
|
@@ -1507,6 +1604,24 @@ models:
|
|
|
1507
1604
|
release_date: 2023-10-05
|
|
1508
1605
|
tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
1509
1606
|
|
|
1607
|
+
- name: microsoft/phi-3-small-8k-instruct
|
|
1608
|
+
display_name: Phi-3 (7B)
|
|
1609
|
+
description: Phi-3-Small-8K-Instruct is a lightweight model trained with synthetic data and filtered publicly available website data with a focus on high-quality and reasoning dense properties. ([paper](https://arxiv.org/abs/2404.14219), [blog](https://azure.microsoft.com/en-us/blog/new-models-added-to-the-phi-3-family-available-on-microsoft-azure/))
|
|
1610
|
+
creator_organization_name: Microsoft
|
|
1611
|
+
access: open
|
|
1612
|
+
num_parameters: 7000000000
|
|
1613
|
+
release_date: 2024-05-21
|
|
1614
|
+
tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
1615
|
+
|
|
1616
|
+
- name: microsoft/phi-3-medium-4k-instruct
|
|
1617
|
+
display_name: Phi-3 (14B)
|
|
1618
|
+
description: Phi-3-Medium-4K-Instruct is a lightweight model trained with synthetic data and filtered publicly available website data with a focus on high-quality and reasoning dense properties. ([paper](https://arxiv.org/abs/2404.14219), [blog](https://azure.microsoft.com/en-us/blog/new-models-added-to-the-phi-3-family-available-on-microsoft-azure/))
|
|
1619
|
+
creator_organization_name: Microsoft
|
|
1620
|
+
access: open
|
|
1621
|
+
num_parameters: 14000000000
|
|
1622
|
+
release_date: 2024-05-21
|
|
1623
|
+
tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
1624
|
+
|
|
1510
1625
|
# KAIST AI
|
|
1511
1626
|
- name: kaistai/prometheus-vision-13b-v1.0-hf
|
|
1512
1627
|
display_name: LLaVA + Vicuna-v1.5 (13B)
|
|
@@ -1663,7 +1778,7 @@ models:
|
|
|
1663
1778
|
num_parameters: 46700000000
|
|
1664
1779
|
# Blog post: https://mistral.ai/news/mixtral-of-experts/
|
|
1665
1780
|
release_date: 2023-12-11
|
|
1666
|
-
tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG
|
|
1781
|
+
tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
1667
1782
|
|
|
1668
1783
|
- name: mistralai/mixtral-8x22b
|
|
1669
1784
|
display_name: Mixtral (8x22B)
|
|
@@ -1694,13 +1809,11 @@ models:
|
|
|
1694
1809
|
|
|
1695
1810
|
- name: mistralai/mistral-small-2402
|
|
1696
1811
|
display_name: Mistral Small (2402)
|
|
1697
|
-
|
|
1698
|
-
description: TBD
|
|
1812
|
+
description: Mistral Small is a multilingual model with a 32K tokens context window and function-calling capabilities. ([blog](https://mistral.ai/news/mistral-large/))
|
|
1699
1813
|
creator_organization_name: Mistral AI
|
|
1700
1814
|
access: limited
|
|
1701
|
-
# Blog post: https://mistral.ai/news/mistral-large/
|
|
1702
1815
|
release_date: 2023-02-26
|
|
1703
|
-
tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG
|
|
1816
|
+
tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
1704
1817
|
|
|
1705
1818
|
- name: mistralai/mistral-medium-2312
|
|
1706
1819
|
display_name: Mistral Medium (2312)
|
|
@@ -1708,18 +1821,32 @@ models:
|
|
|
1708
1821
|
creator_organization_name: Mistral AI
|
|
1709
1822
|
access: limited
|
|
1710
1823
|
release_date: 2023-12-11
|
|
1711
|
-
tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG
|
|
1824
|
+
tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
1712
1825
|
|
|
1713
1826
|
- name: mistralai/mistral-large-2402
|
|
1714
1827
|
display_name: Mistral Large (2402)
|
|
1715
|
-
|
|
1716
|
-
description: TBD
|
|
1828
|
+
description: Mistral Large is a multilingual model with a 32K tokens context window and function-calling capabilities. ([blog](https://mistral.ai/news/mistral-large/))
|
|
1717
1829
|
creator_organization_name: Mistral AI
|
|
1718
1830
|
access: limited
|
|
1719
|
-
# Blog post: https://mistral.ai/news/mistral-large/
|
|
1720
1831
|
release_date: 2023-02-26
|
|
1721
|
-
tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG
|
|
1832
|
+
tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
1722
1833
|
|
|
1834
|
+
- name: mistralai/mistral-large-2407
|
|
1835
|
+
display_name: Mistral Large 2 (2407)
|
|
1836
|
+
description: Mistral Large 2 is a 123 billion parameter model that has a 128k context window and supports dozens of languages and 80+ coding languages. ([blog](https://mistral.ai/news/mistral-large-2407/))
|
|
1837
|
+
creator_organization_name: Mistral AI
|
|
1838
|
+
access: open
|
|
1839
|
+
num_parameters: 123000000000
|
|
1840
|
+
release_date: 2023-07-24
|
|
1841
|
+
tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
1842
|
+
|
|
1843
|
+
- name: mistralai/open-mistral-nemo-2407
|
|
1844
|
+
display_name: Mistral NeMo (2402)
|
|
1845
|
+
description: Mistral NeMo is a multilingual 12B model with a large context window of 128K tokens. ([blog](https://mistral.ai/news/mistral-nemo/))
|
|
1846
|
+
creator_organization_name: Mistral AI
|
|
1847
|
+
access: open
|
|
1848
|
+
release_date: 2024-07-18
|
|
1849
|
+
tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
1723
1850
|
|
|
1724
1851
|
# MosaicML
|
|
1725
1852
|
- name: mosaicml/mpt-7b
|
|
@@ -1798,7 +1925,13 @@ models:
|
|
|
1798
1925
|
release_date: 2019-09-17 # paper date
|
|
1799
1926
|
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG, BUGGY_TEMP_0_TAG]
|
|
1800
1927
|
|
|
1801
|
-
|
|
1928
|
+
- name: nvidia/nemotron-4-340b-instruct
|
|
1929
|
+
display_name: Nemotron-4 Instruct (340B)
|
|
1930
|
+
description: Nemotron-4 Instruct (340B) is an open weights model sized to fit on a single DGX H100 with 8 GPUs when deployed in FP8 precision. 98% of the data used for model alignment was synthetically generated ([paper](https://arxiv.org/abs/2406.11704)).
|
|
1931
|
+
creator_organization_name: NVIDIA
|
|
1932
|
+
access: open
|
|
1933
|
+
release_date: 2024-06-17
|
|
1934
|
+
tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
1802
1935
|
|
|
1803
1936
|
# OpenAI
|
|
1804
1937
|
|
|
@@ -1836,97 +1969,95 @@ models:
|
|
|
1836
1969
|
|
|
1837
1970
|
# DEPRECATED: Announced on July 06 2023 that these models will be shut down on January 04 2024.
|
|
1838
1971
|
|
|
1839
|
-
- name: openai/davinci
|
|
1972
|
+
- name: openai/davinci
|
|
1840
1973
|
display_name: davinci (175B)
|
|
1841
1974
|
description: Original GPT-3 (175B parameters) autoregressive language model ([paper](https://arxiv.org/pdf/2005.14165.pdf), [docs](https://beta.openai.com/docs/model-index-for-researchers)).
|
|
1842
1975
|
creator_organization_name: OpenAI
|
|
1843
1976
|
access: limited
|
|
1844
1977
|
num_parameters: 175000000000
|
|
1845
1978
|
release_date: 2020-05-28
|
|
1846
|
-
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
1979
|
+
tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
1847
1980
|
|
|
1848
|
-
- name: openai/curie
|
|
1981
|
+
- name: openai/curie
|
|
1849
1982
|
display_name: curie (6.7B)
|
|
1850
1983
|
description: Original GPT-3 (6.7B parameters) autoregressive language model ([paper](https://arxiv.org/pdf/2005.14165.pdf), [docs](https://beta.openai.com/docs/model-index-for-researchers)).
|
|
1851
1984
|
creator_organization_name: OpenAI
|
|
1852
1985
|
access: limited
|
|
1853
1986
|
num_parameters: 6700000000
|
|
1854
1987
|
release_date: 2020-05-28
|
|
1855
|
-
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
1988
|
+
tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
1856
1989
|
|
|
1857
|
-
- name: openai/babbage
|
|
1990
|
+
- name: openai/babbage
|
|
1858
1991
|
display_name: babbage (1.3B)
|
|
1859
1992
|
description: Original GPT-3 (1.3B parameters) autoregressive language model ([paper](https://arxiv.org/pdf/2005.14165.pdf), [docs](https://beta.openai.com/docs/model-index-for-researchers)).
|
|
1860
1993
|
creator_organization_name: OpenAI
|
|
1861
1994
|
access: limited
|
|
1862
1995
|
num_parameters: 1300000000
|
|
1863
1996
|
release_date: 2020-05-28
|
|
1864
|
-
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
1997
|
+
tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
1865
1998
|
|
|
1866
|
-
- name: openai/ada
|
|
1999
|
+
- name: openai/ada
|
|
1867
2000
|
display_name: ada (350M)
|
|
1868
2001
|
description: Original GPT-3 (350M parameters) autoregressive language model ([paper](https://arxiv.org/pdf/2005.14165.pdf), [docs](https://beta.openai.com/docs/model-index-for-researchers)).
|
|
1869
2002
|
creator_organization_name: OpenAI
|
|
1870
2003
|
access: limited
|
|
1871
2004
|
num_parameters: 350000000
|
|
1872
2005
|
release_date: 2020-05-28
|
|
1873
|
-
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
2006
|
+
tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
1874
2007
|
|
|
1875
|
-
- name: openai/text-davinci-003
|
|
2008
|
+
- name: openai/text-davinci-003
|
|
1876
2009
|
display_name: GPT-3.5 (text-davinci-003)
|
|
1877
2010
|
description: text-davinci-003 model that involves reinforcement learning (PPO) with reward models. Derived from text-davinci-002 ([docs](https://beta.openai.com/docs/model-index-for-researchers)).
|
|
1878
2011
|
creator_organization_name: OpenAI
|
|
1879
2012
|
access: limited
|
|
1880
2013
|
num_parameters: 175000000000
|
|
1881
2014
|
release_date: 2022-11-28
|
|
1882
|
-
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
2015
|
+
tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
1883
2016
|
|
|
1884
|
-
|
|
1885
|
-
# https://github.com/stanford-crfm/benchmarking/issues/359
|
|
1886
|
-
- name: openai/text-davinci-002 # DEPRECATED
|
|
2017
|
+
- name: openai/text-davinci-002
|
|
1887
2018
|
display_name: GPT-3.5 (text-davinci-002)
|
|
1888
2019
|
description: text-davinci-002 model that involves supervised fine-tuning on human-written demonstrations. Derived from code-davinci-002 ([docs](https://beta.openai.com/docs/model-index-for-researchers)).
|
|
1889
2020
|
creator_organization_name: OpenAI
|
|
1890
2021
|
access: limited
|
|
1891
2022
|
num_parameters: 175000000000
|
|
1892
2023
|
release_date: 2022-01-27
|
|
1893
|
-
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
2024
|
+
tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
1894
2025
|
|
|
1895
|
-
- name: openai/text-davinci-001
|
|
2026
|
+
- name: openai/text-davinci-001
|
|
1896
2027
|
display_name: GPT-3.5 (text-davinci-001)
|
|
1897
2028
|
description: text-davinci-001 model that involves supervised fine-tuning on human-written demonstrations ([docs](https://beta.openai.com/docs/model-index-for-researchers)).
|
|
1898
2029
|
creator_organization_name: OpenAI
|
|
1899
2030
|
access: limited
|
|
1900
2031
|
num_parameters: 175000000000
|
|
1901
2032
|
release_date: 2022-01-27
|
|
1902
|
-
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
2033
|
+
tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
1903
2034
|
|
|
1904
|
-
- name: openai/text-curie-001
|
|
2035
|
+
- name: openai/text-curie-001
|
|
1905
2036
|
display_name: text-curie-001
|
|
1906
2037
|
description: text-curie-001 model that involves supervised fine-tuning on human-written demonstrations ([docs](https://beta.openai.com/docs/model-index-for-researchers)).
|
|
1907
2038
|
creator_organization_name: OpenAI
|
|
1908
2039
|
access: limited
|
|
1909
2040
|
num_parameters: 6700000000
|
|
1910
2041
|
release_date: 2022-01-27
|
|
1911
|
-
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
2042
|
+
tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
1912
2043
|
|
|
1913
|
-
- name: openai/text-babbage-001
|
|
2044
|
+
- name: openai/text-babbage-001
|
|
1914
2045
|
display_name: text-babbage-001
|
|
1915
2046
|
description: text-babbage-001 model that involves supervised fine-tuning on human-written demonstrations ([docs](https://beta.openai.com/docs/model-index-for-researchers)).
|
|
1916
2047
|
creator_organization_name: OpenAI
|
|
1917
2048
|
access: limited
|
|
1918
2049
|
num_parameters: 1300000000
|
|
1919
2050
|
release_date: 2022-01-27
|
|
1920
|
-
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
2051
|
+
tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
1921
2052
|
|
|
1922
|
-
- name: openai/text-ada-001
|
|
2053
|
+
- name: openai/text-ada-001
|
|
1923
2054
|
display_name: text-ada-001
|
|
1924
2055
|
description: text-ada-001 model that involves supervised fine-tuning on human-written demonstrations ([docs](https://beta.openai.com/docs/model-index-for-researchers)).
|
|
1925
2056
|
creator_organization_name: OpenAI
|
|
1926
2057
|
access: limited
|
|
1927
2058
|
num_parameters: 350000000
|
|
1928
2059
|
release_date: 2022-01-27
|
|
1929
|
-
tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
2060
|
+
tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
1930
2061
|
|
|
1931
2062
|
|
|
1932
2063
|
## GPT 3.5 Turbo Models
|
|
@@ -2044,16 +2175,32 @@ models:
|
|
|
2044
2175
|
creator_organization_name: OpenAI
|
|
2045
2176
|
access: limited
|
|
2046
2177
|
release_date: 2024-04-09
|
|
2047
|
-
tags: [TEXT_MODEL_TAG, OPENAI_CHATGPT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
2178
|
+
tags: [TEXT_MODEL_TAG, VISION_LANGUAGE_MODEL_TAG, OPENAI_CHATGPT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
2048
2179
|
|
|
2049
2180
|
- name: openai/gpt-4o-2024-05-13
|
|
2050
2181
|
display_name: GPT-4o (2024-05-13)
|
|
2051
|
-
description: GPT-4o (2024-05-13) is a large multimodal model that accepts as input any combination of text, audio, and image and generates any combination of text, audio, and image outputs.
|
|
2182
|
+
description: GPT-4o (2024-05-13) is a large multimodal model that accepts as input any combination of text, audio, and image and generates any combination of text, audio, and image outputs. ([blog](https://openai.com/index/hello-gpt-4o/))
|
|
2052
2183
|
creator_organization_name: OpenAI
|
|
2053
2184
|
access: limited
|
|
2054
2185
|
release_date: 2024-04-09
|
|
2055
2186
|
tags: [TEXT_MODEL_TAG, VISION_LANGUAGE_MODEL_TAG, OPENAI_CHATGPT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
2056
2187
|
|
|
2188
|
+
- name: openai/gpt-4o-2024-08-06
|
|
2189
|
+
display_name: GPT-4o (2024-08-06)
|
|
2190
|
+
description: GPT-4o (2024-08-06) is a large multimodal model that accepts as input any combination of text, audio, and image and generates any combination of text, audio, and image outputs. ([blog](https://openai.com/index/introducing-structured-outputs-in-the-api/))
|
|
2191
|
+
creator_organization_name: OpenAI
|
|
2192
|
+
access: limited
|
|
2193
|
+
release_date: 2024-08-06
|
|
2194
|
+
tags: [TEXT_MODEL_TAG, VISION_LANGUAGE_MODEL_TAG, OPENAI_CHATGPT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
2195
|
+
|
|
2196
|
+
- name: openai/gpt-4o-mini-2024-07-18
|
|
2197
|
+
display_name: GPT-4o mini (2024-07-18)
|
|
2198
|
+
description: GPT-4o mini (2024-07-18) is a multimodal model with a context window of 128K tokens and improved handling of non-English text. ([blog](https://openai.com/index/gpt-4o-mini-advancing-cost-efficient-intelligence/))
|
|
2199
|
+
creator_organization_name: OpenAI
|
|
2200
|
+
access: limited
|
|
2201
|
+
release_date: 2024-07-18
|
|
2202
|
+
tags: [TEXT_MODEL_TAG, VISION_LANGUAGE_MODEL_TAG, OPENAI_CHATGPT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
2203
|
+
|
|
2057
2204
|
- name: openai/gpt-4-vision-preview
|
|
2058
2205
|
# According to https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4, this model has pointed gpt-4-1106-vision-preview.
|
|
2059
2206
|
display_name: GPT-4V (1106 preview)
|
|
@@ -2074,30 +2221,30 @@ models:
|
|
|
2074
2221
|
## Codex Models
|
|
2075
2222
|
# DEPRECATED: Codex models have been shut down on March 23 2023.
|
|
2076
2223
|
|
|
2077
|
-
- name: openai/code-davinci-002
|
|
2224
|
+
- name: openai/code-davinci-002
|
|
2078
2225
|
display_name: code-davinci-002
|
|
2079
2226
|
description: Codex-style model that is designed for pure code-completion tasks ([docs](https://beta.openai.com/docs/models/codex)).
|
|
2080
2227
|
creator_organization_name: OpenAI
|
|
2081
2228
|
access: limited
|
|
2082
2229
|
release_date: 2021-07-01 # TODO: Find correct date (this is for v1)
|
|
2083
|
-
tags: [CODE_MODEL_TAG]
|
|
2230
|
+
tags: [DEPRECATED_MODEL_TAG, CODE_MODEL_TAG]
|
|
2084
2231
|
|
|
2085
|
-
- name: openai/code-davinci-001
|
|
2232
|
+
- name: openai/code-davinci-001
|
|
2086
2233
|
display_name: code-davinci-001
|
|
2087
2234
|
description: code-davinci-001 model
|
|
2088
2235
|
creator_organization_name: OpenAI
|
|
2089
2236
|
access: limited
|
|
2090
2237
|
release_date: 2021-07-01 # Paper date
|
|
2091
|
-
tags: [CODE_MODEL_TAG]
|
|
2238
|
+
tags: [DEPRECATED_MODEL_TAG, CODE_MODEL_TAG]
|
|
2092
2239
|
|
|
2093
|
-
- name: openai/code-cushman-001
|
|
2240
|
+
- name: openai/code-cushman-001
|
|
2094
2241
|
display_name: code-cushman-001 (12B)
|
|
2095
2242
|
description: Codex-style model that is a stronger, multilingual version of the Codex (12B) model in the [Codex paper](https://arxiv.org/pdf/2107.03374.pdf).
|
|
2096
2243
|
creator_organization_name: OpenAI
|
|
2097
2244
|
access: limited
|
|
2098
2245
|
num_parameters: 12000000000
|
|
2099
2246
|
release_date: 2021-07-01 # Paper date
|
|
2100
|
-
tags: [CODE_MODEL_TAG]
|
|
2247
|
+
tags: [DEPRECATED_MODEL_TAG, CODE_MODEL_TAG]
|
|
2101
2248
|
|
|
2102
2249
|
|
|
2103
2250
|
## Text Similarity Models
|
|
@@ -2107,41 +2254,41 @@ models:
|
|
|
2107
2254
|
# DEPRECATED: Announced on July 06 2023 that first generation embeddings models
|
|
2108
2255
|
# will be shut down on January 04 2024.
|
|
2109
2256
|
|
|
2110
|
-
- name: openai/text-similarity-davinci-001
|
|
2257
|
+
- name: openai/text-similarity-davinci-001
|
|
2111
2258
|
display_name: text-similarity-davinci-001
|
|
2112
2259
|
description: Embedding model that is designed for text similarity tasks ([docs](https://openai.com/blog/introducing-text-and-code-embeddings)).
|
|
2113
2260
|
creator_organization_name: OpenAI
|
|
2114
2261
|
access: limited
|
|
2115
2262
|
num_parameters: 175000000000
|
|
2116
2263
|
release_date: 2022-01-25 # Blog post date
|
|
2117
|
-
tags: [TEXT_SIMILARITY_MODEL_TAG]
|
|
2264
|
+
tags: [DEPRECATED_MODEL_TAG, TEXT_SIMILARITY_MODEL_TAG]
|
|
2118
2265
|
|
|
2119
|
-
- name: openai/text-similarity-curie-001
|
|
2266
|
+
- name: openai/text-similarity-curie-001
|
|
2120
2267
|
display_name: text-similarity-curie-001
|
|
2121
2268
|
description: Embedding model that is designed for text similarity tasks ([docs](https://openai.com/blog/introducing-text-and-code-embeddings)).
|
|
2122
2269
|
creator_organization_name: OpenAI
|
|
2123
2270
|
access: limited
|
|
2124
2271
|
num_parameters: 6700000000
|
|
2125
2272
|
release_date: 2022-01-25 # Blog post date
|
|
2126
|
-
tags: [TEXT_SIMILARITY_MODEL_TAG]
|
|
2273
|
+
tags: [DEPRECATED_MODEL_TAG, TEXT_SIMILARITY_MODEL_TAG]
|
|
2127
2274
|
|
|
2128
|
-
- name: openai/text-similarity-babbage-001
|
|
2275
|
+
- name: openai/text-similarity-babbage-001
|
|
2129
2276
|
display_name: text-similarity-babbage-001
|
|
2130
2277
|
description: Embedding model that is designed for text similarity tasks ([docs](https://openai.com/blog/introducing-text-and-code-embeddings)).
|
|
2131
2278
|
creator_organization_name: OpenAI
|
|
2132
2279
|
access: limited
|
|
2133
2280
|
num_parameters: 1300000000
|
|
2134
2281
|
release_date: 2022-01-25 # Blog post date
|
|
2135
|
-
tags: [TEXT_SIMILARITY_MODEL_TAG]
|
|
2282
|
+
tags: [DEPRECATED_MODEL_TAG, TEXT_SIMILARITY_MODEL_TAG]
|
|
2136
2283
|
|
|
2137
|
-
- name: openai/text-similarity-ada-001
|
|
2284
|
+
- name: openai/text-similarity-ada-001
|
|
2138
2285
|
display_name: text-similarity-ada-001
|
|
2139
2286
|
description: Embedding model that is designed for text similarity tasks ([docs](https://openai.com/blog/introducing-text-and-code-embeddings)).
|
|
2140
2287
|
creator_organization_name: OpenAI
|
|
2141
2288
|
access: limited
|
|
2142
2289
|
num_parameters: 350000000
|
|
2143
2290
|
release_date: 2022-01-25 # Blog post date
|
|
2144
|
-
tags: [TEXT_SIMILARITY_MODEL_TAG]
|
|
2291
|
+
tags: [DEPRECATED_MODEL_TAG, TEXT_SIMILARITY_MODEL_TAG]
|
|
2145
2292
|
|
|
2146
2293
|
- name: openai/text-embedding-ada-002
|
|
2147
2294
|
display_name: text-embedding-ada-002
|
|
@@ -2197,6 +2344,34 @@ models:
|
|
|
2197
2344
|
release_date: 2023-11-06
|
|
2198
2345
|
tags: [TEXT_TO_IMAGE_MODEL_TAG]
|
|
2199
2346
|
|
|
2347
|
+
# OpenThaiGPT
|
|
2348
|
+
- name: openthaigpt/openthaigpt-1.0.0-7b-chat
|
|
2349
|
+
display_name: OpenThaiGPT v1.0.0 (7B)
|
|
2350
|
+
description: OpenThaiGPT v1.0.0 (7B) is a Thai language chat model based on Llama 2 that has been specifically fine-tuned for Thai instructions and enhanced by incorporating over 10,000 of the most commonly used Thai words into the dictionary. ([blog post](https://openthaigpt.aieat.or.th/openthaigpt-1.0.0-less-than-8-apr-2024-greater-than))
|
|
2351
|
+
creator_organization_name: OpenThaiGPT
|
|
2352
|
+
access: open
|
|
2353
|
+
num_parameters: 7000000000
|
|
2354
|
+
release_date: 2024-04-08
|
|
2355
|
+
tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
2356
|
+
|
|
2357
|
+
- name: openthaigpt/openthaigpt-1.0.0-13b-chat
|
|
2358
|
+
display_name: OpenThaiGPT v1.0.0 (13B)
|
|
2359
|
+
description: OpenThaiGPT v1.0.0 (13B) is a Thai language chat model based on Llama 2 that has been specifically fine-tuned for Thai instructions and enhanced by incorporating over 10,000 of the most commonly used Thai words into the dictionary. ([blog post](https://openthaigpt.aieat.or.th/openthaigpt-1.0.0-less-than-8-apr-2024-greater-than))
|
|
2360
|
+
creator_organization_name: OpenThaiGPT
|
|
2361
|
+
access: open
|
|
2362
|
+
num_parameters: 13000000000
|
|
2363
|
+
release_date: 2024-04-08
|
|
2364
|
+
tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
2365
|
+
|
|
2366
|
+
- name: openthaigpt/openthaigpt-1.0.0-70b-chat
|
|
2367
|
+
display_name: OpenThaiGPT v1.0.0 (70B)
|
|
2368
|
+
description: OpenThaiGPT v1.0.0 (70B) is a Thai language chat model based on Llama 2 that has been specifically fine-tuned for Thai instructions and enhanced by incorporating over 10,000 of the most commonly used Thai words into the dictionary. ([blog post](https://openthaigpt.aieat.or.th/openthaigpt-1.0.0-less-than-8-apr-2024-greater-than))
|
|
2369
|
+
creator_organization_name: OpenThaiGPT
|
|
2370
|
+
access: open
|
|
2371
|
+
num_parameters: 70000000000
|
|
2372
|
+
release_date: 2024-04-08
|
|
2373
|
+
tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
2374
|
+
|
|
2200
2375
|
# Qwen
|
|
2201
2376
|
|
|
2202
2377
|
- name: qwen/qwen-7b
|
|
@@ -2311,7 +2486,7 @@ models:
|
|
|
2311
2486
|
access: open
|
|
2312
2487
|
num_parameters: 7000000000
|
|
2313
2488
|
release_date: 2024-04-04
|
|
2314
|
-
tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG
|
|
2489
|
+
tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
2315
2490
|
|
|
2316
2491
|
- name: sail/sailor-7b-chat
|
|
2317
2492
|
display_name: Sailor Chat (7B)
|
|
@@ -2329,7 +2504,7 @@ models:
|
|
|
2329
2504
|
access: open
|
|
2330
2505
|
num_parameters: 14000000000
|
|
2331
2506
|
release_date: 2024-04-04
|
|
2332
|
-
tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG
|
|
2507
|
+
tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
2333
2508
|
|
|
2334
2509
|
- name: sail/sailor-14b-chat
|
|
2335
2510
|
display_name: Sailor Chat (14B)
|
|
@@ -2350,10 +2525,74 @@ models:
|
|
|
2350
2525
|
release_date: 2022-03-25
|
|
2351
2526
|
tags: [] # TODO: add tags
|
|
2352
2527
|
|
|
2528
|
+
# SambaNova
|
|
2529
|
+
- name: sambanova/sambalingo-thai-base
|
|
2530
|
+
display_name: SambaLingo-Thai-Base
|
|
2531
|
+
description: SambaLingo-Thai-Base is a pretrained bi-lingual Thai and English model that adapts Llama 2 (7B) to Thai by training on 38 billion tokens from the Thai split of the Cultura-X dataset. ([paper](https://arxiv.org/abs/2404.05829))
|
|
2532
|
+
creator_organization_name: SambaLingo
|
|
2533
|
+
access: open
|
|
2534
|
+
num_parameters: 7000000000
|
|
2535
|
+
release_date: 2024-04-08
|
|
2536
|
+
tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
2537
|
+
|
|
2538
|
+
- name: sambanova/sambalingo-thai-chat
|
|
2539
|
+
display_name: SambaLingo-Thai-Chat
|
|
2540
|
+
description: SambaLingo-Thai-Chat is a chat model trained using direct preference optimization on SambaLingo-Thai-Base. SambaLingo-Thai-Base adapts Llama 2 (7B) to Thai by training on 38 billion tokens from the Thai split of the Cultura-X dataset. ([paper](https://arxiv.org/abs/2404.05829))
|
|
2541
|
+
creator_organization_name: SambaLingo
|
|
2542
|
+
access: open
|
|
2543
|
+
num_parameters: 7000000000
|
|
2544
|
+
release_date: 2024-04-08
|
|
2545
|
+
tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
2546
|
+
|
|
2547
|
+
- name: sambanova/sambalingo-thai-base-70b
|
|
2548
|
+
display_name: SambaLingo-Thai-Base-70B
|
|
2549
|
+
description: SambaLingo-Thai-Base-70B is a pretrained bi-lingual Thai and English model that adapts Llama 2 (70B) to Thai by training on 26 billion tokens from the Thai split of the Cultura-X dataset. ([paper](https://arxiv.org/abs/2404.05829))
|
|
2550
|
+
creator_organization_name: SambaLingo
|
|
2551
|
+
access: open
|
|
2552
|
+
num_parameters: 70000000000
|
|
2553
|
+
release_date: 2024-04-08
|
|
2554
|
+
tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
2555
|
+
|
|
2556
|
+
- name: sambanova/sambalingo-thai-chat-70b
|
|
2557
|
+
display_name: SambaLingo-Thai-Chat-70B
|
|
2558
|
+
description: SambaLingo-Thai-Chat-70B is a chat model trained using direct preference optimization on SambaLingo-Thai-Base-70B. SambaLingo-Thai-Base-70B adapts Llama 2 (7B) to Thai by training on 26 billion tokens from the Thai split of the Cultura-X dataset. ([paper](https://arxiv.org/abs/2404.05829))
|
|
2559
|
+
creator_organization_name: SambaLingo
|
|
2560
|
+
access: open
|
|
2561
|
+
num_parameters: 70000000000
|
|
2562
|
+
release_date: 2024-04-08
|
|
2563
|
+
tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
2564
|
+
|
|
2353
2565
|
# SCB10X
|
|
2566
|
+
- name: scb10x/typhoon-7b
|
|
2567
|
+
display_name: Typhoon (7B)
|
|
2568
|
+
description: Typhoon (7B) is pretrained Thai large language model with 7 billion parameters based on Mistral 7B. ([paper](https://arxiv.org/abs/2312.13951))
|
|
2569
|
+
creator_organization_name: SCB10X
|
|
2570
|
+
access: open
|
|
2571
|
+
num_parameters: 7000000000
|
|
2572
|
+
release_date: 2023-12-21
|
|
2573
|
+
tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
2574
|
+
|
|
2575
|
+
- name: scb10x/typhoon-v1.5-8b
|
|
2576
|
+
display_name: Typhoon v1.5 (8B)
|
|
2577
|
+
description: Typhoon v1.5 (8B) is a pretrained Thai large language model with 8 billion parameters based on Llama 3 8B. ([blog](https://blog.opentyphoon.ai/typhoon-1-5-release-a9364cb8e8d7))
|
|
2578
|
+
creator_organization_name: SCB10X
|
|
2579
|
+
access: open
|
|
2580
|
+
num_parameters: 8000000000
|
|
2581
|
+
release_date: 2024-05-08
|
|
2582
|
+
tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
2583
|
+
|
|
2584
|
+
- name: scb10x/typhoon-v1.5-8b-instruct
|
|
2585
|
+
display_name: Typhoon v1.5 Instruct (8B)
|
|
2586
|
+
description: Typhoon v1.5 Instruct (8B) is a pretrained Thai large language model with 8 billion parameters based on Llama 3 8B. ([blog](https://blog.opentyphoon.ai/typhoon-1-5-release-a9364cb8e8d7))
|
|
2587
|
+
creator_organization_name: SCB10X
|
|
2588
|
+
access: open
|
|
2589
|
+
num_parameters: 8000000000
|
|
2590
|
+
release_date: 2024-05-08
|
|
2591
|
+
tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
2592
|
+
|
|
2354
2593
|
- name: scb10x/typhoon-v1.5-72b
|
|
2355
2594
|
display_name: Typhoon v1.5 (72B)
|
|
2356
|
-
description: Typhoon v1.5 (72B) is pretrained Thai large language model with 72 billion parameters based on Qwen1.5-72B. ([blog](https://blog.opentyphoon.ai/typhoon-1-5-release-a9364cb8e8d7))
|
|
2595
|
+
description: Typhoon v1.5 (72B) is a pretrained Thai large language model with 72 billion parameters based on Qwen1.5-72B. ([blog](https://blog.opentyphoon.ai/typhoon-1-5-release-a9364cb8e8d7))
|
|
2357
2596
|
creator_organization_name: SCB10X
|
|
2358
2597
|
access: open
|
|
2359
2598
|
num_parameters: 72000000000
|
|
@@ -2362,13 +2601,50 @@ models:
|
|
|
2362
2601
|
|
|
2363
2602
|
- name: scb10x/typhoon-v1.5-72b-instruct
|
|
2364
2603
|
display_name: Typhoon v1.5 Instruct (72B)
|
|
2365
|
-
description: Typhoon v1.5 Instruct (72B) is pretrained Thai large language model with 72 billion parameters based on Qwen1.5-72B. ([blog](https://blog.opentyphoon.ai/typhoon-1-5-release-a9364cb8e8d7))
|
|
2604
|
+
description: Typhoon v1.5 Instruct (72B) is a pretrained Thai large language model with 72 billion parameters based on Qwen1.5-72B. ([blog](https://blog.opentyphoon.ai/typhoon-1-5-release-a9364cb8e8d7))
|
|
2366
2605
|
creator_organization_name: SCB10X
|
|
2367
2606
|
access: open
|
|
2368
2607
|
num_parameters: 72000000000
|
|
2369
2608
|
release_date: 2024-05-08
|
|
2370
2609
|
tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
2371
2610
|
|
|
2611
|
+
- name: scb10x/llama-3-typhoon-v1.5x-8b-instruct
|
|
2612
|
+
display_name: Typhoon 1.5X instruct (8B)
|
|
2613
|
+
description: Llama-3-Typhoon-1.5X-8B-instruct is a 8 billion parameter instruct model designed for the Thai language based on Llama 3 Instruct. It utilizes the task-arithmetic model editing technique. ([blog](https://blog.opentyphoon.ai/typhoon-1-5x-our-experiment-designed-for-application-use-cases-7b85d9e9845c))
|
|
2614
|
+
creator_organization_name: SCB10X
|
|
2615
|
+
access: open
|
|
2616
|
+
num_parameters: 8000000000
|
|
2617
|
+
release_date: 2024-05-29
|
|
2618
|
+
tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
2619
|
+
|
|
2620
|
+
- name: scb10x/llama-3-typhoon-v1.5x-70b-instruct
|
|
2621
|
+
display_name: Typhoon 1.5X instruct (70B)
|
|
2622
|
+
description: Llama-3-Typhoon-1.5X-70B-instruct is a 70 billion parameter instruct model designed for the Thai language based on Llama 3 Instruct. It utilizes the task-arithmetic model editing technique. ([blog](https://blog.opentyphoon.ai/typhoon-1-5x-our-experiment-designed-for-application-use-cases-7b85d9e9845c))
|
|
2623
|
+
creator_organization_name: SCB10X
|
|
2624
|
+
access: open
|
|
2625
|
+
num_parameters: 70000000000
|
|
2626
|
+
release_date: 2024-05-29
|
|
2627
|
+
tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
|
|
2628
|
+
|
|
2629
|
+
# Alibaba DAMO Academy
|
|
2630
|
+
- name: damo/seallm-7b-v2
|
|
2631
|
+
display_name: SeaLLM v2 (7B)
|
|
2632
|
+
description: SeaLLM v2 is a multilingual LLM for Southeast Asian (SEA) languages trained from Mistral (7B). ([website](https://damo-nlp-sg.github.io/SeaLLMs/))
|
|
2633
|
+
creator_organization_name: Alibaba DAMO Academy
|
|
2634
|
+
access: open
|
|
2635
|
+
num_parameters: 7000000000
|
|
2636
|
+
release_date: 2024-02-02
|
|
2637
|
+
tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
2638
|
+
|
|
2639
|
+
- name: damo/seallm-7b-v2.5
|
|
2640
|
+
display_name: SeaLLM v2.5 (7B)
|
|
2641
|
+
description: SeaLLM is a multilingual LLM for Southeast Asian (SEA) languages trained from Gemma (7B). ([website](https://damo-nlp-sg.github.io/SeaLLMs/))
|
|
2642
|
+
creator_organization_name: Alibaba DAMO Academy
|
|
2643
|
+
access: open
|
|
2644
|
+
num_parameters: 7000000000
|
|
2645
|
+
release_date: 2024-04-12
|
|
2646
|
+
tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
2647
|
+
|
|
2372
2648
|
# Snowflake
|
|
2373
2649
|
- name: snowflake/snowflake-arctic-instruct
|
|
2374
2650
|
display_name: Arctic Instruct
|
|
@@ -2538,7 +2814,7 @@ models:
|
|
|
2538
2814
|
release_date: 2022-08-04
|
|
2539
2815
|
# Inference with echo=True is not feasible -- in the prompt encoding phase, they use
|
|
2540
2816
|
# bidirectional attention and do not perform predictions on them.
|
|
2541
|
-
tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, ABLATION_MODEL_TAG, NO_NEWLINES_TAG]
|
|
2817
|
+
tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, ABLATION_MODEL_TAG, NO_NEWLINES_TAG]
|
|
2542
2818
|
|
|
2543
2819
|
- name: tsinghua/codegeex # NOT SUPPORTED
|
|
2544
2820
|
display_name: CodeGeeX (13B)
|
|
@@ -2573,7 +2849,6 @@ models:
|
|
|
2573
2849
|
tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
2574
2850
|
|
|
2575
2851
|
- name: writer/palmyra-instruct-30
|
|
2576
|
-
deprecated: true # Internal error
|
|
2577
2852
|
display_name: InstructPalmyra (30B)
|
|
2578
2853
|
description: InstructPalmyra (30B parameters) is trained using reinforcement learning techniques based on feedback from humans.
|
|
2579
2854
|
creator_organization_name: Writer
|
|
@@ -2581,10 +2856,9 @@ models:
|
|
|
2581
2856
|
num_parameters: 30000000000
|
|
2582
2857
|
release_date: 2023-02-16
|
|
2583
2858
|
# Does not support echo
|
|
2584
|
-
tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
2859
|
+
tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
2585
2860
|
|
|
2586
2861
|
- name: writer/palmyra-e
|
|
2587
|
-
deprecated: true # Internal error
|
|
2588
2862
|
display_name: Palmyra E (30B)
|
|
2589
2863
|
description: Palmyra E (30B)
|
|
2590
2864
|
creator_organization_name: Writer
|
|
@@ -2592,7 +2866,7 @@ models:
|
|
|
2592
2866
|
num_parameters: 30000000000
|
|
2593
2867
|
release_date: 2023-03-03
|
|
2594
2868
|
# Does not support echo
|
|
2595
|
-
tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
2869
|
+
tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG]
|
|
2596
2870
|
|
|
2597
2871
|
- name: writer/silk-road
|
|
2598
2872
|
display_name: Silk Road (35B)
|