crfm-helm 0.5.2__py3-none-any.whl → 0.5.3__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of crfm-helm might be problematic. Click here for more details.

Files changed (184) hide show
  1. {crfm_helm-0.5.2.dist-info → crfm_helm-0.5.3.dist-info}/METADATA +29 -55
  2. {crfm_helm-0.5.2.dist-info → crfm_helm-0.5.3.dist-info}/RECORD +146 -134
  3. {crfm_helm-0.5.2.dist-info → crfm_helm-0.5.3.dist-info}/WHEEL +1 -1
  4. helm/benchmark/adaptation/adapters/multiple_choice_joint_adapter.py +12 -5
  5. helm/benchmark/adaptation/adapters/test_generation_adapter.py +12 -12
  6. helm/benchmark/adaptation/adapters/test_language_modeling_adapter.py +8 -8
  7. helm/benchmark/adaptation/adapters/test_multiple_choice_joint_adapter.py +77 -9
  8. helm/benchmark/adaptation/common_adapter_specs.py +2 -0
  9. helm/benchmark/annotation/anthropic_red_team_annotator.py +70 -0
  10. helm/benchmark/annotation/call_center_annotator.py +247 -0
  11. helm/benchmark/annotation/financebench_annotator.py +79 -0
  12. helm/benchmark/annotation/harm_bench_annotator.py +68 -0
  13. helm/benchmark/annotation/{image2structure → image2struct}/latex_compiler_annotator.py +2 -2
  14. helm/benchmark/annotation/{image2structure → image2struct}/lilypond_compiler_annotator.py +5 -3
  15. helm/benchmark/annotation/{image2structure → image2struct}/webpage_compiler_annotator.py +5 -5
  16. helm/benchmark/annotation/live_qa_annotator.py +32 -45
  17. helm/benchmark/annotation/medication_qa_annotator.py +31 -44
  18. helm/benchmark/annotation/model_as_judge.py +45 -0
  19. helm/benchmark/annotation/simple_safety_tests_annotator.py +64 -0
  20. helm/benchmark/annotation/xstest_annotator.py +110 -0
  21. helm/benchmark/metrics/annotation_metrics.py +108 -0
  22. helm/benchmark/metrics/bhasa_metrics.py +188 -0
  23. helm/benchmark/metrics/bhasa_metrics_specs.py +10 -0
  24. helm/benchmark/metrics/code_metrics_helper.py +11 -1
  25. helm/benchmark/metrics/safety_metrics.py +57 -0
  26. helm/benchmark/metrics/summac/model_summac.py +3 -3
  27. helm/benchmark/metrics/tokens/test_ai21_token_cost_estimator.py +2 -2
  28. helm/benchmark/metrics/tokens/test_openai_token_cost_estimator.py +4 -4
  29. helm/benchmark/metrics/vision_language/image_metrics.py +1 -1
  30. helm/benchmark/metrics/vision_language/image_utils.py +1 -1
  31. helm/benchmark/model_metadata_registry.py +3 -3
  32. helm/benchmark/presentation/test_run_entry.py +1 -0
  33. helm/benchmark/run.py +15 -0
  34. helm/benchmark/run_expander.py +56 -30
  35. helm/benchmark/run_specs/bhasa_run_specs.py +638 -0
  36. helm/benchmark/run_specs/call_center_run_specs.py +152 -0
  37. helm/benchmark/run_specs/decodingtrust_run_specs.py +8 -8
  38. helm/benchmark/run_specs/experimental_run_specs.py +52 -0
  39. helm/benchmark/run_specs/finance_run_specs.py +78 -1
  40. helm/benchmark/run_specs/safety_run_specs.py +154 -0
  41. helm/benchmark/run_specs/vlm_run_specs.py +92 -21
  42. helm/benchmark/scenarios/anthropic_red_team_scenario.py +71 -0
  43. helm/benchmark/scenarios/banking77_scenario.py +51 -0
  44. helm/benchmark/scenarios/bhasa_scenario.py +1798 -0
  45. helm/benchmark/scenarios/call_center_scenario.py +84 -0
  46. helm/benchmark/scenarios/decodingtrust_stereotype_bias_scenario.py +2 -1
  47. helm/benchmark/scenarios/ewok_scenario.py +116 -0
  48. helm/benchmark/scenarios/fin_qa_scenario.py +2 -0
  49. helm/benchmark/scenarios/financebench_scenario.py +53 -0
  50. helm/benchmark/scenarios/harm_bench_scenario.py +59 -0
  51. helm/benchmark/scenarios/scenario.py +1 -1
  52. helm/benchmark/scenarios/simple_safety_tests_scenario.py +33 -0
  53. helm/benchmark/scenarios/test_commonsense_scenario.py +21 -0
  54. helm/benchmark/scenarios/test_ewok_scenario.py +25 -0
  55. helm/benchmark/scenarios/test_financebench_scenario.py +26 -0
  56. helm/benchmark/scenarios/test_gsm_scenario.py +31 -0
  57. helm/benchmark/scenarios/test_legalbench_scenario.py +30 -0
  58. helm/benchmark/scenarios/test_math_scenario.py +2 -8
  59. helm/benchmark/scenarios/test_med_qa_scenario.py +30 -0
  60. helm/benchmark/scenarios/test_mmlu_scenario.py +33 -0
  61. helm/benchmark/scenarios/test_narrativeqa_scenario.py +73 -0
  62. helm/benchmark/scenarios/thai_exam_scenario.py +4 -4
  63. helm/benchmark/scenarios/vision_language/a_okvqa_scenario.py +1 -1
  64. helm/benchmark/scenarios/vision_language/bingo_scenario.py +2 -2
  65. helm/benchmark/scenarios/vision_language/crossmodal_3600_scenario.py +2 -1
  66. helm/benchmark/scenarios/vision_language/exams_v_scenario.py +104 -0
  67. helm/benchmark/scenarios/vision_language/fair_face_scenario.py +136 -0
  68. helm/benchmark/scenarios/vision_language/flickr30k_scenario.py +1 -1
  69. helm/benchmark/scenarios/vision_language/gqa_scenario.py +2 -2
  70. helm/benchmark/scenarios/vision_language/hateful_memes_scenario.py +1 -1
  71. helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/chart2csv_scenario.py +1 -1
  72. helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/latex_scenario.py +3 -3
  73. helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/musicsheet_scenario.py +1 -1
  74. helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/utils_latex.py +31 -39
  75. helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/webpage/driver.py +1 -1
  76. helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/webpage/utils.py +1 -1
  77. helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/webpage_scenario.py +41 -12
  78. helm/benchmark/scenarios/vision_language/math_vista_scenario.py +1 -1
  79. helm/benchmark/scenarios/vision_language/mementos_scenario.py +3 -3
  80. helm/benchmark/scenarios/vision_language/mm_safety_bench_scenario.py +2 -2
  81. helm/benchmark/scenarios/vision_language/mme_scenario.py +21 -18
  82. helm/benchmark/scenarios/vision_language/mmmu_scenario.py +1 -1
  83. helm/benchmark/scenarios/vision_language/pairs_scenario.py +1 -1
  84. helm/benchmark/scenarios/vision_language/pope_scenario.py +2 -1
  85. helm/benchmark/scenarios/vision_language/real_world_qa_scenario.py +57 -0
  86. helm/benchmark/scenarios/vision_language/seed_bench_scenario.py +7 -5
  87. helm/benchmark/scenarios/vision_language/unicorn_scenario.py +2 -2
  88. helm/benchmark/scenarios/vision_language/vibe_eval_scenario.py +6 -3
  89. helm/benchmark/scenarios/vision_language/viz_wiz_scenario.py +1 -1
  90. helm/benchmark/scenarios/vision_language/vqa_scenario.py +3 -1
  91. helm/benchmark/scenarios/xstest_scenario.py +35 -0
  92. helm/benchmark/server.py +1 -6
  93. helm/benchmark/static/schema_air_bench.yaml +750 -750
  94. helm/benchmark/static/schema_bhasa.yaml +709 -0
  95. helm/benchmark/static/schema_call_center.yaml +232 -0
  96. helm/benchmark/static/schema_cleva.yaml +768 -0
  97. helm/benchmark/static/schema_decodingtrust.yaml +444 -0
  98. helm/benchmark/static/schema_ewok.yaml +367 -0
  99. helm/benchmark/static/schema_finance.yaml +55 -9
  100. helm/benchmark/static/{schema_image2structure.yaml → schema_image2struct.yaml} +231 -90
  101. helm/benchmark/static/schema_safety.yaml +247 -0
  102. helm/benchmark/static/schema_tables.yaml +124 -7
  103. helm/benchmark/static/schema_thai.yaml +21 -0
  104. helm/benchmark/static/schema_vhelm.yaml +96 -91
  105. helm/benchmark/static_build/assets/accenture-6f97eeda.png +0 -0
  106. helm/benchmark/static_build/assets/aisingapore-6dfc9acf.png +0 -0
  107. helm/benchmark/static_build/assets/cresta-9e22b983.png +0 -0
  108. helm/benchmark/static_build/assets/cuhk-8c5631e9.png +0 -0
  109. helm/benchmark/static_build/assets/index-05c76bb1.css +1 -0
  110. helm/benchmark/static_build/assets/index-58f97dcd.js +10 -0
  111. helm/benchmark/static_build/assets/scb10x-204bd786.png +0 -0
  112. helm/benchmark/static_build/assets/wellsfargo-a86a6c4a.png +0 -0
  113. helm/benchmark/static_build/index.html +2 -2
  114. helm/benchmark/window_services/test_openai_window_service.py +8 -8
  115. helm/clients/ai21_client.py +71 -1
  116. helm/clients/anthropic_client.py +7 -19
  117. helm/clients/huggingface_client.py +38 -37
  118. helm/clients/nvidia_nim_client.py +35 -0
  119. helm/clients/openai_client.py +2 -3
  120. helm/clients/palmyra_client.py +25 -0
  121. helm/clients/perspective_api_client.py +11 -6
  122. helm/clients/test_client.py +4 -6
  123. helm/clients/vision_language/open_flamingo_client.py +1 -2
  124. helm/clients/vision_language/palmyra_vision_client.py +28 -13
  125. helm/common/images_utils.py +6 -0
  126. helm/common/mongo_key_value_store.py +2 -1
  127. helm/common/request.py +16 -0
  128. helm/config/model_deployments.yaml +315 -332
  129. helm/config/model_metadata.yaml +384 -110
  130. helm/config/tokenizer_configs.yaml +116 -11
  131. helm/proxy/example_queries.py +14 -21
  132. helm/proxy/services/server_service.py +1 -2
  133. helm/proxy/token_counters/test_auto_token_counter.py +2 -2
  134. helm/tokenizers/ai21_tokenizer.py +51 -59
  135. helm/tokenizers/cohere_tokenizer.py +0 -75
  136. helm/tokenizers/huggingface_tokenizer.py +0 -1
  137. helm/tokenizers/test_ai21_tokenizer.py +48 -0
  138. helm/benchmark/static/benchmarking.css +0 -156
  139. helm/benchmark/static/benchmarking.js +0 -1705
  140. helm/benchmark/static/config.js +0 -3
  141. helm/benchmark/static/general.js +0 -122
  142. helm/benchmark/static/images/crfm-logo.png +0 -0
  143. helm/benchmark/static/images/helm-logo-simple.png +0 -0
  144. helm/benchmark/static/images/helm-logo.png +0 -0
  145. helm/benchmark/static/images/language-model-helm.png +0 -0
  146. helm/benchmark/static/images/organizations/ai21.png +0 -0
  147. helm/benchmark/static/images/organizations/anthropic.png +0 -0
  148. helm/benchmark/static/images/organizations/bigscience.png +0 -0
  149. helm/benchmark/static/images/organizations/cohere.png +0 -0
  150. helm/benchmark/static/images/organizations/eleutherai.png +0 -0
  151. helm/benchmark/static/images/organizations/google.png +0 -0
  152. helm/benchmark/static/images/organizations/meta.png +0 -0
  153. helm/benchmark/static/images/organizations/microsoft.png +0 -0
  154. helm/benchmark/static/images/organizations/nvidia.png +0 -0
  155. helm/benchmark/static/images/organizations/openai.png +0 -0
  156. helm/benchmark/static/images/organizations/together.png +0 -0
  157. helm/benchmark/static/images/organizations/tsinghua-keg.png +0 -0
  158. helm/benchmark/static/images/organizations/yandex.png +0 -0
  159. helm/benchmark/static/images/scenarios-by-metrics.png +0 -0
  160. helm/benchmark/static/images/taxonomy-scenarios.png +0 -0
  161. helm/benchmark/static/index.html +0 -68
  162. helm/benchmark/static/info-icon.png +0 -0
  163. helm/benchmark/static/json-urls.js +0 -69
  164. helm/benchmark/static/plot-captions.js +0 -27
  165. helm/benchmark/static/utils.js +0 -285
  166. helm/benchmark/static_build/assets/index-30dbceba.js +0 -10
  167. helm/benchmark/static_build/assets/index-66b02d40.css +0 -1
  168. helm/benchmark/window_services/ai21_window_service.py +0 -247
  169. helm/benchmark/window_services/cohere_window_service.py +0 -101
  170. helm/benchmark/window_services/test_ai21_window_service.py +0 -163
  171. helm/benchmark/window_services/test_cohere_window_service.py +0 -75
  172. helm/benchmark/window_services/test_cohere_window_service_utils.py +0 -8328
  173. helm/benchmark/window_services/test_ice_window_service.py +0 -327
  174. helm/tokenizers/ice_tokenizer.py +0 -30
  175. helm/tokenizers/test_ice_tokenizer.py +0 -57
  176. {crfm_helm-0.5.2.dist-info → crfm_helm-0.5.3.dist-info}/LICENSE +0 -0
  177. {crfm_helm-0.5.2.dist-info → crfm_helm-0.5.3.dist-info}/entry_points.txt +0 -0
  178. {crfm_helm-0.5.2.dist-info → crfm_helm-0.5.3.dist-info}/top_level.txt +0 -0
  179. /helm/benchmark/annotation/{image2structure → image2struct}/__init__.py +0 -0
  180. /helm/benchmark/annotation/{image2structure → image2struct}/image_compiler_annotator.py +0 -0
  181. /helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/__init__.py +0 -0
  182. /helm/benchmark/scenarios/vision_language/{image2structure/image2structure_scenario.py → image2struct/image2struct_scenario.py} +0 -0
  183. /helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/webpage/__init__.py +0 -0
  184. /helm/benchmark/scenarios/vision_language/{image2structure → image2struct}/webpage/jekyll_server.py +0 -0
@@ -31,50 +31,41 @@ models:
31
31
 
32
32
 
33
33
  # AI21 Labs
34
- - name: ai21/j1-jumbo # DEPRECATED
34
+ - name: ai21/j1-jumbo
35
35
  display_name: J1-Jumbo v1 (178B)
36
36
  description: Jurassic-1 Jumbo (178B parameters) ([docs](https://studio.ai21.com/docs/jurassic1-language-models/), [tech report](https://uploads-ssl.webflow.com/60fd4503684b466578c0d307/61138924626a6981ee09caf6_jurassic_tech_paper.pdf)).
37
37
  creator_organization_name: AI21 Labs
38
38
  access: limited
39
39
  num_parameters: 178000000000
40
40
  release_date: 2021-08-11
41
- tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
41
+ tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
42
42
 
43
- - name: ai21/j1-large # DEPRECATED
43
+ - name: ai21/j1-large
44
44
  display_name: J1-Large v1 (7.5B)
45
45
  description: Jurassic-1 Large (7.5B parameters) ([docs](https://studio.ai21.com/docs/jurassic1-language-models/), [tech report](https://uploads-ssl.webflow.com/60fd4503684b466578c0d307/61138924626a6981ee09caf6_jurassic_tech_paper.pdf)).
46
46
  creator_organization_name: AI21 Labs
47
47
  access: limited
48
48
  num_parameters: 7500000000
49
49
  release_date: 2021-08-11
50
- tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
50
+ tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
51
51
 
52
- - name: ai21/j1-grande # DEPRECATED
52
+ - name: ai21/j1-grande
53
53
  display_name: J1-Grande v1 (17B)
54
54
  description: Jurassic-1 Grande (17B parameters) with a "few tweaks" to the training process ([docs](https://studio.ai21.com/docs/jurassic1-language-models/), [tech report](https://uploads-ssl.webflow.com/60fd4503684b466578c0d307/61138924626a6981ee09caf6_jurassic_tech_paper.pdf)).
55
55
  creator_organization_name: AI21 Labs
56
56
  access: limited
57
57
  num_parameters: 17000000000
58
58
  release_date: 2022-05-03
59
- tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
59
+ tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
60
60
 
61
- - name: ai21/j1-grande-v2-beta # DEPRECATED
61
+ - name: ai21/j1-grande-v2-beta
62
62
  display_name: J1-Grande v2 beta (17B)
63
63
  description: Jurassic-1 Grande v2 beta (17B parameters)
64
64
  creator_organization_name: AI21 Labs
65
65
  access: limited
66
66
  num_parameters: 17000000000
67
67
  release_date: 2022-10-28
68
- tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
69
-
70
- - name: ai21/j2-jumbo
71
- display_name: Jurassic-2 Jumbo (178B)
72
- description: Jurassic-2 Jumbo (178B parameters) ([docs](https://www.ai21.com/blog/introducing-j2))
73
- creator_organization_name: AI21 Labs
74
- access: limited
75
- num_parameters: 178000000000
76
- release_date: 2023-03-09
77
- tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
68
+ tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
78
69
 
79
70
  - name: ai21/j2-large
80
71
  display_name: Jurassic-2 Large (7.5B)
@@ -83,7 +74,7 @@ models:
83
74
  access: limited
84
75
  num_parameters: 7500000000
85
76
  release_date: 2023-03-09
86
- tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
77
+ tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
87
78
 
88
79
  - name: ai21/j2-grande
89
80
  display_name: Jurassic-2 Grande (17B)
@@ -92,13 +83,48 @@ models:
92
83
  access: limited
93
84
  num_parameters: 17000000000
94
85
  release_date: 2023-03-09
95
- tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
86
+ tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
87
+
88
+ - name: ai21/j2-jumbo
89
+ display_name: Jurassic-2 Jumbo (178B)
90
+ description: Jurassic-2 Jumbo (178B parameters) ([docs](https://www.ai21.com/blog/introducing-j2))
91
+ creator_organization_name: AI21 Labs
92
+ access: limited
93
+ num_parameters: 178000000000
94
+ release_date: 2023-03-09
95
+ tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
96
96
 
97
97
  # TODO(1524): Change AI21 model names
98
98
  # - j2-jumbo -> j2-ultra
99
99
  # - j2-grande -> j2-mid
100
100
  # - j2-large -> j2-light
101
101
 
102
+ - name: ai21/jamba-instruct
103
+ display_name: Jamba Instruct
104
+ description: Jamba Instruct is an instruction tuned version of Jamba, which uses a hybrid Transformer-Mamba mixture-of-experts (MoE) architecture that interleaves blocks of Transformer and Mamba layers. ([blog](https://www.ai21.com/blog/announcing-jamba-instruct))
105
+ creator_organization_name: AI21 Labs
106
+ access: limited
107
+ num_parameters: 52000000000
108
+ release_date: 2024-05-02
109
+ tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
110
+
111
+ - name: ai21/jamba-1.5-mini
112
+ display_name: Jamba 1.5 Mini
113
+ description: Jamba 1.5 Mini is a long-context, hybrid SSM-Transformer instruction following foundation model that is optimized for function calling, structured output, and grounded generation. ([blog](https://www.ai21.com/blog/announcing-jamba-model-family))
114
+ creator_organization_name: AI21 Labs
115
+ access: open
116
+ num_parameters: 51600000000
117
+ release_date: 2024-08-22
118
+ tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
119
+
120
+ - name: ai21/jamba-1.5-large
121
+ display_name: Jamba 1.5 Large
122
+ description: Jamba 1.5 Large is a long-context, hybrid SSM-Transformer instruction following foundation model that is optimized for function calling, structured output, and grounded generation. ([blog](https://www.ai21.com/blog/announcing-jamba-model-family))
123
+ creator_organization_name: AI21 Labs
124
+ access: open
125
+ num_parameters: 399000000000
126
+ release_date: 2024-08-22
127
+ tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
102
128
 
103
129
  # AI Singapore
104
130
  - name: aisingapore/sea-lion-7b
@@ -117,7 +143,7 @@ models:
117
143
  access: open
118
144
  num_parameters: 7000000000
119
145
  release_date: 2023-02-24
120
- tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
146
+ tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
121
147
 
122
148
 
123
149
  # Aleph Alpha
@@ -272,7 +298,14 @@ models:
272
298
  release_date: 2024-03-04 # https://www.anthropic.com/news/claude-3-family
273
299
  tags: [ANTHROPIC_CLAUDE_3_MODEL_TAG, TEXT_MODEL_TAG, VISION_LANGUAGE_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
274
300
 
275
- # DEPRECATED: Please do not use.
301
+ - name: anthropic/claude-3-5-sonnet-20240620
302
+ display_name: Claude 3.5 Sonnet (20240620)
303
+ description: Claude 3.5 Sonnet is a Claude 3 family model which outperforms Claude 3 Opus while operating faster and at a lower cost. ([blog](https://www.anthropic.com/news/claude-3-5-sonnet))
304
+ creator_organization_name: Anthropic
305
+ access: limited
306
+ release_date: 2024-06-20
307
+ tags: [ANTHROPIC_CLAUDE_3_MODEL_TAG, TEXT_MODEL_TAG, VISION_LANGUAGE_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
308
+
276
309
  - name: anthropic/stanford-online-all-v4-s3
277
310
  display_name: Anthropic-LM v4-s3 (52B)
278
311
  description: A 52B parameter language model, trained using reinforcement learning from human feedback [paper](https://arxiv.org/pdf/2204.05862.pdf).
@@ -280,7 +313,7 @@ models:
280
313
  access: closed
281
314
  num_parameters: 52000000000
282
315
  release_date: 2021-12-01
283
- tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG, ABLATION_MODEL_TAG]
316
+ tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG, ABLATION_MODEL_TAG]
284
317
 
285
318
 
286
319
 
@@ -401,16 +434,16 @@ models:
401
434
  access: limited
402
435
  num_parameters: 52400000000
403
436
  release_date: 2022-06-09
404
- tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
437
+ tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
405
438
 
406
- - name: cohere/large-20220720 # DEPRECATED
439
+ - name: cohere/large-20220720
407
440
  display_name: Cohere large v20220720 (13.1B)
408
441
  description: Cohere large v20220720 (13.1B parameters), which is deprecated by Cohere as of December 2, 2022.
409
442
  creator_organization_name: Cohere
410
443
  access: limited
411
444
  num_parameters: 13100000000
412
445
  release_date: 2022-07-20
413
- tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
446
+ tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
414
447
 
415
448
  - name: cohere/medium-20220720
416
449
  display_name: Cohere medium v20220720 (6.1B)
@@ -419,16 +452,16 @@ models:
419
452
  access: limited
420
453
  num_parameters: 6100000000
421
454
  release_date: 2022-07-20
422
- tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
455
+ tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
423
456
 
424
- - name: cohere/small-20220720 # DEPRECATED
457
+ - name: cohere/small-20220720
425
458
  display_name: Cohere small v20220720 (410M)
426
459
  description: Cohere small v20220720 (410M parameters), which is deprecated by Cohere as of December 2, 2022.
427
460
  creator_organization_name: Cohere
428
461
  access: limited
429
462
  num_parameters: 410000000
430
463
  release_date: 2022-07-20
431
- tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
464
+ tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
432
465
 
433
466
  - name: cohere/xlarge-20221108
434
467
  display_name: Cohere xlarge v20221108 (52.4B)
@@ -437,34 +470,34 @@ models:
437
470
  access: limited
438
471
  num_parameters: 52400000000
439
472
  release_date: 2022-11-08
440
- tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
473
+ tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
441
474
 
442
- - name: cohere/medium-20221108 # DEPRECATED
475
+ - name: cohere/medium-20221108
443
476
  display_name: Cohere medium v20221108 (6.1B)
444
477
  description: Cohere medium v20221108 (6.1B parameters)
445
478
  creator_organization_name: Cohere
446
479
  access: limited
447
480
  num_parameters: 6100000000
448
481
  release_date: 2022-11-08
449
- tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
482
+ tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
450
483
 
451
- - name: cohere/command-medium-beta # DEPRECATED
484
+ - name: cohere/command-medium-beta
452
485
  display_name: Command beta (6.1B)
453
- description: Cohere Command beta (6.1B parameters) is fine-tuned from the medium model to respond well with instruction-like prompts ([details](https://docs.cohere.ai/docs/command-beta)).
486
+ description: Command beta (6.1B parameters) is fine-tuned from the medium model to respond well with instruction-like prompts ([details](https://docs.cohere.ai/docs/command-beta)).
454
487
  creator_organization_name: Cohere
455
488
  access: limited
456
489
  num_parameters: 6100000000
457
490
  release_date: 2022-11-08
458
- tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
491
+ tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
459
492
 
460
- - name: cohere/command-xlarge-beta # DEPRECATED
493
+ - name: cohere/command-xlarge-beta
461
494
  display_name: Command beta (52.4B)
462
- description: Cohere Command beta (52.4B parameters) is fine-tuned from the XL model to respond well with instruction-like prompts ([details](https://docs.cohere.ai/docs/command-beta)).
495
+ description: Command beta (52.4B parameters) is fine-tuned from the XL model to respond well with instruction-like prompts ([details](https://docs.cohere.ai/docs/command-beta)).
463
496
  creator_organization_name: Cohere
464
497
  access: limited
465
498
  num_parameters: 52400000000
466
499
  release_date: 2022-11-08
467
- tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
500
+ tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
468
501
 
469
502
  - name: cohere/command
470
503
  display_name: Command
@@ -837,8 +870,7 @@ models:
837
870
 
838
871
  - name: google/gemma-2b
839
872
  display_name: Gemma (2B)
840
- # TODO: Fill in Gemma description.
841
- description: TBD
873
+ description: Gemma is a family of lightweight, open models built from the research and technology that Google used to create the Gemini models. ([model card](https://www.kaggle.com/models/google/gemma), [blog post](https://blog.google/technology/developers/gemma-open-models/))
842
874
  creator_organization_name: Google
843
875
  access: open
844
876
  release_date: 2024-02-21
@@ -846,8 +878,7 @@ models:
846
878
 
847
879
  - name: google/gemma-2b-it
848
880
  display_name: Gemma Instruct (2B)
849
- # TODO: Fill in Gemma description.
850
- description: TBD
881
+ description: Gemma is a family of lightweight, open models built from the research and technology that Google used to create the Gemini models. ([model card](https://www.kaggle.com/models/google/gemma), [blog post](https://blog.google/technology/developers/gemma-open-models/))
851
882
  creator_organization_name: Google
852
883
  access: open
853
884
  release_date: 2024-02-21
@@ -855,8 +886,7 @@ models:
855
886
 
856
887
  - name: google/gemma-7b
857
888
  display_name: Gemma (7B)
858
- # TODO: Fill in Gemma description.
859
- description: TBD
889
+ description: Gemma is a family of lightweight, open models built from the research and technology that Google used to create the Gemini models. ([model card](https://www.kaggle.com/models/google/gemma), [blog post](https://blog.google/technology/developers/gemma-open-models/))
860
890
  creator_organization_name: Google
861
891
  access: open
862
892
  release_date: 2024-02-21
@@ -864,12 +894,42 @@ models:
864
894
 
865
895
  - name: google/gemma-7b-it
866
896
  display_name: Gemma Instruct (7B)
867
- # TODO: Fill in Gemma description.
868
- description: TBD
897
+ description: Gemma is a family of lightweight, open models built from the research and technology that Google used to create the Gemini models. ([model card](https://www.kaggle.com/models/google/gemma), [blog post](https://blog.google/technology/developers/gemma-open-models/))
869
898
  creator_organization_name: Google
870
899
  access: open
871
900
  release_date: 2024-02-21
872
- # TODO: Add OUTPUT_FORMAT_INSTRUCTIONS_TAG tag
901
+ tags: [TEXT_MODEL_TAG, GOOGLE_GEMMA_INSTRUCT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
902
+
903
+ - name: google/gemma-2-9b
904
+ display_name: Gemma 2 (9B)
905
+ description: Gemma is a family of lightweight, open models built from the research and technology that Google used to create the Gemini models. ([model card](https://www.kaggle.com/models/google/gemma), [blog post](https://blog.google/technology/developers/google-gemma-2/))
906
+ creator_organization_name: Google
907
+ access: open
908
+ release_date: 2024-06-27
909
+ tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG]
910
+
911
+ - name: google/gemma-2-9b-it
912
+ display_name: Gemma 2 Instruct (9B)
913
+ description: Gemma is a family of lightweight, open models built from the research and technology that Google used to create the Gemini models. ([model card](https://www.kaggle.com/models/google/gemma), [blog post](https://blog.google/technology/developers/google-gemma-2/))
914
+ creator_organization_name: Google
915
+ access: open
916
+ release_date: 2024-06-27
917
+ tags: [TEXT_MODEL_TAG, GOOGLE_GEMMA_INSTRUCT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
918
+
919
+ - name: google/gemma-2-27b
920
+ display_name: Gemma 2 (27B)
921
+ description: Gemma is a family of lightweight, open models built from the research and technology that Google used to create the Gemini models. ([model card](https://www.kaggle.com/models/google/gemma), [blog post](https://blog.google/technology/developers/google-gemma-2/))
922
+ creator_organization_name: Google
923
+ access: open
924
+ release_date: 2024-06-27
925
+ tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG]
926
+
927
+ - name: google/gemma-2-27b-it
928
+ display_name: Gemma 2 Instruct (27B)
929
+ description: Gemma is a family of lightweight, open models built from the research and technology that Google used to create the Gemini models. ([model card](https://www.kaggle.com/models/google/gemma), [blog post](https://blog.google/technology/developers/google-gemma-2/))
930
+ creator_organization_name: Google
931
+ access: open
932
+ release_date: 2024-06-27
873
933
  tags: [TEXT_MODEL_TAG, GOOGLE_GEMMA_INSTRUCT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
874
934
 
875
935
  - name: google/paligemma-3b-mix-224
@@ -1361,7 +1421,7 @@ models:
1361
1421
 
1362
1422
  - name: meta/llama-3-8b
1363
1423
  display_name: Llama 3 (8B)
1364
- description: Llama 3 is a family of language models that have been trained on more than 15 trillion tokens, and use Grouped-Query Attention (GQA) for improved inference scalability.
1424
+ description: Llama 3 is a family of language models that have been trained on more than 15 trillion tokens, and use Grouped-Query Attention (GQA) for improved inference scalability. ([paper](https://ai.meta.com/research/publications/the-llama-3-herd-of-models/)
1365
1425
  creator_organization_name: Meta
1366
1426
  access: open
1367
1427
  num_parameters: 8000000000
@@ -1370,16 +1430,43 @@ models:
1370
1430
 
1371
1431
  - name: meta/llama-3-70b
1372
1432
  display_name: Llama 3 (70B)
1373
- description: Llama 3 is a family of language models that have been trained on more than 15 trillion tokens, and use Grouped-Query Attention (GQA) for improved inference scalability.
1433
+ description: Llama 3 is a family of language models that have been trained on more than 15 trillion tokens, and use Grouped-Query Attention (GQA) for improved inference scalability. ([paper](https://ai.meta.com/research/publications/the-llama-3-herd-of-models/)
1374
1434
  creator_organization_name: Meta
1375
1435
  access: open
1376
1436
  num_parameters: 70000000000
1377
1437
  release_date: 2024-04-18
1378
1438
  tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG]
1379
1439
 
1440
+ - name: meta/llama-3.1-8b-instruct-turbo
1441
+ display_name: Llama 3.1 Instruct Turbo (8B)
1442
+ description: Llama 3.1 (8B) is part of the Llama 3 family of dense Transformer models that that natively support multilinguality, coding, reasoning, and tool usage. ([paper](https://ai.meta.com/research/publications/the-llama-3-herd-of-models/), [blog](https://ai.meta.com/blog/meta-llama-3-1/)) Turbo is Together's implementation, providing a near negligible difference in quality from the reference implementation with faster performance and lower cost, currently using FP8 quantization. ([blog](https://www.together.ai/blog/llama-31-quality))
1443
+ creator_organization_name: Meta
1444
+ access: open
1445
+ num_parameters: 8000000000
1446
+ release_date: 2024-07-23
1447
+ tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG]
1448
+
1449
+ - name: meta/llama-3.1-70b-instruct-turbo
1450
+ display_name: Llama 3.1 Instruct Turbo (70B)
1451
+ description: Llama 3.1 (70B) is part of the Llama 3 family of dense Transformer models that that natively support multilinguality, coding, reasoning, and tool usage. ([paper](https://ai.meta.com/research/publications/the-llama-3-herd-of-models/), [blog](https://ai.meta.com/blog/meta-llama-3-1/)) Turbo is Together's implementation, providing a near negligible difference in quality from the reference implementation with faster performance and lower cost, currently using FP8 quantization. ([blog](https://www.together.ai/blog/llama-31-quality))
1452
+ creator_organization_name: Meta
1453
+ access: open
1454
+ num_parameters: 70000000000
1455
+ release_date: 2024-07-23
1456
+ tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG]
1457
+
1458
+ - name: meta/llama-3.1-405b-instruct-turbo
1459
+ display_name: Llama 3.1 Instruct Turbo (405B)
1460
+ description: Llama 3.1 (405B) is part of the Llama 3 family of dense Transformer models that that natively support multilinguality, coding, reasoning, and tool usage. ([paper](https://ai.meta.com/research/publications/the-llama-3-herd-of-models/), [blog](https://ai.meta.com/blog/meta-llama-3-1/)) Turbo is Together's implementation, providing a near negligible difference in quality from the reference implementation with faster performance and lower cost, currently using FP8 quantization. ([blog](https://www.together.ai/blog/llama-31-quality))
1461
+ creator_organization_name: Meta
1462
+ access: open
1463
+ num_parameters: 405000000000
1464
+ release_date: 2024-07-23
1465
+ tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG]
1466
+
1380
1467
  - name: meta/llama-3-8b-chat
1381
1468
  display_name: Llama 3 Instruct (8B)
1382
- description: Llama 3 is a family of language models that have been trained on more than 15 trillion tokens, and use Grouped-Query Attention (GQA) for improved inference scalability. It used SFT, rejection sampling, PPO and DPO for post-training.
1469
+ description: Llama 3 is a family of language models that have been trained on more than 15 trillion tokens, and use Grouped-Query Attention (GQA) for improved inference scalability. It used SFT, rejection sampling, PPO and DPO for post-training. ([paper](https://ai.meta.com/research/publications/the-llama-3-herd-of-models/)
1383
1470
  creator_organization_name: Meta
1384
1471
  access: open
1385
1472
  num_parameters: 8000000000
@@ -1388,7 +1475,7 @@ models:
1388
1475
 
1389
1476
  - name: meta/llama-3-70b-chat
1390
1477
  display_name: Llama 3 Instruct (70B)
1391
- description: Llama 3 is a family of language models that have been trained on more than 15 trillion tokens, and use Grouped-Query Attention (GQA) for improved inference scalability. It used SFT, rejection sampling, PPO and DPO for post-training.
1478
+ description: Llama 3 is a family of language models that have been trained on more than 15 trillion tokens, and use Grouped-Query Attention (GQA) for improved inference scalability. It used SFT, rejection sampling, PPO and DPO for post-training. ([paper](https://ai.meta.com/research/publications/the-llama-3-herd-of-models/)
1392
1479
  creator_organization_name: Meta
1393
1480
  access: open
1394
1481
  num_parameters: 70000000000
@@ -1413,6 +1500,16 @@ models:
1413
1500
  release_date: 2024-04-18
1414
1501
  tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
1415
1502
 
1503
+ - name: meta/llama-guard-3-8b
1504
+ display_name: Llama Guard 3 (8B)
1505
+ description: Llama Guard 3 is an 8B parameter Llama 3.1-based LLM safeguard model. Similar to Llama Guard, it can be used for classifying content in both LLM inputs (prompt classification) and in LLM responses (response classification). It acts as an LLM – it generates text in its output that indicates whether a given prompt or response is safe or unsafe, and if unsafe, it also lists the content categories violated.
1506
+ creator_organization_name: Meta
1507
+ access: open
1508
+ num_parameters: 8000000000
1509
+ release_date: 2024-07-23
1510
+ tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
1511
+
1512
+
1416
1513
 
1417
1514
 
1418
1515
 
@@ -1424,7 +1521,7 @@ models:
1424
1521
  access: closed
1425
1522
  num_parameters: 530000000000
1426
1523
  release_date: 2022-01-28
1427
- tags: [] # deprecated text model
1524
+ tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG]
1428
1525
 
1429
1526
  - name: microsoft/TNLGv2_7B
1430
1527
  display_name: TNLG v2 (6.7B)
@@ -1433,7 +1530,7 @@ models:
1433
1530
  access: closed
1434
1531
  num_parameters: 6700000000
1435
1532
  release_date: 2022-01-28
1436
- tags: [] # deprecated text model
1533
+ tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG]
1437
1534
 
1438
1535
  - name: microsoft/llava-1.5-7b-hf
1439
1536
  display_name: LLaVA 1.5 (7B)
@@ -1507,6 +1604,24 @@ models:
1507
1604
  release_date: 2023-10-05
1508
1605
  tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG]
1509
1606
 
1607
+ - name: microsoft/phi-3-small-8k-instruct
1608
+ display_name: Phi-3 (7B)
1609
+ description: Phi-3-Small-8K-Instruct is a lightweight model trained with synthetic data and filtered publicly available website data with a focus on high-quality and reasoning dense properties. ([paper](https://arxiv.org/abs/2404.14219), [blog](https://azure.microsoft.com/en-us/blog/new-models-added-to-the-phi-3-family-available-on-microsoft-azure/))
1610
+ creator_organization_name: Microsoft
1611
+ access: open
1612
+ num_parameters: 7000000000
1613
+ release_date: 2024-05-21
1614
+ tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
1615
+
1616
+ - name: microsoft/phi-3-medium-4k-instruct
1617
+ display_name: Phi-3 (14B)
1618
+ description: Phi-3-Medium-4K-Instruct is a lightweight model trained with synthetic data and filtered publicly available website data with a focus on high-quality and reasoning dense properties. ([paper](https://arxiv.org/abs/2404.14219), [blog](https://azure.microsoft.com/en-us/blog/new-models-added-to-the-phi-3-family-available-on-microsoft-azure/))
1619
+ creator_organization_name: Microsoft
1620
+ access: open
1621
+ num_parameters: 14000000000
1622
+ release_date: 2024-05-21
1623
+ tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
1624
+
1510
1625
  # KAIST AI
1511
1626
  - name: kaistai/prometheus-vision-13b-v1.0-hf
1512
1627
  display_name: LLaVA + Vicuna-v1.5 (13B)
@@ -1663,7 +1778,7 @@ models:
1663
1778
  num_parameters: 46700000000
1664
1779
  # Blog post: https://mistral.ai/news/mixtral-of-experts/
1665
1780
  release_date: 2023-12-11
1666
- tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG, MISTRAL_MODEL_TAG]
1781
+ tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
1667
1782
 
1668
1783
  - name: mistralai/mixtral-8x22b
1669
1784
  display_name: Mixtral (8x22B)
@@ -1694,13 +1809,11 @@ models:
1694
1809
 
1695
1810
  - name: mistralai/mistral-small-2402
1696
1811
  display_name: Mistral Small (2402)
1697
- # TODO: Fill in description
1698
- description: TBD
1812
+ description: Mistral Small is a multilingual model with a 32K tokens context window and function-calling capabilities. ([blog](https://mistral.ai/news/mistral-large/))
1699
1813
  creator_organization_name: Mistral AI
1700
1814
  access: limited
1701
- # Blog post: https://mistral.ai/news/mistral-large/
1702
1815
  release_date: 2023-02-26
1703
- tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG, MISTRAL_MODEL_TAG]
1816
+ tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
1704
1817
 
1705
1818
  - name: mistralai/mistral-medium-2312
1706
1819
  display_name: Mistral Medium (2312)
@@ -1708,18 +1821,32 @@ models:
1708
1821
  creator_organization_name: Mistral AI
1709
1822
  access: limited
1710
1823
  release_date: 2023-12-11
1711
- tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG, MISTRAL_MODEL_TAG]
1824
+ tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
1712
1825
 
1713
1826
  - name: mistralai/mistral-large-2402
1714
1827
  display_name: Mistral Large (2402)
1715
- # TODO: Fill in description
1716
- description: TBD
1828
+ description: Mistral Large is a multilingual model with a 32K tokens context window and function-calling capabilities. ([blog](https://mistral.ai/news/mistral-large/))
1717
1829
  creator_organization_name: Mistral AI
1718
1830
  access: limited
1719
- # Blog post: https://mistral.ai/news/mistral-large/
1720
1831
  release_date: 2023-02-26
1721
- tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG, MISTRAL_MODEL_TAG]
1832
+ tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
1722
1833
 
1834
+ - name: mistralai/mistral-large-2407
1835
+ display_name: Mistral Large 2 (2407)
1836
+ description: Mistral Large 2 is a 123 billion parameter model that has a 128k context window and supports dozens of languages and 80+ coding languages. ([blog](https://mistral.ai/news/mistral-large-2407/))
1837
+ creator_organization_name: Mistral AI
1838
+ access: open
1839
+ num_parameters: 123000000000
1840
+ release_date: 2023-07-24
1841
+ tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
1842
+
1843
+ - name: mistralai/open-mistral-nemo-2407
1844
+ display_name: Mistral NeMo (2402)
1845
+ description: Mistral NeMo is a multilingual 12B model with a large context window of 128K tokens. ([blog](https://mistral.ai/news/mistral-nemo/))
1846
+ creator_organization_name: Mistral AI
1847
+ access: open
1848
+ release_date: 2024-07-18
1849
+ tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
1723
1850
 
1724
1851
  # MosaicML
1725
1852
  - name: mosaicml/mpt-7b
@@ -1798,7 +1925,13 @@ models:
1798
1925
  release_date: 2019-09-17 # paper date
1799
1926
  tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG, BUGGY_TEMP_0_TAG]
1800
1927
 
1801
-
1928
+ - name: nvidia/nemotron-4-340b-instruct
1929
+ display_name: Nemotron-4 Instruct (340B)
1930
+ description: Nemotron-4 Instruct (340B) is an open weights model sized to fit on a single DGX H100 with 8 GPUs when deployed in FP8 precision. 98% of the data used for model alignment was synthetically generated ([paper](https://arxiv.org/abs/2406.11704)).
1931
+ creator_organization_name: NVIDIA
1932
+ access: open
1933
+ release_date: 2024-06-17
1934
+ tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
1802
1935
 
1803
1936
  # OpenAI
1804
1937
 
@@ -1836,97 +1969,95 @@ models:
1836
1969
 
1837
1970
  # DEPRECATED: Announced on July 06 2023 that these models will be shut down on January 04 2024.
1838
1971
 
1839
- - name: openai/davinci # DEPRECATED
1972
+ - name: openai/davinci
1840
1973
  display_name: davinci (175B)
1841
1974
  description: Original GPT-3 (175B parameters) autoregressive language model ([paper](https://arxiv.org/pdf/2005.14165.pdf), [docs](https://beta.openai.com/docs/model-index-for-researchers)).
1842
1975
  creator_organization_name: OpenAI
1843
1976
  access: limited
1844
1977
  num_parameters: 175000000000
1845
1978
  release_date: 2020-05-28
1846
- tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
1979
+ tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
1847
1980
 
1848
- - name: openai/curie # DEPRECATED
1981
+ - name: openai/curie
1849
1982
  display_name: curie (6.7B)
1850
1983
  description: Original GPT-3 (6.7B parameters) autoregressive language model ([paper](https://arxiv.org/pdf/2005.14165.pdf), [docs](https://beta.openai.com/docs/model-index-for-researchers)).
1851
1984
  creator_organization_name: OpenAI
1852
1985
  access: limited
1853
1986
  num_parameters: 6700000000
1854
1987
  release_date: 2020-05-28
1855
- tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
1988
+ tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
1856
1989
 
1857
- - name: openai/babbage # DEPRECATED
1990
+ - name: openai/babbage
1858
1991
  display_name: babbage (1.3B)
1859
1992
  description: Original GPT-3 (1.3B parameters) autoregressive language model ([paper](https://arxiv.org/pdf/2005.14165.pdf), [docs](https://beta.openai.com/docs/model-index-for-researchers)).
1860
1993
  creator_organization_name: OpenAI
1861
1994
  access: limited
1862
1995
  num_parameters: 1300000000
1863
1996
  release_date: 2020-05-28
1864
- tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
1997
+ tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
1865
1998
 
1866
- - name: openai/ada # DEPRECATED
1999
+ - name: openai/ada
1867
2000
  display_name: ada (350M)
1868
2001
  description: Original GPT-3 (350M parameters) autoregressive language model ([paper](https://arxiv.org/pdf/2005.14165.pdf), [docs](https://beta.openai.com/docs/model-index-for-researchers)).
1869
2002
  creator_organization_name: OpenAI
1870
2003
  access: limited
1871
2004
  num_parameters: 350000000
1872
2005
  release_date: 2020-05-28
1873
- tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
2006
+ tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
1874
2007
 
1875
- - name: openai/text-davinci-003 # DEPRECATED
2008
+ - name: openai/text-davinci-003
1876
2009
  display_name: GPT-3.5 (text-davinci-003)
1877
2010
  description: text-davinci-003 model that involves reinforcement learning (PPO) with reward models. Derived from text-davinci-002 ([docs](https://beta.openai.com/docs/model-index-for-researchers)).
1878
2011
  creator_organization_name: OpenAI
1879
2012
  access: limited
1880
2013
  num_parameters: 175000000000
1881
2014
  release_date: 2022-11-28
1882
- tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
2015
+ tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
1883
2016
 
1884
- # TODO: text-davinci-002 supports insertion. Support insertion in our framework.
1885
- # https://github.com/stanford-crfm/benchmarking/issues/359
1886
- - name: openai/text-davinci-002 # DEPRECATED
2017
+ - name: openai/text-davinci-002
1887
2018
  display_name: GPT-3.5 (text-davinci-002)
1888
2019
  description: text-davinci-002 model that involves supervised fine-tuning on human-written demonstrations. Derived from code-davinci-002 ([docs](https://beta.openai.com/docs/model-index-for-researchers)).
1889
2020
  creator_organization_name: OpenAI
1890
2021
  access: limited
1891
2022
  num_parameters: 175000000000
1892
2023
  release_date: 2022-01-27
1893
- tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
2024
+ tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
1894
2025
 
1895
- - name: openai/text-davinci-001 # DEPRECATED
2026
+ - name: openai/text-davinci-001
1896
2027
  display_name: GPT-3.5 (text-davinci-001)
1897
2028
  description: text-davinci-001 model that involves supervised fine-tuning on human-written demonstrations ([docs](https://beta.openai.com/docs/model-index-for-researchers)).
1898
2029
  creator_organization_name: OpenAI
1899
2030
  access: limited
1900
2031
  num_parameters: 175000000000
1901
2032
  release_date: 2022-01-27
1902
- tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
2033
+ tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
1903
2034
 
1904
- - name: openai/text-curie-001 # DEPRECATED
2035
+ - name: openai/text-curie-001
1905
2036
  display_name: text-curie-001
1906
2037
  description: text-curie-001 model that involves supervised fine-tuning on human-written demonstrations ([docs](https://beta.openai.com/docs/model-index-for-researchers)).
1907
2038
  creator_organization_name: OpenAI
1908
2039
  access: limited
1909
2040
  num_parameters: 6700000000
1910
2041
  release_date: 2022-01-27
1911
- tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
2042
+ tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
1912
2043
 
1913
- - name: openai/text-babbage-001 # DEPRECATED
2044
+ - name: openai/text-babbage-001
1914
2045
  display_name: text-babbage-001
1915
2046
  description: text-babbage-001 model that involves supervised fine-tuning on human-written demonstrations ([docs](https://beta.openai.com/docs/model-index-for-researchers)).
1916
2047
  creator_organization_name: OpenAI
1917
2048
  access: limited
1918
2049
  num_parameters: 1300000000
1919
2050
  release_date: 2022-01-27
1920
- tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
2051
+ tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
1921
2052
 
1922
- - name: openai/text-ada-001 # DEPRECATED
2053
+ - name: openai/text-ada-001
1923
2054
  display_name: text-ada-001
1924
2055
  description: text-ada-001 model that involves supervised fine-tuning on human-written demonstrations ([docs](https://beta.openai.com/docs/model-index-for-researchers)).
1925
2056
  creator_organization_name: OpenAI
1926
2057
  access: limited
1927
2058
  num_parameters: 350000000
1928
2059
  release_date: 2022-01-27
1929
- tags: [TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
2060
+ tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG]
1930
2061
 
1931
2062
 
1932
2063
  ## GPT 3.5 Turbo Models
@@ -2044,16 +2175,32 @@ models:
2044
2175
  creator_organization_name: OpenAI
2045
2176
  access: limited
2046
2177
  release_date: 2024-04-09
2047
- tags: [TEXT_MODEL_TAG, OPENAI_CHATGPT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
2178
+ tags: [TEXT_MODEL_TAG, VISION_LANGUAGE_MODEL_TAG, OPENAI_CHATGPT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
2048
2179
 
2049
2180
  - name: openai/gpt-4o-2024-05-13
2050
2181
  display_name: GPT-4o (2024-05-13)
2051
- description: GPT-4o (2024-05-13) is a large multimodal model that accepts as input any combination of text, audio, and image and generates any combination of text, audio, and image outputs.
2182
+ description: GPT-4o (2024-05-13) is a large multimodal model that accepts as input any combination of text, audio, and image and generates any combination of text, audio, and image outputs. ([blog](https://openai.com/index/hello-gpt-4o/))
2052
2183
  creator_organization_name: OpenAI
2053
2184
  access: limited
2054
2185
  release_date: 2024-04-09
2055
2186
  tags: [TEXT_MODEL_TAG, VISION_LANGUAGE_MODEL_TAG, OPENAI_CHATGPT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
2056
2187
 
2188
+ - name: openai/gpt-4o-2024-08-06
2189
+ display_name: GPT-4o (2024-08-06)
2190
+ description: GPT-4o (2024-08-06) is a large multimodal model that accepts as input any combination of text, audio, and image and generates any combination of text, audio, and image outputs. ([blog](https://openai.com/index/introducing-structured-outputs-in-the-api/))
2191
+ creator_organization_name: OpenAI
2192
+ access: limited
2193
+ release_date: 2024-08-06
2194
+ tags: [TEXT_MODEL_TAG, VISION_LANGUAGE_MODEL_TAG, OPENAI_CHATGPT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
2195
+
2196
+ - name: openai/gpt-4o-mini-2024-07-18
2197
+ display_name: GPT-4o mini (2024-07-18)
2198
+ description: GPT-4o mini (2024-07-18) is a multimodal model with a context window of 128K tokens and improved handling of non-English text. ([blog](https://openai.com/index/gpt-4o-mini-advancing-cost-efficient-intelligence/))
2199
+ creator_organization_name: OpenAI
2200
+ access: limited
2201
+ release_date: 2024-07-18
2202
+ tags: [TEXT_MODEL_TAG, VISION_LANGUAGE_MODEL_TAG, OPENAI_CHATGPT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
2203
+
2057
2204
  - name: openai/gpt-4-vision-preview
2058
2205
  # According to https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4, this model has pointed gpt-4-1106-vision-preview.
2059
2206
  display_name: GPT-4V (1106 preview)
@@ -2074,30 +2221,30 @@ models:
2074
2221
  ## Codex Models
2075
2222
  # DEPRECATED: Codex models have been shut down on March 23 2023.
2076
2223
 
2077
- - name: openai/code-davinci-002 # DEPRECATED
2224
+ - name: openai/code-davinci-002
2078
2225
  display_name: code-davinci-002
2079
2226
  description: Codex-style model that is designed for pure code-completion tasks ([docs](https://beta.openai.com/docs/models/codex)).
2080
2227
  creator_organization_name: OpenAI
2081
2228
  access: limited
2082
2229
  release_date: 2021-07-01 # TODO: Find correct date (this is for v1)
2083
- tags: [CODE_MODEL_TAG]
2230
+ tags: [DEPRECATED_MODEL_TAG, CODE_MODEL_TAG]
2084
2231
 
2085
- - name: openai/code-davinci-001 # DEPRECATED
2232
+ - name: openai/code-davinci-001
2086
2233
  display_name: code-davinci-001
2087
2234
  description: code-davinci-001 model
2088
2235
  creator_organization_name: OpenAI
2089
2236
  access: limited
2090
2237
  release_date: 2021-07-01 # Paper date
2091
- tags: [CODE_MODEL_TAG]
2238
+ tags: [DEPRECATED_MODEL_TAG, CODE_MODEL_TAG]
2092
2239
 
2093
- - name: openai/code-cushman-001 # DEPRECATED
2240
+ - name: openai/code-cushman-001
2094
2241
  display_name: code-cushman-001 (12B)
2095
2242
  description: Codex-style model that is a stronger, multilingual version of the Codex (12B) model in the [Codex paper](https://arxiv.org/pdf/2107.03374.pdf).
2096
2243
  creator_organization_name: OpenAI
2097
2244
  access: limited
2098
2245
  num_parameters: 12000000000
2099
2246
  release_date: 2021-07-01 # Paper date
2100
- tags: [CODE_MODEL_TAG]
2247
+ tags: [DEPRECATED_MODEL_TAG, CODE_MODEL_TAG]
2101
2248
 
2102
2249
 
2103
2250
  ## Text Similarity Models
@@ -2107,41 +2254,41 @@ models:
2107
2254
  # DEPRECATED: Announced on July 06 2023 that first generation embeddings models
2108
2255
  # will be shut down on January 04 2024.
2109
2256
 
2110
- - name: openai/text-similarity-davinci-001 # DEPRECATED
2257
+ - name: openai/text-similarity-davinci-001
2111
2258
  display_name: text-similarity-davinci-001
2112
2259
  description: Embedding model that is designed for text similarity tasks ([docs](https://openai.com/blog/introducing-text-and-code-embeddings)).
2113
2260
  creator_organization_name: OpenAI
2114
2261
  access: limited
2115
2262
  num_parameters: 175000000000
2116
2263
  release_date: 2022-01-25 # Blog post date
2117
- tags: [TEXT_SIMILARITY_MODEL_TAG]
2264
+ tags: [DEPRECATED_MODEL_TAG, TEXT_SIMILARITY_MODEL_TAG]
2118
2265
 
2119
- - name: openai/text-similarity-curie-001 # DEPRECATED
2266
+ - name: openai/text-similarity-curie-001
2120
2267
  display_name: text-similarity-curie-001
2121
2268
  description: Embedding model that is designed for text similarity tasks ([docs](https://openai.com/blog/introducing-text-and-code-embeddings)).
2122
2269
  creator_organization_name: OpenAI
2123
2270
  access: limited
2124
2271
  num_parameters: 6700000000
2125
2272
  release_date: 2022-01-25 # Blog post date
2126
- tags: [TEXT_SIMILARITY_MODEL_TAG]
2273
+ tags: [DEPRECATED_MODEL_TAG, TEXT_SIMILARITY_MODEL_TAG]
2127
2274
 
2128
- - name: openai/text-similarity-babbage-001 # DEPRECATED
2275
+ - name: openai/text-similarity-babbage-001
2129
2276
  display_name: text-similarity-babbage-001
2130
2277
  description: Embedding model that is designed for text similarity tasks ([docs](https://openai.com/blog/introducing-text-and-code-embeddings)).
2131
2278
  creator_organization_name: OpenAI
2132
2279
  access: limited
2133
2280
  num_parameters: 1300000000
2134
2281
  release_date: 2022-01-25 # Blog post date
2135
- tags: [TEXT_SIMILARITY_MODEL_TAG]
2282
+ tags: [DEPRECATED_MODEL_TAG, TEXT_SIMILARITY_MODEL_TAG]
2136
2283
 
2137
- - name: openai/text-similarity-ada-001 # DEPRECATED
2284
+ - name: openai/text-similarity-ada-001
2138
2285
  display_name: text-similarity-ada-001
2139
2286
  description: Embedding model that is designed for text similarity tasks ([docs](https://openai.com/blog/introducing-text-and-code-embeddings)).
2140
2287
  creator_organization_name: OpenAI
2141
2288
  access: limited
2142
2289
  num_parameters: 350000000
2143
2290
  release_date: 2022-01-25 # Blog post date
2144
- tags: [TEXT_SIMILARITY_MODEL_TAG]
2291
+ tags: [DEPRECATED_MODEL_TAG, TEXT_SIMILARITY_MODEL_TAG]
2145
2292
 
2146
2293
  - name: openai/text-embedding-ada-002
2147
2294
  display_name: text-embedding-ada-002
@@ -2197,6 +2344,34 @@ models:
2197
2344
  release_date: 2023-11-06
2198
2345
  tags: [TEXT_TO_IMAGE_MODEL_TAG]
2199
2346
 
2347
+ # OpenThaiGPT
2348
+ - name: openthaigpt/openthaigpt-1.0.0-7b-chat
2349
+ display_name: OpenThaiGPT v1.0.0 (7B)
2350
+ description: OpenThaiGPT v1.0.0 (7B) is a Thai language chat model based on Llama 2 that has been specifically fine-tuned for Thai instructions and enhanced by incorporating over 10,000 of the most commonly used Thai words into the dictionary. ([blog post](https://openthaigpt.aieat.or.th/openthaigpt-1.0.0-less-than-8-apr-2024-greater-than))
2351
+ creator_organization_name: OpenThaiGPT
2352
+ access: open
2353
+ num_parameters: 7000000000
2354
+ release_date: 2024-04-08
2355
+ tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
2356
+
2357
+ - name: openthaigpt/openthaigpt-1.0.0-13b-chat
2358
+ display_name: OpenThaiGPT v1.0.0 (13B)
2359
+ description: OpenThaiGPT v1.0.0 (13B) is a Thai language chat model based on Llama 2 that has been specifically fine-tuned for Thai instructions and enhanced by incorporating over 10,000 of the most commonly used Thai words into the dictionary. ([blog post](https://openthaigpt.aieat.or.th/openthaigpt-1.0.0-less-than-8-apr-2024-greater-than))
2360
+ creator_organization_name: OpenThaiGPT
2361
+ access: open
2362
+ num_parameters: 13000000000
2363
+ release_date: 2024-04-08
2364
+ tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
2365
+
2366
+ - name: openthaigpt/openthaigpt-1.0.0-70b-chat
2367
+ display_name: OpenThaiGPT v1.0.0 (70B)
2368
+ description: OpenThaiGPT v1.0.0 (70B) is a Thai language chat model based on Llama 2 that has been specifically fine-tuned for Thai instructions and enhanced by incorporating over 10,000 of the most commonly used Thai words into the dictionary. ([blog post](https://openthaigpt.aieat.or.th/openthaigpt-1.0.0-less-than-8-apr-2024-greater-than))
2369
+ creator_organization_name: OpenThaiGPT
2370
+ access: open
2371
+ num_parameters: 70000000000
2372
+ release_date: 2024-04-08
2373
+ tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
2374
+
2200
2375
  # Qwen
2201
2376
 
2202
2377
  - name: qwen/qwen-7b
@@ -2311,7 +2486,7 @@ models:
2311
2486
  access: open
2312
2487
  num_parameters: 7000000000
2313
2488
  release_date: 2024-04-04
2314
- tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
2489
+ tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG]
2315
2490
 
2316
2491
  - name: sail/sailor-7b-chat
2317
2492
  display_name: Sailor Chat (7B)
@@ -2329,7 +2504,7 @@ models:
2329
2504
  access: open
2330
2505
  num_parameters: 14000000000
2331
2506
  release_date: 2024-04-04
2332
- tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
2507
+ tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG]
2333
2508
 
2334
2509
  - name: sail/sailor-14b-chat
2335
2510
  display_name: Sailor Chat (14B)
@@ -2350,10 +2525,74 @@ models:
2350
2525
  release_date: 2022-03-25
2351
2526
  tags: [] # TODO: add tags
2352
2527
 
2528
+ # SambaNova
2529
+ - name: sambanova/sambalingo-thai-base
2530
+ display_name: SambaLingo-Thai-Base
2531
+ description: SambaLingo-Thai-Base is a pretrained bi-lingual Thai and English model that adapts Llama 2 (7B) to Thai by training on 38 billion tokens from the Thai split of the Cultura-X dataset. ([paper](https://arxiv.org/abs/2404.05829))
2532
+ creator_organization_name: SambaLingo
2533
+ access: open
2534
+ num_parameters: 7000000000
2535
+ release_date: 2024-04-08
2536
+ tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG]
2537
+
2538
+ - name: sambanova/sambalingo-thai-chat
2539
+ display_name: SambaLingo-Thai-Chat
2540
+ description: SambaLingo-Thai-Chat is a chat model trained using direct preference optimization on SambaLingo-Thai-Base. SambaLingo-Thai-Base adapts Llama 2 (7B) to Thai by training on 38 billion tokens from the Thai split of the Cultura-X dataset. ([paper](https://arxiv.org/abs/2404.05829))
2541
+ creator_organization_name: SambaLingo
2542
+ access: open
2543
+ num_parameters: 7000000000
2544
+ release_date: 2024-04-08
2545
+ tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
2546
+
2547
+ - name: sambanova/sambalingo-thai-base-70b
2548
+ display_name: SambaLingo-Thai-Base-70B
2549
+ description: SambaLingo-Thai-Base-70B is a pretrained bi-lingual Thai and English model that adapts Llama 2 (70B) to Thai by training on 26 billion tokens from the Thai split of the Cultura-X dataset. ([paper](https://arxiv.org/abs/2404.05829))
2550
+ creator_organization_name: SambaLingo
2551
+ access: open
2552
+ num_parameters: 70000000000
2553
+ release_date: 2024-04-08
2554
+ tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG]
2555
+
2556
+ - name: sambanova/sambalingo-thai-chat-70b
2557
+ display_name: SambaLingo-Thai-Chat-70B
2558
+ description: SambaLingo-Thai-Chat-70B is a chat model trained using direct preference optimization on SambaLingo-Thai-Base-70B. SambaLingo-Thai-Base-70B adapts Llama 2 (7B) to Thai by training on 26 billion tokens from the Thai split of the Cultura-X dataset. ([paper](https://arxiv.org/abs/2404.05829))
2559
+ creator_organization_name: SambaLingo
2560
+ access: open
2561
+ num_parameters: 70000000000
2562
+ release_date: 2024-04-08
2563
+ tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
2564
+
2353
2565
  # SCB10X
2566
+ - name: scb10x/typhoon-7b
2567
+ display_name: Typhoon (7B)
2568
+ description: Typhoon (7B) is pretrained Thai large language model with 7 billion parameters based on Mistral 7B. ([paper](https://arxiv.org/abs/2312.13951))
2569
+ creator_organization_name: SCB10X
2570
+ access: open
2571
+ num_parameters: 7000000000
2572
+ release_date: 2023-12-21
2573
+ tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG]
2574
+
2575
+ - name: scb10x/typhoon-v1.5-8b
2576
+ display_name: Typhoon v1.5 (8B)
2577
+ description: Typhoon v1.5 (8B) is a pretrained Thai large language model with 8 billion parameters based on Llama 3 8B. ([blog](https://blog.opentyphoon.ai/typhoon-1-5-release-a9364cb8e8d7))
2578
+ creator_organization_name: SCB10X
2579
+ access: open
2580
+ num_parameters: 8000000000
2581
+ release_date: 2024-05-08
2582
+ tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG]
2583
+
2584
+ - name: scb10x/typhoon-v1.5-8b-instruct
2585
+ display_name: Typhoon v1.5 Instruct (8B)
2586
+ description: Typhoon v1.5 Instruct (8B) is a pretrained Thai large language model with 8 billion parameters based on Llama 3 8B. ([blog](https://blog.opentyphoon.ai/typhoon-1-5-release-a9364cb8e8d7))
2587
+ creator_organization_name: SCB10X
2588
+ access: open
2589
+ num_parameters: 8000000000
2590
+ release_date: 2024-05-08
2591
+ tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
2592
+
2354
2593
  - name: scb10x/typhoon-v1.5-72b
2355
2594
  display_name: Typhoon v1.5 (72B)
2356
- description: Typhoon v1.5 (72B) is pretrained Thai large language model with 72 billion parameters based on Qwen1.5-72B. ([blog](https://blog.opentyphoon.ai/typhoon-1-5-release-a9364cb8e8d7))
2595
+ description: Typhoon v1.5 (72B) is a pretrained Thai large language model with 72 billion parameters based on Qwen1.5-72B. ([blog](https://blog.opentyphoon.ai/typhoon-1-5-release-a9364cb8e8d7))
2357
2596
  creator_organization_name: SCB10X
2358
2597
  access: open
2359
2598
  num_parameters: 72000000000
@@ -2362,13 +2601,50 @@ models:
2362
2601
 
2363
2602
  - name: scb10x/typhoon-v1.5-72b-instruct
2364
2603
  display_name: Typhoon v1.5 Instruct (72B)
2365
- description: Typhoon v1.5 Instruct (72B) is pretrained Thai large language model with 72 billion parameters based on Qwen1.5-72B. ([blog](https://blog.opentyphoon.ai/typhoon-1-5-release-a9364cb8e8d7))
2604
+ description: Typhoon v1.5 Instruct (72B) is a pretrained Thai large language model with 72 billion parameters based on Qwen1.5-72B. ([blog](https://blog.opentyphoon.ai/typhoon-1-5-release-a9364cb8e8d7))
2366
2605
  creator_organization_name: SCB10X
2367
2606
  access: open
2368
2607
  num_parameters: 72000000000
2369
2608
  release_date: 2024-05-08
2370
2609
  tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
2371
2610
 
2611
+ - name: scb10x/llama-3-typhoon-v1.5x-8b-instruct
2612
+ display_name: Typhoon 1.5X instruct (8B)
2613
+ description: Llama-3-Typhoon-1.5X-8B-instruct is a 8 billion parameter instruct model designed for the Thai language based on Llama 3 Instruct. It utilizes the task-arithmetic model editing technique. ([blog](https://blog.opentyphoon.ai/typhoon-1-5x-our-experiment-designed-for-application-use-cases-7b85d9e9845c))
2614
+ creator_organization_name: SCB10X
2615
+ access: open
2616
+ num_parameters: 8000000000
2617
+ release_date: 2024-05-29
2618
+ tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
2619
+
2620
+ - name: scb10x/llama-3-typhoon-v1.5x-70b-instruct
2621
+ display_name: Typhoon 1.5X instruct (70B)
2622
+ description: Llama-3-Typhoon-1.5X-70B-instruct is a 70 billion parameter instruct model designed for the Thai language based on Llama 3 Instruct. It utilizes the task-arithmetic model editing technique. ([blog](https://blog.opentyphoon.ai/typhoon-1-5x-our-experiment-designed-for-application-use-cases-7b85d9e9845c))
2623
+ creator_organization_name: SCB10X
2624
+ access: open
2625
+ num_parameters: 70000000000
2626
+ release_date: 2024-05-29
2627
+ tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG, INSTRUCTION_FOLLOWING_MODEL_TAG]
2628
+
2629
+ # Alibaba DAMO Academy
2630
+ - name: damo/seallm-7b-v2
2631
+ display_name: SeaLLM v2 (7B)
2632
+ description: SeaLLM v2 is a multilingual LLM for Southeast Asian (SEA) languages trained from Mistral (7B). ([website](https://damo-nlp-sg.github.io/SeaLLMs/))
2633
+ creator_organization_name: Alibaba DAMO Academy
2634
+ access: open
2635
+ num_parameters: 7000000000
2636
+ release_date: 2024-02-02
2637
+ tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG]
2638
+
2639
+ - name: damo/seallm-7b-v2.5
2640
+ display_name: SeaLLM v2.5 (7B)
2641
+ description: SeaLLM is a multilingual LLM for Southeast Asian (SEA) languages trained from Gemma (7B). ([website](https://damo-nlp-sg.github.io/SeaLLMs/))
2642
+ creator_organization_name: Alibaba DAMO Academy
2643
+ access: open
2644
+ num_parameters: 7000000000
2645
+ release_date: 2024-04-12
2646
+ tags: [TEXT_MODEL_TAG, PARTIAL_FUNCTIONALITY_TEXT_MODEL_TAG]
2647
+
2372
2648
  # Snowflake
2373
2649
  - name: snowflake/snowflake-arctic-instruct
2374
2650
  display_name: Arctic Instruct
@@ -2538,7 +2814,7 @@ models:
2538
2814
  release_date: 2022-08-04
2539
2815
  # Inference with echo=True is not feasible -- in the prompt encoding phase, they use
2540
2816
  # bidirectional attention and do not perform predictions on them.
2541
- tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, ABLATION_MODEL_TAG, NO_NEWLINES_TAG]
2817
+ tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG, ABLATION_MODEL_TAG, NO_NEWLINES_TAG]
2542
2818
 
2543
2819
  - name: tsinghua/codegeex # NOT SUPPORTED
2544
2820
  display_name: CodeGeeX (13B)
@@ -2573,7 +2849,6 @@ models:
2573
2849
  tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG]
2574
2850
 
2575
2851
  - name: writer/palmyra-instruct-30
2576
- deprecated: true # Internal error
2577
2852
  display_name: InstructPalmyra (30B)
2578
2853
  description: InstructPalmyra (30B parameters) is trained using reinforcement learning techniques based on feedback from humans.
2579
2854
  creator_organization_name: Writer
@@ -2581,10 +2856,9 @@ models:
2581
2856
  num_parameters: 30000000000
2582
2857
  release_date: 2023-02-16
2583
2858
  # Does not support echo
2584
- tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG]
2859
+ tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG]
2585
2860
 
2586
2861
  - name: writer/palmyra-e
2587
- deprecated: true # Internal error
2588
2862
  display_name: Palmyra E (30B)
2589
2863
  description: Palmyra E (30B)
2590
2864
  creator_organization_name: Writer
@@ -2592,7 +2866,7 @@ models:
2592
2866
  num_parameters: 30000000000
2593
2867
  release_date: 2023-03-03
2594
2868
  # Does not support echo
2595
- tags: [TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG]
2869
+ tags: [DEPRECATED_MODEL_TAG, TEXT_MODEL_TAG, LIMITED_FUNCTIONALITY_TEXT_MODEL_TAG]
2596
2870
 
2597
2871
  - name: writer/silk-road
2598
2872
  display_name: Silk Road (35B)