crfm-helm 0.4.0__py3-none-any.whl → 0.5.1__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of crfm-helm might be problematic. Click here for more details.
- {crfm_helm-0.4.0.dist-info → crfm_helm-0.5.1.dist-info}/METADATA +138 -31
- crfm_helm-0.5.1.dist-info/RECORD +654 -0
- {crfm_helm-0.4.0.dist-info → crfm_helm-0.5.1.dist-info}/WHEEL +1 -1
- helm/benchmark/adaptation/adapter_spec.py +31 -3
- helm/benchmark/adaptation/adapters/adapter.py +2 -2
- helm/benchmark/adaptation/adapters/adapter_factory.py +24 -27
- helm/benchmark/adaptation/adapters/generation_adapter.py +1 -0
- helm/benchmark/adaptation/adapters/in_context_learning_adapter.py +20 -4
- helm/benchmark/adaptation/adapters/language_modeling_adapter.py +2 -3
- helm/benchmark/adaptation/adapters/multimodal/in_context_learning_multimodal_adapter.py +1 -0
- helm/benchmark/adaptation/adapters/multimodal/multimodal_prompt.py +7 -0
- helm/benchmark/adaptation/adapters/multimodal/multiple_choice_joint_multimodal_adapter.py +104 -0
- helm/benchmark/adaptation/adapters/multimodal/test_in_context_learning_multimodal_adapter.py +2 -1
- helm/benchmark/adaptation/adapters/multimodal/test_multimodal_prompt.py +2 -0
- helm/benchmark/adaptation/adapters/test_adapter.py +2 -1
- helm/benchmark/adaptation/adapters/test_generation_adapter.py +32 -8
- helm/benchmark/adaptation/adapters/test_language_modeling_adapter.py +7 -19
- helm/benchmark/adaptation/adapters/test_multiple_choice_joint_adapter.py +60 -6
- helm/benchmark/adaptation/common_adapter_specs.py +376 -0
- helm/benchmark/adaptation/request_state.py +6 -1
- helm/benchmark/adaptation/scenario_state.py +6 -2
- helm/benchmark/annotation/annotator.py +43 -0
- helm/benchmark/annotation/annotator_factory.py +61 -0
- helm/benchmark/annotation/image2structure/image_compiler_annotator.py +88 -0
- helm/benchmark/annotation/image2structure/latex_compiler_annotator.py +59 -0
- helm/benchmark/annotation/image2structure/lilypond_compiler_annotator.py +84 -0
- helm/benchmark/annotation/image2structure/webpage_compiler_annotator.py +132 -0
- helm/benchmark/annotation/test_annotator_factory.py +26 -0
- helm/benchmark/annotation/test_dummy_annotator.py +44 -0
- helm/benchmark/annotation_executor.py +124 -0
- helm/benchmark/augmentations/data_augmenter.py +0 -2
- helm/benchmark/augmentations/gender_perturbation.py +1 -1
- helm/benchmark/augmentations/perturbation.py +25 -3
- helm/benchmark/augmentations/perturbation_description.py +1 -1
- helm/benchmark/augmentations/suffix_perturbation.py +29 -0
- helm/benchmark/augmentations/test_perturbation.py +41 -7
- helm/benchmark/augmentations/translate_perturbation.py +30 -0
- helm/benchmark/config_registry.py +7 -1
- helm/benchmark/executor.py +46 -16
- helm/benchmark/huggingface_registration.py +20 -7
- helm/benchmark/metrics/basic_metrics.py +169 -664
- helm/benchmark/metrics/bbq_metrics.py +3 -4
- helm/benchmark/metrics/bias_metrics.py +6 -6
- helm/benchmark/metrics/classification_metrics.py +11 -8
- helm/benchmark/metrics/cleva_accuracy_metrics.py +8 -5
- helm/benchmark/metrics/cleva_harms_metrics.py +2 -2
- helm/benchmark/metrics/code_metrics_helper.py +0 -2
- helm/benchmark/metrics/common_metric_specs.py +167 -0
- helm/benchmark/metrics/decodingtrust_fairness_metrics.py +72 -0
- helm/benchmark/metrics/decodingtrust_ood_knowledge_metrics.py +66 -0
- helm/benchmark/metrics/decodingtrust_privacy_metrics.py +101 -0
- helm/benchmark/metrics/decodingtrust_stereotype_bias_metrics.py +202 -0
- helm/benchmark/metrics/disinformation_metrics.py +4 -110
- helm/benchmark/metrics/dry_run_metrics.py +2 -2
- helm/benchmark/metrics/efficiency_metrics.py +213 -0
- helm/benchmark/metrics/evaluate_instances_metric.py +59 -0
- helm/benchmark/metrics/evaluate_reference_metrics.py +392 -0
- helm/benchmark/metrics/image_generation/aesthetics_metrics.py +54 -0
- helm/benchmark/metrics/image_generation/aesthetics_scorer.py +66 -0
- helm/benchmark/metrics/image_generation/clip_score_metrics.py +73 -0
- helm/benchmark/metrics/image_generation/denoised_runtime_metric.py +42 -0
- helm/benchmark/metrics/image_generation/detection_metrics.py +57 -0
- helm/benchmark/metrics/image_generation/detectors/base_detector.py +8 -0
- helm/benchmark/metrics/image_generation/detectors/vitdet.py +178 -0
- helm/benchmark/metrics/image_generation/efficiency_metrics.py +41 -0
- helm/benchmark/metrics/image_generation/fidelity_metrics.py +168 -0
- helm/benchmark/metrics/image_generation/fractal_dimension/__init__.py +0 -0
- helm/benchmark/metrics/image_generation/fractal_dimension/fractal_dimension_util.py +63 -0
- helm/benchmark/metrics/image_generation/fractal_dimension/test_fractal_dimension_util.py +33 -0
- helm/benchmark/metrics/image_generation/fractal_dimension_metric.py +50 -0
- helm/benchmark/metrics/image_generation/gender_metrics.py +58 -0
- helm/benchmark/metrics/image_generation/image_critique_metrics.py +284 -0
- helm/benchmark/metrics/image_generation/lpips_metrics.py +82 -0
- helm/benchmark/metrics/image_generation/multi_scale_ssim_metrics.py +82 -0
- helm/benchmark/metrics/image_generation/nsfw_detector.py +96 -0
- helm/benchmark/metrics/image_generation/nsfw_metrics.py +103 -0
- helm/benchmark/metrics/image_generation/nudity_metrics.py +38 -0
- helm/benchmark/metrics/image_generation/photorealism_critique_metrics.py +153 -0
- helm/benchmark/metrics/image_generation/psnr_metrics.py +78 -0
- helm/benchmark/metrics/image_generation/q16/__init__.py +0 -0
- helm/benchmark/metrics/image_generation/q16/q16_toxicity_detector.py +90 -0
- helm/benchmark/metrics/image_generation/q16/test_q16.py +18 -0
- helm/benchmark/metrics/image_generation/q16_toxicity_metrics.py +48 -0
- helm/benchmark/metrics/image_generation/skin_tone_metrics.py +164 -0
- helm/benchmark/metrics/image_generation/uiqi_metrics.py +92 -0
- helm/benchmark/metrics/image_generation/watermark/__init__.py +0 -0
- helm/benchmark/metrics/image_generation/watermark/test_watermark_detector.py +16 -0
- helm/benchmark/metrics/image_generation/watermark/watermark_detector.py +87 -0
- helm/benchmark/metrics/image_generation/watermark_metrics.py +48 -0
- helm/benchmark/metrics/instruction_following_critique_metrics.py +3 -1
- helm/benchmark/metrics/language_modeling_metrics.py +99 -0
- helm/benchmark/metrics/machine_translation_metrics.py +89 -0
- helm/benchmark/metrics/metric.py +93 -172
- helm/benchmark/metrics/metric_name.py +0 -1
- helm/benchmark/metrics/metric_service.py +16 -0
- helm/benchmark/metrics/paraphrase_generation_metrics.py +3 -4
- helm/benchmark/metrics/ranking_metrics.py +2 -2
- helm/benchmark/metrics/reference_metric.py +148 -0
- helm/benchmark/metrics/summac/model_summac.py +0 -2
- helm/benchmark/metrics/summarization_metrics.py +2 -2
- helm/benchmark/metrics/test_classification_metrics.py +8 -5
- helm/benchmark/metrics/test_disinformation_metrics.py +78 -0
- helm/benchmark/metrics/{test_basic_metrics.py → test_evaluate_reference_metrics.py} +5 -1
- helm/benchmark/metrics/test_metric.py +2 -2
- helm/benchmark/metrics/tokens/gooseai_token_cost_estimator.py +10 -2
- helm/benchmark/metrics/toxicity_metrics.py +1 -1
- helm/benchmark/metrics/toxicity_utils.py +23 -0
- helm/benchmark/metrics/unitxt_metrics.py +81 -0
- helm/benchmark/metrics/vision_language/__init__.py +0 -0
- helm/benchmark/metrics/vision_language/emd_utils.py +341 -0
- helm/benchmark/metrics/vision_language/image_metrics.py +575 -0
- helm/benchmark/metrics/vision_language/image_utils.py +100 -0
- helm/benchmark/model_deployment_registry.py +74 -0
- helm/benchmark/model_metadata_registry.py +41 -1
- helm/benchmark/multi_gpu_runner.py +133 -0
- helm/benchmark/presentation/create_plots.py +8 -7
- helm/benchmark/presentation/run_display.py +26 -10
- helm/benchmark/presentation/schema.py +15 -40
- helm/benchmark/presentation/summarize.py +119 -79
- helm/benchmark/presentation/table.py +8 -8
- helm/benchmark/presentation/test_contamination.py +2 -2
- helm/benchmark/presentation/test_run_entry.py +1 -2
- helm/benchmark/presentation/test_summarize.py +3 -3
- helm/benchmark/run.py +54 -26
- helm/benchmark/run_expander.py +205 -35
- helm/benchmark/run_spec.py +93 -0
- helm/benchmark/run_spec_factory.py +163 -0
- helm/benchmark/run_specs/__init__.py +0 -0
- helm/benchmark/run_specs/classic_run_specs.py +1510 -0
- helm/benchmark/run_specs/cleva_run_specs.py +277 -0
- helm/benchmark/run_specs/decodingtrust_run_specs.py +314 -0
- helm/benchmark/run_specs/heim_run_specs.py +623 -0
- helm/benchmark/run_specs/instruction_following_run_specs.py +129 -0
- helm/benchmark/run_specs/lite_run_specs.py +307 -0
- helm/benchmark/run_specs/simple_run_specs.py +104 -0
- helm/benchmark/run_specs/unitxt_run_specs.py +42 -0
- helm/benchmark/run_specs/vlm_run_specs.py +757 -0
- helm/benchmark/runner.py +51 -57
- helm/benchmark/runner_config_registry.py +21 -0
- helm/benchmark/scenarios/bbq_scenario.py +1 -1
- helm/benchmark/scenarios/bold_scenario.py +2 -2
- helm/benchmark/scenarios/code_scenario.py +1 -0
- helm/benchmark/scenarios/decodingtrust_adv_demonstration_scenario.py +169 -0
- helm/benchmark/scenarios/decodingtrust_adv_robustness_scenario.py +121 -0
- helm/benchmark/scenarios/decodingtrust_fairness_scenario.py +77 -0
- helm/benchmark/scenarios/decodingtrust_machine_ethics_scenario.py +324 -0
- helm/benchmark/scenarios/decodingtrust_ood_robustness_scenario.py +204 -0
- helm/benchmark/scenarios/decodingtrust_privacy_scenario.py +559 -0
- helm/benchmark/scenarios/decodingtrust_stereotype_bias_scenario.py +67 -0
- helm/benchmark/scenarios/decodingtrust_toxicity_prompts_scenario.py +78 -0
- helm/benchmark/scenarios/dialogue_scenarios.py +0 -1
- helm/benchmark/scenarios/image_generation/__init__.py +0 -0
- helm/benchmark/scenarios/image_generation/common_syntactic_processes_scenario.py +105 -0
- helm/benchmark/scenarios/image_generation/cub200_scenario.py +95 -0
- helm/benchmark/scenarios/image_generation/daily_dalle_scenario.py +124 -0
- helm/benchmark/scenarios/image_generation/demographic_stereotypes_scenario.py +82 -0
- helm/benchmark/scenarios/image_generation/detection_scenario.py +83 -0
- helm/benchmark/scenarios/image_generation/draw_bench_scenario.py +74 -0
- helm/benchmark/scenarios/image_generation/i2p_scenario.py +57 -0
- helm/benchmark/scenarios/image_generation/landing_page_scenario.py +46 -0
- helm/benchmark/scenarios/image_generation/logos_scenario.py +223 -0
- helm/benchmark/scenarios/image_generation/magazine_cover_scenario.py +91 -0
- helm/benchmark/scenarios/image_generation/mental_disorders_scenario.py +46 -0
- helm/benchmark/scenarios/image_generation/mscoco_scenario.py +91 -0
- helm/benchmark/scenarios/image_generation/paint_skills_scenario.py +72 -0
- helm/benchmark/scenarios/image_generation/parti_prompts_scenario.py +94 -0
- helm/benchmark/scenarios/image_generation/radiology_scenario.py +42 -0
- helm/benchmark/scenarios/image_generation/relational_understanding_scenario.py +52 -0
- helm/benchmark/scenarios/image_generation/time_most_significant_historical_figures_scenario.py +124 -0
- helm/benchmark/scenarios/image_generation/winoground_scenario.py +62 -0
- helm/benchmark/scenarios/imdb_scenario.py +0 -1
- helm/benchmark/scenarios/legalbench_scenario.py +6 -2
- helm/benchmark/scenarios/live_qa_scenario.py +94 -0
- helm/benchmark/scenarios/lm_entry_scenario.py +185 -0
- helm/benchmark/scenarios/math_scenario.py +19 -2
- helm/benchmark/scenarios/medication_qa_scenario.py +60 -0
- helm/benchmark/scenarios/numeracy_scenario.py +1 -1
- helm/benchmark/scenarios/opinions_qa_scenario.py +0 -4
- helm/benchmark/scenarios/scenario.py +4 -0
- helm/benchmark/scenarios/simple_scenarios.py +122 -1
- helm/benchmark/scenarios/test_math_scenario.py +6 -0
- helm/benchmark/scenarios/test_scenario.py +6 -3
- helm/benchmark/scenarios/test_simple_scenarios.py +50 -0
- helm/benchmark/scenarios/thai_exam_scenario.py +135 -0
- helm/benchmark/scenarios/unitxt_scenario.py +56 -0
- helm/benchmark/scenarios/verifiability_judgment_scenario.py +3 -1
- helm/benchmark/scenarios/vicuna_scenario.py +1 -1
- helm/benchmark/scenarios/vision_language/a_okvqa_scenario.py +83 -0
- helm/benchmark/scenarios/vision_language/bingo_scenario.py +103 -0
- helm/benchmark/scenarios/vision_language/crossmodal_3600_scenario.py +134 -0
- helm/benchmark/scenarios/vision_language/flickr30k_scenario.py +74 -0
- helm/benchmark/scenarios/vision_language/gqa_scenario.py +91 -0
- helm/benchmark/scenarios/vision_language/hateful_memes_scenario.py +94 -0
- helm/benchmark/scenarios/vision_language/heim_human_eval_scenario.py +113 -0
- helm/benchmark/scenarios/vision_language/image2structure/__init__.py +0 -0
- helm/benchmark/scenarios/vision_language/image2structure/chart2csv_scenario.py +55 -0
- helm/benchmark/scenarios/vision_language/image2structure/image2structure_scenario.py +214 -0
- helm/benchmark/scenarios/vision_language/image2structure/latex_scenario.py +25 -0
- helm/benchmark/scenarios/vision_language/image2structure/musicsheet_scenario.py +20 -0
- helm/benchmark/scenarios/vision_language/image2structure/utils_latex.py +347 -0
- helm/benchmark/scenarios/vision_language/image2structure/webpage/__init__.py +0 -0
- helm/benchmark/scenarios/vision_language/image2structure/webpage/driver.py +84 -0
- helm/benchmark/scenarios/vision_language/image2structure/webpage/jekyll_server.py +182 -0
- helm/benchmark/scenarios/vision_language/image2structure/webpage/utils.py +31 -0
- helm/benchmark/scenarios/vision_language/image2structure/webpage_scenario.py +225 -0
- helm/benchmark/scenarios/vision_language/math_vista_scenario.py +117 -0
- helm/benchmark/scenarios/vision_language/mementos_scenario.py +124 -0
- helm/benchmark/scenarios/vision_language/mm_safety_bench_scenario.py +103 -0
- helm/benchmark/scenarios/vision_language/mme_scenario.py +145 -0
- helm/benchmark/scenarios/vision_language/mmmu_scenario.py +187 -0
- helm/benchmark/scenarios/vision_language/mscoco_captioning_scenario.py +92 -0
- helm/benchmark/scenarios/vision_language/mscoco_categorization_scenario.py +117 -0
- helm/benchmark/scenarios/vision_language/multipanelvqa_scenario.py +169 -0
- helm/benchmark/scenarios/vision_language/originality_scenario.py +35 -0
- helm/benchmark/scenarios/vision_language/pairs_scenario.py +246 -0
- helm/benchmark/scenarios/vision_language/pope_scenario.py +104 -0
- helm/benchmark/scenarios/vision_language/seed_bench_scenario.py +129 -0
- helm/benchmark/scenarios/vision_language/unicorn_scenario.py +108 -0
- helm/benchmark/scenarios/vision_language/viz_wiz_scenario.py +3 -4
- helm/benchmark/scenarios/vision_language/vqa_scenario.py +5 -3
- helm/benchmark/scenarios/wmt_14_scenario.py +1 -1
- helm/benchmark/server.py +24 -1
- helm/benchmark/slurm_runner.py +70 -49
- helm/benchmark/static/benchmarking.js +1 -1
- helm/benchmark/static/schema_classic.yaml +258 -1066
- helm/benchmark/static/schema_image2structure.yaml +304 -0
- helm/benchmark/static/schema_instruction_following.yaml +210 -0
- helm/benchmark/static/schema_lite.yaml +2 -227
- helm/benchmark/static/schema_mmlu.yaml +1507 -0
- helm/benchmark/static/schema_unitxt.yaml +428 -0
- helm/benchmark/static/schema_vhelm_lite.yaml +164 -0
- helm/benchmark/static/schema_vlm.yaml +823 -0
- helm/benchmark/static_build/assets/01-694cb9b7.png +0 -0
- helm/benchmark/static_build/assets/ai21-0eb91ec3.png +0 -0
- helm/benchmark/static_build/assets/aleph-alpha-7ce10034.png +0 -0
- helm/benchmark/static_build/assets/anthropic-70d8bc39.png +0 -0
- helm/benchmark/static_build/assets/bigscience-7f0400c0.png +0 -0
- helm/benchmark/static_build/assets/cohere-3550c6cb.png +0 -0
- helm/benchmark/static_build/assets/crfm-logo-74391ab8.png +0 -0
- helm/benchmark/static_build/assets/eleutherai-b9451114.png +0 -0
- helm/benchmark/static_build/assets/google-06d997ad.png +0 -0
- helm/benchmark/static_build/assets/heim-logo-3e5e3aa4.png +0 -0
- helm/benchmark/static_build/assets/helm-logo-simple-2ed5400b.png +0 -0
- helm/benchmark/static_build/assets/helmhero-28e90f4d.png +0 -0
- helm/benchmark/static_build/assets/index-737eef9e.js +10 -0
- helm/benchmark/static_build/assets/index-878a1094.css +1 -0
- helm/benchmark/static_build/assets/meta-5580e9f1.png +0 -0
- helm/benchmark/static_build/assets/microsoft-f5ee5016.png +0 -0
- helm/benchmark/static_build/assets/mistral-18e1be23.png +0 -0
- helm/benchmark/static_build/assets/nvidia-86fa75c1.png +0 -0
- helm/benchmark/static_build/assets/openai-3f8653e4.png +0 -0
- helm/benchmark/static_build/assets/react-d4a0b69b.js +85 -0
- helm/benchmark/static_build/assets/recharts-6d337683.js +97 -0
- helm/benchmark/static_build/assets/tii-24de195c.png +0 -0
- helm/benchmark/static_build/assets/together-a665a35b.png +0 -0
- helm/benchmark/static_build/assets/tremor-54a99cc4.js +10 -0
- helm/benchmark/static_build/assets/tsinghua-keg-97d4b395.png +0 -0
- helm/benchmark/static_build/assets/vhelm-framework-cde7618a.png +0 -0
- helm/benchmark/static_build/assets/vhelm-model-6d812526.png +0 -0
- helm/benchmark/static_build/assets/yandex-38e09d70.png +0 -0
- helm/benchmark/static_build/config.js +4 -0
- helm/benchmark/static_build/index.html +20 -0
- helm/benchmark/test_data_preprocessor.py +3 -3
- helm/benchmark/test_run_expander.py +1 -1
- helm/benchmark/window_services/ai21_window_service.py +22 -33
- helm/benchmark/window_services/cohere_window_service.py +1 -63
- helm/benchmark/window_services/default_window_service.py +2 -44
- helm/benchmark/window_services/encoder_decoder_window_service.py +0 -11
- helm/benchmark/window_services/ice_window_service.py +0 -34
- helm/benchmark/window_services/image_generation/__init__.py +0 -0
- helm/benchmark/window_services/image_generation/clip_window_service.py +15 -0
- helm/benchmark/window_services/image_generation/lexica_search_window_service.py +9 -0
- helm/benchmark/window_services/image_generation/openai_dalle_window_service.py +9 -0
- helm/benchmark/window_services/image_generation/test_clip_window_service.py +29 -0
- helm/benchmark/window_services/image_generation/test_openai_dalle_window_service.py +30 -0
- helm/benchmark/window_services/local_window_service.py +21 -4
- helm/benchmark/window_services/test_anthropic_window_service.py +2 -1
- helm/benchmark/window_services/test_bloom_window_service.py +2 -1
- helm/benchmark/window_services/test_cohere_window_service.py +2 -1
- helm/benchmark/window_services/test_flan_t5_window_service.py +2 -1
- helm/benchmark/window_services/test_gpt2_window_service.py +2 -2
- helm/benchmark/window_services/test_gpt4_window_service.py +2 -1
- helm/benchmark/window_services/test_gptj_window_service.py +3 -2
- helm/benchmark/window_services/test_gptneox_window_service.py +3 -2
- helm/benchmark/window_services/test_ice_window_service.py +2 -1
- helm/benchmark/window_services/test_openai_window_service.py +2 -1
- helm/benchmark/window_services/test_opt_window_service.py +3 -2
- helm/benchmark/window_services/test_palmyra_window_service.py +2 -1
- helm/benchmark/window_services/test_t0pp_window_service.py +2 -1
- helm/benchmark/window_services/test_t511b_window_service.py +2 -1
- helm/benchmark/window_services/test_ul2_window_service.py +2 -1
- helm/benchmark/window_services/test_utils.py +3 -2
- helm/benchmark/window_services/test_yalm_window_service.py +2 -1
- helm/benchmark/window_services/window_service.py +42 -0
- helm/benchmark/window_services/window_service_factory.py +4 -1
- helm/benchmark/window_services/yalm_window_service.py +0 -27
- helm/clients/__init__.py +0 -0
- helm/{proxy/clients → clients}/ai21_client.py +3 -9
- helm/clients/aleph_alpha_client.py +112 -0
- helm/{proxy/clients → clients}/anthropic_client.py +233 -18
- helm/{proxy/clients → clients}/auto_client.py +59 -31
- helm/clients/bedrock_client.py +128 -0
- helm/clients/bedrock_utils.py +72 -0
- helm/{proxy/clients → clients}/client.py +65 -7
- helm/clients/clip_score_client.py +49 -0
- helm/clients/clip_scorers/__init__.py +0 -0
- helm/clients/clip_scorers/base_clip_scorer.py +18 -0
- helm/clients/clip_scorers/clip_scorer.py +50 -0
- helm/clients/clip_scorers/multilingual_clip_scorer.py +50 -0
- helm/{proxy/clients → clients}/cohere_client.py +4 -11
- helm/clients/gcs_client.py +82 -0
- helm/{proxy/clients → clients}/google_client.py +5 -5
- helm/clients/google_translate_client.py +35 -0
- helm/{proxy/clients → clients}/http_model_client.py +5 -7
- helm/{proxy/clients → clients}/huggingface_client.py +43 -64
- helm/clients/image_generation/__init__.py +0 -0
- helm/clients/image_generation/adobe_vision_client.py +78 -0
- helm/clients/image_generation/aleph_alpha_image_generation_client.py +98 -0
- helm/clients/image_generation/cogview2/__init__.py +0 -0
- helm/clients/image_generation/cogview2/coglm_strategy.py +96 -0
- helm/clients/image_generation/cogview2/coglm_utils.py +82 -0
- helm/clients/image_generation/cogview2/sr_pipeline/__init__.py +15 -0
- helm/clients/image_generation/cogview2/sr_pipeline/direct_sr.py +96 -0
- helm/clients/image_generation/cogview2/sr_pipeline/dsr_model.py +254 -0
- helm/clients/image_generation/cogview2/sr_pipeline/dsr_sampling.py +190 -0
- helm/clients/image_generation/cogview2/sr_pipeline/iterative_sr.py +141 -0
- helm/clients/image_generation/cogview2/sr_pipeline/itersr_model.py +269 -0
- helm/clients/image_generation/cogview2/sr_pipeline/itersr_sampling.py +120 -0
- helm/clients/image_generation/cogview2/sr_pipeline/sr_group.py +42 -0
- helm/clients/image_generation/cogview2_client.py +191 -0
- helm/clients/image_generation/dalle2_client.py +192 -0
- helm/clients/image_generation/dalle3_client.py +108 -0
- helm/clients/image_generation/dalle_mini/__init__.py +3 -0
- helm/clients/image_generation/dalle_mini/data.py +442 -0
- helm/clients/image_generation/dalle_mini/model/__init__.py +5 -0
- helm/clients/image_generation/dalle_mini/model/configuration.py +175 -0
- helm/clients/image_generation/dalle_mini/model/modeling.py +1834 -0
- helm/clients/image_generation/dalle_mini/model/partitions.py +84 -0
- helm/clients/image_generation/dalle_mini/model/processor.py +63 -0
- helm/clients/image_generation/dalle_mini/model/text.py +251 -0
- helm/clients/image_generation/dalle_mini/model/tokenizer.py +9 -0
- helm/clients/image_generation/dalle_mini/model/utils.py +29 -0
- helm/clients/image_generation/dalle_mini/vqgan_jax/__init__.py +1 -0
- helm/clients/image_generation/dalle_mini/vqgan_jax/configuration_vqgan.py +40 -0
- helm/clients/image_generation/dalle_mini/vqgan_jax/convert_pt_model_to_jax.py +107 -0
- helm/clients/image_generation/dalle_mini/vqgan_jax/modeling_flax_vqgan.py +610 -0
- helm/clients/image_generation/dalle_mini_client.py +190 -0
- helm/clients/image_generation/deep_floyd_client.py +78 -0
- helm/clients/image_generation/huggingface_diffusers_client.py +249 -0
- helm/clients/image_generation/image_generation_client_utils.py +9 -0
- helm/clients/image_generation/lexica_client.py +86 -0
- helm/clients/image_generation/mindalle/__init__.py +0 -0
- helm/clients/image_generation/mindalle/models/__init__.py +216 -0
- helm/clients/image_generation/mindalle/models/stage1/__init__.py +0 -0
- helm/clients/image_generation/mindalle/models/stage1/layers.py +312 -0
- helm/clients/image_generation/mindalle/models/stage1/vqgan.py +103 -0
- helm/clients/image_generation/mindalle/models/stage2/__init__.py +0 -0
- helm/clients/image_generation/mindalle/models/stage2/layers.py +144 -0
- helm/clients/image_generation/mindalle/models/stage2/transformer.py +268 -0
- helm/clients/image_generation/mindalle/models/tokenizer.py +30 -0
- helm/clients/image_generation/mindalle/utils/__init__.py +3 -0
- helm/clients/image_generation/mindalle/utils/config.py +129 -0
- helm/clients/image_generation/mindalle/utils/sampling.py +149 -0
- helm/clients/image_generation/mindalle/utils/utils.py +89 -0
- helm/clients/image_generation/mindalle_client.py +115 -0
- helm/clients/image_generation/nudity_check_client.py +64 -0
- helm/clients/image_generation/together_image_generation_client.py +111 -0
- helm/{proxy/clients → clients}/lit_gpt_client.py +4 -4
- helm/{proxy/clients → clients}/megatron_client.py +5 -5
- helm/clients/mistral_client.py +134 -0
- helm/clients/moderation_api_client.py +109 -0
- helm/clients/open_lm_client.py +43 -0
- helm/clients/openai_client.py +301 -0
- helm/{proxy/clients → clients}/palmyra_client.py +6 -8
- helm/{proxy/clients → clients}/perspective_api_client.py +7 -8
- helm/clients/simple_client.py +64 -0
- helm/{proxy/clients → clients}/test_auto_client.py +13 -15
- helm/clients/test_client.py +100 -0
- helm/{proxy/clients → clients}/test_huggingface_client.py +15 -16
- helm/clients/test_simple_client.py +19 -0
- helm/{proxy/clients → clients}/test_together_client.py +20 -8
- helm/{proxy/clients → clients}/together_client.py +104 -73
- helm/clients/vertexai_client.py +400 -0
- helm/clients/vision_language/__init__.py +0 -0
- helm/clients/vision_language/huggingface_vision2seq_client.py +145 -0
- helm/clients/vision_language/huggingface_vlm_client.py +111 -0
- helm/{proxy/clients → clients}/vision_language/idefics_client.py +54 -49
- helm/clients/vision_language/open_flamingo/__init__.py +2 -0
- helm/clients/vision_language/open_flamingo/src/__init__.py +0 -0
- helm/clients/vision_language/open_flamingo/src/factory.py +147 -0
- helm/clients/vision_language/open_flamingo/src/flamingo.py +337 -0
- helm/clients/vision_language/open_flamingo/src/flamingo_lm.py +155 -0
- helm/clients/vision_language/open_flamingo/src/helpers.py +267 -0
- helm/clients/vision_language/open_flamingo/src/utils.py +47 -0
- helm/clients/vision_language/open_flamingo_client.py +155 -0
- helm/clients/vision_language/qwen_vlm_client.py +171 -0
- helm/clients/vllm_client.py +46 -0
- helm/common/cache.py +16 -4
- helm/common/cache_backend_config.py +47 -0
- helm/common/clip_score_request.py +41 -0
- helm/common/file_caches/__init__.py +0 -0
- helm/common/file_caches/file_cache.py +16 -0
- helm/common/file_caches/local_file_cache.py +61 -0
- helm/common/file_caches/test_local_file_cache.py +25 -0
- helm/common/file_upload_request.py +27 -0
- helm/common/general.py +1 -1
- helm/common/image_generation_parameters.py +25 -0
- helm/common/images_utils.py +33 -3
- helm/common/key_value_store.py +35 -4
- helm/common/media_object.py +13 -0
- helm/common/moderations_api_request.py +71 -0
- helm/common/mongo_key_value_store.py +3 -3
- helm/common/multimodal_request_utils.py +31 -0
- helm/common/nudity_check_request.py +29 -0
- helm/common/request.py +15 -17
- helm/common/test_general.py +6 -0
- helm/common/tokenization_request.py +1 -1
- helm/config/model_deployments.yaml +1159 -538
- helm/config/model_metadata.yaml +868 -41
- helm/config/tokenizer_configs.yaml +149 -43
- helm/proxy/accounts.py +31 -4
- helm/proxy/critique/mechanical_turk_critique_importer.py +3 -0
- helm/proxy/critique/model_critique_client.py +8 -6
- helm/proxy/example_queries.py +29 -17
- helm/proxy/server.py +70 -5
- helm/proxy/services/remote_service.py +31 -0
- helm/proxy/services/server_service.py +96 -16
- helm/proxy/services/service.py +30 -0
- helm/proxy/services/test_remote_service.py +4 -3
- helm/proxy/services/test_service.py +0 -12
- helm/proxy/test_accounts.py +32 -0
- helm/proxy/token_counters/auto_token_counter.py +37 -37
- helm/proxy/token_counters/test_auto_token_counter.py +164 -0
- helm/proxy/token_counters/token_counter.py +3 -5
- helm/tokenizers/__init__.py +0 -0
- helm/{proxy/tokenizers → tokenizers}/ai21_tokenizer.py +3 -3
- helm/{proxy/tokenizers → tokenizers}/anthropic_tokenizer.py +1 -1
- helm/{proxy/tokenizers → tokenizers}/auto_tokenizer.py +6 -9
- helm/{proxy/tokenizers → tokenizers}/cohere_tokenizer.py +1 -1
- helm/{proxy/tokenizers → tokenizers}/http_model_tokenizer.py +3 -3
- helm/{proxy/tokenizers → tokenizers}/huggingface_tokenizer.py +7 -26
- helm/tokenizers/simple_tokenizer.py +33 -0
- helm/{proxy/tokenizers → tokenizers}/test_anthropic_tokenizer.py +1 -1
- helm/{proxy/tokenizers → tokenizers}/test_huggingface_tokenizer.py +3 -0
- helm/tokenizers/test_simple_tokenizer.py +33 -0
- helm/{proxy/tokenizers → tokenizers}/vertexai_tokenizer.py +1 -1
- helm/{proxy/tokenizers → tokenizers}/yalm_tokenizer.py +5 -3
- helm/tokenizers/yalm_tokenizer_data/__init__.py +0 -0
- helm/tokenizers/yalm_tokenizer_data/voc_100b.sp +0 -0
- helm/{proxy/tokenizers → tokenizers}/yalm_tokenizer_data/yalm_tokenizer.py +1 -1
- crfm_helm-0.4.0.dist-info/RECORD +0 -397
- helm/benchmark/run_specs.py +0 -2762
- helm/benchmark/test_model_deployment_definition.py +0 -92
- helm/benchmark/test_model_properties.py +0 -1570
- helm/benchmark/vlm_run_specs.py +0 -97
- helm/benchmark/window_services/flan_t5_window_service.py +0 -29
- helm/benchmark/window_services/gpt2_window_service.py +0 -32
- helm/benchmark/window_services/huggingface_window_service.py +0 -60
- helm/benchmark/window_services/t0pp_window_service.py +0 -35
- helm/benchmark/window_services/t511b_window_service.py +0 -30
- helm/benchmark/window_services/test_mt_nlg_window_service.py +0 -48
- helm/benchmark/window_services/ul2_window_service.py +0 -30
- helm/benchmark/window_services/wider_ai21_window_service.py +0 -24
- helm/common/cache_utils.py +0 -14
- helm/proxy/clients/aleph_alpha_client.py +0 -95
- helm/proxy/clients/goose_ai_client.py +0 -99
- helm/proxy/clients/microsoft_client.py +0 -180
- helm/proxy/clients/openai_client.py +0 -206
- helm/proxy/clients/simple_client.py +0 -60
- helm/proxy/clients/test_client.py +0 -49
- helm/proxy/clients/vertexai_client.py +0 -115
- helm/proxy/token_counters/ai21_token_counter.py +0 -20
- helm/proxy/token_counters/cohere_token_counter.py +0 -13
- helm/proxy/token_counters/free_token_counter.py +0 -12
- helm/proxy/token_counters/gooseai_token_counter.py +0 -24
- helm/proxy/token_counters/openai_token_counter.py +0 -22
- helm/proxy/token_counters/test_ai21_token_counter.py +0 -88
- helm/proxy/token_counters/test_openai_token_counter.py +0 -81
- helm/proxy/tokenizers/simple_tokenizer.py +0 -32
- {crfm_helm-0.4.0.dist-info → crfm_helm-0.5.1.dist-info}/LICENSE +0 -0
- {crfm_helm-0.4.0.dist-info → crfm_helm-0.5.1.dist-info}/entry_points.txt +0 -0
- {crfm_helm-0.4.0.dist-info → crfm_helm-0.5.1.dist-info}/top_level.txt +0 -0
- /helm/{proxy/clients → benchmark/annotation}/__init__.py +0 -0
- /helm/{proxy/clients/vision_language → benchmark/annotation/image2structure}/__init__.py +0 -0
- /helm/{proxy/tokenizers → benchmark/metrics/image_generation}/__init__.py +0 -0
- /helm/{proxy/tokenizers/yalm_tokenizer_data → benchmark/metrics/image_generation/detectors}/__init__.py +0 -0
- /helm/{proxy/clients → clients}/ai21_utils.py +0 -0
- /helm/{proxy/clients → clients}/cohere_utils.py +0 -0
- /helm/{proxy/clients → clients}/lit_gpt_generate.py +0 -0
- /helm/{proxy/clients → clients}/toxicity_classifier_client.py +0 -0
- /helm/{proxy/tokenizers → tokenizers}/aleph_alpha_tokenizer.py +0 -0
- /helm/{proxy/tokenizers → tokenizers}/caching_tokenizer.py +0 -0
- /helm/{proxy/tokenizers → tokenizers}/ice_tokenizer.py +0 -0
- /helm/{proxy/tokenizers → tokenizers}/lit_gpt_tokenizer.py +0 -0
- /helm/{proxy/tokenizers → tokenizers}/test_ice_tokenizer.py +0 -0
- /helm/{proxy/tokenizers → tokenizers}/test_yalm_tokenizer.py +0 -0
- /helm/{proxy/tokenizers → tokenizers}/tiktoken_tokenizer.py +0 -0
- /helm/{proxy/tokenizers → tokenizers}/tokenizer.py +0 -0
- /helm/{proxy/tokenizers → tokenizers}/yalm_tokenizer_data/test_yalm_tokenizer.py +0 -0
|
Binary file
|
|
Binary file
|
|
Binary file
|
|
Binary file
|
|
@@ -0,0 +1,20 @@
|
|
|
1
|
+
<!doctype html>
|
|
2
|
+
<html lang="en">
|
|
3
|
+
<head>
|
|
4
|
+
<meta charset="UTF-8" />
|
|
5
|
+
<link rel="icon" type="image/svg+xml" href="./helm.svg" />
|
|
6
|
+
<meta name="viewport" content="width=device-width, initial-scale=1.0" />
|
|
7
|
+
<title>Holistic Evaluation of Language Models (HELM)</title>
|
|
8
|
+
<meta name="description" content="The Holistic Evaluation of Language Models (HELM) serves as a living benchmark for transparency in language models. Providing broad coverage and recognizing incompleteness, multi-metric measurements, and standardization. All data and analysis are freely accessible on the website for exploration and study." />
|
|
9
|
+
<script type="text/javascript" src="./config.js"></script>
|
|
10
|
+
<script type="module" crossorigin src="./assets/index-737eef9e.js"></script>
|
|
11
|
+
<link rel="modulepreload" crossorigin href="./assets/react-d4a0b69b.js">
|
|
12
|
+
<link rel="modulepreload" crossorigin href="./assets/recharts-6d337683.js">
|
|
13
|
+
<link rel="modulepreload" crossorigin href="./assets/tremor-54a99cc4.js">
|
|
14
|
+
<link rel="stylesheet" href="./assets/index-878a1094.css">
|
|
15
|
+
</head>
|
|
16
|
+
<body class="block">
|
|
17
|
+
<div id="root"></div>
|
|
18
|
+
|
|
19
|
+
</body>
|
|
20
|
+
</html>
|
|
@@ -4,14 +4,14 @@ from typing import List
|
|
|
4
4
|
from helm.benchmark.augmentations.data_augmenter import DataAugmenterSpec
|
|
5
5
|
from helm.benchmark.augmentations.perturbation import PerturbationSpec
|
|
6
6
|
from helm.benchmark.data_preprocessor import DataPreprocessor
|
|
7
|
-
from helm.benchmark.run_specs import
|
|
7
|
+
from helm.benchmark.run_specs.simple_run_specs import get_simple1_spec
|
|
8
8
|
from helm.benchmark.scenarios.scenario import create_scenario, Instance, Scenario, with_instance_ids
|
|
9
9
|
|
|
10
10
|
|
|
11
11
|
def test_data_preprocessor():
|
|
12
12
|
# Test that each Instance is given a unique ID and is preserved through data augmentation
|
|
13
13
|
data_preprocessor = DataPreprocessor(DataAugmenterSpec())
|
|
14
|
-
scenario: Scenario = create_scenario(
|
|
14
|
+
scenario: Scenario = create_scenario(get_simple1_spec().scenario_spec)
|
|
15
15
|
instances = with_instance_ids(scenario.get_instances(output_path=""))
|
|
16
16
|
instances: List[Instance] = data_preprocessor.preprocess(instances)
|
|
17
17
|
for i, instance in enumerate(instances):
|
|
@@ -32,7 +32,7 @@ def test_data_preprocessor_with_data_augmentation():
|
|
|
32
32
|
should_include_original_eval=True,
|
|
33
33
|
)
|
|
34
34
|
data_preprocessor = DataPreprocessor(data_augmenter_spec)
|
|
35
|
-
scenario: Scenario = create_scenario(
|
|
35
|
+
scenario: Scenario = create_scenario(get_simple1_spec().scenario_spec)
|
|
36
36
|
instances = with_instance_ids(scenario.get_instances(output_path=""))
|
|
37
37
|
instances: List[Instance] = data_preprocessor.preprocess(instances)
|
|
38
38
|
assert len(instances) == 10 + 10 + 10 # original train + original eval + perturbed eval
|
|
@@ -2,7 +2,7 @@ import unittest
|
|
|
2
2
|
|
|
3
3
|
from helm.benchmark.adaptation.adapter_spec import AdapterSpec
|
|
4
4
|
from helm.benchmark.run_expander import IncreaseMaxTokensRunExpander
|
|
5
|
-
from helm.benchmark.
|
|
5
|
+
from helm.benchmark.run_spec import RunSpec
|
|
6
6
|
from helm.benchmark.scenarios.scenario import ScenarioSpec
|
|
7
7
|
|
|
8
8
|
|
|
@@ -9,12 +9,11 @@ from helm.common.tokenization_request import (
|
|
|
9
9
|
TokenizationToken,
|
|
10
10
|
TextRange,
|
|
11
11
|
)
|
|
12
|
-
from .window_service import
|
|
12
|
+
from .window_service import ConfigurableWindowService, EncodeResult, WindowService
|
|
13
13
|
from .tokenizer_service import TokenizerService
|
|
14
|
-
from .gpt2_window_service import GPT2WindowService
|
|
15
14
|
|
|
16
15
|
|
|
17
|
-
class AI21WindowService(
|
|
16
|
+
class AI21WindowService(ConfigurableWindowService):
|
|
18
17
|
"""Tokenizes by making a request to the proxy server with REST endpoint: `/api/tokenize`."""
|
|
19
18
|
|
|
20
19
|
# AI21's tokenizer API rejects a tokenization request if the input sequence is too long, so
|
|
@@ -32,39 +31,29 @@ class AI21WindowService(WindowService):
|
|
|
32
31
|
"AI21 only gave API access to their tokenizer, so this method is not supported."
|
|
33
32
|
)
|
|
34
33
|
|
|
35
|
-
def __init__(
|
|
34
|
+
def __init__(
|
|
35
|
+
self,
|
|
36
|
+
gpt2_window_service: WindowService,
|
|
37
|
+
service: TokenizerService,
|
|
38
|
+
tokenizer_name: str,
|
|
39
|
+
max_sequence_length: int,
|
|
40
|
+
max_request_length: Optional[int] = None,
|
|
41
|
+
max_sequence_and_generated_tokens_length: Optional[int] = None,
|
|
42
|
+
end_of_text_token: Optional[str] = None,
|
|
43
|
+
prefix_token: Optional[str] = None,
|
|
44
|
+
):
|
|
45
|
+
super().__init__(
|
|
46
|
+
tokenizer_name=tokenizer_name,
|
|
47
|
+
max_sequence_length=max_sequence_length,
|
|
48
|
+
max_request_length=max_request_length,
|
|
49
|
+
max_sequence_and_generated_tokens_length=max_sequence_and_generated_tokens_length,
|
|
50
|
+
end_of_text_token=end_of_text_token,
|
|
51
|
+
prefix_token=prefix_token,
|
|
52
|
+
)
|
|
36
53
|
# We need the `TokenizerService` to make requests to the server.
|
|
37
54
|
self.service: TokenizerService = service
|
|
38
55
|
# As explained above, we need a `GPT2WindowService` to help tokenize long text sequences.
|
|
39
|
-
self.gpt2_window_service:
|
|
40
|
-
|
|
41
|
-
@property
|
|
42
|
-
def tokenizer_name(self) -> str:
|
|
43
|
-
"""Name of the tokenizer to use when sending a request."""
|
|
44
|
-
return "ai21/j1"
|
|
45
|
-
|
|
46
|
-
@property
|
|
47
|
-
def max_sequence_length(self) -> int:
|
|
48
|
-
"""
|
|
49
|
-
The max token length of the model in. The AI21 server automatically prepends a token to every prompt,
|
|
50
|
-
so the actual max sequence length is 2048-1 = 2047.
|
|
51
|
-
"""
|
|
52
|
-
return 2047
|
|
53
|
-
|
|
54
|
-
@property
|
|
55
|
-
def max_request_length(self) -> int:
|
|
56
|
-
"""The max sequence length is the same as the max request length for AI21."""
|
|
57
|
-
return self.max_sequence_length
|
|
58
|
-
|
|
59
|
-
@property
|
|
60
|
-
def end_of_text_token(self) -> str:
|
|
61
|
-
# TODO: I'm not sure what their end of text token is. I don't think it's documented.
|
|
62
|
-
return " "
|
|
63
|
-
|
|
64
|
-
@property
|
|
65
|
-
def prefix_token(self) -> str:
|
|
66
|
-
"""AI21 tokenizers do no have a prefix token"""
|
|
67
|
-
return ""
|
|
56
|
+
self.gpt2_window_service: WindowService = gpt2_window_service
|
|
68
57
|
|
|
69
58
|
def encode(self, text: str, truncation: bool = False, max_length: Optional[int] = None) -> EncodeResult:
|
|
70
59
|
"""
|
|
@@ -1,8 +1,7 @@
|
|
|
1
1
|
from typing import List, Optional
|
|
2
2
|
|
|
3
|
-
from helm.
|
|
3
|
+
from helm.tokenizers.cohere_tokenizer import CohereTokenizer
|
|
4
4
|
from .local_window_service import LocalWindowService
|
|
5
|
-
from .tokenizer_service import TokenizerService
|
|
6
5
|
from .window_service import EncodeResult
|
|
7
6
|
from helm.common.tokenization_request import (
|
|
8
7
|
TokenizationRequest,
|
|
@@ -12,47 +11,6 @@ from helm.common.tokenization_request import (
|
|
|
12
11
|
|
|
13
12
|
|
|
14
13
|
class CohereWindowService(LocalWindowService):
|
|
15
|
-
def __init__(self, service: TokenizerService):
|
|
16
|
-
super().__init__(service)
|
|
17
|
-
|
|
18
|
-
@property
|
|
19
|
-
def tokenizer_name(self) -> str:
|
|
20
|
-
return "cohere/cohere"
|
|
21
|
-
|
|
22
|
-
@property
|
|
23
|
-
def max_sequence_length(self) -> int:
|
|
24
|
-
"""
|
|
25
|
-
The max length of the model input. Similar to MT-NLG, Cohere does not predict the logprob of
|
|
26
|
-
the first input token so `max_sequence_length` is one token shorter than `max_request_length`.
|
|
27
|
-
"""
|
|
28
|
-
return self.max_request_length - 1
|
|
29
|
-
|
|
30
|
-
@property
|
|
31
|
-
def max_request_length(self) -> int:
|
|
32
|
-
"""
|
|
33
|
-
The max request length of the model. For Cohere, this is the same as the `max_sequence_length`.
|
|
34
|
-
If we exceed the `max_sequence_length`, we get the following error:
|
|
35
|
-
|
|
36
|
-
Request failed with too many tokens: total number of tokens (prompt and prediction) cannot
|
|
37
|
-
exceed 2048 - received 2049. Try using a shorter prompt or a smaller max_tokens value.
|
|
38
|
-
"""
|
|
39
|
-
return 2048
|
|
40
|
-
|
|
41
|
-
@property
|
|
42
|
-
def end_of_text_token(self) -> str:
|
|
43
|
-
"""
|
|
44
|
-
The end of text token. Cohere does not have one.
|
|
45
|
-
"""
|
|
46
|
-
return ""
|
|
47
|
-
|
|
48
|
-
@property
|
|
49
|
-
def prefix_token(self) -> str:
|
|
50
|
-
"""
|
|
51
|
-
The prefix token. Cohere does not return the log prob for the first token when `echo_prompt` is True.
|
|
52
|
-
"""
|
|
53
|
-
# Cohere recommended ":", but we can try out different values
|
|
54
|
-
return ":"
|
|
55
|
-
|
|
56
14
|
def encode(self, text: str, truncation: bool = False, max_length: Optional[int] = None) -> EncodeResult:
|
|
57
15
|
"""
|
|
58
16
|
Encodes the input text to tokens.
|
|
@@ -141,23 +99,3 @@ class CohereWindowService(LocalWindowService):
|
|
|
141
99
|
result = result[:-1]
|
|
142
100
|
|
|
143
101
|
return result
|
|
144
|
-
|
|
145
|
-
|
|
146
|
-
class CohereCommandWindowService(CohereWindowService):
|
|
147
|
-
def __init__(self, service: TokenizerService):
|
|
148
|
-
super().__init__(service)
|
|
149
|
-
|
|
150
|
-
@property
|
|
151
|
-
def max_request_length(self) -> int:
|
|
152
|
-
"""
|
|
153
|
-
The max request length of the model. For Cohere, this is the same as the `max_sequence_length`.
|
|
154
|
-
If we exceed the `max_sequence_length`, we get the following error:
|
|
155
|
-
|
|
156
|
-
Request failed with too many tokens: total number of tokens (prompt and prediction) cannot
|
|
157
|
-
exceed 2048 - received 2049. Try using a shorter prompt or a smaller max_tokens value.
|
|
158
|
-
|
|
159
|
-
For the Command model, in rare situations, the co.tokenize returns a shorter list of tokens
|
|
160
|
-
than the co.generate. This causes sequence length errors for rare inputs. Cohere's advice is
|
|
161
|
-
to reduce the sequence length to 2020 to avoid these issues.
|
|
162
|
-
"""
|
|
163
|
-
return 2020
|
|
@@ -1,48 +1,6 @@
|
|
|
1
|
-
from typing import Optional
|
|
2
|
-
from .window_service import INT_MAX
|
|
3
1
|
from .local_window_service import LocalWindowService
|
|
4
|
-
from .tokenizer_service import TokenizerService
|
|
5
2
|
|
|
6
3
|
|
|
7
4
|
class DefaultWindowService(LocalWindowService):
|
|
8
|
-
|
|
9
|
-
|
|
10
|
-
service: TokenizerService,
|
|
11
|
-
tokenizer_name: str,
|
|
12
|
-
max_sequence_length: int,
|
|
13
|
-
max_request_length: Optional[int] = None,
|
|
14
|
-
max_sequence_and_generated_tokens_length: Optional[int] = None,
|
|
15
|
-
end_of_text_token: Optional[str] = None,
|
|
16
|
-
prefix_token: Optional[str] = None,
|
|
17
|
-
):
|
|
18
|
-
super().__init__(service)
|
|
19
|
-
self._tokenizer_name = tokenizer_name
|
|
20
|
-
self._max_sequence_length = max_sequence_length
|
|
21
|
-
self._max_request_length = max_request_length or max_sequence_length
|
|
22
|
-
self._max_sequence_and_generated_tokens_length = max_sequence_and_generated_tokens_length or INT_MAX
|
|
23
|
-
self._end_of_text_token = end_of_text_token or ""
|
|
24
|
-
self._prefix_token = prefix_token or ""
|
|
25
|
-
|
|
26
|
-
@property
|
|
27
|
-
def tokenizer_name(self) -> str:
|
|
28
|
-
return self._tokenizer_name
|
|
29
|
-
|
|
30
|
-
@property
|
|
31
|
-
def max_sequence_length(self) -> int:
|
|
32
|
-
return self._max_sequence_length
|
|
33
|
-
|
|
34
|
-
@property
|
|
35
|
-
def max_request_length(self) -> int:
|
|
36
|
-
return self._max_request_length
|
|
37
|
-
|
|
38
|
-
@property
|
|
39
|
-
def max_sequence_and_generated_tokens_length(self) -> int:
|
|
40
|
-
return self._max_sequence_and_generated_tokens_length
|
|
41
|
-
|
|
42
|
-
@property
|
|
43
|
-
def end_of_text_token(self) -> str:
|
|
44
|
-
return self._end_of_text_token
|
|
45
|
-
|
|
46
|
-
@property
|
|
47
|
-
def prefix_token(self) -> str:
|
|
48
|
-
return self._prefix_token
|
|
5
|
+
# TODO: Delete this WindowService.
|
|
6
|
+
pass
|
|
@@ -2,20 +2,9 @@ from abc import ABC
|
|
|
2
2
|
|
|
3
3
|
from helm.common.hierarchical_logger import hlog
|
|
4
4
|
from .local_window_service import LocalWindowService
|
|
5
|
-
from .tokenizer_service import TokenizerService
|
|
6
5
|
|
|
7
6
|
|
|
8
7
|
class EncoderDecoderWindowService(LocalWindowService, ABC):
|
|
9
|
-
def __init__(self, service: TokenizerService):
|
|
10
|
-
super().__init__(service)
|
|
11
|
-
|
|
12
|
-
@property
|
|
13
|
-
def max_request_length(self) -> int:
|
|
14
|
-
"""
|
|
15
|
-
Return the max request length. We set the max requests length to be `max_sequence_length`.
|
|
16
|
-
"""
|
|
17
|
-
return self.max_sequence_length
|
|
18
|
-
|
|
19
8
|
@property
|
|
20
9
|
def max_output_length(self) -> int:
|
|
21
10
|
"""
|
|
@@ -1,41 +1,7 @@
|
|
|
1
1
|
from .local_window_service import LocalWindowService
|
|
2
|
-
from .tokenizer_service import TokenizerService
|
|
3
2
|
|
|
4
3
|
|
|
5
4
|
class ICEWindowService(LocalWindowService):
|
|
6
|
-
def __init__(self, service: TokenizerService):
|
|
7
|
-
super().__init__(service)
|
|
8
|
-
|
|
9
|
-
@property
|
|
10
|
-
def tokenizer_name(self) -> str:
|
|
11
|
-
return "TsinghuaKEG/ice"
|
|
12
|
-
|
|
13
|
-
@property
|
|
14
|
-
def max_sequence_length(self) -> int:
|
|
15
|
-
"""
|
|
16
|
-
The max length of the model input.
|
|
17
|
-
According to https://github.com/THUDM/GLM-130B, the max sequence length is 2048.
|
|
18
|
-
"""
|
|
19
|
-
return 2048
|
|
20
|
-
|
|
21
|
-
@property
|
|
22
|
-
def max_request_length(self) -> int:
|
|
23
|
-
return self.max_sequence_length + 1
|
|
24
|
-
|
|
25
|
-
@property
|
|
26
|
-
def end_of_text_token(self) -> str:
|
|
27
|
-
"""The end of text token."""
|
|
28
|
-
# Followed up in https://github.com/THUDM/icetk/issues/1
|
|
29
|
-
return "</s>"
|
|
30
|
-
|
|
31
|
-
@property
|
|
32
|
-
def prefix_token(self) -> str:
|
|
33
|
-
"""
|
|
34
|
-
The prefix token.
|
|
35
|
-
Inference with echo=True is not feasible, so just set it to the empty string.
|
|
36
|
-
"""
|
|
37
|
-
return ""
|
|
38
|
-
|
|
39
5
|
def truncate_from_right(self, text: str, expected_completion_token_length: int = 0) -> str:
|
|
40
6
|
"""
|
|
41
7
|
Truncates text from the right to fit within the context window given by `max_request_length`
|
|
File without changes
|
|
@@ -0,0 +1,15 @@
|
|
|
1
|
+
from abc import ABC
|
|
2
|
+
|
|
3
|
+
from helm.benchmark.window_services.local_window_service import LocalWindowService
|
|
4
|
+
|
|
5
|
+
|
|
6
|
+
class CLIPWindowService(LocalWindowService, ABC):
|
|
7
|
+
def truncate_from_right(self, text: str, expected_completion_token_length: int = 0) -> str:
|
|
8
|
+
result: str = self.decode(self.encode(text, truncation=True, max_length=self.max_request_length).tokens)
|
|
9
|
+
|
|
10
|
+
# HACK: For the vast majority of cases, the above logic works, but there are a few where the
|
|
11
|
+
# token count exceeds `max_length` by 1.
|
|
12
|
+
while not self.fits_within_context_window(result):
|
|
13
|
+
result = result[:-1]
|
|
14
|
+
|
|
15
|
+
return result
|
|
@@ -0,0 +1,9 @@
|
|
|
1
|
+
from .clip_window_service import CLIPWindowService
|
|
2
|
+
|
|
3
|
+
|
|
4
|
+
class LexicaSearchWindowService(CLIPWindowService):
|
|
5
|
+
def fits_within_context_window(self, text: str, expected_completion_token_length: int = 0) -> bool:
|
|
6
|
+
return len(text) <= self.max_sequence_length
|
|
7
|
+
|
|
8
|
+
def truncate_from_right(self, text: str, expected_completion_token_length: int = 0) -> str:
|
|
9
|
+
return text[: self.max_sequence_length]
|
|
@@ -0,0 +1,9 @@
|
|
|
1
|
+
from .clip_window_service import CLIPWindowService
|
|
2
|
+
|
|
3
|
+
|
|
4
|
+
class OpenAIDALLEWindowService(CLIPWindowService):
|
|
5
|
+
def fits_within_context_window(self, text: str, expected_completion_token_length: int = 0) -> bool:
|
|
6
|
+
return len(text) <= self.max_sequence_length
|
|
7
|
+
|
|
8
|
+
def truncate_from_right(self, text: str, expected_completion_token_length: int = 0) -> str:
|
|
9
|
+
return text[: self.max_sequence_length]
|
|
@@ -0,0 +1,29 @@
|
|
|
1
|
+
import shutil
|
|
2
|
+
import tempfile
|
|
3
|
+
|
|
4
|
+
from helm.benchmark.window_services.tokenizer_service import TokenizerService
|
|
5
|
+
from helm.benchmark.window_services.test_utils import get_tokenizer_service
|
|
6
|
+
from helm.benchmark.window_services.window_service_factory import WindowServiceFactory
|
|
7
|
+
from helm.common.cache_backend_config import BlackHoleCacheBackendConfig
|
|
8
|
+
|
|
9
|
+
|
|
10
|
+
class TestCLIPWindowService:
|
|
11
|
+
def setup_method(self):
|
|
12
|
+
self.path: str = tempfile.mkdtemp()
|
|
13
|
+
service: TokenizerService = get_tokenizer_service(self.path, BlackHoleCacheBackendConfig())
|
|
14
|
+
self.window_service = WindowServiceFactory.get_window_service("huggingface/dreamlike-photoreal-v2-0", service)
|
|
15
|
+
|
|
16
|
+
def teardown_method(self, method):
|
|
17
|
+
shutil.rmtree(self.path)
|
|
18
|
+
|
|
19
|
+
def test_truncate_from_right(self):
|
|
20
|
+
example_text: str = (
|
|
21
|
+
"an instqrumemnt used for cutting cloth , paper , axdz othr thdin mteroial , "
|
|
22
|
+
"consamistng of two blades lad one on tvopb of the other and fhastned in tle mixdqdjle "
|
|
23
|
+
"so as to bllow them txo be pened and closed by thumb and fitngesr inserted tgrough rings on"
|
|
24
|
+
)
|
|
25
|
+
assert not self.window_service.fits_within_context_window(example_text)
|
|
26
|
+
|
|
27
|
+
# Truncate and ensure it fits within the context window
|
|
28
|
+
truncated_prompt: str = self.window_service.truncate_from_right(example_text)
|
|
29
|
+
assert self.window_service.fits_within_context_window(truncated_prompt)
|
|
@@ -0,0 +1,30 @@
|
|
|
1
|
+
import shutil
|
|
2
|
+
import tempfile
|
|
3
|
+
|
|
4
|
+
from helm.benchmark.window_services.tokenizer_service import TokenizerService
|
|
5
|
+
from helm.clients.image_generation.dalle2_client import DALLE2Client
|
|
6
|
+
from helm.benchmark.window_services.test_utils import get_tokenizer_service, TEST_PROMPT
|
|
7
|
+
from helm.benchmark.window_services.window_service_factory import WindowServiceFactory
|
|
8
|
+
from helm.common.cache_backend_config import BlackHoleCacheBackendConfig
|
|
9
|
+
|
|
10
|
+
|
|
11
|
+
class TestOpenAIDALLEWindowService:
|
|
12
|
+
def setup_method(self):
|
|
13
|
+
self.path: str = tempfile.mkdtemp()
|
|
14
|
+
service: TokenizerService = get_tokenizer_service(self.path, BlackHoleCacheBackendConfig())
|
|
15
|
+
self.window_service = WindowServiceFactory.get_window_service("openai/dall-e-2", service)
|
|
16
|
+
|
|
17
|
+
def teardown_method(self, method):
|
|
18
|
+
shutil.rmtree(self.path)
|
|
19
|
+
|
|
20
|
+
def test_fits_within_context_window(self):
|
|
21
|
+
assert self.window_service.fits_within_context_window(TEST_PROMPT)
|
|
22
|
+
|
|
23
|
+
def test_truncate_from_right(self):
|
|
24
|
+
long_prompt: str = TEST_PROMPT * 10
|
|
25
|
+
assert not self.window_service.fits_within_context_window(long_prompt)
|
|
26
|
+
|
|
27
|
+
# Truncate and ensure it fits within the context window
|
|
28
|
+
truncated_long_prompt: str = self.window_service.truncate_from_right(long_prompt)
|
|
29
|
+
assert len(truncated_long_prompt) == DALLE2Client.MAX_PROMPT_LENGTH
|
|
30
|
+
assert self.window_service.fits_within_context_window(truncated_long_prompt)
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
from abc import ABC
|
|
2
2
|
from typing import List, Optional, cast
|
|
3
3
|
|
|
4
|
-
from .window_service import
|
|
4
|
+
from .window_service import ConfigurableWindowService, EncodeResult
|
|
5
5
|
from .tokenizer_service import TokenizerService
|
|
6
6
|
from helm.common.tokenization_request import (
|
|
7
7
|
DecodeRequest,
|
|
@@ -10,11 +10,28 @@ from helm.common.tokenization_request import (
|
|
|
10
10
|
TokenizationRequestResult,
|
|
11
11
|
TokenizationToken,
|
|
12
12
|
)
|
|
13
|
-
from helm.
|
|
13
|
+
from helm.clients.client import cleanup_tokens
|
|
14
14
|
|
|
15
15
|
|
|
16
|
-
class LocalWindowService(
|
|
17
|
-
def __init__(
|
|
16
|
+
class LocalWindowService(ConfigurableWindowService, ABC):
|
|
17
|
+
def __init__(
|
|
18
|
+
self,
|
|
19
|
+
service: TokenizerService,
|
|
20
|
+
tokenizer_name: str,
|
|
21
|
+
max_sequence_length: int,
|
|
22
|
+
max_request_length: Optional[int] = None,
|
|
23
|
+
max_sequence_and_generated_tokens_length: Optional[int] = None,
|
|
24
|
+
end_of_text_token: Optional[str] = None,
|
|
25
|
+
prefix_token: Optional[str] = None,
|
|
26
|
+
):
|
|
27
|
+
super().__init__(
|
|
28
|
+
tokenizer_name=tokenizer_name,
|
|
29
|
+
max_sequence_length=max_sequence_length,
|
|
30
|
+
max_request_length=max_request_length,
|
|
31
|
+
max_sequence_and_generated_tokens_length=max_sequence_and_generated_tokens_length,
|
|
32
|
+
end_of_text_token=end_of_text_token,
|
|
33
|
+
prefix_token=prefix_token,
|
|
34
|
+
)
|
|
18
35
|
self.service: TokenizerService = service
|
|
19
36
|
|
|
20
37
|
def encode(self, text: str, truncation: bool = False, max_length: Optional[int] = None) -> EncodeResult:
|
|
@@ -2,6 +2,7 @@ import shutil
|
|
|
2
2
|
import tempfile
|
|
3
3
|
from typing import List
|
|
4
4
|
|
|
5
|
+
from helm.common.cache_backend_config import BlackHoleCacheBackendConfig
|
|
5
6
|
from .tokenizer_service import TokenizerService
|
|
6
7
|
from .window_service_factory import WindowServiceFactory
|
|
7
8
|
from .test_utils import get_tokenizer_service, TEST_PROMPT
|
|
@@ -120,7 +121,7 @@ class TestAnthropicWindowService:
|
|
|
120
121
|
|
|
121
122
|
def setup_method(self):
|
|
122
123
|
self.path: str = tempfile.mkdtemp()
|
|
123
|
-
service: TokenizerService = get_tokenizer_service(self.path)
|
|
124
|
+
service: TokenizerService = get_tokenizer_service(self.path, BlackHoleCacheBackendConfig())
|
|
124
125
|
self.window_service = WindowServiceFactory.get_window_service("anthropic/claude-v1.3", service)
|
|
125
126
|
|
|
126
127
|
def teardown_method(self, method):
|
|
@@ -2,6 +2,7 @@ import shutil
|
|
|
2
2
|
import tempfile
|
|
3
3
|
from typing import List
|
|
4
4
|
|
|
5
|
+
from helm.common.cache_backend_config import BlackHoleCacheBackendConfig
|
|
5
6
|
from .tokenizer_service import TokenizerService
|
|
6
7
|
from .window_service_factory import WindowServiceFactory
|
|
7
8
|
from .test_utils import get_tokenizer_service, TEST_PROMPT
|
|
@@ -64,7 +65,7 @@ class TestBloomWindowService:
|
|
|
64
65
|
|
|
65
66
|
def setup_method(self):
|
|
66
67
|
self.path: str = tempfile.mkdtemp()
|
|
67
|
-
service: TokenizerService = get_tokenizer_service(self.path)
|
|
68
|
+
service: TokenizerService = get_tokenizer_service(self.path, BlackHoleCacheBackendConfig())
|
|
68
69
|
self.window_service = WindowServiceFactory.get_window_service("together/bloom", service)
|
|
69
70
|
|
|
70
71
|
def teardown_method(self, method):
|
|
@@ -6,6 +6,7 @@ from typing import List
|
|
|
6
6
|
|
|
7
7
|
from sqlitedict import SqliteDict
|
|
8
8
|
|
|
9
|
+
from helm.common.cache_backend_config import SqliteCacheBackendConfig
|
|
9
10
|
from helm.common.general import ensure_directory_exists
|
|
10
11
|
from .test_cohere_window_service_utils import REQUESTS_TO_RESPONSES, TEST_PROMPT, TOKENIZED_PROMPT
|
|
11
12
|
from .tokenizer_service import TokenizerService
|
|
@@ -30,7 +31,7 @@ class TestCohereWindowService:
|
|
|
30
31
|
with open(os.path.join(cls.path, "credentials.conf"), "w") as f:
|
|
31
32
|
f.write("cohereApiKey: secret")
|
|
32
33
|
|
|
33
|
-
service: TokenizerService = get_tokenizer_service(cls.path)
|
|
34
|
+
service: TokenizerService = get_tokenizer_service(cls.path, SqliteCacheBackendConfig(cache_path))
|
|
34
35
|
cls.window_service = WindowServiceFactory.get_window_service("cohere/xlarge-20220609", service)
|
|
35
36
|
cls.prompt: str = TEST_PROMPT
|
|
36
37
|
cls.tokenized_prompt: List[str] = TOKENIZED_PROMPT
|
|
@@ -1,5 +1,6 @@
|
|
|
1
1
|
import tempfile
|
|
2
2
|
|
|
3
|
+
from helm.common.cache_backend_config import BlackHoleCacheBackendConfig
|
|
3
4
|
from helm.benchmark.window_services.test_t511b_window_service import TestT511bWindowService
|
|
4
5
|
from helm.benchmark.window_services.window_service_factory import TokenizerService, WindowServiceFactory
|
|
5
6
|
from helm.benchmark.window_services.test_utils import get_tokenizer_service
|
|
@@ -8,5 +9,5 @@ from helm.benchmark.window_services.test_utils import get_tokenizer_service
|
|
|
8
9
|
class TestFlanT5WindowService(TestT511bWindowService):
|
|
9
10
|
def setup_method(self):
|
|
10
11
|
self.path: str = tempfile.mkdtemp()
|
|
11
|
-
service: TokenizerService = get_tokenizer_service(self.path)
|
|
12
|
+
service: TokenizerService = get_tokenizer_service(self.path, BlackHoleCacheBackendConfig())
|
|
12
13
|
self.window_service = WindowServiceFactory.get_window_service("together/flan-t5-xxl", service)
|
|
@@ -2,7 +2,7 @@ import shutil
|
|
|
2
2
|
import tempfile
|
|
3
3
|
|
|
4
4
|
from helm.benchmark.window_services.tokenizer_service import TokenizerService
|
|
5
|
-
|
|
5
|
+
from helm.common.cache_backend_config import BlackHoleCacheBackendConfig
|
|
6
6
|
from .test_utils import get_tokenizer_service, TEST_PROMPT, GPT2_TEST_TOKENS, GPT2_TEST_TOKEN_IDS
|
|
7
7
|
from .window_service_factory import WindowServiceFactory
|
|
8
8
|
|
|
@@ -10,7 +10,7 @@ from .window_service_factory import WindowServiceFactory
|
|
|
10
10
|
class TestGPT2WindowService:
|
|
11
11
|
def setup_method(self):
|
|
12
12
|
self.path: str = tempfile.mkdtemp()
|
|
13
|
-
service: TokenizerService = get_tokenizer_service(self.path)
|
|
13
|
+
service: TokenizerService = get_tokenizer_service(self.path, BlackHoleCacheBackendConfig())
|
|
14
14
|
self.window_service = WindowServiceFactory.get_window_service("huggingface/gpt2", service)
|
|
15
15
|
|
|
16
16
|
def teardown_method(self, method):
|
|
@@ -1,6 +1,7 @@
|
|
|
1
1
|
import shutil
|
|
2
2
|
import tempfile
|
|
3
3
|
|
|
4
|
+
from helm.common.cache_backend_config import BlackHoleCacheBackendConfig
|
|
4
5
|
from .test_utils import get_tokenizer_service, TEST_PROMPT, GPT4_TEST_TOKEN_IDS, GPT4_TEST_TOKENS
|
|
5
6
|
from .tokenizer_service import TokenizerService
|
|
6
7
|
from .window_service_factory import WindowServiceFactory
|
|
@@ -9,7 +10,7 @@ from .window_service_factory import WindowServiceFactory
|
|
|
9
10
|
class TestOpenAIWindowService:
|
|
10
11
|
def setup_method(self):
|
|
11
12
|
self.path: str = tempfile.mkdtemp()
|
|
12
|
-
service: TokenizerService = get_tokenizer_service(self.path)
|
|
13
|
+
service: TokenizerService = get_tokenizer_service(self.path, BlackHoleCacheBackendConfig())
|
|
13
14
|
self.window_service = WindowServiceFactory.get_window_service("openai/gpt-3.5-turbo-0301", service)
|
|
14
15
|
|
|
15
16
|
def teardown_method(self, method):
|
|
@@ -1,6 +1,7 @@
|
|
|
1
1
|
import shutil
|
|
2
2
|
import tempfile
|
|
3
3
|
|
|
4
|
+
from helm.common.cache_backend_config import BlackHoleCacheBackendConfig
|
|
4
5
|
from .tokenizer_service import TokenizerService
|
|
5
6
|
from .window_service_factory import WindowServiceFactory
|
|
6
7
|
from .test_utils import get_tokenizer_service, GPT2_TEST_TOKENS, GPT2_TEST_TOKEN_IDS, TEST_PROMPT
|
|
@@ -9,8 +10,8 @@ from .test_utils import get_tokenizer_service, GPT2_TEST_TOKENS, GPT2_TEST_TOKEN
|
|
|
9
10
|
class TestGPTJWindowService:
|
|
10
11
|
def setup_method(self):
|
|
11
12
|
self.path: str = tempfile.mkdtemp()
|
|
12
|
-
service: TokenizerService = get_tokenizer_service(self.path)
|
|
13
|
-
self.window_service = WindowServiceFactory.get_window_service("
|
|
13
|
+
service: TokenizerService = get_tokenizer_service(self.path, BlackHoleCacheBackendConfig())
|
|
14
|
+
self.window_service = WindowServiceFactory.get_window_service("huggingface/gpt-j-6b", service)
|
|
14
15
|
|
|
15
16
|
def teardown_method(self, method):
|
|
16
17
|
shutil.rmtree(self.path)
|
|
@@ -2,6 +2,7 @@ import shutil
|
|
|
2
2
|
import tempfile
|
|
3
3
|
from typing import List
|
|
4
4
|
|
|
5
|
+
from helm.common.cache_backend_config import BlackHoleCacheBackendConfig
|
|
5
6
|
from .tokenizer_service import TokenizerService
|
|
6
7
|
from .window_service_factory import WindowServiceFactory
|
|
7
8
|
from .test_utils import get_tokenizer_service, TEST_PROMPT
|
|
@@ -65,8 +66,8 @@ class TestGPTNeoXWindowService:
|
|
|
65
66
|
|
|
66
67
|
def setup_method(self):
|
|
67
68
|
self.path: str = tempfile.mkdtemp()
|
|
68
|
-
service: TokenizerService = get_tokenizer_service(self.path)
|
|
69
|
-
self.window_service = WindowServiceFactory.get_window_service("
|
|
69
|
+
service: TokenizerService = get_tokenizer_service(self.path, BlackHoleCacheBackendConfig())
|
|
70
|
+
self.window_service = WindowServiceFactory.get_window_service("huggingface/gpt-neox-20b", service)
|
|
70
71
|
|
|
71
72
|
def teardown_method(self, method):
|
|
72
73
|
shutil.rmtree(self.path)
|
|
@@ -2,6 +2,7 @@ import shutil
|
|
|
2
2
|
import tempfile
|
|
3
3
|
from typing import List
|
|
4
4
|
|
|
5
|
+
from helm.common.cache_backend_config import BlackHoleCacheBackendConfig
|
|
5
6
|
from .tokenizer_service import TokenizerService
|
|
6
7
|
from .window_service_factory import WindowServiceFactory
|
|
7
8
|
from .test_utils import get_tokenizer_service, TEST_PROMPT
|
|
@@ -64,7 +65,7 @@ class TestICEWindowService:
|
|
|
64
65
|
|
|
65
66
|
def setup_method(self):
|
|
66
67
|
self.path: str = tempfile.mkdtemp()
|
|
67
|
-
service: TokenizerService = get_tokenizer_service(self.path)
|
|
68
|
+
service: TokenizerService = get_tokenizer_service(self.path, BlackHoleCacheBackendConfig())
|
|
68
69
|
self.window_service = WindowServiceFactory.get_window_service("together/glm", service)
|
|
69
70
|
|
|
70
71
|
def teardown_method(self, method):
|