@synsci/cli-darwin-x64 1.1.49
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/bin/skills/accelerate/SKILL.md +332 -0
- package/bin/skills/accelerate/references/custom-plugins.md +453 -0
- package/bin/skills/accelerate/references/megatron-integration.md +489 -0
- package/bin/skills/accelerate/references/performance.md +525 -0
- package/bin/skills/audiocraft/SKILL.md +564 -0
- package/bin/skills/audiocraft/references/advanced-usage.md +666 -0
- package/bin/skills/audiocraft/references/troubleshooting.md +504 -0
- package/bin/skills/autogpt/SKILL.md +403 -0
- package/bin/skills/autogpt/references/advanced-usage.md +535 -0
- package/bin/skills/autogpt/references/troubleshooting.md +420 -0
- package/bin/skills/awq/SKILL.md +310 -0
- package/bin/skills/awq/references/advanced-usage.md +324 -0
- package/bin/skills/awq/references/troubleshooting.md +344 -0
- package/bin/skills/axolotl/SKILL.md +158 -0
- package/bin/skills/axolotl/references/api.md +5548 -0
- package/bin/skills/axolotl/references/dataset-formats.md +1029 -0
- package/bin/skills/axolotl/references/index.md +15 -0
- package/bin/skills/axolotl/references/other.md +3563 -0
- package/bin/skills/bigcode-evaluation-harness/SKILL.md +405 -0
- package/bin/skills/bigcode-evaluation-harness/references/benchmarks.md +393 -0
- package/bin/skills/bigcode-evaluation-harness/references/custom-tasks.md +424 -0
- package/bin/skills/bigcode-evaluation-harness/references/issues.md +394 -0
- package/bin/skills/bitsandbytes/SKILL.md +411 -0
- package/bin/skills/bitsandbytes/references/memory-optimization.md +521 -0
- package/bin/skills/bitsandbytes/references/qlora-training.md +521 -0
- package/bin/skills/bitsandbytes/references/quantization-formats.md +447 -0
- package/bin/skills/blip-2/SKILL.md +564 -0
- package/bin/skills/blip-2/references/advanced-usage.md +680 -0
- package/bin/skills/blip-2/references/troubleshooting.md +526 -0
- package/bin/skills/chroma/SKILL.md +406 -0
- package/bin/skills/chroma/references/integration.md +38 -0
- package/bin/skills/clip/SKILL.md +253 -0
- package/bin/skills/clip/references/applications.md +207 -0
- package/bin/skills/constitutional-ai/SKILL.md +290 -0
- package/bin/skills/crewai/SKILL.md +498 -0
- package/bin/skills/crewai/references/flows.md +438 -0
- package/bin/skills/crewai/references/tools.md +429 -0
- package/bin/skills/crewai/references/troubleshooting.md +480 -0
- package/bin/skills/deepspeed/SKILL.md +141 -0
- package/bin/skills/deepspeed/references/08.md +17 -0
- package/bin/skills/deepspeed/references/09.md +173 -0
- package/bin/skills/deepspeed/references/2020.md +378 -0
- package/bin/skills/deepspeed/references/2023.md +279 -0
- package/bin/skills/deepspeed/references/assets.md +179 -0
- package/bin/skills/deepspeed/references/index.md +35 -0
- package/bin/skills/deepspeed/references/mii.md +118 -0
- package/bin/skills/deepspeed/references/other.md +1191 -0
- package/bin/skills/deepspeed/references/tutorials.md +6554 -0
- package/bin/skills/dspy/SKILL.md +590 -0
- package/bin/skills/dspy/references/examples.md +663 -0
- package/bin/skills/dspy/references/modules.md +475 -0
- package/bin/skills/dspy/references/optimizers.md +566 -0
- package/bin/skills/faiss/SKILL.md +221 -0
- package/bin/skills/faiss/references/index_types.md +280 -0
- package/bin/skills/flash-attention/SKILL.md +367 -0
- package/bin/skills/flash-attention/references/benchmarks.md +215 -0
- package/bin/skills/flash-attention/references/transformers-integration.md +293 -0
- package/bin/skills/gguf/SKILL.md +427 -0
- package/bin/skills/gguf/references/advanced-usage.md +504 -0
- package/bin/skills/gguf/references/troubleshooting.md +442 -0
- package/bin/skills/gptq/SKILL.md +450 -0
- package/bin/skills/gptq/references/calibration.md +337 -0
- package/bin/skills/gptq/references/integration.md +129 -0
- package/bin/skills/gptq/references/troubleshooting.md +95 -0
- package/bin/skills/grpo-rl-training/README.md +97 -0
- package/bin/skills/grpo-rl-training/SKILL.md +572 -0
- package/bin/skills/grpo-rl-training/examples/reward_functions_library.py +393 -0
- package/bin/skills/grpo-rl-training/templates/basic_grpo_training.py +228 -0
- package/bin/skills/guidance/SKILL.md +572 -0
- package/bin/skills/guidance/references/backends.md +554 -0
- package/bin/skills/guidance/references/constraints.md +674 -0
- package/bin/skills/guidance/references/examples.md +767 -0
- package/bin/skills/hqq/SKILL.md +445 -0
- package/bin/skills/hqq/references/advanced-usage.md +528 -0
- package/bin/skills/hqq/references/troubleshooting.md +503 -0
- package/bin/skills/hugging-face-cli/SKILL.md +191 -0
- package/bin/skills/hugging-face-cli/references/commands.md +954 -0
- package/bin/skills/hugging-face-cli/references/examples.md +374 -0
- package/bin/skills/hugging-face-datasets/SKILL.md +547 -0
- package/bin/skills/hugging-face-datasets/examples/diverse_training_examples.json +239 -0
- package/bin/skills/hugging-face-datasets/examples/system_prompt_template.txt +196 -0
- package/bin/skills/hugging-face-datasets/examples/training_examples.json +176 -0
- package/bin/skills/hugging-face-datasets/scripts/dataset_manager.py +522 -0
- package/bin/skills/hugging-face-datasets/scripts/sql_manager.py +844 -0
- package/bin/skills/hugging-face-datasets/templates/chat.json +55 -0
- package/bin/skills/hugging-face-datasets/templates/classification.json +62 -0
- package/bin/skills/hugging-face-datasets/templates/completion.json +51 -0
- package/bin/skills/hugging-face-datasets/templates/custom.json +75 -0
- package/bin/skills/hugging-face-datasets/templates/qa.json +54 -0
- package/bin/skills/hugging-face-datasets/templates/tabular.json +81 -0
- package/bin/skills/hugging-face-evaluation/SKILL.md +656 -0
- package/bin/skills/hugging-face-evaluation/examples/USAGE_EXAMPLES.md +382 -0
- package/bin/skills/hugging-face-evaluation/examples/artificial_analysis_to_hub.py +141 -0
- package/bin/skills/hugging-face-evaluation/examples/example_readme_tables.md +135 -0
- package/bin/skills/hugging-face-evaluation/examples/metric_mapping.json +50 -0
- package/bin/skills/hugging-face-evaluation/requirements.txt +20 -0
- package/bin/skills/hugging-face-evaluation/scripts/evaluation_manager.py +1374 -0
- package/bin/skills/hugging-face-evaluation/scripts/inspect_eval_uv.py +104 -0
- package/bin/skills/hugging-face-evaluation/scripts/inspect_vllm_uv.py +317 -0
- package/bin/skills/hugging-face-evaluation/scripts/lighteval_vllm_uv.py +303 -0
- package/bin/skills/hugging-face-evaluation/scripts/run_eval_job.py +98 -0
- package/bin/skills/hugging-face-evaluation/scripts/run_vllm_eval_job.py +331 -0
- package/bin/skills/hugging-face-evaluation/scripts/test_extraction.py +206 -0
- package/bin/skills/hugging-face-jobs/SKILL.md +1041 -0
- package/bin/skills/hugging-face-jobs/index.html +216 -0
- package/bin/skills/hugging-face-jobs/references/hardware_guide.md +336 -0
- package/bin/skills/hugging-face-jobs/references/hub_saving.md +352 -0
- package/bin/skills/hugging-face-jobs/references/token_usage.md +546 -0
- package/bin/skills/hugging-face-jobs/references/troubleshooting.md +475 -0
- package/bin/skills/hugging-face-jobs/scripts/cot-self-instruct.py +718 -0
- package/bin/skills/hugging-face-jobs/scripts/finepdfs-stats.py +546 -0
- package/bin/skills/hugging-face-jobs/scripts/generate-responses.py +587 -0
- package/bin/skills/hugging-face-model-trainer/SKILL.md +711 -0
- package/bin/skills/hugging-face-model-trainer/references/gguf_conversion.md +296 -0
- package/bin/skills/hugging-face-model-trainer/references/hardware_guide.md +283 -0
- package/bin/skills/hugging-face-model-trainer/references/hub_saving.md +364 -0
- package/bin/skills/hugging-face-model-trainer/references/reliability_principles.md +371 -0
- package/bin/skills/hugging-face-model-trainer/references/trackio_guide.md +189 -0
- package/bin/skills/hugging-face-model-trainer/references/training_methods.md +150 -0
- package/bin/skills/hugging-face-model-trainer/references/training_patterns.md +203 -0
- package/bin/skills/hugging-face-model-trainer/references/troubleshooting.md +282 -0
- package/bin/skills/hugging-face-model-trainer/scripts/convert_to_gguf.py +424 -0
- package/bin/skills/hugging-face-model-trainer/scripts/dataset_inspector.py +417 -0
- package/bin/skills/hugging-face-model-trainer/scripts/estimate_cost.py +150 -0
- package/bin/skills/hugging-face-model-trainer/scripts/train_dpo_example.py +106 -0
- package/bin/skills/hugging-face-model-trainer/scripts/train_grpo_example.py +89 -0
- package/bin/skills/hugging-face-model-trainer/scripts/train_sft_example.py +122 -0
- package/bin/skills/hugging-face-paper-publisher/SKILL.md +627 -0
- package/bin/skills/hugging-face-paper-publisher/examples/example_usage.md +327 -0
- package/bin/skills/hugging-face-paper-publisher/references/quick_reference.md +216 -0
- package/bin/skills/hugging-face-paper-publisher/scripts/paper_manager.py +508 -0
- package/bin/skills/hugging-face-paper-publisher/templates/arxiv.md +299 -0
- package/bin/skills/hugging-face-paper-publisher/templates/ml-report.md +358 -0
- package/bin/skills/hugging-face-paper-publisher/templates/modern.md +319 -0
- package/bin/skills/hugging-face-paper-publisher/templates/standard.md +201 -0
- package/bin/skills/hugging-face-tool-builder/SKILL.md +115 -0
- package/bin/skills/hugging-face-tool-builder/references/baseline_hf_api.py +57 -0
- package/bin/skills/hugging-face-tool-builder/references/baseline_hf_api.sh +40 -0
- package/bin/skills/hugging-face-tool-builder/references/baseline_hf_api.tsx +57 -0
- package/bin/skills/hugging-face-tool-builder/references/find_models_by_paper.sh +230 -0
- package/bin/skills/hugging-face-tool-builder/references/hf_enrich_models.sh +96 -0
- package/bin/skills/hugging-face-tool-builder/references/hf_model_card_frontmatter.sh +188 -0
- package/bin/skills/hugging-face-tool-builder/references/hf_model_papers_auth.sh +171 -0
- package/bin/skills/hugging-face-trackio/SKILL.md +65 -0
- package/bin/skills/hugging-face-trackio/references/logging_metrics.md +206 -0
- package/bin/skills/hugging-face-trackio/references/retrieving_metrics.md +223 -0
- package/bin/skills/huggingface-tokenizers/SKILL.md +516 -0
- package/bin/skills/huggingface-tokenizers/references/algorithms.md +653 -0
- package/bin/skills/huggingface-tokenizers/references/integration.md +637 -0
- package/bin/skills/huggingface-tokenizers/references/pipeline.md +723 -0
- package/bin/skills/huggingface-tokenizers/references/training.md +565 -0
- package/bin/skills/instructor/SKILL.md +740 -0
- package/bin/skills/instructor/references/examples.md +107 -0
- package/bin/skills/instructor/references/providers.md +70 -0
- package/bin/skills/instructor/references/validation.md +606 -0
- package/bin/skills/knowledge-distillation/SKILL.md +458 -0
- package/bin/skills/knowledge-distillation/references/minillm.md +334 -0
- package/bin/skills/lambda-labs/SKILL.md +545 -0
- package/bin/skills/lambda-labs/references/advanced-usage.md +611 -0
- package/bin/skills/lambda-labs/references/troubleshooting.md +530 -0
- package/bin/skills/langchain/SKILL.md +480 -0
- package/bin/skills/langchain/references/agents.md +499 -0
- package/bin/skills/langchain/references/integration.md +562 -0
- package/bin/skills/langchain/references/rag.md +600 -0
- package/bin/skills/langsmith/SKILL.md +422 -0
- package/bin/skills/langsmith/references/advanced-usage.md +548 -0
- package/bin/skills/langsmith/references/troubleshooting.md +537 -0
- package/bin/skills/litgpt/SKILL.md +469 -0
- package/bin/skills/litgpt/references/custom-models.md +568 -0
- package/bin/skills/litgpt/references/distributed-training.md +451 -0
- package/bin/skills/litgpt/references/supported-models.md +336 -0
- package/bin/skills/litgpt/references/training-recipes.md +619 -0
- package/bin/skills/llama-cpp/SKILL.md +258 -0
- package/bin/skills/llama-cpp/references/optimization.md +89 -0
- package/bin/skills/llama-cpp/references/quantization.md +213 -0
- package/bin/skills/llama-cpp/references/server.md +125 -0
- package/bin/skills/llama-factory/SKILL.md +80 -0
- package/bin/skills/llama-factory/references/_images.md +23 -0
- package/bin/skills/llama-factory/references/advanced.md +1055 -0
- package/bin/skills/llama-factory/references/getting_started.md +349 -0
- package/bin/skills/llama-factory/references/index.md +19 -0
- package/bin/skills/llama-factory/references/other.md +31 -0
- package/bin/skills/llamaguard/SKILL.md +337 -0
- package/bin/skills/llamaindex/SKILL.md +569 -0
- package/bin/skills/llamaindex/references/agents.md +83 -0
- package/bin/skills/llamaindex/references/data_connectors.md +108 -0
- package/bin/skills/llamaindex/references/query_engines.md +406 -0
- package/bin/skills/llava/SKILL.md +304 -0
- package/bin/skills/llava/references/training.md +197 -0
- package/bin/skills/lm-evaluation-harness/SKILL.md +490 -0
- package/bin/skills/lm-evaluation-harness/references/api-evaluation.md +490 -0
- package/bin/skills/lm-evaluation-harness/references/benchmark-guide.md +488 -0
- package/bin/skills/lm-evaluation-harness/references/custom-tasks.md +602 -0
- package/bin/skills/lm-evaluation-harness/references/distributed-eval.md +519 -0
- package/bin/skills/long-context/SKILL.md +536 -0
- package/bin/skills/long-context/references/extension_methods.md +468 -0
- package/bin/skills/long-context/references/fine_tuning.md +611 -0
- package/bin/skills/long-context/references/rope.md +402 -0
- package/bin/skills/mamba/SKILL.md +260 -0
- package/bin/skills/mamba/references/architecture-details.md +206 -0
- package/bin/skills/mamba/references/benchmarks.md +255 -0
- package/bin/skills/mamba/references/training-guide.md +388 -0
- package/bin/skills/megatron-core/SKILL.md +366 -0
- package/bin/skills/megatron-core/references/benchmarks.md +249 -0
- package/bin/skills/megatron-core/references/parallelism-guide.md +404 -0
- package/bin/skills/megatron-core/references/production-examples.md +473 -0
- package/bin/skills/megatron-core/references/training-recipes.md +547 -0
- package/bin/skills/miles/SKILL.md +315 -0
- package/bin/skills/miles/references/api-reference.md +141 -0
- package/bin/skills/miles/references/troubleshooting.md +352 -0
- package/bin/skills/mlflow/SKILL.md +704 -0
- package/bin/skills/mlflow/references/deployment.md +744 -0
- package/bin/skills/mlflow/references/model-registry.md +770 -0
- package/bin/skills/mlflow/references/tracking.md +680 -0
- package/bin/skills/modal/SKILL.md +341 -0
- package/bin/skills/modal/references/advanced-usage.md +503 -0
- package/bin/skills/modal/references/troubleshooting.md +494 -0
- package/bin/skills/model-merging/SKILL.md +539 -0
- package/bin/skills/model-merging/references/evaluation.md +462 -0
- package/bin/skills/model-merging/references/examples.md +428 -0
- package/bin/skills/model-merging/references/methods.md +352 -0
- package/bin/skills/model-pruning/SKILL.md +495 -0
- package/bin/skills/model-pruning/references/wanda.md +347 -0
- package/bin/skills/moe-training/SKILL.md +526 -0
- package/bin/skills/moe-training/references/architectures.md +432 -0
- package/bin/skills/moe-training/references/inference.md +348 -0
- package/bin/skills/moe-training/references/training.md +425 -0
- package/bin/skills/nanogpt/SKILL.md +290 -0
- package/bin/skills/nanogpt/references/architecture.md +382 -0
- package/bin/skills/nanogpt/references/data.md +476 -0
- package/bin/skills/nanogpt/references/training.md +564 -0
- package/bin/skills/nemo-curator/SKILL.md +383 -0
- package/bin/skills/nemo-curator/references/deduplication.md +87 -0
- package/bin/skills/nemo-curator/references/filtering.md +102 -0
- package/bin/skills/nemo-evaluator/SKILL.md +494 -0
- package/bin/skills/nemo-evaluator/references/adapter-system.md +340 -0
- package/bin/skills/nemo-evaluator/references/configuration.md +447 -0
- package/bin/skills/nemo-evaluator/references/custom-benchmarks.md +315 -0
- package/bin/skills/nemo-evaluator/references/execution-backends.md +361 -0
- package/bin/skills/nemo-guardrails/SKILL.md +297 -0
- package/bin/skills/nnsight/SKILL.md +436 -0
- package/bin/skills/nnsight/references/README.md +78 -0
- package/bin/skills/nnsight/references/api.md +344 -0
- package/bin/skills/nnsight/references/tutorials.md +300 -0
- package/bin/skills/openrlhf/SKILL.md +249 -0
- package/bin/skills/openrlhf/references/algorithm-comparison.md +404 -0
- package/bin/skills/openrlhf/references/custom-rewards.md +530 -0
- package/bin/skills/openrlhf/references/hybrid-engine.md +287 -0
- package/bin/skills/openrlhf/references/multi-node-training.md +454 -0
- package/bin/skills/outlines/SKILL.md +652 -0
- package/bin/skills/outlines/references/backends.md +615 -0
- package/bin/skills/outlines/references/examples.md +773 -0
- package/bin/skills/outlines/references/json_generation.md +652 -0
- package/bin/skills/peft/SKILL.md +431 -0
- package/bin/skills/peft/references/advanced-usage.md +514 -0
- package/bin/skills/peft/references/troubleshooting.md +480 -0
- package/bin/skills/phoenix/SKILL.md +475 -0
- package/bin/skills/phoenix/references/advanced-usage.md +619 -0
- package/bin/skills/phoenix/references/troubleshooting.md +538 -0
- package/bin/skills/pinecone/SKILL.md +358 -0
- package/bin/skills/pinecone/references/deployment.md +181 -0
- package/bin/skills/pytorch-fsdp/SKILL.md +126 -0
- package/bin/skills/pytorch-fsdp/references/index.md +7 -0
- package/bin/skills/pytorch-fsdp/references/other.md +4249 -0
- package/bin/skills/pytorch-lightning/SKILL.md +346 -0
- package/bin/skills/pytorch-lightning/references/callbacks.md +436 -0
- package/bin/skills/pytorch-lightning/references/distributed.md +490 -0
- package/bin/skills/pytorch-lightning/references/hyperparameter-tuning.md +556 -0
- package/bin/skills/pyvene/SKILL.md +473 -0
- package/bin/skills/pyvene/references/README.md +73 -0
- package/bin/skills/pyvene/references/api.md +383 -0
- package/bin/skills/pyvene/references/tutorials.md +376 -0
- package/bin/skills/qdrant/SKILL.md +493 -0
- package/bin/skills/qdrant/references/advanced-usage.md +648 -0
- package/bin/skills/qdrant/references/troubleshooting.md +631 -0
- package/bin/skills/ray-data/SKILL.md +326 -0
- package/bin/skills/ray-data/references/integration.md +82 -0
- package/bin/skills/ray-data/references/transformations.md +83 -0
- package/bin/skills/ray-train/SKILL.md +406 -0
- package/bin/skills/ray-train/references/multi-node.md +628 -0
- package/bin/skills/rwkv/SKILL.md +260 -0
- package/bin/skills/rwkv/references/architecture-details.md +344 -0
- package/bin/skills/rwkv/references/rwkv7.md +386 -0
- package/bin/skills/rwkv/references/state-management.md +369 -0
- package/bin/skills/saelens/SKILL.md +386 -0
- package/bin/skills/saelens/references/README.md +70 -0
- package/bin/skills/saelens/references/api.md +333 -0
- package/bin/skills/saelens/references/tutorials.md +318 -0
- package/bin/skills/segment-anything/SKILL.md +500 -0
- package/bin/skills/segment-anything/references/advanced-usage.md +589 -0
- package/bin/skills/segment-anything/references/troubleshooting.md +484 -0
- package/bin/skills/sentence-transformers/SKILL.md +255 -0
- package/bin/skills/sentence-transformers/references/models.md +123 -0
- package/bin/skills/sentencepiece/SKILL.md +235 -0
- package/bin/skills/sentencepiece/references/algorithms.md +200 -0
- package/bin/skills/sentencepiece/references/training.md +304 -0
- package/bin/skills/sglang/SKILL.md +442 -0
- package/bin/skills/sglang/references/deployment.md +490 -0
- package/bin/skills/sglang/references/radix-attention.md +413 -0
- package/bin/skills/sglang/references/structured-generation.md +541 -0
- package/bin/skills/simpo/SKILL.md +219 -0
- package/bin/skills/simpo/references/datasets.md +478 -0
- package/bin/skills/simpo/references/hyperparameters.md +452 -0
- package/bin/skills/simpo/references/loss-functions.md +350 -0
- package/bin/skills/skypilot/SKILL.md +509 -0
- package/bin/skills/skypilot/references/advanced-usage.md +491 -0
- package/bin/skills/skypilot/references/troubleshooting.md +570 -0
- package/bin/skills/slime/SKILL.md +464 -0
- package/bin/skills/slime/references/api-reference.md +392 -0
- package/bin/skills/slime/references/troubleshooting.md +386 -0
- package/bin/skills/speculative-decoding/SKILL.md +467 -0
- package/bin/skills/speculative-decoding/references/lookahead.md +309 -0
- package/bin/skills/speculative-decoding/references/medusa.md +350 -0
- package/bin/skills/stable-diffusion/SKILL.md +519 -0
- package/bin/skills/stable-diffusion/references/advanced-usage.md +716 -0
- package/bin/skills/stable-diffusion/references/troubleshooting.md +555 -0
- package/bin/skills/tensorboard/SKILL.md +629 -0
- package/bin/skills/tensorboard/references/integrations.md +638 -0
- package/bin/skills/tensorboard/references/profiling.md +545 -0
- package/bin/skills/tensorboard/references/visualization.md +620 -0
- package/bin/skills/tensorrt-llm/SKILL.md +187 -0
- package/bin/skills/tensorrt-llm/references/multi-gpu.md +298 -0
- package/bin/skills/tensorrt-llm/references/optimization.md +242 -0
- package/bin/skills/tensorrt-llm/references/serving.md +470 -0
- package/bin/skills/tinker/SKILL.md +362 -0
- package/bin/skills/tinker/references/api-reference.md +168 -0
- package/bin/skills/tinker/references/getting-started.md +157 -0
- package/bin/skills/tinker/references/loss-functions.md +163 -0
- package/bin/skills/tinker/references/models-and-lora.md +139 -0
- package/bin/skills/tinker/references/recipes.md +280 -0
- package/bin/skills/tinker/references/reinforcement-learning.md +212 -0
- package/bin/skills/tinker/references/rendering.md +243 -0
- package/bin/skills/tinker/references/supervised-learning.md +232 -0
- package/bin/skills/tinker-training-cost/SKILL.md +187 -0
- package/bin/skills/tinker-training-cost/scripts/calculate_cost.py +123 -0
- package/bin/skills/torchforge/SKILL.md +433 -0
- package/bin/skills/torchforge/references/api-reference.md +327 -0
- package/bin/skills/torchforge/references/troubleshooting.md +409 -0
- package/bin/skills/torchtitan/SKILL.md +358 -0
- package/bin/skills/torchtitan/references/checkpoint.md +181 -0
- package/bin/skills/torchtitan/references/custom-models.md +258 -0
- package/bin/skills/torchtitan/references/float8.md +133 -0
- package/bin/skills/torchtitan/references/fsdp.md +126 -0
- package/bin/skills/transformer-lens/SKILL.md +346 -0
- package/bin/skills/transformer-lens/references/README.md +54 -0
- package/bin/skills/transformer-lens/references/api.md +362 -0
- package/bin/skills/transformer-lens/references/tutorials.md +339 -0
- package/bin/skills/trl-fine-tuning/SKILL.md +455 -0
- package/bin/skills/trl-fine-tuning/references/dpo-variants.md +227 -0
- package/bin/skills/trl-fine-tuning/references/online-rl.md +82 -0
- package/bin/skills/trl-fine-tuning/references/reward-modeling.md +122 -0
- package/bin/skills/trl-fine-tuning/references/sft-training.md +168 -0
- package/bin/skills/unsloth/SKILL.md +80 -0
- package/bin/skills/unsloth/references/index.md +7 -0
- package/bin/skills/unsloth/references/llms-full.md +16799 -0
- package/bin/skills/unsloth/references/llms-txt.md +12044 -0
- package/bin/skills/unsloth/references/llms.md +82 -0
- package/bin/skills/verl/SKILL.md +391 -0
- package/bin/skills/verl/references/api-reference.md +301 -0
- package/bin/skills/verl/references/troubleshooting.md +391 -0
- package/bin/skills/vllm/SKILL.md +364 -0
- package/bin/skills/vllm/references/optimization.md +226 -0
- package/bin/skills/vllm/references/quantization.md +284 -0
- package/bin/skills/vllm/references/server-deployment.md +255 -0
- package/bin/skills/vllm/references/troubleshooting.md +447 -0
- package/bin/skills/weights-and-biases/SKILL.md +590 -0
- package/bin/skills/weights-and-biases/references/artifacts.md +584 -0
- package/bin/skills/weights-and-biases/references/integrations.md +700 -0
- package/bin/skills/weights-and-biases/references/sweeps.md +847 -0
- package/bin/skills/whisper/SKILL.md +317 -0
- package/bin/skills/whisper/references/languages.md +189 -0
- package/bin/synsc +0 -0
- package/package.json +10 -0
|
@@ -0,0 +1,104 @@
|
|
|
1
|
+
# /// script
|
|
2
|
+
# requires-python = ">=3.10"
|
|
3
|
+
# dependencies = [
|
|
4
|
+
# "inspect-ai>=0.3.0",
|
|
5
|
+
# "inspect-evals",
|
|
6
|
+
# "openai",
|
|
7
|
+
# ]
|
|
8
|
+
# ///
|
|
9
|
+
|
|
10
|
+
"""
|
|
11
|
+
Entry point script for running inspect-ai evaluations via `hf jobs uv run`.
|
|
12
|
+
"""
|
|
13
|
+
|
|
14
|
+
from __future__ import annotations
|
|
15
|
+
|
|
16
|
+
import argparse
|
|
17
|
+
import os
|
|
18
|
+
import subprocess
|
|
19
|
+
import sys
|
|
20
|
+
from pathlib import Path
|
|
21
|
+
from typing import Optional
|
|
22
|
+
|
|
23
|
+
|
|
24
|
+
def _inspect_evals_tasks_root() -> Optional[Path]:
|
|
25
|
+
"""Return the installed inspect_evals package path if available."""
|
|
26
|
+
try:
|
|
27
|
+
import inspect_evals
|
|
28
|
+
|
|
29
|
+
return Path(inspect_evals.__file__).parent
|
|
30
|
+
except Exception:
|
|
31
|
+
return None
|
|
32
|
+
|
|
33
|
+
|
|
34
|
+
def _normalize_task(task: str) -> str:
|
|
35
|
+
"""Allow lighteval-style `suite|task|shots` strings by keeping the task name."""
|
|
36
|
+
if "|" in task:
|
|
37
|
+
parts = task.split("|")
|
|
38
|
+
if len(parts) >= 2 and parts[1]:
|
|
39
|
+
return parts[1]
|
|
40
|
+
return task
|
|
41
|
+
|
|
42
|
+
|
|
43
|
+
def main() -> None:
|
|
44
|
+
parser = argparse.ArgumentParser(description="Inspect-ai job runner")
|
|
45
|
+
parser.add_argument("--model", required=True, help="Model ID on Hugging Face Hub")
|
|
46
|
+
parser.add_argument("--task", required=True, help="inspect-ai task to execute")
|
|
47
|
+
parser.add_argument("--limit", type=int, default=None, help="Limit number of samples to evaluate")
|
|
48
|
+
parser.add_argument(
|
|
49
|
+
"--tasks-root",
|
|
50
|
+
default=None,
|
|
51
|
+
help="Optional path to inspect task files. Defaults to the installed inspect_evals package.",
|
|
52
|
+
)
|
|
53
|
+
parser.add_argument(
|
|
54
|
+
"--sandbox",
|
|
55
|
+
default="local",
|
|
56
|
+
help="Sandbox backend to use (default: local for HF jobs without Docker).",
|
|
57
|
+
)
|
|
58
|
+
args = parser.parse_args()
|
|
59
|
+
|
|
60
|
+
# Ensure downstream libraries can read the token passed as a secret
|
|
61
|
+
hf_token = os.getenv("HF_TOKEN")
|
|
62
|
+
if hf_token:
|
|
63
|
+
os.environ.setdefault("HUGGING_FACE_HUB_TOKEN", hf_token)
|
|
64
|
+
os.environ.setdefault("HF_HUB_TOKEN", hf_token)
|
|
65
|
+
|
|
66
|
+
task = _normalize_task(args.task)
|
|
67
|
+
tasks_root = Path(args.tasks_root) if args.tasks_root else _inspect_evals_tasks_root()
|
|
68
|
+
if tasks_root and not tasks_root.exists():
|
|
69
|
+
tasks_root = None
|
|
70
|
+
|
|
71
|
+
cmd = [
|
|
72
|
+
"inspect",
|
|
73
|
+
"eval",
|
|
74
|
+
task,
|
|
75
|
+
"--model",
|
|
76
|
+
f"hf-inference-providers/{args.model}",
|
|
77
|
+
"--log-level",
|
|
78
|
+
"info",
|
|
79
|
+
# Reduce batch size to avoid OOM errors (default is 32)
|
|
80
|
+
"--max-connections",
|
|
81
|
+
"1",
|
|
82
|
+
# Set a small positive temperature (HF doesn't allow temperature=0)
|
|
83
|
+
"--temperature",
|
|
84
|
+
"0.001",
|
|
85
|
+
]
|
|
86
|
+
|
|
87
|
+
if args.sandbox:
|
|
88
|
+
cmd.extend(["--sandbox", args.sandbox])
|
|
89
|
+
|
|
90
|
+
if args.limit:
|
|
91
|
+
cmd.extend(["--limit", str(args.limit)])
|
|
92
|
+
|
|
93
|
+
try:
|
|
94
|
+
subprocess.run(cmd, check=True, cwd=tasks_root)
|
|
95
|
+
print("Evaluation complete.")
|
|
96
|
+
except subprocess.CalledProcessError as exc:
|
|
97
|
+
location = f" (cwd={tasks_root})" if tasks_root else ""
|
|
98
|
+
print(f"Evaluation failed with exit code {exc.returncode}{location}", file=sys.stderr)
|
|
99
|
+
raise
|
|
100
|
+
|
|
101
|
+
|
|
102
|
+
if __name__ == "__main__":
|
|
103
|
+
main()
|
|
104
|
+
|
|
@@ -0,0 +1,317 @@
|
|
|
1
|
+
# /// script
|
|
2
|
+
# requires-python = ">=3.10"
|
|
3
|
+
# dependencies = [
|
|
4
|
+
# "inspect-ai>=0.3.0",
|
|
5
|
+
# "inspect-evals",
|
|
6
|
+
# "vllm>=0.4.0",
|
|
7
|
+
# "torch>=2.0.0",
|
|
8
|
+
# "transformers>=4.40.0",
|
|
9
|
+
# ]
|
|
10
|
+
# ///
|
|
11
|
+
|
|
12
|
+
"""
|
|
13
|
+
Entry point script for running inspect-ai evaluations with vLLM or HuggingFace Transformers backend.
|
|
14
|
+
|
|
15
|
+
This script runs evaluations on custom HuggingFace models using local GPU inference,
|
|
16
|
+
separate from inference provider scripts (which use external APIs).
|
|
17
|
+
|
|
18
|
+
Usage (standalone):
|
|
19
|
+
python inspect_vllm_uv.py --model "meta-llama/Llama-3.2-1B" --task "mmlu"
|
|
20
|
+
|
|
21
|
+
Usage (via HF Jobs):
|
|
22
|
+
hf jobs uv run inspect_vllm_uv.py \\
|
|
23
|
+
--flavor a10g-small \\
|
|
24
|
+
--secret HF_TOKEN=$HF_TOKEN \\
|
|
25
|
+
-- --model "meta-llama/Llama-3.2-1B" --task "mmlu"
|
|
26
|
+
|
|
27
|
+
Model backends:
|
|
28
|
+
- vllm: Fast inference with vLLM (recommended for large models)
|
|
29
|
+
- hf: HuggingFace Transformers backend (broader model compatibility)
|
|
30
|
+
"""
|
|
31
|
+
|
|
32
|
+
from __future__ import annotations
|
|
33
|
+
|
|
34
|
+
import argparse
|
|
35
|
+
import os
|
|
36
|
+
import subprocess
|
|
37
|
+
import sys
|
|
38
|
+
from typing import Optional
|
|
39
|
+
|
|
40
|
+
|
|
41
|
+
def setup_environment() -> None:
|
|
42
|
+
"""Configure environment variables for HuggingFace authentication."""
|
|
43
|
+
hf_token = os.getenv("HF_TOKEN")
|
|
44
|
+
if hf_token:
|
|
45
|
+
os.environ.setdefault("HUGGING_FACE_HUB_TOKEN", hf_token)
|
|
46
|
+
os.environ.setdefault("HF_HUB_TOKEN", hf_token)
|
|
47
|
+
|
|
48
|
+
|
|
49
|
+
def run_inspect_vllm(
|
|
50
|
+
model_id: str,
|
|
51
|
+
task: str,
|
|
52
|
+
limit: Optional[int] = None,
|
|
53
|
+
max_connections: int = 4,
|
|
54
|
+
temperature: float = 0.0,
|
|
55
|
+
tensor_parallel_size: int = 1,
|
|
56
|
+
gpu_memory_utilization: float = 0.8,
|
|
57
|
+
dtype: str = "auto",
|
|
58
|
+
trust_remote_code: bool = False,
|
|
59
|
+
log_level: str = "info",
|
|
60
|
+
) -> None:
|
|
61
|
+
"""
|
|
62
|
+
Run inspect-ai evaluation with vLLM backend.
|
|
63
|
+
|
|
64
|
+
Args:
|
|
65
|
+
model_id: HuggingFace model ID
|
|
66
|
+
task: inspect-ai task to execute (e.g., "mmlu", "gsm8k")
|
|
67
|
+
limit: Limit number of samples to evaluate
|
|
68
|
+
max_connections: Maximum concurrent connections
|
|
69
|
+
temperature: Sampling temperature
|
|
70
|
+
tensor_parallel_size: Number of GPUs for tensor parallelism
|
|
71
|
+
gpu_memory_utilization: GPU memory fraction
|
|
72
|
+
dtype: Data type (auto, float16, bfloat16)
|
|
73
|
+
trust_remote_code: Allow remote code execution
|
|
74
|
+
log_level: Logging level
|
|
75
|
+
"""
|
|
76
|
+
setup_environment()
|
|
77
|
+
|
|
78
|
+
model_spec = f"vllm/{model_id}"
|
|
79
|
+
cmd = [
|
|
80
|
+
"inspect",
|
|
81
|
+
"eval",
|
|
82
|
+
task,
|
|
83
|
+
"--model",
|
|
84
|
+
model_spec,
|
|
85
|
+
"--log-level",
|
|
86
|
+
log_level,
|
|
87
|
+
"--max-connections",
|
|
88
|
+
str(max_connections),
|
|
89
|
+
]
|
|
90
|
+
|
|
91
|
+
# vLLM supports temperature=0 unlike HF inference providers
|
|
92
|
+
cmd.extend(["--temperature", str(temperature)])
|
|
93
|
+
|
|
94
|
+
# Older inspect-ai CLI versions do not support --model-args; rely on defaults
|
|
95
|
+
# and let vLLM choose sensible settings for small models.
|
|
96
|
+
if tensor_parallel_size != 1:
|
|
97
|
+
cmd.extend(["--tensor-parallel-size", str(tensor_parallel_size)])
|
|
98
|
+
if gpu_memory_utilization != 0.8:
|
|
99
|
+
cmd.extend(["--gpu-memory-utilization", str(gpu_memory_utilization)])
|
|
100
|
+
if dtype != "auto":
|
|
101
|
+
cmd.extend(["--dtype", dtype])
|
|
102
|
+
if trust_remote_code:
|
|
103
|
+
cmd.append("--trust-remote-code")
|
|
104
|
+
|
|
105
|
+
if limit:
|
|
106
|
+
cmd.extend(["--limit", str(limit)])
|
|
107
|
+
|
|
108
|
+
print(f"Running: {' '.join(cmd)}")
|
|
109
|
+
|
|
110
|
+
try:
|
|
111
|
+
subprocess.run(cmd, check=True)
|
|
112
|
+
print("Evaluation complete.")
|
|
113
|
+
except subprocess.CalledProcessError as exc:
|
|
114
|
+
print(f"Evaluation failed with exit code {exc.returncode}", file=sys.stderr)
|
|
115
|
+
sys.exit(exc.returncode)
|
|
116
|
+
|
|
117
|
+
|
|
118
|
+
def run_inspect_hf(
|
|
119
|
+
model_id: str,
|
|
120
|
+
task: str,
|
|
121
|
+
limit: Optional[int] = None,
|
|
122
|
+
max_connections: int = 1,
|
|
123
|
+
temperature: float = 0.001,
|
|
124
|
+
device: str = "auto",
|
|
125
|
+
dtype: str = "auto",
|
|
126
|
+
trust_remote_code: bool = False,
|
|
127
|
+
log_level: str = "info",
|
|
128
|
+
) -> None:
|
|
129
|
+
"""
|
|
130
|
+
Run inspect-ai evaluation with HuggingFace Transformers backend.
|
|
131
|
+
|
|
132
|
+
Use this when vLLM doesn't support the model architecture.
|
|
133
|
+
|
|
134
|
+
Args:
|
|
135
|
+
model_id: HuggingFace model ID
|
|
136
|
+
task: inspect-ai task to execute
|
|
137
|
+
limit: Limit number of samples
|
|
138
|
+
max_connections: Maximum concurrent connections (keep low for memory)
|
|
139
|
+
temperature: Sampling temperature
|
|
140
|
+
device: Device to use (auto, cuda, cpu)
|
|
141
|
+
dtype: Data type
|
|
142
|
+
trust_remote_code: Allow remote code execution
|
|
143
|
+
log_level: Logging level
|
|
144
|
+
"""
|
|
145
|
+
setup_environment()
|
|
146
|
+
|
|
147
|
+
model_spec = f"hf/{model_id}"
|
|
148
|
+
|
|
149
|
+
cmd = [
|
|
150
|
+
"inspect",
|
|
151
|
+
"eval",
|
|
152
|
+
task,
|
|
153
|
+
"--model",
|
|
154
|
+
model_spec,
|
|
155
|
+
"--log-level",
|
|
156
|
+
log_level,
|
|
157
|
+
"--max-connections",
|
|
158
|
+
str(max_connections),
|
|
159
|
+
"--temperature",
|
|
160
|
+
str(temperature),
|
|
161
|
+
]
|
|
162
|
+
|
|
163
|
+
if device != "auto":
|
|
164
|
+
cmd.extend(["--device", device])
|
|
165
|
+
if dtype != "auto":
|
|
166
|
+
cmd.extend(["--dtype", dtype])
|
|
167
|
+
if trust_remote_code:
|
|
168
|
+
cmd.append("--trust-remote-code")
|
|
169
|
+
|
|
170
|
+
if limit:
|
|
171
|
+
cmd.extend(["--limit", str(limit)])
|
|
172
|
+
|
|
173
|
+
print(f"Running: {' '.join(cmd)}")
|
|
174
|
+
|
|
175
|
+
try:
|
|
176
|
+
subprocess.run(cmd, check=True)
|
|
177
|
+
print("Evaluation complete.")
|
|
178
|
+
except subprocess.CalledProcessError as exc:
|
|
179
|
+
print(f"Evaluation failed with exit code {exc.returncode}", file=sys.stderr)
|
|
180
|
+
sys.exit(exc.returncode)
|
|
181
|
+
|
|
182
|
+
|
|
183
|
+
def main() -> None:
|
|
184
|
+
parser = argparse.ArgumentParser(
|
|
185
|
+
description="Run inspect-ai evaluations with vLLM or HuggingFace Transformers on custom models",
|
|
186
|
+
formatter_class=argparse.RawDescriptionHelpFormatter,
|
|
187
|
+
epilog="""
|
|
188
|
+
Examples:
|
|
189
|
+
# Run MMLU with vLLM backend
|
|
190
|
+
python inspect_vllm_uv.py --model meta-llama/Llama-3.2-1B --task mmlu
|
|
191
|
+
|
|
192
|
+
# Run with HuggingFace Transformers backend
|
|
193
|
+
python inspect_vllm_uv.py --model meta-llama/Llama-3.2-1B --task mmlu --backend hf
|
|
194
|
+
|
|
195
|
+
# Run with limited samples for testing
|
|
196
|
+
python inspect_vllm_uv.py --model meta-llama/Llama-3.2-1B --task mmlu --limit 10
|
|
197
|
+
|
|
198
|
+
# Run on multiple GPUs with tensor parallelism
|
|
199
|
+
python inspect_vllm_uv.py --model meta-llama/Llama-3.2-70B --task mmlu --tensor-parallel-size 4
|
|
200
|
+
|
|
201
|
+
Available tasks (from inspect-evals):
|
|
202
|
+
- mmlu: Massive Multitask Language Understanding
|
|
203
|
+
- gsm8k: Grade School Math
|
|
204
|
+
- hellaswag: Common sense reasoning
|
|
205
|
+
- arc_challenge: AI2 Reasoning Challenge
|
|
206
|
+
- truthfulqa: TruthfulQA benchmark
|
|
207
|
+
- winogrande: Winograd Schema Challenge
|
|
208
|
+
- humaneval: Code generation (HumanEval)
|
|
209
|
+
|
|
210
|
+
Via HF Jobs:
|
|
211
|
+
hf jobs uv run inspect_vllm_uv.py \\
|
|
212
|
+
--flavor a10g-small \\
|
|
213
|
+
--secret HF_TOKEN=$HF_TOKEN \\
|
|
214
|
+
-- --model meta-llama/Llama-3.2-1B --task mmlu
|
|
215
|
+
""",
|
|
216
|
+
)
|
|
217
|
+
|
|
218
|
+
parser.add_argument(
|
|
219
|
+
"--model",
|
|
220
|
+
required=True,
|
|
221
|
+
help="HuggingFace model ID (e.g., meta-llama/Llama-3.2-1B)",
|
|
222
|
+
)
|
|
223
|
+
parser.add_argument(
|
|
224
|
+
"--task",
|
|
225
|
+
required=True,
|
|
226
|
+
help="inspect-ai task to execute (e.g., mmlu, gsm8k)",
|
|
227
|
+
)
|
|
228
|
+
parser.add_argument(
|
|
229
|
+
"--backend",
|
|
230
|
+
choices=["vllm", "hf"],
|
|
231
|
+
default="vllm",
|
|
232
|
+
help="Model backend (default: vllm)",
|
|
233
|
+
)
|
|
234
|
+
parser.add_argument(
|
|
235
|
+
"--limit",
|
|
236
|
+
type=int,
|
|
237
|
+
default=None,
|
|
238
|
+
help="Limit number of samples to evaluate",
|
|
239
|
+
)
|
|
240
|
+
parser.add_argument(
|
|
241
|
+
"--max-connections",
|
|
242
|
+
type=int,
|
|
243
|
+
default=None,
|
|
244
|
+
help="Maximum concurrent connections (default: 4 for vllm, 1 for hf)",
|
|
245
|
+
)
|
|
246
|
+
parser.add_argument(
|
|
247
|
+
"--temperature",
|
|
248
|
+
type=float,
|
|
249
|
+
default=None,
|
|
250
|
+
help="Sampling temperature (default: 0.0 for vllm, 0.001 for hf)",
|
|
251
|
+
)
|
|
252
|
+
parser.add_argument(
|
|
253
|
+
"--tensor-parallel-size",
|
|
254
|
+
type=int,
|
|
255
|
+
default=1,
|
|
256
|
+
help="Number of GPUs for tensor parallelism (vLLM only, default: 1)",
|
|
257
|
+
)
|
|
258
|
+
parser.add_argument(
|
|
259
|
+
"--gpu-memory-utilization",
|
|
260
|
+
type=float,
|
|
261
|
+
default=0.8,
|
|
262
|
+
help="GPU memory fraction to use (vLLM only, default: 0.8)",
|
|
263
|
+
)
|
|
264
|
+
parser.add_argument(
|
|
265
|
+
"--dtype",
|
|
266
|
+
default="auto",
|
|
267
|
+
choices=["auto", "float16", "bfloat16", "float32"],
|
|
268
|
+
help="Data type for model weights (default: auto)",
|
|
269
|
+
)
|
|
270
|
+
parser.add_argument(
|
|
271
|
+
"--device",
|
|
272
|
+
default="auto",
|
|
273
|
+
help="Device for HF backend (auto, cuda, cpu)",
|
|
274
|
+
)
|
|
275
|
+
parser.add_argument(
|
|
276
|
+
"--trust-remote-code",
|
|
277
|
+
action="store_true",
|
|
278
|
+
help="Allow executing remote code from model repository",
|
|
279
|
+
)
|
|
280
|
+
parser.add_argument(
|
|
281
|
+
"--log-level",
|
|
282
|
+
default="info",
|
|
283
|
+
choices=["debug", "info", "warning", "error"],
|
|
284
|
+
help="Logging level (default: info)",
|
|
285
|
+
)
|
|
286
|
+
|
|
287
|
+
args = parser.parse_args()
|
|
288
|
+
|
|
289
|
+
if args.backend == "vllm":
|
|
290
|
+
run_inspect_vllm(
|
|
291
|
+
model_id=args.model,
|
|
292
|
+
task=args.task,
|
|
293
|
+
limit=args.limit,
|
|
294
|
+
max_connections=args.max_connections or 4,
|
|
295
|
+
temperature=args.temperature if args.temperature is not None else 0.0,
|
|
296
|
+
tensor_parallel_size=args.tensor_parallel_size,
|
|
297
|
+
gpu_memory_utilization=args.gpu_memory_utilization,
|
|
298
|
+
dtype=args.dtype,
|
|
299
|
+
trust_remote_code=args.trust_remote_code,
|
|
300
|
+
log_level=args.log_level,
|
|
301
|
+
)
|
|
302
|
+
else:
|
|
303
|
+
run_inspect_hf(
|
|
304
|
+
model_id=args.model,
|
|
305
|
+
task=args.task,
|
|
306
|
+
limit=args.limit,
|
|
307
|
+
max_connections=args.max_connections or 1,
|
|
308
|
+
temperature=args.temperature if args.temperature is not None else 0.001,
|
|
309
|
+
device=args.device,
|
|
310
|
+
dtype=args.dtype,
|
|
311
|
+
trust_remote_code=args.trust_remote_code,
|
|
312
|
+
log_level=args.log_level,
|
|
313
|
+
)
|
|
314
|
+
|
|
315
|
+
|
|
316
|
+
if __name__ == "__main__":
|
|
317
|
+
main()
|