@synsci/cli-darwin-x64 1.1.49
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/bin/skills/accelerate/SKILL.md +332 -0
- package/bin/skills/accelerate/references/custom-plugins.md +453 -0
- package/bin/skills/accelerate/references/megatron-integration.md +489 -0
- package/bin/skills/accelerate/references/performance.md +525 -0
- package/bin/skills/audiocraft/SKILL.md +564 -0
- package/bin/skills/audiocraft/references/advanced-usage.md +666 -0
- package/bin/skills/audiocraft/references/troubleshooting.md +504 -0
- package/bin/skills/autogpt/SKILL.md +403 -0
- package/bin/skills/autogpt/references/advanced-usage.md +535 -0
- package/bin/skills/autogpt/references/troubleshooting.md +420 -0
- package/bin/skills/awq/SKILL.md +310 -0
- package/bin/skills/awq/references/advanced-usage.md +324 -0
- package/bin/skills/awq/references/troubleshooting.md +344 -0
- package/bin/skills/axolotl/SKILL.md +158 -0
- package/bin/skills/axolotl/references/api.md +5548 -0
- package/bin/skills/axolotl/references/dataset-formats.md +1029 -0
- package/bin/skills/axolotl/references/index.md +15 -0
- package/bin/skills/axolotl/references/other.md +3563 -0
- package/bin/skills/bigcode-evaluation-harness/SKILL.md +405 -0
- package/bin/skills/bigcode-evaluation-harness/references/benchmarks.md +393 -0
- package/bin/skills/bigcode-evaluation-harness/references/custom-tasks.md +424 -0
- package/bin/skills/bigcode-evaluation-harness/references/issues.md +394 -0
- package/bin/skills/bitsandbytes/SKILL.md +411 -0
- package/bin/skills/bitsandbytes/references/memory-optimization.md +521 -0
- package/bin/skills/bitsandbytes/references/qlora-training.md +521 -0
- package/bin/skills/bitsandbytes/references/quantization-formats.md +447 -0
- package/bin/skills/blip-2/SKILL.md +564 -0
- package/bin/skills/blip-2/references/advanced-usage.md +680 -0
- package/bin/skills/blip-2/references/troubleshooting.md +526 -0
- package/bin/skills/chroma/SKILL.md +406 -0
- package/bin/skills/chroma/references/integration.md +38 -0
- package/bin/skills/clip/SKILL.md +253 -0
- package/bin/skills/clip/references/applications.md +207 -0
- package/bin/skills/constitutional-ai/SKILL.md +290 -0
- package/bin/skills/crewai/SKILL.md +498 -0
- package/bin/skills/crewai/references/flows.md +438 -0
- package/bin/skills/crewai/references/tools.md +429 -0
- package/bin/skills/crewai/references/troubleshooting.md +480 -0
- package/bin/skills/deepspeed/SKILL.md +141 -0
- package/bin/skills/deepspeed/references/08.md +17 -0
- package/bin/skills/deepspeed/references/09.md +173 -0
- package/bin/skills/deepspeed/references/2020.md +378 -0
- package/bin/skills/deepspeed/references/2023.md +279 -0
- package/bin/skills/deepspeed/references/assets.md +179 -0
- package/bin/skills/deepspeed/references/index.md +35 -0
- package/bin/skills/deepspeed/references/mii.md +118 -0
- package/bin/skills/deepspeed/references/other.md +1191 -0
- package/bin/skills/deepspeed/references/tutorials.md +6554 -0
- package/bin/skills/dspy/SKILL.md +590 -0
- package/bin/skills/dspy/references/examples.md +663 -0
- package/bin/skills/dspy/references/modules.md +475 -0
- package/bin/skills/dspy/references/optimizers.md +566 -0
- package/bin/skills/faiss/SKILL.md +221 -0
- package/bin/skills/faiss/references/index_types.md +280 -0
- package/bin/skills/flash-attention/SKILL.md +367 -0
- package/bin/skills/flash-attention/references/benchmarks.md +215 -0
- package/bin/skills/flash-attention/references/transformers-integration.md +293 -0
- package/bin/skills/gguf/SKILL.md +427 -0
- package/bin/skills/gguf/references/advanced-usage.md +504 -0
- package/bin/skills/gguf/references/troubleshooting.md +442 -0
- package/bin/skills/gptq/SKILL.md +450 -0
- package/bin/skills/gptq/references/calibration.md +337 -0
- package/bin/skills/gptq/references/integration.md +129 -0
- package/bin/skills/gptq/references/troubleshooting.md +95 -0
- package/bin/skills/grpo-rl-training/README.md +97 -0
- package/bin/skills/grpo-rl-training/SKILL.md +572 -0
- package/bin/skills/grpo-rl-training/examples/reward_functions_library.py +393 -0
- package/bin/skills/grpo-rl-training/templates/basic_grpo_training.py +228 -0
- package/bin/skills/guidance/SKILL.md +572 -0
- package/bin/skills/guidance/references/backends.md +554 -0
- package/bin/skills/guidance/references/constraints.md +674 -0
- package/bin/skills/guidance/references/examples.md +767 -0
- package/bin/skills/hqq/SKILL.md +445 -0
- package/bin/skills/hqq/references/advanced-usage.md +528 -0
- package/bin/skills/hqq/references/troubleshooting.md +503 -0
- package/bin/skills/hugging-face-cli/SKILL.md +191 -0
- package/bin/skills/hugging-face-cli/references/commands.md +954 -0
- package/bin/skills/hugging-face-cli/references/examples.md +374 -0
- package/bin/skills/hugging-face-datasets/SKILL.md +547 -0
- package/bin/skills/hugging-face-datasets/examples/diverse_training_examples.json +239 -0
- package/bin/skills/hugging-face-datasets/examples/system_prompt_template.txt +196 -0
- package/bin/skills/hugging-face-datasets/examples/training_examples.json +176 -0
- package/bin/skills/hugging-face-datasets/scripts/dataset_manager.py +522 -0
- package/bin/skills/hugging-face-datasets/scripts/sql_manager.py +844 -0
- package/bin/skills/hugging-face-datasets/templates/chat.json +55 -0
- package/bin/skills/hugging-face-datasets/templates/classification.json +62 -0
- package/bin/skills/hugging-face-datasets/templates/completion.json +51 -0
- package/bin/skills/hugging-face-datasets/templates/custom.json +75 -0
- package/bin/skills/hugging-face-datasets/templates/qa.json +54 -0
- package/bin/skills/hugging-face-datasets/templates/tabular.json +81 -0
- package/bin/skills/hugging-face-evaluation/SKILL.md +656 -0
- package/bin/skills/hugging-face-evaluation/examples/USAGE_EXAMPLES.md +382 -0
- package/bin/skills/hugging-face-evaluation/examples/artificial_analysis_to_hub.py +141 -0
- package/bin/skills/hugging-face-evaluation/examples/example_readme_tables.md +135 -0
- package/bin/skills/hugging-face-evaluation/examples/metric_mapping.json +50 -0
- package/bin/skills/hugging-face-evaluation/requirements.txt +20 -0
- package/bin/skills/hugging-face-evaluation/scripts/evaluation_manager.py +1374 -0
- package/bin/skills/hugging-face-evaluation/scripts/inspect_eval_uv.py +104 -0
- package/bin/skills/hugging-face-evaluation/scripts/inspect_vllm_uv.py +317 -0
- package/bin/skills/hugging-face-evaluation/scripts/lighteval_vllm_uv.py +303 -0
- package/bin/skills/hugging-face-evaluation/scripts/run_eval_job.py +98 -0
- package/bin/skills/hugging-face-evaluation/scripts/run_vllm_eval_job.py +331 -0
- package/bin/skills/hugging-face-evaluation/scripts/test_extraction.py +206 -0
- package/bin/skills/hugging-face-jobs/SKILL.md +1041 -0
- package/bin/skills/hugging-face-jobs/index.html +216 -0
- package/bin/skills/hugging-face-jobs/references/hardware_guide.md +336 -0
- package/bin/skills/hugging-face-jobs/references/hub_saving.md +352 -0
- package/bin/skills/hugging-face-jobs/references/token_usage.md +546 -0
- package/bin/skills/hugging-face-jobs/references/troubleshooting.md +475 -0
- package/bin/skills/hugging-face-jobs/scripts/cot-self-instruct.py +718 -0
- package/bin/skills/hugging-face-jobs/scripts/finepdfs-stats.py +546 -0
- package/bin/skills/hugging-face-jobs/scripts/generate-responses.py +587 -0
- package/bin/skills/hugging-face-model-trainer/SKILL.md +711 -0
- package/bin/skills/hugging-face-model-trainer/references/gguf_conversion.md +296 -0
- package/bin/skills/hugging-face-model-trainer/references/hardware_guide.md +283 -0
- package/bin/skills/hugging-face-model-trainer/references/hub_saving.md +364 -0
- package/bin/skills/hugging-face-model-trainer/references/reliability_principles.md +371 -0
- package/bin/skills/hugging-face-model-trainer/references/trackio_guide.md +189 -0
- package/bin/skills/hugging-face-model-trainer/references/training_methods.md +150 -0
- package/bin/skills/hugging-face-model-trainer/references/training_patterns.md +203 -0
- package/bin/skills/hugging-face-model-trainer/references/troubleshooting.md +282 -0
- package/bin/skills/hugging-face-model-trainer/scripts/convert_to_gguf.py +424 -0
- package/bin/skills/hugging-face-model-trainer/scripts/dataset_inspector.py +417 -0
- package/bin/skills/hugging-face-model-trainer/scripts/estimate_cost.py +150 -0
- package/bin/skills/hugging-face-model-trainer/scripts/train_dpo_example.py +106 -0
- package/bin/skills/hugging-face-model-trainer/scripts/train_grpo_example.py +89 -0
- package/bin/skills/hugging-face-model-trainer/scripts/train_sft_example.py +122 -0
- package/bin/skills/hugging-face-paper-publisher/SKILL.md +627 -0
- package/bin/skills/hugging-face-paper-publisher/examples/example_usage.md +327 -0
- package/bin/skills/hugging-face-paper-publisher/references/quick_reference.md +216 -0
- package/bin/skills/hugging-face-paper-publisher/scripts/paper_manager.py +508 -0
- package/bin/skills/hugging-face-paper-publisher/templates/arxiv.md +299 -0
- package/bin/skills/hugging-face-paper-publisher/templates/ml-report.md +358 -0
- package/bin/skills/hugging-face-paper-publisher/templates/modern.md +319 -0
- package/bin/skills/hugging-face-paper-publisher/templates/standard.md +201 -0
- package/bin/skills/hugging-face-tool-builder/SKILL.md +115 -0
- package/bin/skills/hugging-face-tool-builder/references/baseline_hf_api.py +57 -0
- package/bin/skills/hugging-face-tool-builder/references/baseline_hf_api.sh +40 -0
- package/bin/skills/hugging-face-tool-builder/references/baseline_hf_api.tsx +57 -0
- package/bin/skills/hugging-face-tool-builder/references/find_models_by_paper.sh +230 -0
- package/bin/skills/hugging-face-tool-builder/references/hf_enrich_models.sh +96 -0
- package/bin/skills/hugging-face-tool-builder/references/hf_model_card_frontmatter.sh +188 -0
- package/bin/skills/hugging-face-tool-builder/references/hf_model_papers_auth.sh +171 -0
- package/bin/skills/hugging-face-trackio/SKILL.md +65 -0
- package/bin/skills/hugging-face-trackio/references/logging_metrics.md +206 -0
- package/bin/skills/hugging-face-trackio/references/retrieving_metrics.md +223 -0
- package/bin/skills/huggingface-tokenizers/SKILL.md +516 -0
- package/bin/skills/huggingface-tokenizers/references/algorithms.md +653 -0
- package/bin/skills/huggingface-tokenizers/references/integration.md +637 -0
- package/bin/skills/huggingface-tokenizers/references/pipeline.md +723 -0
- package/bin/skills/huggingface-tokenizers/references/training.md +565 -0
- package/bin/skills/instructor/SKILL.md +740 -0
- package/bin/skills/instructor/references/examples.md +107 -0
- package/bin/skills/instructor/references/providers.md +70 -0
- package/bin/skills/instructor/references/validation.md +606 -0
- package/bin/skills/knowledge-distillation/SKILL.md +458 -0
- package/bin/skills/knowledge-distillation/references/minillm.md +334 -0
- package/bin/skills/lambda-labs/SKILL.md +545 -0
- package/bin/skills/lambda-labs/references/advanced-usage.md +611 -0
- package/bin/skills/lambda-labs/references/troubleshooting.md +530 -0
- package/bin/skills/langchain/SKILL.md +480 -0
- package/bin/skills/langchain/references/agents.md +499 -0
- package/bin/skills/langchain/references/integration.md +562 -0
- package/bin/skills/langchain/references/rag.md +600 -0
- package/bin/skills/langsmith/SKILL.md +422 -0
- package/bin/skills/langsmith/references/advanced-usage.md +548 -0
- package/bin/skills/langsmith/references/troubleshooting.md +537 -0
- package/bin/skills/litgpt/SKILL.md +469 -0
- package/bin/skills/litgpt/references/custom-models.md +568 -0
- package/bin/skills/litgpt/references/distributed-training.md +451 -0
- package/bin/skills/litgpt/references/supported-models.md +336 -0
- package/bin/skills/litgpt/references/training-recipes.md +619 -0
- package/bin/skills/llama-cpp/SKILL.md +258 -0
- package/bin/skills/llama-cpp/references/optimization.md +89 -0
- package/bin/skills/llama-cpp/references/quantization.md +213 -0
- package/bin/skills/llama-cpp/references/server.md +125 -0
- package/bin/skills/llama-factory/SKILL.md +80 -0
- package/bin/skills/llama-factory/references/_images.md +23 -0
- package/bin/skills/llama-factory/references/advanced.md +1055 -0
- package/bin/skills/llama-factory/references/getting_started.md +349 -0
- package/bin/skills/llama-factory/references/index.md +19 -0
- package/bin/skills/llama-factory/references/other.md +31 -0
- package/bin/skills/llamaguard/SKILL.md +337 -0
- package/bin/skills/llamaindex/SKILL.md +569 -0
- package/bin/skills/llamaindex/references/agents.md +83 -0
- package/bin/skills/llamaindex/references/data_connectors.md +108 -0
- package/bin/skills/llamaindex/references/query_engines.md +406 -0
- package/bin/skills/llava/SKILL.md +304 -0
- package/bin/skills/llava/references/training.md +197 -0
- package/bin/skills/lm-evaluation-harness/SKILL.md +490 -0
- package/bin/skills/lm-evaluation-harness/references/api-evaluation.md +490 -0
- package/bin/skills/lm-evaluation-harness/references/benchmark-guide.md +488 -0
- package/bin/skills/lm-evaluation-harness/references/custom-tasks.md +602 -0
- package/bin/skills/lm-evaluation-harness/references/distributed-eval.md +519 -0
- package/bin/skills/long-context/SKILL.md +536 -0
- package/bin/skills/long-context/references/extension_methods.md +468 -0
- package/bin/skills/long-context/references/fine_tuning.md +611 -0
- package/bin/skills/long-context/references/rope.md +402 -0
- package/bin/skills/mamba/SKILL.md +260 -0
- package/bin/skills/mamba/references/architecture-details.md +206 -0
- package/bin/skills/mamba/references/benchmarks.md +255 -0
- package/bin/skills/mamba/references/training-guide.md +388 -0
- package/bin/skills/megatron-core/SKILL.md +366 -0
- package/bin/skills/megatron-core/references/benchmarks.md +249 -0
- package/bin/skills/megatron-core/references/parallelism-guide.md +404 -0
- package/bin/skills/megatron-core/references/production-examples.md +473 -0
- package/bin/skills/megatron-core/references/training-recipes.md +547 -0
- package/bin/skills/miles/SKILL.md +315 -0
- package/bin/skills/miles/references/api-reference.md +141 -0
- package/bin/skills/miles/references/troubleshooting.md +352 -0
- package/bin/skills/mlflow/SKILL.md +704 -0
- package/bin/skills/mlflow/references/deployment.md +744 -0
- package/bin/skills/mlflow/references/model-registry.md +770 -0
- package/bin/skills/mlflow/references/tracking.md +680 -0
- package/bin/skills/modal/SKILL.md +341 -0
- package/bin/skills/modal/references/advanced-usage.md +503 -0
- package/bin/skills/modal/references/troubleshooting.md +494 -0
- package/bin/skills/model-merging/SKILL.md +539 -0
- package/bin/skills/model-merging/references/evaluation.md +462 -0
- package/bin/skills/model-merging/references/examples.md +428 -0
- package/bin/skills/model-merging/references/methods.md +352 -0
- package/bin/skills/model-pruning/SKILL.md +495 -0
- package/bin/skills/model-pruning/references/wanda.md +347 -0
- package/bin/skills/moe-training/SKILL.md +526 -0
- package/bin/skills/moe-training/references/architectures.md +432 -0
- package/bin/skills/moe-training/references/inference.md +348 -0
- package/bin/skills/moe-training/references/training.md +425 -0
- package/bin/skills/nanogpt/SKILL.md +290 -0
- package/bin/skills/nanogpt/references/architecture.md +382 -0
- package/bin/skills/nanogpt/references/data.md +476 -0
- package/bin/skills/nanogpt/references/training.md +564 -0
- package/bin/skills/nemo-curator/SKILL.md +383 -0
- package/bin/skills/nemo-curator/references/deduplication.md +87 -0
- package/bin/skills/nemo-curator/references/filtering.md +102 -0
- package/bin/skills/nemo-evaluator/SKILL.md +494 -0
- package/bin/skills/nemo-evaluator/references/adapter-system.md +340 -0
- package/bin/skills/nemo-evaluator/references/configuration.md +447 -0
- package/bin/skills/nemo-evaluator/references/custom-benchmarks.md +315 -0
- package/bin/skills/nemo-evaluator/references/execution-backends.md +361 -0
- package/bin/skills/nemo-guardrails/SKILL.md +297 -0
- package/bin/skills/nnsight/SKILL.md +436 -0
- package/bin/skills/nnsight/references/README.md +78 -0
- package/bin/skills/nnsight/references/api.md +344 -0
- package/bin/skills/nnsight/references/tutorials.md +300 -0
- package/bin/skills/openrlhf/SKILL.md +249 -0
- package/bin/skills/openrlhf/references/algorithm-comparison.md +404 -0
- package/bin/skills/openrlhf/references/custom-rewards.md +530 -0
- package/bin/skills/openrlhf/references/hybrid-engine.md +287 -0
- package/bin/skills/openrlhf/references/multi-node-training.md +454 -0
- package/bin/skills/outlines/SKILL.md +652 -0
- package/bin/skills/outlines/references/backends.md +615 -0
- package/bin/skills/outlines/references/examples.md +773 -0
- package/bin/skills/outlines/references/json_generation.md +652 -0
- package/bin/skills/peft/SKILL.md +431 -0
- package/bin/skills/peft/references/advanced-usage.md +514 -0
- package/bin/skills/peft/references/troubleshooting.md +480 -0
- package/bin/skills/phoenix/SKILL.md +475 -0
- package/bin/skills/phoenix/references/advanced-usage.md +619 -0
- package/bin/skills/phoenix/references/troubleshooting.md +538 -0
- package/bin/skills/pinecone/SKILL.md +358 -0
- package/bin/skills/pinecone/references/deployment.md +181 -0
- package/bin/skills/pytorch-fsdp/SKILL.md +126 -0
- package/bin/skills/pytorch-fsdp/references/index.md +7 -0
- package/bin/skills/pytorch-fsdp/references/other.md +4249 -0
- package/bin/skills/pytorch-lightning/SKILL.md +346 -0
- package/bin/skills/pytorch-lightning/references/callbacks.md +436 -0
- package/bin/skills/pytorch-lightning/references/distributed.md +490 -0
- package/bin/skills/pytorch-lightning/references/hyperparameter-tuning.md +556 -0
- package/bin/skills/pyvene/SKILL.md +473 -0
- package/bin/skills/pyvene/references/README.md +73 -0
- package/bin/skills/pyvene/references/api.md +383 -0
- package/bin/skills/pyvene/references/tutorials.md +376 -0
- package/bin/skills/qdrant/SKILL.md +493 -0
- package/bin/skills/qdrant/references/advanced-usage.md +648 -0
- package/bin/skills/qdrant/references/troubleshooting.md +631 -0
- package/bin/skills/ray-data/SKILL.md +326 -0
- package/bin/skills/ray-data/references/integration.md +82 -0
- package/bin/skills/ray-data/references/transformations.md +83 -0
- package/bin/skills/ray-train/SKILL.md +406 -0
- package/bin/skills/ray-train/references/multi-node.md +628 -0
- package/bin/skills/rwkv/SKILL.md +260 -0
- package/bin/skills/rwkv/references/architecture-details.md +344 -0
- package/bin/skills/rwkv/references/rwkv7.md +386 -0
- package/bin/skills/rwkv/references/state-management.md +369 -0
- package/bin/skills/saelens/SKILL.md +386 -0
- package/bin/skills/saelens/references/README.md +70 -0
- package/bin/skills/saelens/references/api.md +333 -0
- package/bin/skills/saelens/references/tutorials.md +318 -0
- package/bin/skills/segment-anything/SKILL.md +500 -0
- package/bin/skills/segment-anything/references/advanced-usage.md +589 -0
- package/bin/skills/segment-anything/references/troubleshooting.md +484 -0
- package/bin/skills/sentence-transformers/SKILL.md +255 -0
- package/bin/skills/sentence-transformers/references/models.md +123 -0
- package/bin/skills/sentencepiece/SKILL.md +235 -0
- package/bin/skills/sentencepiece/references/algorithms.md +200 -0
- package/bin/skills/sentencepiece/references/training.md +304 -0
- package/bin/skills/sglang/SKILL.md +442 -0
- package/bin/skills/sglang/references/deployment.md +490 -0
- package/bin/skills/sglang/references/radix-attention.md +413 -0
- package/bin/skills/sglang/references/structured-generation.md +541 -0
- package/bin/skills/simpo/SKILL.md +219 -0
- package/bin/skills/simpo/references/datasets.md +478 -0
- package/bin/skills/simpo/references/hyperparameters.md +452 -0
- package/bin/skills/simpo/references/loss-functions.md +350 -0
- package/bin/skills/skypilot/SKILL.md +509 -0
- package/bin/skills/skypilot/references/advanced-usage.md +491 -0
- package/bin/skills/skypilot/references/troubleshooting.md +570 -0
- package/bin/skills/slime/SKILL.md +464 -0
- package/bin/skills/slime/references/api-reference.md +392 -0
- package/bin/skills/slime/references/troubleshooting.md +386 -0
- package/bin/skills/speculative-decoding/SKILL.md +467 -0
- package/bin/skills/speculative-decoding/references/lookahead.md +309 -0
- package/bin/skills/speculative-decoding/references/medusa.md +350 -0
- package/bin/skills/stable-diffusion/SKILL.md +519 -0
- package/bin/skills/stable-diffusion/references/advanced-usage.md +716 -0
- package/bin/skills/stable-diffusion/references/troubleshooting.md +555 -0
- package/bin/skills/tensorboard/SKILL.md +629 -0
- package/bin/skills/tensorboard/references/integrations.md +638 -0
- package/bin/skills/tensorboard/references/profiling.md +545 -0
- package/bin/skills/tensorboard/references/visualization.md +620 -0
- package/bin/skills/tensorrt-llm/SKILL.md +187 -0
- package/bin/skills/tensorrt-llm/references/multi-gpu.md +298 -0
- package/bin/skills/tensorrt-llm/references/optimization.md +242 -0
- package/bin/skills/tensorrt-llm/references/serving.md +470 -0
- package/bin/skills/tinker/SKILL.md +362 -0
- package/bin/skills/tinker/references/api-reference.md +168 -0
- package/bin/skills/tinker/references/getting-started.md +157 -0
- package/bin/skills/tinker/references/loss-functions.md +163 -0
- package/bin/skills/tinker/references/models-and-lora.md +139 -0
- package/bin/skills/tinker/references/recipes.md +280 -0
- package/bin/skills/tinker/references/reinforcement-learning.md +212 -0
- package/bin/skills/tinker/references/rendering.md +243 -0
- package/bin/skills/tinker/references/supervised-learning.md +232 -0
- package/bin/skills/tinker-training-cost/SKILL.md +187 -0
- package/bin/skills/tinker-training-cost/scripts/calculate_cost.py +123 -0
- package/bin/skills/torchforge/SKILL.md +433 -0
- package/bin/skills/torchforge/references/api-reference.md +327 -0
- package/bin/skills/torchforge/references/troubleshooting.md +409 -0
- package/bin/skills/torchtitan/SKILL.md +358 -0
- package/bin/skills/torchtitan/references/checkpoint.md +181 -0
- package/bin/skills/torchtitan/references/custom-models.md +258 -0
- package/bin/skills/torchtitan/references/float8.md +133 -0
- package/bin/skills/torchtitan/references/fsdp.md +126 -0
- package/bin/skills/transformer-lens/SKILL.md +346 -0
- package/bin/skills/transformer-lens/references/README.md +54 -0
- package/bin/skills/transformer-lens/references/api.md +362 -0
- package/bin/skills/transformer-lens/references/tutorials.md +339 -0
- package/bin/skills/trl-fine-tuning/SKILL.md +455 -0
- package/bin/skills/trl-fine-tuning/references/dpo-variants.md +227 -0
- package/bin/skills/trl-fine-tuning/references/online-rl.md +82 -0
- package/bin/skills/trl-fine-tuning/references/reward-modeling.md +122 -0
- package/bin/skills/trl-fine-tuning/references/sft-training.md +168 -0
- package/bin/skills/unsloth/SKILL.md +80 -0
- package/bin/skills/unsloth/references/index.md +7 -0
- package/bin/skills/unsloth/references/llms-full.md +16799 -0
- package/bin/skills/unsloth/references/llms-txt.md +12044 -0
- package/bin/skills/unsloth/references/llms.md +82 -0
- package/bin/skills/verl/SKILL.md +391 -0
- package/bin/skills/verl/references/api-reference.md +301 -0
- package/bin/skills/verl/references/troubleshooting.md +391 -0
- package/bin/skills/vllm/SKILL.md +364 -0
- package/bin/skills/vllm/references/optimization.md +226 -0
- package/bin/skills/vllm/references/quantization.md +284 -0
- package/bin/skills/vllm/references/server-deployment.md +255 -0
- package/bin/skills/vllm/references/troubleshooting.md +447 -0
- package/bin/skills/weights-and-biases/SKILL.md +590 -0
- package/bin/skills/weights-and-biases/references/artifacts.md +584 -0
- package/bin/skills/weights-and-biases/references/integrations.md +700 -0
- package/bin/skills/weights-and-biases/references/sweeps.md +847 -0
- package/bin/skills/whisper/SKILL.md +317 -0
- package/bin/skills/whisper/references/languages.md +189 -0
- package/bin/synsc +0 -0
- package/package.json +10 -0
|
@@ -0,0 +1,442 @@
|
|
|
1
|
+
# GGUF Troubleshooting Guide
|
|
2
|
+
|
|
3
|
+
## Installation Issues
|
|
4
|
+
|
|
5
|
+
### Build Fails
|
|
6
|
+
|
|
7
|
+
**Error**: `make: *** No targets specified and no makefile found`
|
|
8
|
+
|
|
9
|
+
**Fix**:
|
|
10
|
+
```bash
|
|
11
|
+
# Ensure you're in llama.cpp directory
|
|
12
|
+
cd llama.cpp
|
|
13
|
+
make
|
|
14
|
+
```
|
|
15
|
+
|
|
16
|
+
**Error**: `fatal error: cuda_runtime.h: No such file or directory`
|
|
17
|
+
|
|
18
|
+
**Fix**:
|
|
19
|
+
```bash
|
|
20
|
+
# Install CUDA toolkit
|
|
21
|
+
# Ubuntu
|
|
22
|
+
sudo apt install nvidia-cuda-toolkit
|
|
23
|
+
|
|
24
|
+
# Or set CUDA path
|
|
25
|
+
export CUDA_PATH=/usr/local/cuda
|
|
26
|
+
export PATH=$CUDA_PATH/bin:$PATH
|
|
27
|
+
make GGML_CUDA=1
|
|
28
|
+
```
|
|
29
|
+
|
|
30
|
+
### Python Bindings Issues
|
|
31
|
+
|
|
32
|
+
**Error**: `ERROR: Failed building wheel for llama-cpp-python`
|
|
33
|
+
|
|
34
|
+
**Fix**:
|
|
35
|
+
```bash
|
|
36
|
+
# Install build dependencies
|
|
37
|
+
pip install cmake scikit-build-core
|
|
38
|
+
|
|
39
|
+
# For CUDA support
|
|
40
|
+
CMAKE_ARGS="-DGGML_CUDA=on" pip install llama-cpp-python --force-reinstall --no-cache-dir
|
|
41
|
+
|
|
42
|
+
# For Metal (macOS)
|
|
43
|
+
CMAKE_ARGS="-DGGML_METAL=on" pip install llama-cpp-python --force-reinstall --no-cache-dir
|
|
44
|
+
```
|
|
45
|
+
|
|
46
|
+
**Error**: `ImportError: libcudart.so.XX: cannot open shared object file`
|
|
47
|
+
|
|
48
|
+
**Fix**:
|
|
49
|
+
```bash
|
|
50
|
+
# Add CUDA libraries to path
|
|
51
|
+
export LD_LIBRARY_PATH=/usr/local/cuda/lib64:$LD_LIBRARY_PATH
|
|
52
|
+
|
|
53
|
+
# Or reinstall with correct CUDA version
|
|
54
|
+
pip uninstall llama-cpp-python
|
|
55
|
+
CUDACXX=/usr/local/cuda/bin/nvcc CMAKE_ARGS="-DGGML_CUDA=on" pip install llama-cpp-python
|
|
56
|
+
```
|
|
57
|
+
|
|
58
|
+
## Conversion Issues
|
|
59
|
+
|
|
60
|
+
### Model Not Supported
|
|
61
|
+
|
|
62
|
+
**Error**: `KeyError: 'model.embed_tokens.weight'`
|
|
63
|
+
|
|
64
|
+
**Fix**:
|
|
65
|
+
```bash
|
|
66
|
+
# Check model architecture
|
|
67
|
+
python -c "from transformers import AutoConfig; print(AutoConfig.from_pretrained('./model').architectures)"
|
|
68
|
+
|
|
69
|
+
# Use appropriate conversion script
|
|
70
|
+
# For most models:
|
|
71
|
+
python convert_hf_to_gguf.py ./model --outfile model.gguf
|
|
72
|
+
|
|
73
|
+
# For older models, check if legacy script needed
|
|
74
|
+
```
|
|
75
|
+
|
|
76
|
+
### Vocabulary Mismatch
|
|
77
|
+
|
|
78
|
+
**Error**: `RuntimeError: Vocabulary size mismatch`
|
|
79
|
+
|
|
80
|
+
**Fix**:
|
|
81
|
+
```python
|
|
82
|
+
# Ensure tokenizer matches model
|
|
83
|
+
from transformers import AutoTokenizer, AutoModelForCausalLM
|
|
84
|
+
|
|
85
|
+
tokenizer = AutoTokenizer.from_pretrained("./model")
|
|
86
|
+
model = AutoModelForCausalLM.from_pretrained("./model")
|
|
87
|
+
|
|
88
|
+
print(f"Tokenizer vocab size: {len(tokenizer)}")
|
|
89
|
+
print(f"Model vocab size: {model.config.vocab_size}")
|
|
90
|
+
|
|
91
|
+
# If mismatch, resize embeddings before conversion
|
|
92
|
+
model.resize_token_embeddings(len(tokenizer))
|
|
93
|
+
model.save_pretrained("./model-fixed")
|
|
94
|
+
```
|
|
95
|
+
|
|
96
|
+
### Out of Memory During Conversion
|
|
97
|
+
|
|
98
|
+
**Error**: `torch.cuda.OutOfMemoryError` during conversion
|
|
99
|
+
|
|
100
|
+
**Fix**:
|
|
101
|
+
```bash
|
|
102
|
+
# Use CPU for conversion
|
|
103
|
+
CUDA_VISIBLE_DEVICES="" python convert_hf_to_gguf.py ./model --outfile model.gguf
|
|
104
|
+
|
|
105
|
+
# Or use low memory mode
|
|
106
|
+
python convert_hf_to_gguf.py ./model --outfile model.gguf --outtype f16
|
|
107
|
+
```
|
|
108
|
+
|
|
109
|
+
## Quantization Issues
|
|
110
|
+
|
|
111
|
+
### Wrong Output File Size
|
|
112
|
+
|
|
113
|
+
**Problem**: Quantized file is larger than expected
|
|
114
|
+
|
|
115
|
+
**Check**:
|
|
116
|
+
```bash
|
|
117
|
+
# Verify quantization type
|
|
118
|
+
./llama-cli -m model.gguf --verbose
|
|
119
|
+
|
|
120
|
+
# Expected sizes for 7B model:
|
|
121
|
+
# Q4_K_M: ~4.1 GB
|
|
122
|
+
# Q5_K_M: ~4.8 GB
|
|
123
|
+
# Q8_0: ~7.2 GB
|
|
124
|
+
# F16: ~13.5 GB
|
|
125
|
+
```
|
|
126
|
+
|
|
127
|
+
### Quantization Crashes
|
|
128
|
+
|
|
129
|
+
**Error**: `Segmentation fault` during quantization
|
|
130
|
+
|
|
131
|
+
**Fix**:
|
|
132
|
+
```bash
|
|
133
|
+
# Increase stack size
|
|
134
|
+
ulimit -s unlimited
|
|
135
|
+
|
|
136
|
+
# Or use less threads
|
|
137
|
+
./llama-quantize -t 4 model-f16.gguf model-q4.gguf Q4_K_M
|
|
138
|
+
```
|
|
139
|
+
|
|
140
|
+
### Poor Quality After Quantization
|
|
141
|
+
|
|
142
|
+
**Problem**: Model outputs gibberish after quantization
|
|
143
|
+
|
|
144
|
+
**Solutions**:
|
|
145
|
+
|
|
146
|
+
1. **Use importance matrix**:
|
|
147
|
+
```bash
|
|
148
|
+
# Generate imatrix with good calibration data
|
|
149
|
+
./llama-imatrix -m model-f16.gguf \
|
|
150
|
+
-f wiki_sample.txt \
|
|
151
|
+
--chunk 512 \
|
|
152
|
+
-o model.imatrix
|
|
153
|
+
|
|
154
|
+
# Quantize with imatrix
|
|
155
|
+
./llama-quantize --imatrix model.imatrix \
|
|
156
|
+
model-f16.gguf model-q4_k_m.gguf Q4_K_M
|
|
157
|
+
```
|
|
158
|
+
|
|
159
|
+
2. **Try higher precision**:
|
|
160
|
+
```bash
|
|
161
|
+
# Use Q5_K_M or Q6_K instead of Q4
|
|
162
|
+
./llama-quantize model-f16.gguf model-q5_k_m.gguf Q5_K_M
|
|
163
|
+
```
|
|
164
|
+
|
|
165
|
+
3. **Check original model**:
|
|
166
|
+
```bash
|
|
167
|
+
# Test FP16 version first
|
|
168
|
+
./llama-cli -m model-f16.gguf -p "Hello, how are you?" -n 50
|
|
169
|
+
```
|
|
170
|
+
|
|
171
|
+
## Inference Issues
|
|
172
|
+
|
|
173
|
+
### Slow Generation
|
|
174
|
+
|
|
175
|
+
**Problem**: Generation is slower than expected
|
|
176
|
+
|
|
177
|
+
**Solutions**:
|
|
178
|
+
|
|
179
|
+
1. **Enable GPU offload**:
|
|
180
|
+
```bash
|
|
181
|
+
./llama-cli -m model.gguf -ngl 35 -p "Hello"
|
|
182
|
+
```
|
|
183
|
+
|
|
184
|
+
2. **Optimize batch size**:
|
|
185
|
+
```python
|
|
186
|
+
llm = Llama(
|
|
187
|
+
model_path="model.gguf",
|
|
188
|
+
n_batch=512, # Increase for faster prompt processing
|
|
189
|
+
n_gpu_layers=35
|
|
190
|
+
)
|
|
191
|
+
```
|
|
192
|
+
|
|
193
|
+
3. **Use appropriate threads**:
|
|
194
|
+
```bash
|
|
195
|
+
# Match physical cores, not logical
|
|
196
|
+
./llama-cli -m model.gguf -t 8 -p "Hello"
|
|
197
|
+
```
|
|
198
|
+
|
|
199
|
+
4. **Enable Flash Attention** (if supported):
|
|
200
|
+
```bash
|
|
201
|
+
./llama-cli -m model.gguf -ngl 35 --flash-attn -p "Hello"
|
|
202
|
+
```
|
|
203
|
+
|
|
204
|
+
### Out of Memory
|
|
205
|
+
|
|
206
|
+
**Error**: `CUDA out of memory` or system freeze
|
|
207
|
+
|
|
208
|
+
**Solutions**:
|
|
209
|
+
|
|
210
|
+
1. **Reduce GPU layers**:
|
|
211
|
+
```python
|
|
212
|
+
# Start low and increase
|
|
213
|
+
llm = Llama(model_path="model.gguf", n_gpu_layers=10)
|
|
214
|
+
```
|
|
215
|
+
|
|
216
|
+
2. **Use smaller quantization**:
|
|
217
|
+
```bash
|
|
218
|
+
./llama-quantize model-f16.gguf model-q3_k_m.gguf Q3_K_M
|
|
219
|
+
```
|
|
220
|
+
|
|
221
|
+
3. **Reduce context length**:
|
|
222
|
+
```python
|
|
223
|
+
llm = Llama(
|
|
224
|
+
model_path="model.gguf",
|
|
225
|
+
n_ctx=2048, # Reduce from 4096
|
|
226
|
+
n_gpu_layers=35
|
|
227
|
+
)
|
|
228
|
+
```
|
|
229
|
+
|
|
230
|
+
4. **Quantize KV cache**:
|
|
231
|
+
```python
|
|
232
|
+
llm = Llama(
|
|
233
|
+
model_path="model.gguf",
|
|
234
|
+
type_k=2, # Q4_0 for K cache
|
|
235
|
+
type_v=2, # Q4_0 for V cache
|
|
236
|
+
n_gpu_layers=35
|
|
237
|
+
)
|
|
238
|
+
```
|
|
239
|
+
|
|
240
|
+
### Garbage Output
|
|
241
|
+
|
|
242
|
+
**Problem**: Model outputs random characters or nonsense
|
|
243
|
+
|
|
244
|
+
**Diagnose**:
|
|
245
|
+
```python
|
|
246
|
+
# Check model loading
|
|
247
|
+
llm = Llama(model_path="model.gguf", verbose=True)
|
|
248
|
+
|
|
249
|
+
# Test with simple prompt
|
|
250
|
+
output = llm("1+1=", max_tokens=5, temperature=0)
|
|
251
|
+
print(output)
|
|
252
|
+
```
|
|
253
|
+
|
|
254
|
+
**Solutions**:
|
|
255
|
+
|
|
256
|
+
1. **Check model integrity**:
|
|
257
|
+
```bash
|
|
258
|
+
# Verify GGUF file
|
|
259
|
+
./llama-cli -m model.gguf --verbose 2>&1 | head -50
|
|
260
|
+
```
|
|
261
|
+
|
|
262
|
+
2. **Use correct chat format**:
|
|
263
|
+
```python
|
|
264
|
+
llm = Llama(
|
|
265
|
+
model_path="model.gguf",
|
|
266
|
+
chat_format="llama-3" # Match your model: chatml, mistral, etc.
|
|
267
|
+
)
|
|
268
|
+
```
|
|
269
|
+
|
|
270
|
+
3. **Check temperature**:
|
|
271
|
+
```python
|
|
272
|
+
# Use lower temperature for deterministic output
|
|
273
|
+
output = llm("Hello", max_tokens=50, temperature=0.1)
|
|
274
|
+
```
|
|
275
|
+
|
|
276
|
+
### Token Issues
|
|
277
|
+
|
|
278
|
+
**Error**: `RuntimeError: unknown token` or encoding errors
|
|
279
|
+
|
|
280
|
+
**Fix**:
|
|
281
|
+
```python
|
|
282
|
+
# Ensure UTF-8 encoding
|
|
283
|
+
prompt = "Hello, world!".encode('utf-8').decode('utf-8')
|
|
284
|
+
output = llm(prompt, max_tokens=50)
|
|
285
|
+
```
|
|
286
|
+
|
|
287
|
+
## Server Issues
|
|
288
|
+
|
|
289
|
+
### Connection Refused
|
|
290
|
+
|
|
291
|
+
**Error**: `Connection refused` when accessing server
|
|
292
|
+
|
|
293
|
+
**Fix**:
|
|
294
|
+
```bash
|
|
295
|
+
# Bind to all interfaces
|
|
296
|
+
./llama-server -m model.gguf --host 0.0.0.0 --port 8080
|
|
297
|
+
|
|
298
|
+
# Check if port is in use
|
|
299
|
+
lsof -i :8080
|
|
300
|
+
```
|
|
301
|
+
|
|
302
|
+
### Server Crashes Under Load
|
|
303
|
+
|
|
304
|
+
**Problem**: Server crashes with multiple concurrent requests
|
|
305
|
+
|
|
306
|
+
**Solutions**:
|
|
307
|
+
|
|
308
|
+
1. **Limit parallelism**:
|
|
309
|
+
```bash
|
|
310
|
+
./llama-server -m model.gguf \
|
|
311
|
+
--parallel 2 \
|
|
312
|
+
-c 4096 \
|
|
313
|
+
--cont-batching
|
|
314
|
+
```
|
|
315
|
+
|
|
316
|
+
2. **Add request timeout**:
|
|
317
|
+
```bash
|
|
318
|
+
./llama-server -m model.gguf --timeout 300
|
|
319
|
+
```
|
|
320
|
+
|
|
321
|
+
3. **Monitor memory**:
|
|
322
|
+
```bash
|
|
323
|
+
watch -n 1 nvidia-smi # For GPU
|
|
324
|
+
watch -n 1 free -h # For RAM
|
|
325
|
+
```
|
|
326
|
+
|
|
327
|
+
### API Compatibility Issues
|
|
328
|
+
|
|
329
|
+
**Problem**: OpenAI client not working with server
|
|
330
|
+
|
|
331
|
+
**Fix**:
|
|
332
|
+
```python
|
|
333
|
+
from openai import OpenAI
|
|
334
|
+
|
|
335
|
+
# Use correct base URL format
|
|
336
|
+
client = OpenAI(
|
|
337
|
+
base_url="http://localhost:8080/v1", # Include /v1
|
|
338
|
+
api_key="not-needed"
|
|
339
|
+
)
|
|
340
|
+
|
|
341
|
+
# Use correct model name
|
|
342
|
+
response = client.chat.completions.create(
|
|
343
|
+
model="local", # Or the actual model name
|
|
344
|
+
messages=[{"role": "user", "content": "Hello"}]
|
|
345
|
+
)
|
|
346
|
+
```
|
|
347
|
+
|
|
348
|
+
## Apple Silicon Issues
|
|
349
|
+
|
|
350
|
+
### Metal Not Working
|
|
351
|
+
|
|
352
|
+
**Problem**: Metal acceleration not enabled
|
|
353
|
+
|
|
354
|
+
**Check**:
|
|
355
|
+
```bash
|
|
356
|
+
# Verify Metal support
|
|
357
|
+
./llama-cli -m model.gguf --verbose 2>&1 | grep -i metal
|
|
358
|
+
```
|
|
359
|
+
|
|
360
|
+
**Fix**:
|
|
361
|
+
```bash
|
|
362
|
+
# Rebuild with Metal
|
|
363
|
+
make clean
|
|
364
|
+
make GGML_METAL=1
|
|
365
|
+
|
|
366
|
+
# Python bindings
|
|
367
|
+
CMAKE_ARGS="-DGGML_METAL=on" pip install llama-cpp-python --force-reinstall
|
|
368
|
+
```
|
|
369
|
+
|
|
370
|
+
### Incorrect Memory Usage on M1/M2
|
|
371
|
+
|
|
372
|
+
**Problem**: Model uses too much unified memory
|
|
373
|
+
|
|
374
|
+
**Fix**:
|
|
375
|
+
```python
|
|
376
|
+
# Offload all layers for Metal
|
|
377
|
+
llm = Llama(
|
|
378
|
+
model_path="model.gguf",
|
|
379
|
+
n_gpu_layers=99, # Offload everything
|
|
380
|
+
n_threads=1 # Metal handles parallelism
|
|
381
|
+
)
|
|
382
|
+
```
|
|
383
|
+
|
|
384
|
+
## Debugging
|
|
385
|
+
|
|
386
|
+
### Enable Verbose Output
|
|
387
|
+
|
|
388
|
+
```bash
|
|
389
|
+
# CLI verbose mode
|
|
390
|
+
./llama-cli -m model.gguf --verbose -p "Hello" -n 50
|
|
391
|
+
|
|
392
|
+
# Python verbose
|
|
393
|
+
llm = Llama(model_path="model.gguf", verbose=True)
|
|
394
|
+
```
|
|
395
|
+
|
|
396
|
+
### Check Model Metadata
|
|
397
|
+
|
|
398
|
+
```bash
|
|
399
|
+
# View GGUF metadata
|
|
400
|
+
./llama-cli -m model.gguf --verbose 2>&1 | head -100
|
|
401
|
+
```
|
|
402
|
+
|
|
403
|
+
### Validate GGUF File
|
|
404
|
+
|
|
405
|
+
```python
|
|
406
|
+
import struct
|
|
407
|
+
|
|
408
|
+
def validate_gguf(filepath):
|
|
409
|
+
with open(filepath, 'rb') as f:
|
|
410
|
+
magic = f.read(4)
|
|
411
|
+
if magic != b'GGUF':
|
|
412
|
+
print(f"Invalid magic: {magic}")
|
|
413
|
+
return False
|
|
414
|
+
|
|
415
|
+
version = struct.unpack('<I', f.read(4))[0]
|
|
416
|
+
print(f"GGUF version: {version}")
|
|
417
|
+
|
|
418
|
+
tensor_count = struct.unpack('<Q', f.read(8))[0]
|
|
419
|
+
metadata_count = struct.unpack('<Q', f.read(8))[0]
|
|
420
|
+
print(f"Tensors: {tensor_count}, Metadata: {metadata_count}")
|
|
421
|
+
|
|
422
|
+
return True
|
|
423
|
+
|
|
424
|
+
validate_gguf("model.gguf")
|
|
425
|
+
```
|
|
426
|
+
|
|
427
|
+
## Getting Help
|
|
428
|
+
|
|
429
|
+
1. **GitHub Issues**: https://github.com/ggml-org/llama.cpp/issues
|
|
430
|
+
2. **Discussions**: https://github.com/ggml-org/llama.cpp/discussions
|
|
431
|
+
3. **Reddit**: r/LocalLLaMA
|
|
432
|
+
|
|
433
|
+
### Reporting Issues
|
|
434
|
+
|
|
435
|
+
Include:
|
|
436
|
+
- llama.cpp version/commit hash
|
|
437
|
+
- Build command used
|
|
438
|
+
- Model name and quantization
|
|
439
|
+
- Full error message/stack trace
|
|
440
|
+
- Hardware: CPU/GPU model, RAM, VRAM
|
|
441
|
+
- OS version
|
|
442
|
+
- Minimal reproduction steps
|