@synsci/cli-darwin-x64 1.1.49
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/bin/skills/accelerate/SKILL.md +332 -0
- package/bin/skills/accelerate/references/custom-plugins.md +453 -0
- package/bin/skills/accelerate/references/megatron-integration.md +489 -0
- package/bin/skills/accelerate/references/performance.md +525 -0
- package/bin/skills/audiocraft/SKILL.md +564 -0
- package/bin/skills/audiocraft/references/advanced-usage.md +666 -0
- package/bin/skills/audiocraft/references/troubleshooting.md +504 -0
- package/bin/skills/autogpt/SKILL.md +403 -0
- package/bin/skills/autogpt/references/advanced-usage.md +535 -0
- package/bin/skills/autogpt/references/troubleshooting.md +420 -0
- package/bin/skills/awq/SKILL.md +310 -0
- package/bin/skills/awq/references/advanced-usage.md +324 -0
- package/bin/skills/awq/references/troubleshooting.md +344 -0
- package/bin/skills/axolotl/SKILL.md +158 -0
- package/bin/skills/axolotl/references/api.md +5548 -0
- package/bin/skills/axolotl/references/dataset-formats.md +1029 -0
- package/bin/skills/axolotl/references/index.md +15 -0
- package/bin/skills/axolotl/references/other.md +3563 -0
- package/bin/skills/bigcode-evaluation-harness/SKILL.md +405 -0
- package/bin/skills/bigcode-evaluation-harness/references/benchmarks.md +393 -0
- package/bin/skills/bigcode-evaluation-harness/references/custom-tasks.md +424 -0
- package/bin/skills/bigcode-evaluation-harness/references/issues.md +394 -0
- package/bin/skills/bitsandbytes/SKILL.md +411 -0
- package/bin/skills/bitsandbytes/references/memory-optimization.md +521 -0
- package/bin/skills/bitsandbytes/references/qlora-training.md +521 -0
- package/bin/skills/bitsandbytes/references/quantization-formats.md +447 -0
- package/bin/skills/blip-2/SKILL.md +564 -0
- package/bin/skills/blip-2/references/advanced-usage.md +680 -0
- package/bin/skills/blip-2/references/troubleshooting.md +526 -0
- package/bin/skills/chroma/SKILL.md +406 -0
- package/bin/skills/chroma/references/integration.md +38 -0
- package/bin/skills/clip/SKILL.md +253 -0
- package/bin/skills/clip/references/applications.md +207 -0
- package/bin/skills/constitutional-ai/SKILL.md +290 -0
- package/bin/skills/crewai/SKILL.md +498 -0
- package/bin/skills/crewai/references/flows.md +438 -0
- package/bin/skills/crewai/references/tools.md +429 -0
- package/bin/skills/crewai/references/troubleshooting.md +480 -0
- package/bin/skills/deepspeed/SKILL.md +141 -0
- package/bin/skills/deepspeed/references/08.md +17 -0
- package/bin/skills/deepspeed/references/09.md +173 -0
- package/bin/skills/deepspeed/references/2020.md +378 -0
- package/bin/skills/deepspeed/references/2023.md +279 -0
- package/bin/skills/deepspeed/references/assets.md +179 -0
- package/bin/skills/deepspeed/references/index.md +35 -0
- package/bin/skills/deepspeed/references/mii.md +118 -0
- package/bin/skills/deepspeed/references/other.md +1191 -0
- package/bin/skills/deepspeed/references/tutorials.md +6554 -0
- package/bin/skills/dspy/SKILL.md +590 -0
- package/bin/skills/dspy/references/examples.md +663 -0
- package/bin/skills/dspy/references/modules.md +475 -0
- package/bin/skills/dspy/references/optimizers.md +566 -0
- package/bin/skills/faiss/SKILL.md +221 -0
- package/bin/skills/faiss/references/index_types.md +280 -0
- package/bin/skills/flash-attention/SKILL.md +367 -0
- package/bin/skills/flash-attention/references/benchmarks.md +215 -0
- package/bin/skills/flash-attention/references/transformers-integration.md +293 -0
- package/bin/skills/gguf/SKILL.md +427 -0
- package/bin/skills/gguf/references/advanced-usage.md +504 -0
- package/bin/skills/gguf/references/troubleshooting.md +442 -0
- package/bin/skills/gptq/SKILL.md +450 -0
- package/bin/skills/gptq/references/calibration.md +337 -0
- package/bin/skills/gptq/references/integration.md +129 -0
- package/bin/skills/gptq/references/troubleshooting.md +95 -0
- package/bin/skills/grpo-rl-training/README.md +97 -0
- package/bin/skills/grpo-rl-training/SKILL.md +572 -0
- package/bin/skills/grpo-rl-training/examples/reward_functions_library.py +393 -0
- package/bin/skills/grpo-rl-training/templates/basic_grpo_training.py +228 -0
- package/bin/skills/guidance/SKILL.md +572 -0
- package/bin/skills/guidance/references/backends.md +554 -0
- package/bin/skills/guidance/references/constraints.md +674 -0
- package/bin/skills/guidance/references/examples.md +767 -0
- package/bin/skills/hqq/SKILL.md +445 -0
- package/bin/skills/hqq/references/advanced-usage.md +528 -0
- package/bin/skills/hqq/references/troubleshooting.md +503 -0
- package/bin/skills/hugging-face-cli/SKILL.md +191 -0
- package/bin/skills/hugging-face-cli/references/commands.md +954 -0
- package/bin/skills/hugging-face-cli/references/examples.md +374 -0
- package/bin/skills/hugging-face-datasets/SKILL.md +547 -0
- package/bin/skills/hugging-face-datasets/examples/diverse_training_examples.json +239 -0
- package/bin/skills/hugging-face-datasets/examples/system_prompt_template.txt +196 -0
- package/bin/skills/hugging-face-datasets/examples/training_examples.json +176 -0
- package/bin/skills/hugging-face-datasets/scripts/dataset_manager.py +522 -0
- package/bin/skills/hugging-face-datasets/scripts/sql_manager.py +844 -0
- package/bin/skills/hugging-face-datasets/templates/chat.json +55 -0
- package/bin/skills/hugging-face-datasets/templates/classification.json +62 -0
- package/bin/skills/hugging-face-datasets/templates/completion.json +51 -0
- package/bin/skills/hugging-face-datasets/templates/custom.json +75 -0
- package/bin/skills/hugging-face-datasets/templates/qa.json +54 -0
- package/bin/skills/hugging-face-datasets/templates/tabular.json +81 -0
- package/bin/skills/hugging-face-evaluation/SKILL.md +656 -0
- package/bin/skills/hugging-face-evaluation/examples/USAGE_EXAMPLES.md +382 -0
- package/bin/skills/hugging-face-evaluation/examples/artificial_analysis_to_hub.py +141 -0
- package/bin/skills/hugging-face-evaluation/examples/example_readme_tables.md +135 -0
- package/bin/skills/hugging-face-evaluation/examples/metric_mapping.json +50 -0
- package/bin/skills/hugging-face-evaluation/requirements.txt +20 -0
- package/bin/skills/hugging-face-evaluation/scripts/evaluation_manager.py +1374 -0
- package/bin/skills/hugging-face-evaluation/scripts/inspect_eval_uv.py +104 -0
- package/bin/skills/hugging-face-evaluation/scripts/inspect_vllm_uv.py +317 -0
- package/bin/skills/hugging-face-evaluation/scripts/lighteval_vllm_uv.py +303 -0
- package/bin/skills/hugging-face-evaluation/scripts/run_eval_job.py +98 -0
- package/bin/skills/hugging-face-evaluation/scripts/run_vllm_eval_job.py +331 -0
- package/bin/skills/hugging-face-evaluation/scripts/test_extraction.py +206 -0
- package/bin/skills/hugging-face-jobs/SKILL.md +1041 -0
- package/bin/skills/hugging-face-jobs/index.html +216 -0
- package/bin/skills/hugging-face-jobs/references/hardware_guide.md +336 -0
- package/bin/skills/hugging-face-jobs/references/hub_saving.md +352 -0
- package/bin/skills/hugging-face-jobs/references/token_usage.md +546 -0
- package/bin/skills/hugging-face-jobs/references/troubleshooting.md +475 -0
- package/bin/skills/hugging-face-jobs/scripts/cot-self-instruct.py +718 -0
- package/bin/skills/hugging-face-jobs/scripts/finepdfs-stats.py +546 -0
- package/bin/skills/hugging-face-jobs/scripts/generate-responses.py +587 -0
- package/bin/skills/hugging-face-model-trainer/SKILL.md +711 -0
- package/bin/skills/hugging-face-model-trainer/references/gguf_conversion.md +296 -0
- package/bin/skills/hugging-face-model-trainer/references/hardware_guide.md +283 -0
- package/bin/skills/hugging-face-model-trainer/references/hub_saving.md +364 -0
- package/bin/skills/hugging-face-model-trainer/references/reliability_principles.md +371 -0
- package/bin/skills/hugging-face-model-trainer/references/trackio_guide.md +189 -0
- package/bin/skills/hugging-face-model-trainer/references/training_methods.md +150 -0
- package/bin/skills/hugging-face-model-trainer/references/training_patterns.md +203 -0
- package/bin/skills/hugging-face-model-trainer/references/troubleshooting.md +282 -0
- package/bin/skills/hugging-face-model-trainer/scripts/convert_to_gguf.py +424 -0
- package/bin/skills/hugging-face-model-trainer/scripts/dataset_inspector.py +417 -0
- package/bin/skills/hugging-face-model-trainer/scripts/estimate_cost.py +150 -0
- package/bin/skills/hugging-face-model-trainer/scripts/train_dpo_example.py +106 -0
- package/bin/skills/hugging-face-model-trainer/scripts/train_grpo_example.py +89 -0
- package/bin/skills/hugging-face-model-trainer/scripts/train_sft_example.py +122 -0
- package/bin/skills/hugging-face-paper-publisher/SKILL.md +627 -0
- package/bin/skills/hugging-face-paper-publisher/examples/example_usage.md +327 -0
- package/bin/skills/hugging-face-paper-publisher/references/quick_reference.md +216 -0
- package/bin/skills/hugging-face-paper-publisher/scripts/paper_manager.py +508 -0
- package/bin/skills/hugging-face-paper-publisher/templates/arxiv.md +299 -0
- package/bin/skills/hugging-face-paper-publisher/templates/ml-report.md +358 -0
- package/bin/skills/hugging-face-paper-publisher/templates/modern.md +319 -0
- package/bin/skills/hugging-face-paper-publisher/templates/standard.md +201 -0
- package/bin/skills/hugging-face-tool-builder/SKILL.md +115 -0
- package/bin/skills/hugging-face-tool-builder/references/baseline_hf_api.py +57 -0
- package/bin/skills/hugging-face-tool-builder/references/baseline_hf_api.sh +40 -0
- package/bin/skills/hugging-face-tool-builder/references/baseline_hf_api.tsx +57 -0
- package/bin/skills/hugging-face-tool-builder/references/find_models_by_paper.sh +230 -0
- package/bin/skills/hugging-face-tool-builder/references/hf_enrich_models.sh +96 -0
- package/bin/skills/hugging-face-tool-builder/references/hf_model_card_frontmatter.sh +188 -0
- package/bin/skills/hugging-face-tool-builder/references/hf_model_papers_auth.sh +171 -0
- package/bin/skills/hugging-face-trackio/SKILL.md +65 -0
- package/bin/skills/hugging-face-trackio/references/logging_metrics.md +206 -0
- package/bin/skills/hugging-face-trackio/references/retrieving_metrics.md +223 -0
- package/bin/skills/huggingface-tokenizers/SKILL.md +516 -0
- package/bin/skills/huggingface-tokenizers/references/algorithms.md +653 -0
- package/bin/skills/huggingface-tokenizers/references/integration.md +637 -0
- package/bin/skills/huggingface-tokenizers/references/pipeline.md +723 -0
- package/bin/skills/huggingface-tokenizers/references/training.md +565 -0
- package/bin/skills/instructor/SKILL.md +740 -0
- package/bin/skills/instructor/references/examples.md +107 -0
- package/bin/skills/instructor/references/providers.md +70 -0
- package/bin/skills/instructor/references/validation.md +606 -0
- package/bin/skills/knowledge-distillation/SKILL.md +458 -0
- package/bin/skills/knowledge-distillation/references/minillm.md +334 -0
- package/bin/skills/lambda-labs/SKILL.md +545 -0
- package/bin/skills/lambda-labs/references/advanced-usage.md +611 -0
- package/bin/skills/lambda-labs/references/troubleshooting.md +530 -0
- package/bin/skills/langchain/SKILL.md +480 -0
- package/bin/skills/langchain/references/agents.md +499 -0
- package/bin/skills/langchain/references/integration.md +562 -0
- package/bin/skills/langchain/references/rag.md +600 -0
- package/bin/skills/langsmith/SKILL.md +422 -0
- package/bin/skills/langsmith/references/advanced-usage.md +548 -0
- package/bin/skills/langsmith/references/troubleshooting.md +537 -0
- package/bin/skills/litgpt/SKILL.md +469 -0
- package/bin/skills/litgpt/references/custom-models.md +568 -0
- package/bin/skills/litgpt/references/distributed-training.md +451 -0
- package/bin/skills/litgpt/references/supported-models.md +336 -0
- package/bin/skills/litgpt/references/training-recipes.md +619 -0
- package/bin/skills/llama-cpp/SKILL.md +258 -0
- package/bin/skills/llama-cpp/references/optimization.md +89 -0
- package/bin/skills/llama-cpp/references/quantization.md +213 -0
- package/bin/skills/llama-cpp/references/server.md +125 -0
- package/bin/skills/llama-factory/SKILL.md +80 -0
- package/bin/skills/llama-factory/references/_images.md +23 -0
- package/bin/skills/llama-factory/references/advanced.md +1055 -0
- package/bin/skills/llama-factory/references/getting_started.md +349 -0
- package/bin/skills/llama-factory/references/index.md +19 -0
- package/bin/skills/llama-factory/references/other.md +31 -0
- package/bin/skills/llamaguard/SKILL.md +337 -0
- package/bin/skills/llamaindex/SKILL.md +569 -0
- package/bin/skills/llamaindex/references/agents.md +83 -0
- package/bin/skills/llamaindex/references/data_connectors.md +108 -0
- package/bin/skills/llamaindex/references/query_engines.md +406 -0
- package/bin/skills/llava/SKILL.md +304 -0
- package/bin/skills/llava/references/training.md +197 -0
- package/bin/skills/lm-evaluation-harness/SKILL.md +490 -0
- package/bin/skills/lm-evaluation-harness/references/api-evaluation.md +490 -0
- package/bin/skills/lm-evaluation-harness/references/benchmark-guide.md +488 -0
- package/bin/skills/lm-evaluation-harness/references/custom-tasks.md +602 -0
- package/bin/skills/lm-evaluation-harness/references/distributed-eval.md +519 -0
- package/bin/skills/long-context/SKILL.md +536 -0
- package/bin/skills/long-context/references/extension_methods.md +468 -0
- package/bin/skills/long-context/references/fine_tuning.md +611 -0
- package/bin/skills/long-context/references/rope.md +402 -0
- package/bin/skills/mamba/SKILL.md +260 -0
- package/bin/skills/mamba/references/architecture-details.md +206 -0
- package/bin/skills/mamba/references/benchmarks.md +255 -0
- package/bin/skills/mamba/references/training-guide.md +388 -0
- package/bin/skills/megatron-core/SKILL.md +366 -0
- package/bin/skills/megatron-core/references/benchmarks.md +249 -0
- package/bin/skills/megatron-core/references/parallelism-guide.md +404 -0
- package/bin/skills/megatron-core/references/production-examples.md +473 -0
- package/bin/skills/megatron-core/references/training-recipes.md +547 -0
- package/bin/skills/miles/SKILL.md +315 -0
- package/bin/skills/miles/references/api-reference.md +141 -0
- package/bin/skills/miles/references/troubleshooting.md +352 -0
- package/bin/skills/mlflow/SKILL.md +704 -0
- package/bin/skills/mlflow/references/deployment.md +744 -0
- package/bin/skills/mlflow/references/model-registry.md +770 -0
- package/bin/skills/mlflow/references/tracking.md +680 -0
- package/bin/skills/modal/SKILL.md +341 -0
- package/bin/skills/modal/references/advanced-usage.md +503 -0
- package/bin/skills/modal/references/troubleshooting.md +494 -0
- package/bin/skills/model-merging/SKILL.md +539 -0
- package/bin/skills/model-merging/references/evaluation.md +462 -0
- package/bin/skills/model-merging/references/examples.md +428 -0
- package/bin/skills/model-merging/references/methods.md +352 -0
- package/bin/skills/model-pruning/SKILL.md +495 -0
- package/bin/skills/model-pruning/references/wanda.md +347 -0
- package/bin/skills/moe-training/SKILL.md +526 -0
- package/bin/skills/moe-training/references/architectures.md +432 -0
- package/bin/skills/moe-training/references/inference.md +348 -0
- package/bin/skills/moe-training/references/training.md +425 -0
- package/bin/skills/nanogpt/SKILL.md +290 -0
- package/bin/skills/nanogpt/references/architecture.md +382 -0
- package/bin/skills/nanogpt/references/data.md +476 -0
- package/bin/skills/nanogpt/references/training.md +564 -0
- package/bin/skills/nemo-curator/SKILL.md +383 -0
- package/bin/skills/nemo-curator/references/deduplication.md +87 -0
- package/bin/skills/nemo-curator/references/filtering.md +102 -0
- package/bin/skills/nemo-evaluator/SKILL.md +494 -0
- package/bin/skills/nemo-evaluator/references/adapter-system.md +340 -0
- package/bin/skills/nemo-evaluator/references/configuration.md +447 -0
- package/bin/skills/nemo-evaluator/references/custom-benchmarks.md +315 -0
- package/bin/skills/nemo-evaluator/references/execution-backends.md +361 -0
- package/bin/skills/nemo-guardrails/SKILL.md +297 -0
- package/bin/skills/nnsight/SKILL.md +436 -0
- package/bin/skills/nnsight/references/README.md +78 -0
- package/bin/skills/nnsight/references/api.md +344 -0
- package/bin/skills/nnsight/references/tutorials.md +300 -0
- package/bin/skills/openrlhf/SKILL.md +249 -0
- package/bin/skills/openrlhf/references/algorithm-comparison.md +404 -0
- package/bin/skills/openrlhf/references/custom-rewards.md +530 -0
- package/bin/skills/openrlhf/references/hybrid-engine.md +287 -0
- package/bin/skills/openrlhf/references/multi-node-training.md +454 -0
- package/bin/skills/outlines/SKILL.md +652 -0
- package/bin/skills/outlines/references/backends.md +615 -0
- package/bin/skills/outlines/references/examples.md +773 -0
- package/bin/skills/outlines/references/json_generation.md +652 -0
- package/bin/skills/peft/SKILL.md +431 -0
- package/bin/skills/peft/references/advanced-usage.md +514 -0
- package/bin/skills/peft/references/troubleshooting.md +480 -0
- package/bin/skills/phoenix/SKILL.md +475 -0
- package/bin/skills/phoenix/references/advanced-usage.md +619 -0
- package/bin/skills/phoenix/references/troubleshooting.md +538 -0
- package/bin/skills/pinecone/SKILL.md +358 -0
- package/bin/skills/pinecone/references/deployment.md +181 -0
- package/bin/skills/pytorch-fsdp/SKILL.md +126 -0
- package/bin/skills/pytorch-fsdp/references/index.md +7 -0
- package/bin/skills/pytorch-fsdp/references/other.md +4249 -0
- package/bin/skills/pytorch-lightning/SKILL.md +346 -0
- package/bin/skills/pytorch-lightning/references/callbacks.md +436 -0
- package/bin/skills/pytorch-lightning/references/distributed.md +490 -0
- package/bin/skills/pytorch-lightning/references/hyperparameter-tuning.md +556 -0
- package/bin/skills/pyvene/SKILL.md +473 -0
- package/bin/skills/pyvene/references/README.md +73 -0
- package/bin/skills/pyvene/references/api.md +383 -0
- package/bin/skills/pyvene/references/tutorials.md +376 -0
- package/bin/skills/qdrant/SKILL.md +493 -0
- package/bin/skills/qdrant/references/advanced-usage.md +648 -0
- package/bin/skills/qdrant/references/troubleshooting.md +631 -0
- package/bin/skills/ray-data/SKILL.md +326 -0
- package/bin/skills/ray-data/references/integration.md +82 -0
- package/bin/skills/ray-data/references/transformations.md +83 -0
- package/bin/skills/ray-train/SKILL.md +406 -0
- package/bin/skills/ray-train/references/multi-node.md +628 -0
- package/bin/skills/rwkv/SKILL.md +260 -0
- package/bin/skills/rwkv/references/architecture-details.md +344 -0
- package/bin/skills/rwkv/references/rwkv7.md +386 -0
- package/bin/skills/rwkv/references/state-management.md +369 -0
- package/bin/skills/saelens/SKILL.md +386 -0
- package/bin/skills/saelens/references/README.md +70 -0
- package/bin/skills/saelens/references/api.md +333 -0
- package/bin/skills/saelens/references/tutorials.md +318 -0
- package/bin/skills/segment-anything/SKILL.md +500 -0
- package/bin/skills/segment-anything/references/advanced-usage.md +589 -0
- package/bin/skills/segment-anything/references/troubleshooting.md +484 -0
- package/bin/skills/sentence-transformers/SKILL.md +255 -0
- package/bin/skills/sentence-transformers/references/models.md +123 -0
- package/bin/skills/sentencepiece/SKILL.md +235 -0
- package/bin/skills/sentencepiece/references/algorithms.md +200 -0
- package/bin/skills/sentencepiece/references/training.md +304 -0
- package/bin/skills/sglang/SKILL.md +442 -0
- package/bin/skills/sglang/references/deployment.md +490 -0
- package/bin/skills/sglang/references/radix-attention.md +413 -0
- package/bin/skills/sglang/references/structured-generation.md +541 -0
- package/bin/skills/simpo/SKILL.md +219 -0
- package/bin/skills/simpo/references/datasets.md +478 -0
- package/bin/skills/simpo/references/hyperparameters.md +452 -0
- package/bin/skills/simpo/references/loss-functions.md +350 -0
- package/bin/skills/skypilot/SKILL.md +509 -0
- package/bin/skills/skypilot/references/advanced-usage.md +491 -0
- package/bin/skills/skypilot/references/troubleshooting.md +570 -0
- package/bin/skills/slime/SKILL.md +464 -0
- package/bin/skills/slime/references/api-reference.md +392 -0
- package/bin/skills/slime/references/troubleshooting.md +386 -0
- package/bin/skills/speculative-decoding/SKILL.md +467 -0
- package/bin/skills/speculative-decoding/references/lookahead.md +309 -0
- package/bin/skills/speculative-decoding/references/medusa.md +350 -0
- package/bin/skills/stable-diffusion/SKILL.md +519 -0
- package/bin/skills/stable-diffusion/references/advanced-usage.md +716 -0
- package/bin/skills/stable-diffusion/references/troubleshooting.md +555 -0
- package/bin/skills/tensorboard/SKILL.md +629 -0
- package/bin/skills/tensorboard/references/integrations.md +638 -0
- package/bin/skills/tensorboard/references/profiling.md +545 -0
- package/bin/skills/tensorboard/references/visualization.md +620 -0
- package/bin/skills/tensorrt-llm/SKILL.md +187 -0
- package/bin/skills/tensorrt-llm/references/multi-gpu.md +298 -0
- package/bin/skills/tensorrt-llm/references/optimization.md +242 -0
- package/bin/skills/tensorrt-llm/references/serving.md +470 -0
- package/bin/skills/tinker/SKILL.md +362 -0
- package/bin/skills/tinker/references/api-reference.md +168 -0
- package/bin/skills/tinker/references/getting-started.md +157 -0
- package/bin/skills/tinker/references/loss-functions.md +163 -0
- package/bin/skills/tinker/references/models-and-lora.md +139 -0
- package/bin/skills/tinker/references/recipes.md +280 -0
- package/bin/skills/tinker/references/reinforcement-learning.md +212 -0
- package/bin/skills/tinker/references/rendering.md +243 -0
- package/bin/skills/tinker/references/supervised-learning.md +232 -0
- package/bin/skills/tinker-training-cost/SKILL.md +187 -0
- package/bin/skills/tinker-training-cost/scripts/calculate_cost.py +123 -0
- package/bin/skills/torchforge/SKILL.md +433 -0
- package/bin/skills/torchforge/references/api-reference.md +327 -0
- package/bin/skills/torchforge/references/troubleshooting.md +409 -0
- package/bin/skills/torchtitan/SKILL.md +358 -0
- package/bin/skills/torchtitan/references/checkpoint.md +181 -0
- package/bin/skills/torchtitan/references/custom-models.md +258 -0
- package/bin/skills/torchtitan/references/float8.md +133 -0
- package/bin/skills/torchtitan/references/fsdp.md +126 -0
- package/bin/skills/transformer-lens/SKILL.md +346 -0
- package/bin/skills/transformer-lens/references/README.md +54 -0
- package/bin/skills/transformer-lens/references/api.md +362 -0
- package/bin/skills/transformer-lens/references/tutorials.md +339 -0
- package/bin/skills/trl-fine-tuning/SKILL.md +455 -0
- package/bin/skills/trl-fine-tuning/references/dpo-variants.md +227 -0
- package/bin/skills/trl-fine-tuning/references/online-rl.md +82 -0
- package/bin/skills/trl-fine-tuning/references/reward-modeling.md +122 -0
- package/bin/skills/trl-fine-tuning/references/sft-training.md +168 -0
- package/bin/skills/unsloth/SKILL.md +80 -0
- package/bin/skills/unsloth/references/index.md +7 -0
- package/bin/skills/unsloth/references/llms-full.md +16799 -0
- package/bin/skills/unsloth/references/llms-txt.md +12044 -0
- package/bin/skills/unsloth/references/llms.md +82 -0
- package/bin/skills/verl/SKILL.md +391 -0
- package/bin/skills/verl/references/api-reference.md +301 -0
- package/bin/skills/verl/references/troubleshooting.md +391 -0
- package/bin/skills/vllm/SKILL.md +364 -0
- package/bin/skills/vllm/references/optimization.md +226 -0
- package/bin/skills/vllm/references/quantization.md +284 -0
- package/bin/skills/vllm/references/server-deployment.md +255 -0
- package/bin/skills/vllm/references/troubleshooting.md +447 -0
- package/bin/skills/weights-and-biases/SKILL.md +590 -0
- package/bin/skills/weights-and-biases/references/artifacts.md +584 -0
- package/bin/skills/weights-and-biases/references/integrations.md +700 -0
- package/bin/skills/weights-and-biases/references/sweeps.md +847 -0
- package/bin/skills/whisper/SKILL.md +317 -0
- package/bin/skills/whisper/references/languages.md +189 -0
- package/bin/synsc +0 -0
- package/package.json +10 -0
|
@@ -0,0 +1,304 @@
|
|
|
1
|
+
---
|
|
2
|
+
name: llava
|
|
3
|
+
description: Large Language and Vision Assistant. Enables visual instruction tuning and image-based conversations. Combines CLIP vision encoder with Vicuna/LLaMA language models. Supports multi-turn image chat, visual question answering, and instruction following. Use for vision-language chatbots or image understanding tasks. Best for conversational image analysis.
|
|
4
|
+
version: 1.0.0
|
|
5
|
+
author: Synthetic Sciences
|
|
6
|
+
license: MIT
|
|
7
|
+
tags: [LLaVA, Vision-Language, Multimodal, Visual Question Answering, Image Chat, CLIP, Vicuna, Conversational AI, Instruction Tuning, VQA]
|
|
8
|
+
dependencies: [transformers, torch, pillow]
|
|
9
|
+
---
|
|
10
|
+
|
|
11
|
+
# LLaVA - Large Language and Vision Assistant
|
|
12
|
+
|
|
13
|
+
Open-source vision-language model for conversational image understanding.
|
|
14
|
+
|
|
15
|
+
## When to use LLaVA
|
|
16
|
+
|
|
17
|
+
**Use when:**
|
|
18
|
+
- Building vision-language chatbots
|
|
19
|
+
- Visual question answering (VQA)
|
|
20
|
+
- Image description and captioning
|
|
21
|
+
- Multi-turn image conversations
|
|
22
|
+
- Visual instruction following
|
|
23
|
+
- Document understanding with images
|
|
24
|
+
|
|
25
|
+
**Metrics**:
|
|
26
|
+
- **23,000+ GitHub stars**
|
|
27
|
+
- GPT-4V level capabilities (targeted)
|
|
28
|
+
- Apache 2.0 License
|
|
29
|
+
- Multiple model sizes (7B-34B params)
|
|
30
|
+
|
|
31
|
+
**Use alternatives instead**:
|
|
32
|
+
- **GPT-4V**: Highest quality, API-based
|
|
33
|
+
- **CLIP**: Simple zero-shot classification
|
|
34
|
+
- **BLIP-2**: Better for captioning only
|
|
35
|
+
- **Flamingo**: Research, not open-source
|
|
36
|
+
|
|
37
|
+
## Quick start
|
|
38
|
+
|
|
39
|
+
### Installation
|
|
40
|
+
|
|
41
|
+
```bash
|
|
42
|
+
# Clone repository
|
|
43
|
+
git clone https://github.com/haotian-liu/LLaVA
|
|
44
|
+
cd LLaVA
|
|
45
|
+
|
|
46
|
+
# Install
|
|
47
|
+
pip install -e .
|
|
48
|
+
```
|
|
49
|
+
|
|
50
|
+
### Basic usage
|
|
51
|
+
|
|
52
|
+
```python
|
|
53
|
+
from llava.model.builder import load_pretrained_model
|
|
54
|
+
from llava.mm_utils import get_model_name_from_path, process_images, tokenizer_image_token
|
|
55
|
+
from llava.constants import IMAGE_TOKEN_INDEX, DEFAULT_IMAGE_TOKEN
|
|
56
|
+
from llava.conversation import conv_templates
|
|
57
|
+
from PIL import Image
|
|
58
|
+
import torch
|
|
59
|
+
|
|
60
|
+
# Load model
|
|
61
|
+
model_path = "liuhaotian/llava-v1.5-7b"
|
|
62
|
+
tokenizer, model, image_processor, context_len = load_pretrained_model(
|
|
63
|
+
model_path=model_path,
|
|
64
|
+
model_base=None,
|
|
65
|
+
model_name=get_model_name_from_path(model_path)
|
|
66
|
+
)
|
|
67
|
+
|
|
68
|
+
# Load image
|
|
69
|
+
image = Image.open("image.jpg")
|
|
70
|
+
image_tensor = process_images([image], image_processor, model.config)
|
|
71
|
+
image_tensor = image_tensor.to(model.device, dtype=torch.float16)
|
|
72
|
+
|
|
73
|
+
# Create conversation
|
|
74
|
+
conv = conv_templates["llava_v1"].copy()
|
|
75
|
+
conv.append_message(conv.roles[0], DEFAULT_IMAGE_TOKEN + "\nWhat is in this image?")
|
|
76
|
+
conv.append_message(conv.roles[1], None)
|
|
77
|
+
prompt = conv.get_prompt()
|
|
78
|
+
|
|
79
|
+
# Generate response
|
|
80
|
+
input_ids = tokenizer_image_token(prompt, tokenizer, IMAGE_TOKEN_INDEX, return_tensors='pt').unsqueeze(0).to(model.device)
|
|
81
|
+
|
|
82
|
+
with torch.inference_mode():
|
|
83
|
+
output_ids = model.generate(
|
|
84
|
+
input_ids,
|
|
85
|
+
images=image_tensor,
|
|
86
|
+
do_sample=True,
|
|
87
|
+
temperature=0.2,
|
|
88
|
+
max_new_tokens=512
|
|
89
|
+
)
|
|
90
|
+
|
|
91
|
+
response = tokenizer.decode(output_ids[0], skip_special_tokens=True).strip()
|
|
92
|
+
print(response)
|
|
93
|
+
```
|
|
94
|
+
|
|
95
|
+
## Available models
|
|
96
|
+
|
|
97
|
+
| Model | Parameters | VRAM | Quality |
|
|
98
|
+
|-------|------------|------|---------|
|
|
99
|
+
| LLaVA-v1.5-7B | 7B | ~14 GB | Good |
|
|
100
|
+
| LLaVA-v1.5-13B | 13B | ~28 GB | Better |
|
|
101
|
+
| LLaVA-v1.6-34B | 34B | ~70 GB | Best |
|
|
102
|
+
|
|
103
|
+
```python
|
|
104
|
+
# Load different models
|
|
105
|
+
model_7b = "liuhaotian/llava-v1.5-7b"
|
|
106
|
+
model_13b = "liuhaotian/llava-v1.5-13b"
|
|
107
|
+
model_34b = "liuhaotian/llava-v1.6-34b"
|
|
108
|
+
|
|
109
|
+
# 4-bit quantization for lower VRAM
|
|
110
|
+
load_4bit = True # Reduces VRAM by ~4×
|
|
111
|
+
```
|
|
112
|
+
|
|
113
|
+
## CLI usage
|
|
114
|
+
|
|
115
|
+
```bash
|
|
116
|
+
# Single image query
|
|
117
|
+
python -m llava.serve.cli \
|
|
118
|
+
--model-path liuhaotian/llava-v1.5-7b \
|
|
119
|
+
--image-file image.jpg \
|
|
120
|
+
--query "What is in this image?"
|
|
121
|
+
|
|
122
|
+
# Multi-turn conversation
|
|
123
|
+
python -m llava.serve.cli \
|
|
124
|
+
--model-path liuhaotian/llava-v1.5-7b \
|
|
125
|
+
--image-file image.jpg
|
|
126
|
+
# Then type questions interactively
|
|
127
|
+
```
|
|
128
|
+
|
|
129
|
+
## Web UI (Gradio)
|
|
130
|
+
|
|
131
|
+
```bash
|
|
132
|
+
# Launch Gradio interface
|
|
133
|
+
python -m llava.serve.gradio_web_server \
|
|
134
|
+
--model-path liuhaotian/llava-v1.5-7b \
|
|
135
|
+
--load-4bit # Optional: reduce VRAM
|
|
136
|
+
|
|
137
|
+
# Access at http://localhost:7860
|
|
138
|
+
```
|
|
139
|
+
|
|
140
|
+
## Multi-turn conversations
|
|
141
|
+
|
|
142
|
+
```python
|
|
143
|
+
# Initialize conversation
|
|
144
|
+
conv = conv_templates["llava_v1"].copy()
|
|
145
|
+
|
|
146
|
+
# Turn 1
|
|
147
|
+
conv.append_message(conv.roles[0], DEFAULT_IMAGE_TOKEN + "\nWhat is in this image?")
|
|
148
|
+
conv.append_message(conv.roles[1], None)
|
|
149
|
+
response1 = generate(conv, model, image) # "A dog playing in a park"
|
|
150
|
+
|
|
151
|
+
# Turn 2
|
|
152
|
+
conv.messages[-1][1] = response1 # Add previous response
|
|
153
|
+
conv.append_message(conv.roles[0], "What breed is the dog?")
|
|
154
|
+
conv.append_message(conv.roles[1], None)
|
|
155
|
+
response2 = generate(conv, model, image) # "Golden Retriever"
|
|
156
|
+
|
|
157
|
+
# Turn 3
|
|
158
|
+
conv.messages[-1][1] = response2
|
|
159
|
+
conv.append_message(conv.roles[0], "What time of day is it?")
|
|
160
|
+
conv.append_message(conv.roles[1], None)
|
|
161
|
+
response3 = generate(conv, model, image)
|
|
162
|
+
```
|
|
163
|
+
|
|
164
|
+
## Common tasks
|
|
165
|
+
|
|
166
|
+
### Image captioning
|
|
167
|
+
|
|
168
|
+
```python
|
|
169
|
+
question = "Describe this image in detail."
|
|
170
|
+
response = ask(model, image, question)
|
|
171
|
+
```
|
|
172
|
+
|
|
173
|
+
### Visual question answering
|
|
174
|
+
|
|
175
|
+
```python
|
|
176
|
+
question = "How many people are in the image?"
|
|
177
|
+
response = ask(model, image, question)
|
|
178
|
+
```
|
|
179
|
+
|
|
180
|
+
### Object detection (textual)
|
|
181
|
+
|
|
182
|
+
```python
|
|
183
|
+
question = "List all the objects you can see in this image."
|
|
184
|
+
response = ask(model, image, question)
|
|
185
|
+
```
|
|
186
|
+
|
|
187
|
+
### Scene understanding
|
|
188
|
+
|
|
189
|
+
```python
|
|
190
|
+
question = "What is happening in this scene?"
|
|
191
|
+
response = ask(model, image, question)
|
|
192
|
+
```
|
|
193
|
+
|
|
194
|
+
### Document understanding
|
|
195
|
+
|
|
196
|
+
```python
|
|
197
|
+
question = "What is the main topic of this document?"
|
|
198
|
+
response = ask(model, document_image, question)
|
|
199
|
+
```
|
|
200
|
+
|
|
201
|
+
## Training custom model
|
|
202
|
+
|
|
203
|
+
```bash
|
|
204
|
+
# Stage 1: Feature alignment (558K image-caption pairs)
|
|
205
|
+
bash scripts/v1_5/pretrain.sh
|
|
206
|
+
|
|
207
|
+
# Stage 2: Visual instruction tuning (150K instruction data)
|
|
208
|
+
bash scripts/v1_5/finetune.sh
|
|
209
|
+
```
|
|
210
|
+
|
|
211
|
+
## Quantization (reduce VRAM)
|
|
212
|
+
|
|
213
|
+
```python
|
|
214
|
+
# 4-bit quantization
|
|
215
|
+
tokenizer, model, image_processor, context_len = load_pretrained_model(
|
|
216
|
+
model_path="liuhaotian/llava-v1.5-13b",
|
|
217
|
+
model_base=None,
|
|
218
|
+
model_name=get_model_name_from_path("liuhaotian/llava-v1.5-13b"),
|
|
219
|
+
load_4bit=True # Reduces VRAM ~4×
|
|
220
|
+
)
|
|
221
|
+
|
|
222
|
+
# 8-bit quantization
|
|
223
|
+
load_8bit=True # Reduces VRAM ~2×
|
|
224
|
+
```
|
|
225
|
+
|
|
226
|
+
## Best practices
|
|
227
|
+
|
|
228
|
+
1. **Start with 7B model** - Good quality, manageable VRAM
|
|
229
|
+
2. **Use 4-bit quantization** - Reduces VRAM significantly
|
|
230
|
+
3. **GPU required** - CPU inference extremely slow
|
|
231
|
+
4. **Clear prompts** - Specific questions get better answers
|
|
232
|
+
5. **Multi-turn conversations** - Maintain conversation context
|
|
233
|
+
6. **Temperature 0.2-0.7** - Balance creativity/consistency
|
|
234
|
+
7. **max_new_tokens 512-1024** - For detailed responses
|
|
235
|
+
8. **Batch processing** - Process multiple images sequentially
|
|
236
|
+
|
|
237
|
+
## Performance
|
|
238
|
+
|
|
239
|
+
| Model | VRAM (FP16) | VRAM (4-bit) | Speed (tokens/s) |
|
|
240
|
+
|-------|-------------|--------------|------------------|
|
|
241
|
+
| 7B | ~14 GB | ~4 GB | ~20 |
|
|
242
|
+
| 13B | ~28 GB | ~8 GB | ~12 |
|
|
243
|
+
| 34B | ~70 GB | ~18 GB | ~5 |
|
|
244
|
+
|
|
245
|
+
*On A100 GPU*
|
|
246
|
+
|
|
247
|
+
## Benchmarks
|
|
248
|
+
|
|
249
|
+
LLaVA achieves competitive scores on:
|
|
250
|
+
- **VQAv2**: 78.5%
|
|
251
|
+
- **GQA**: 62.0%
|
|
252
|
+
- **MM-Vet**: 35.4%
|
|
253
|
+
- **MMBench**: 64.3%
|
|
254
|
+
|
|
255
|
+
## Limitations
|
|
256
|
+
|
|
257
|
+
1. **Hallucinations** - May describe things not in image
|
|
258
|
+
2. **Spatial reasoning** - Struggles with precise locations
|
|
259
|
+
3. **Small text** - Difficulty reading fine print
|
|
260
|
+
4. **Object counting** - Imprecise for many objects
|
|
261
|
+
5. **VRAM requirements** - Need powerful GPU
|
|
262
|
+
6. **Inference speed** - Slower than CLIP
|
|
263
|
+
|
|
264
|
+
## Integration with frameworks
|
|
265
|
+
|
|
266
|
+
### LangChain
|
|
267
|
+
|
|
268
|
+
```python
|
|
269
|
+
from langchain.llms.base import LLM
|
|
270
|
+
|
|
271
|
+
class LLaVALLM(LLM):
|
|
272
|
+
def _call(self, prompt, stop=None):
|
|
273
|
+
# Custom LLaVA inference
|
|
274
|
+
return response
|
|
275
|
+
|
|
276
|
+
llm = LLaVALLM()
|
|
277
|
+
```
|
|
278
|
+
|
|
279
|
+
### Gradio App
|
|
280
|
+
|
|
281
|
+
```python
|
|
282
|
+
import gradio as gr
|
|
283
|
+
|
|
284
|
+
def chat(image, text, history):
|
|
285
|
+
response = ask_llava(model, image, text)
|
|
286
|
+
return response
|
|
287
|
+
|
|
288
|
+
demo = gr.ChatInterface(
|
|
289
|
+
chat,
|
|
290
|
+
additional_inputs=[gr.Image(type="pil")],
|
|
291
|
+
title="LLaVA Chat"
|
|
292
|
+
)
|
|
293
|
+
demo.launch()
|
|
294
|
+
```
|
|
295
|
+
|
|
296
|
+
## Resources
|
|
297
|
+
|
|
298
|
+
- **GitHub**: https://github.com/haotian-liu/LLaVA ⭐ 23,000+
|
|
299
|
+
- **Paper**: https://arxiv.org/abs/2304.08485
|
|
300
|
+
- **Demo**: https://llava.hliu.cc
|
|
301
|
+
- **Models**: https://huggingface.co/liuhaotian
|
|
302
|
+
- **License**: Apache 2.0
|
|
303
|
+
|
|
304
|
+
|
|
@@ -0,0 +1,197 @@
|
|
|
1
|
+
# LLaVA Training Guide
|
|
2
|
+
|
|
3
|
+
Guide to training and fine-tuning LLaVA models.
|
|
4
|
+
|
|
5
|
+
## Training stages
|
|
6
|
+
|
|
7
|
+
### Stage 1: Feature alignment (Pretraining)
|
|
8
|
+
|
|
9
|
+
**Purpose**: Align vision encoder with language model
|
|
10
|
+
|
|
11
|
+
**Data**: 558K image-caption pairs (CC3M subset)
|
|
12
|
+
|
|
13
|
+
```bash
|
|
14
|
+
# Download pretrained projector or train from scratch
|
|
15
|
+
bash scripts/v1_5/pretrain.sh
|
|
16
|
+
```
|
|
17
|
+
|
|
18
|
+
**Configuration:**
|
|
19
|
+
- Base model: Vicuna-7B or LLaMA-2-7B
|
|
20
|
+
- Vision encoder: CLIP ViT-L/14
|
|
21
|
+
- Training time: ~20 hours on 8× A100
|
|
22
|
+
|
|
23
|
+
### Stage 2: Visual instruction tuning
|
|
24
|
+
|
|
25
|
+
**Purpose**: Teach model to follow visual instructions
|
|
26
|
+
|
|
27
|
+
**Data**: 150K GPT-generated multimodal instruction data
|
|
28
|
+
|
|
29
|
+
```bash
|
|
30
|
+
# Fine-tune with instruction data
|
|
31
|
+
bash scripts/v1_5/finetune.sh
|
|
32
|
+
```
|
|
33
|
+
|
|
34
|
+
**Configuration:**
|
|
35
|
+
- Epochs: 1
|
|
36
|
+
- Batch size: 128 (across 8 GPUs)
|
|
37
|
+
- Learning rate: 2e-5
|
|
38
|
+
- Training time: ~24 hours on 8× A100
|
|
39
|
+
|
|
40
|
+
## Data format
|
|
41
|
+
|
|
42
|
+
### Instruction data format
|
|
43
|
+
|
|
44
|
+
```json
|
|
45
|
+
[
|
|
46
|
+
{
|
|
47
|
+
"id": "001",
|
|
48
|
+
"image": "path/to/image.jpg",
|
|
49
|
+
"conversations": [
|
|
50
|
+
{
|
|
51
|
+
"from": "human",
|
|
52
|
+
"value": "<image>\nWhat is in this image?"
|
|
53
|
+
},
|
|
54
|
+
{
|
|
55
|
+
"from": "gpt",
|
|
56
|
+
"value": "The image shows a dog playing in a park."
|
|
57
|
+
},
|
|
58
|
+
{
|
|
59
|
+
"from": "human",
|
|
60
|
+
"value": "What breed is the dog?"
|
|
61
|
+
},
|
|
62
|
+
{
|
|
63
|
+
"from": "gpt",
|
|
64
|
+
"value": "It appears to be a Golden Retriever."
|
|
65
|
+
}
|
|
66
|
+
]
|
|
67
|
+
}
|
|
68
|
+
]
|
|
69
|
+
```
|
|
70
|
+
|
|
71
|
+
## Fine-tuning on custom data
|
|
72
|
+
|
|
73
|
+
### Prepare your data
|
|
74
|
+
|
|
75
|
+
```python
|
|
76
|
+
import json
|
|
77
|
+
|
|
78
|
+
# Create instruction data
|
|
79
|
+
data = []
|
|
80
|
+
for image_path, qa_pairs in your_dataset:
|
|
81
|
+
conversations = []
|
|
82
|
+
for q, a in qa_pairs:
|
|
83
|
+
conversations.append({"from": "human", "value": f"<image>\n{q}"})
|
|
84
|
+
conversations.append({"from": "gpt", "value": a})
|
|
85
|
+
|
|
86
|
+
data.append({
|
|
87
|
+
"id": str(len(data)),
|
|
88
|
+
"image": image_path,
|
|
89
|
+
"conversations": conversations
|
|
90
|
+
})
|
|
91
|
+
|
|
92
|
+
# Save
|
|
93
|
+
with open("custom_data.json", "w") as f:
|
|
94
|
+
json.dump(data, f, indent=2)
|
|
95
|
+
```
|
|
96
|
+
|
|
97
|
+
### Fine-tune script
|
|
98
|
+
|
|
99
|
+
```bash
|
|
100
|
+
#!/bin/bash
|
|
101
|
+
|
|
102
|
+
# Set paths
|
|
103
|
+
DATA_PATH="custom_data.json"
|
|
104
|
+
IMAGE_FOLDER="path/to/images"
|
|
105
|
+
MODEL_PATH="liuhaotian/llava-v1.5-7b"
|
|
106
|
+
OUTPUT_DIR="./checkpoints/llava-custom"
|
|
107
|
+
|
|
108
|
+
# Fine-tune
|
|
109
|
+
deepspeed llava/train/train_mem.py \
|
|
110
|
+
--deepspeed ./scripts/zero2.json \
|
|
111
|
+
--model_name_or_path $MODEL_PATH \
|
|
112
|
+
--version v1 \
|
|
113
|
+
--data_path $DATA_PATH \
|
|
114
|
+
--image_folder $IMAGE_FOLDER \
|
|
115
|
+
--vision_tower openai/clip-vit-large-patch14-336 \
|
|
116
|
+
--mm_projector_type mlp2x_gelu \
|
|
117
|
+
--mm_vision_select_layer -2 \
|
|
118
|
+
--mm_use_im_start_end False \
|
|
119
|
+
--mm_use_im_patch_token False \
|
|
120
|
+
--image_aspect_ratio pad \
|
|
121
|
+
--group_by_modality_length True \
|
|
122
|
+
--bf16 True \
|
|
123
|
+
--output_dir $OUTPUT_DIR \
|
|
124
|
+
--num_train_epochs 1 \
|
|
125
|
+
--per_device_train_batch_size 16 \
|
|
126
|
+
--per_device_eval_batch_size 4 \
|
|
127
|
+
--gradient_accumulation_steps 1 \
|
|
128
|
+
--evaluation_strategy "no" \
|
|
129
|
+
--save_strategy "steps" \
|
|
130
|
+
--save_steps 50000 \
|
|
131
|
+
--save_total_limit 1 \
|
|
132
|
+
--learning_rate 2e-5 \
|
|
133
|
+
--weight_decay 0. \
|
|
134
|
+
--warmup_ratio 0.03 \
|
|
135
|
+
--lr_scheduler_type "cosine" \
|
|
136
|
+
--logging_steps 1 \
|
|
137
|
+
--tf32 True \
|
|
138
|
+
--model_max_length 2048 \
|
|
139
|
+
--gradient_checkpointing True \
|
|
140
|
+
--dataloader_num_workers 4 \
|
|
141
|
+
--lazy_preprocess True \
|
|
142
|
+
--report_to wandb
|
|
143
|
+
```
|
|
144
|
+
|
|
145
|
+
## LoRA fine-tuning (memory efficient)
|
|
146
|
+
|
|
147
|
+
```python
|
|
148
|
+
from peft import LoraConfig, get_peft_model
|
|
149
|
+
|
|
150
|
+
# LoRA config
|
|
151
|
+
lora_config = LoraConfig(
|
|
152
|
+
r=8, # LoRA rank
|
|
153
|
+
lora_alpha=16,
|
|
154
|
+
target_modules=["q_proj", "v_proj"],
|
|
155
|
+
lora_dropout=0.05,
|
|
156
|
+
bias="none",
|
|
157
|
+
task_type="CAUSAL_LM"
|
|
158
|
+
)
|
|
159
|
+
|
|
160
|
+
# Apply LoRA
|
|
161
|
+
model = get_peft_model(base_model, lora_config)
|
|
162
|
+
|
|
163
|
+
# Train with much lower memory
|
|
164
|
+
```
|
|
165
|
+
|
|
166
|
+
## Hardware requirements
|
|
167
|
+
|
|
168
|
+
### Full fine-tuning
|
|
169
|
+
|
|
170
|
+
- **7B model**: 8× A100 (40GB)
|
|
171
|
+
- **13B model**: 8× A100 (80GB)
|
|
172
|
+
- **Training time**: 20-48 hours
|
|
173
|
+
|
|
174
|
+
### LoRA fine-tuning
|
|
175
|
+
|
|
176
|
+
- **7B model**: 1× A100 (40GB)
|
|
177
|
+
- **13B model**: 2× A100 (40GB)
|
|
178
|
+
- **Training time**: 10-24 hours
|
|
179
|
+
|
|
180
|
+
## Best practices
|
|
181
|
+
|
|
182
|
+
1. **Start with pretrained** - Don't train from scratch
|
|
183
|
+
2. **Use LoRA for efficiency** - 10× less memory
|
|
184
|
+
3. **Quality over quantity** - 1K high-quality > 10K low-quality
|
|
185
|
+
4. **Multi-turn conversations** - More engaging than single Q&A
|
|
186
|
+
5. **Diverse images** - Cover different scenarios
|
|
187
|
+
6. **Clear instructions** - Specific questions get better answers
|
|
188
|
+
7. **Monitor loss** - Should decrease smoothly
|
|
189
|
+
8. **Save checkpoints** - Training can fail
|
|
190
|
+
9. **Test regularly** - Validate on held-out set
|
|
191
|
+
10. **Use DeepSpeed** - For multi-GPU training
|
|
192
|
+
|
|
193
|
+
## Resources
|
|
194
|
+
|
|
195
|
+
- **Training script**: https://github.com/haotian-liu/LLaVA/tree/main/scripts
|
|
196
|
+
- **Data format**: https://github.com/haotian-liu/LLaVA/blob/main/docs/Data.md
|
|
197
|
+
- **Paper**: https://arxiv.org/abs/2304.08485
|