@synsci/cli-darwin-x64 1.1.49
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/bin/skills/accelerate/SKILL.md +332 -0
- package/bin/skills/accelerate/references/custom-plugins.md +453 -0
- package/bin/skills/accelerate/references/megatron-integration.md +489 -0
- package/bin/skills/accelerate/references/performance.md +525 -0
- package/bin/skills/audiocraft/SKILL.md +564 -0
- package/bin/skills/audiocraft/references/advanced-usage.md +666 -0
- package/bin/skills/audiocraft/references/troubleshooting.md +504 -0
- package/bin/skills/autogpt/SKILL.md +403 -0
- package/bin/skills/autogpt/references/advanced-usage.md +535 -0
- package/bin/skills/autogpt/references/troubleshooting.md +420 -0
- package/bin/skills/awq/SKILL.md +310 -0
- package/bin/skills/awq/references/advanced-usage.md +324 -0
- package/bin/skills/awq/references/troubleshooting.md +344 -0
- package/bin/skills/axolotl/SKILL.md +158 -0
- package/bin/skills/axolotl/references/api.md +5548 -0
- package/bin/skills/axolotl/references/dataset-formats.md +1029 -0
- package/bin/skills/axolotl/references/index.md +15 -0
- package/bin/skills/axolotl/references/other.md +3563 -0
- package/bin/skills/bigcode-evaluation-harness/SKILL.md +405 -0
- package/bin/skills/bigcode-evaluation-harness/references/benchmarks.md +393 -0
- package/bin/skills/bigcode-evaluation-harness/references/custom-tasks.md +424 -0
- package/bin/skills/bigcode-evaluation-harness/references/issues.md +394 -0
- package/bin/skills/bitsandbytes/SKILL.md +411 -0
- package/bin/skills/bitsandbytes/references/memory-optimization.md +521 -0
- package/bin/skills/bitsandbytes/references/qlora-training.md +521 -0
- package/bin/skills/bitsandbytes/references/quantization-formats.md +447 -0
- package/bin/skills/blip-2/SKILL.md +564 -0
- package/bin/skills/blip-2/references/advanced-usage.md +680 -0
- package/bin/skills/blip-2/references/troubleshooting.md +526 -0
- package/bin/skills/chroma/SKILL.md +406 -0
- package/bin/skills/chroma/references/integration.md +38 -0
- package/bin/skills/clip/SKILL.md +253 -0
- package/bin/skills/clip/references/applications.md +207 -0
- package/bin/skills/constitutional-ai/SKILL.md +290 -0
- package/bin/skills/crewai/SKILL.md +498 -0
- package/bin/skills/crewai/references/flows.md +438 -0
- package/bin/skills/crewai/references/tools.md +429 -0
- package/bin/skills/crewai/references/troubleshooting.md +480 -0
- package/bin/skills/deepspeed/SKILL.md +141 -0
- package/bin/skills/deepspeed/references/08.md +17 -0
- package/bin/skills/deepspeed/references/09.md +173 -0
- package/bin/skills/deepspeed/references/2020.md +378 -0
- package/bin/skills/deepspeed/references/2023.md +279 -0
- package/bin/skills/deepspeed/references/assets.md +179 -0
- package/bin/skills/deepspeed/references/index.md +35 -0
- package/bin/skills/deepspeed/references/mii.md +118 -0
- package/bin/skills/deepspeed/references/other.md +1191 -0
- package/bin/skills/deepspeed/references/tutorials.md +6554 -0
- package/bin/skills/dspy/SKILL.md +590 -0
- package/bin/skills/dspy/references/examples.md +663 -0
- package/bin/skills/dspy/references/modules.md +475 -0
- package/bin/skills/dspy/references/optimizers.md +566 -0
- package/bin/skills/faiss/SKILL.md +221 -0
- package/bin/skills/faiss/references/index_types.md +280 -0
- package/bin/skills/flash-attention/SKILL.md +367 -0
- package/bin/skills/flash-attention/references/benchmarks.md +215 -0
- package/bin/skills/flash-attention/references/transformers-integration.md +293 -0
- package/bin/skills/gguf/SKILL.md +427 -0
- package/bin/skills/gguf/references/advanced-usage.md +504 -0
- package/bin/skills/gguf/references/troubleshooting.md +442 -0
- package/bin/skills/gptq/SKILL.md +450 -0
- package/bin/skills/gptq/references/calibration.md +337 -0
- package/bin/skills/gptq/references/integration.md +129 -0
- package/bin/skills/gptq/references/troubleshooting.md +95 -0
- package/bin/skills/grpo-rl-training/README.md +97 -0
- package/bin/skills/grpo-rl-training/SKILL.md +572 -0
- package/bin/skills/grpo-rl-training/examples/reward_functions_library.py +393 -0
- package/bin/skills/grpo-rl-training/templates/basic_grpo_training.py +228 -0
- package/bin/skills/guidance/SKILL.md +572 -0
- package/bin/skills/guidance/references/backends.md +554 -0
- package/bin/skills/guidance/references/constraints.md +674 -0
- package/bin/skills/guidance/references/examples.md +767 -0
- package/bin/skills/hqq/SKILL.md +445 -0
- package/bin/skills/hqq/references/advanced-usage.md +528 -0
- package/bin/skills/hqq/references/troubleshooting.md +503 -0
- package/bin/skills/hugging-face-cli/SKILL.md +191 -0
- package/bin/skills/hugging-face-cli/references/commands.md +954 -0
- package/bin/skills/hugging-face-cli/references/examples.md +374 -0
- package/bin/skills/hugging-face-datasets/SKILL.md +547 -0
- package/bin/skills/hugging-face-datasets/examples/diverse_training_examples.json +239 -0
- package/bin/skills/hugging-face-datasets/examples/system_prompt_template.txt +196 -0
- package/bin/skills/hugging-face-datasets/examples/training_examples.json +176 -0
- package/bin/skills/hugging-face-datasets/scripts/dataset_manager.py +522 -0
- package/bin/skills/hugging-face-datasets/scripts/sql_manager.py +844 -0
- package/bin/skills/hugging-face-datasets/templates/chat.json +55 -0
- package/bin/skills/hugging-face-datasets/templates/classification.json +62 -0
- package/bin/skills/hugging-face-datasets/templates/completion.json +51 -0
- package/bin/skills/hugging-face-datasets/templates/custom.json +75 -0
- package/bin/skills/hugging-face-datasets/templates/qa.json +54 -0
- package/bin/skills/hugging-face-datasets/templates/tabular.json +81 -0
- package/bin/skills/hugging-face-evaluation/SKILL.md +656 -0
- package/bin/skills/hugging-face-evaluation/examples/USAGE_EXAMPLES.md +382 -0
- package/bin/skills/hugging-face-evaluation/examples/artificial_analysis_to_hub.py +141 -0
- package/bin/skills/hugging-face-evaluation/examples/example_readme_tables.md +135 -0
- package/bin/skills/hugging-face-evaluation/examples/metric_mapping.json +50 -0
- package/bin/skills/hugging-face-evaluation/requirements.txt +20 -0
- package/bin/skills/hugging-face-evaluation/scripts/evaluation_manager.py +1374 -0
- package/bin/skills/hugging-face-evaluation/scripts/inspect_eval_uv.py +104 -0
- package/bin/skills/hugging-face-evaluation/scripts/inspect_vllm_uv.py +317 -0
- package/bin/skills/hugging-face-evaluation/scripts/lighteval_vllm_uv.py +303 -0
- package/bin/skills/hugging-face-evaluation/scripts/run_eval_job.py +98 -0
- package/bin/skills/hugging-face-evaluation/scripts/run_vllm_eval_job.py +331 -0
- package/bin/skills/hugging-face-evaluation/scripts/test_extraction.py +206 -0
- package/bin/skills/hugging-face-jobs/SKILL.md +1041 -0
- package/bin/skills/hugging-face-jobs/index.html +216 -0
- package/bin/skills/hugging-face-jobs/references/hardware_guide.md +336 -0
- package/bin/skills/hugging-face-jobs/references/hub_saving.md +352 -0
- package/bin/skills/hugging-face-jobs/references/token_usage.md +546 -0
- package/bin/skills/hugging-face-jobs/references/troubleshooting.md +475 -0
- package/bin/skills/hugging-face-jobs/scripts/cot-self-instruct.py +718 -0
- package/bin/skills/hugging-face-jobs/scripts/finepdfs-stats.py +546 -0
- package/bin/skills/hugging-face-jobs/scripts/generate-responses.py +587 -0
- package/bin/skills/hugging-face-model-trainer/SKILL.md +711 -0
- package/bin/skills/hugging-face-model-trainer/references/gguf_conversion.md +296 -0
- package/bin/skills/hugging-face-model-trainer/references/hardware_guide.md +283 -0
- package/bin/skills/hugging-face-model-trainer/references/hub_saving.md +364 -0
- package/bin/skills/hugging-face-model-trainer/references/reliability_principles.md +371 -0
- package/bin/skills/hugging-face-model-trainer/references/trackio_guide.md +189 -0
- package/bin/skills/hugging-face-model-trainer/references/training_methods.md +150 -0
- package/bin/skills/hugging-face-model-trainer/references/training_patterns.md +203 -0
- package/bin/skills/hugging-face-model-trainer/references/troubleshooting.md +282 -0
- package/bin/skills/hugging-face-model-trainer/scripts/convert_to_gguf.py +424 -0
- package/bin/skills/hugging-face-model-trainer/scripts/dataset_inspector.py +417 -0
- package/bin/skills/hugging-face-model-trainer/scripts/estimate_cost.py +150 -0
- package/bin/skills/hugging-face-model-trainer/scripts/train_dpo_example.py +106 -0
- package/bin/skills/hugging-face-model-trainer/scripts/train_grpo_example.py +89 -0
- package/bin/skills/hugging-face-model-trainer/scripts/train_sft_example.py +122 -0
- package/bin/skills/hugging-face-paper-publisher/SKILL.md +627 -0
- package/bin/skills/hugging-face-paper-publisher/examples/example_usage.md +327 -0
- package/bin/skills/hugging-face-paper-publisher/references/quick_reference.md +216 -0
- package/bin/skills/hugging-face-paper-publisher/scripts/paper_manager.py +508 -0
- package/bin/skills/hugging-face-paper-publisher/templates/arxiv.md +299 -0
- package/bin/skills/hugging-face-paper-publisher/templates/ml-report.md +358 -0
- package/bin/skills/hugging-face-paper-publisher/templates/modern.md +319 -0
- package/bin/skills/hugging-face-paper-publisher/templates/standard.md +201 -0
- package/bin/skills/hugging-face-tool-builder/SKILL.md +115 -0
- package/bin/skills/hugging-face-tool-builder/references/baseline_hf_api.py +57 -0
- package/bin/skills/hugging-face-tool-builder/references/baseline_hf_api.sh +40 -0
- package/bin/skills/hugging-face-tool-builder/references/baseline_hf_api.tsx +57 -0
- package/bin/skills/hugging-face-tool-builder/references/find_models_by_paper.sh +230 -0
- package/bin/skills/hugging-face-tool-builder/references/hf_enrich_models.sh +96 -0
- package/bin/skills/hugging-face-tool-builder/references/hf_model_card_frontmatter.sh +188 -0
- package/bin/skills/hugging-face-tool-builder/references/hf_model_papers_auth.sh +171 -0
- package/bin/skills/hugging-face-trackio/SKILL.md +65 -0
- package/bin/skills/hugging-face-trackio/references/logging_metrics.md +206 -0
- package/bin/skills/hugging-face-trackio/references/retrieving_metrics.md +223 -0
- package/bin/skills/huggingface-tokenizers/SKILL.md +516 -0
- package/bin/skills/huggingface-tokenizers/references/algorithms.md +653 -0
- package/bin/skills/huggingface-tokenizers/references/integration.md +637 -0
- package/bin/skills/huggingface-tokenizers/references/pipeline.md +723 -0
- package/bin/skills/huggingface-tokenizers/references/training.md +565 -0
- package/bin/skills/instructor/SKILL.md +740 -0
- package/bin/skills/instructor/references/examples.md +107 -0
- package/bin/skills/instructor/references/providers.md +70 -0
- package/bin/skills/instructor/references/validation.md +606 -0
- package/bin/skills/knowledge-distillation/SKILL.md +458 -0
- package/bin/skills/knowledge-distillation/references/minillm.md +334 -0
- package/bin/skills/lambda-labs/SKILL.md +545 -0
- package/bin/skills/lambda-labs/references/advanced-usage.md +611 -0
- package/bin/skills/lambda-labs/references/troubleshooting.md +530 -0
- package/bin/skills/langchain/SKILL.md +480 -0
- package/bin/skills/langchain/references/agents.md +499 -0
- package/bin/skills/langchain/references/integration.md +562 -0
- package/bin/skills/langchain/references/rag.md +600 -0
- package/bin/skills/langsmith/SKILL.md +422 -0
- package/bin/skills/langsmith/references/advanced-usage.md +548 -0
- package/bin/skills/langsmith/references/troubleshooting.md +537 -0
- package/bin/skills/litgpt/SKILL.md +469 -0
- package/bin/skills/litgpt/references/custom-models.md +568 -0
- package/bin/skills/litgpt/references/distributed-training.md +451 -0
- package/bin/skills/litgpt/references/supported-models.md +336 -0
- package/bin/skills/litgpt/references/training-recipes.md +619 -0
- package/bin/skills/llama-cpp/SKILL.md +258 -0
- package/bin/skills/llama-cpp/references/optimization.md +89 -0
- package/bin/skills/llama-cpp/references/quantization.md +213 -0
- package/bin/skills/llama-cpp/references/server.md +125 -0
- package/bin/skills/llama-factory/SKILL.md +80 -0
- package/bin/skills/llama-factory/references/_images.md +23 -0
- package/bin/skills/llama-factory/references/advanced.md +1055 -0
- package/bin/skills/llama-factory/references/getting_started.md +349 -0
- package/bin/skills/llama-factory/references/index.md +19 -0
- package/bin/skills/llama-factory/references/other.md +31 -0
- package/bin/skills/llamaguard/SKILL.md +337 -0
- package/bin/skills/llamaindex/SKILL.md +569 -0
- package/bin/skills/llamaindex/references/agents.md +83 -0
- package/bin/skills/llamaindex/references/data_connectors.md +108 -0
- package/bin/skills/llamaindex/references/query_engines.md +406 -0
- package/bin/skills/llava/SKILL.md +304 -0
- package/bin/skills/llava/references/training.md +197 -0
- package/bin/skills/lm-evaluation-harness/SKILL.md +490 -0
- package/bin/skills/lm-evaluation-harness/references/api-evaluation.md +490 -0
- package/bin/skills/lm-evaluation-harness/references/benchmark-guide.md +488 -0
- package/bin/skills/lm-evaluation-harness/references/custom-tasks.md +602 -0
- package/bin/skills/lm-evaluation-harness/references/distributed-eval.md +519 -0
- package/bin/skills/long-context/SKILL.md +536 -0
- package/bin/skills/long-context/references/extension_methods.md +468 -0
- package/bin/skills/long-context/references/fine_tuning.md +611 -0
- package/bin/skills/long-context/references/rope.md +402 -0
- package/bin/skills/mamba/SKILL.md +260 -0
- package/bin/skills/mamba/references/architecture-details.md +206 -0
- package/bin/skills/mamba/references/benchmarks.md +255 -0
- package/bin/skills/mamba/references/training-guide.md +388 -0
- package/bin/skills/megatron-core/SKILL.md +366 -0
- package/bin/skills/megatron-core/references/benchmarks.md +249 -0
- package/bin/skills/megatron-core/references/parallelism-guide.md +404 -0
- package/bin/skills/megatron-core/references/production-examples.md +473 -0
- package/bin/skills/megatron-core/references/training-recipes.md +547 -0
- package/bin/skills/miles/SKILL.md +315 -0
- package/bin/skills/miles/references/api-reference.md +141 -0
- package/bin/skills/miles/references/troubleshooting.md +352 -0
- package/bin/skills/mlflow/SKILL.md +704 -0
- package/bin/skills/mlflow/references/deployment.md +744 -0
- package/bin/skills/mlflow/references/model-registry.md +770 -0
- package/bin/skills/mlflow/references/tracking.md +680 -0
- package/bin/skills/modal/SKILL.md +341 -0
- package/bin/skills/modal/references/advanced-usage.md +503 -0
- package/bin/skills/modal/references/troubleshooting.md +494 -0
- package/bin/skills/model-merging/SKILL.md +539 -0
- package/bin/skills/model-merging/references/evaluation.md +462 -0
- package/bin/skills/model-merging/references/examples.md +428 -0
- package/bin/skills/model-merging/references/methods.md +352 -0
- package/bin/skills/model-pruning/SKILL.md +495 -0
- package/bin/skills/model-pruning/references/wanda.md +347 -0
- package/bin/skills/moe-training/SKILL.md +526 -0
- package/bin/skills/moe-training/references/architectures.md +432 -0
- package/bin/skills/moe-training/references/inference.md +348 -0
- package/bin/skills/moe-training/references/training.md +425 -0
- package/bin/skills/nanogpt/SKILL.md +290 -0
- package/bin/skills/nanogpt/references/architecture.md +382 -0
- package/bin/skills/nanogpt/references/data.md +476 -0
- package/bin/skills/nanogpt/references/training.md +564 -0
- package/bin/skills/nemo-curator/SKILL.md +383 -0
- package/bin/skills/nemo-curator/references/deduplication.md +87 -0
- package/bin/skills/nemo-curator/references/filtering.md +102 -0
- package/bin/skills/nemo-evaluator/SKILL.md +494 -0
- package/bin/skills/nemo-evaluator/references/adapter-system.md +340 -0
- package/bin/skills/nemo-evaluator/references/configuration.md +447 -0
- package/bin/skills/nemo-evaluator/references/custom-benchmarks.md +315 -0
- package/bin/skills/nemo-evaluator/references/execution-backends.md +361 -0
- package/bin/skills/nemo-guardrails/SKILL.md +297 -0
- package/bin/skills/nnsight/SKILL.md +436 -0
- package/bin/skills/nnsight/references/README.md +78 -0
- package/bin/skills/nnsight/references/api.md +344 -0
- package/bin/skills/nnsight/references/tutorials.md +300 -0
- package/bin/skills/openrlhf/SKILL.md +249 -0
- package/bin/skills/openrlhf/references/algorithm-comparison.md +404 -0
- package/bin/skills/openrlhf/references/custom-rewards.md +530 -0
- package/bin/skills/openrlhf/references/hybrid-engine.md +287 -0
- package/bin/skills/openrlhf/references/multi-node-training.md +454 -0
- package/bin/skills/outlines/SKILL.md +652 -0
- package/bin/skills/outlines/references/backends.md +615 -0
- package/bin/skills/outlines/references/examples.md +773 -0
- package/bin/skills/outlines/references/json_generation.md +652 -0
- package/bin/skills/peft/SKILL.md +431 -0
- package/bin/skills/peft/references/advanced-usage.md +514 -0
- package/bin/skills/peft/references/troubleshooting.md +480 -0
- package/bin/skills/phoenix/SKILL.md +475 -0
- package/bin/skills/phoenix/references/advanced-usage.md +619 -0
- package/bin/skills/phoenix/references/troubleshooting.md +538 -0
- package/bin/skills/pinecone/SKILL.md +358 -0
- package/bin/skills/pinecone/references/deployment.md +181 -0
- package/bin/skills/pytorch-fsdp/SKILL.md +126 -0
- package/bin/skills/pytorch-fsdp/references/index.md +7 -0
- package/bin/skills/pytorch-fsdp/references/other.md +4249 -0
- package/bin/skills/pytorch-lightning/SKILL.md +346 -0
- package/bin/skills/pytorch-lightning/references/callbacks.md +436 -0
- package/bin/skills/pytorch-lightning/references/distributed.md +490 -0
- package/bin/skills/pytorch-lightning/references/hyperparameter-tuning.md +556 -0
- package/bin/skills/pyvene/SKILL.md +473 -0
- package/bin/skills/pyvene/references/README.md +73 -0
- package/bin/skills/pyvene/references/api.md +383 -0
- package/bin/skills/pyvene/references/tutorials.md +376 -0
- package/bin/skills/qdrant/SKILL.md +493 -0
- package/bin/skills/qdrant/references/advanced-usage.md +648 -0
- package/bin/skills/qdrant/references/troubleshooting.md +631 -0
- package/bin/skills/ray-data/SKILL.md +326 -0
- package/bin/skills/ray-data/references/integration.md +82 -0
- package/bin/skills/ray-data/references/transformations.md +83 -0
- package/bin/skills/ray-train/SKILL.md +406 -0
- package/bin/skills/ray-train/references/multi-node.md +628 -0
- package/bin/skills/rwkv/SKILL.md +260 -0
- package/bin/skills/rwkv/references/architecture-details.md +344 -0
- package/bin/skills/rwkv/references/rwkv7.md +386 -0
- package/bin/skills/rwkv/references/state-management.md +369 -0
- package/bin/skills/saelens/SKILL.md +386 -0
- package/bin/skills/saelens/references/README.md +70 -0
- package/bin/skills/saelens/references/api.md +333 -0
- package/bin/skills/saelens/references/tutorials.md +318 -0
- package/bin/skills/segment-anything/SKILL.md +500 -0
- package/bin/skills/segment-anything/references/advanced-usage.md +589 -0
- package/bin/skills/segment-anything/references/troubleshooting.md +484 -0
- package/bin/skills/sentence-transformers/SKILL.md +255 -0
- package/bin/skills/sentence-transformers/references/models.md +123 -0
- package/bin/skills/sentencepiece/SKILL.md +235 -0
- package/bin/skills/sentencepiece/references/algorithms.md +200 -0
- package/bin/skills/sentencepiece/references/training.md +304 -0
- package/bin/skills/sglang/SKILL.md +442 -0
- package/bin/skills/sglang/references/deployment.md +490 -0
- package/bin/skills/sglang/references/radix-attention.md +413 -0
- package/bin/skills/sglang/references/structured-generation.md +541 -0
- package/bin/skills/simpo/SKILL.md +219 -0
- package/bin/skills/simpo/references/datasets.md +478 -0
- package/bin/skills/simpo/references/hyperparameters.md +452 -0
- package/bin/skills/simpo/references/loss-functions.md +350 -0
- package/bin/skills/skypilot/SKILL.md +509 -0
- package/bin/skills/skypilot/references/advanced-usage.md +491 -0
- package/bin/skills/skypilot/references/troubleshooting.md +570 -0
- package/bin/skills/slime/SKILL.md +464 -0
- package/bin/skills/slime/references/api-reference.md +392 -0
- package/bin/skills/slime/references/troubleshooting.md +386 -0
- package/bin/skills/speculative-decoding/SKILL.md +467 -0
- package/bin/skills/speculative-decoding/references/lookahead.md +309 -0
- package/bin/skills/speculative-decoding/references/medusa.md +350 -0
- package/bin/skills/stable-diffusion/SKILL.md +519 -0
- package/bin/skills/stable-diffusion/references/advanced-usage.md +716 -0
- package/bin/skills/stable-diffusion/references/troubleshooting.md +555 -0
- package/bin/skills/tensorboard/SKILL.md +629 -0
- package/bin/skills/tensorboard/references/integrations.md +638 -0
- package/bin/skills/tensorboard/references/profiling.md +545 -0
- package/bin/skills/tensorboard/references/visualization.md +620 -0
- package/bin/skills/tensorrt-llm/SKILL.md +187 -0
- package/bin/skills/tensorrt-llm/references/multi-gpu.md +298 -0
- package/bin/skills/tensorrt-llm/references/optimization.md +242 -0
- package/bin/skills/tensorrt-llm/references/serving.md +470 -0
- package/bin/skills/tinker/SKILL.md +362 -0
- package/bin/skills/tinker/references/api-reference.md +168 -0
- package/bin/skills/tinker/references/getting-started.md +157 -0
- package/bin/skills/tinker/references/loss-functions.md +163 -0
- package/bin/skills/tinker/references/models-and-lora.md +139 -0
- package/bin/skills/tinker/references/recipes.md +280 -0
- package/bin/skills/tinker/references/reinforcement-learning.md +212 -0
- package/bin/skills/tinker/references/rendering.md +243 -0
- package/bin/skills/tinker/references/supervised-learning.md +232 -0
- package/bin/skills/tinker-training-cost/SKILL.md +187 -0
- package/bin/skills/tinker-training-cost/scripts/calculate_cost.py +123 -0
- package/bin/skills/torchforge/SKILL.md +433 -0
- package/bin/skills/torchforge/references/api-reference.md +327 -0
- package/bin/skills/torchforge/references/troubleshooting.md +409 -0
- package/bin/skills/torchtitan/SKILL.md +358 -0
- package/bin/skills/torchtitan/references/checkpoint.md +181 -0
- package/bin/skills/torchtitan/references/custom-models.md +258 -0
- package/bin/skills/torchtitan/references/float8.md +133 -0
- package/bin/skills/torchtitan/references/fsdp.md +126 -0
- package/bin/skills/transformer-lens/SKILL.md +346 -0
- package/bin/skills/transformer-lens/references/README.md +54 -0
- package/bin/skills/transformer-lens/references/api.md +362 -0
- package/bin/skills/transformer-lens/references/tutorials.md +339 -0
- package/bin/skills/trl-fine-tuning/SKILL.md +455 -0
- package/bin/skills/trl-fine-tuning/references/dpo-variants.md +227 -0
- package/bin/skills/trl-fine-tuning/references/online-rl.md +82 -0
- package/bin/skills/trl-fine-tuning/references/reward-modeling.md +122 -0
- package/bin/skills/trl-fine-tuning/references/sft-training.md +168 -0
- package/bin/skills/unsloth/SKILL.md +80 -0
- package/bin/skills/unsloth/references/index.md +7 -0
- package/bin/skills/unsloth/references/llms-full.md +16799 -0
- package/bin/skills/unsloth/references/llms-txt.md +12044 -0
- package/bin/skills/unsloth/references/llms.md +82 -0
- package/bin/skills/verl/SKILL.md +391 -0
- package/bin/skills/verl/references/api-reference.md +301 -0
- package/bin/skills/verl/references/troubleshooting.md +391 -0
- package/bin/skills/vllm/SKILL.md +364 -0
- package/bin/skills/vllm/references/optimization.md +226 -0
- package/bin/skills/vllm/references/quantization.md +284 -0
- package/bin/skills/vllm/references/server-deployment.md +255 -0
- package/bin/skills/vllm/references/troubleshooting.md +447 -0
- package/bin/skills/weights-and-biases/SKILL.md +590 -0
- package/bin/skills/weights-and-biases/references/artifacts.md +584 -0
- package/bin/skills/weights-and-biases/references/integrations.md +700 -0
- package/bin/skills/weights-and-biases/references/sweeps.md +847 -0
- package/bin/skills/whisper/SKILL.md +317 -0
- package/bin/skills/whisper/references/languages.md +189 -0
- package/bin/synsc +0 -0
- package/package.json +10 -0
|
@@ -0,0 +1,500 @@
|
|
|
1
|
+
---
|
|
2
|
+
name: segment-anything-model
|
|
3
|
+
description: Foundation model for image segmentation with zero-shot transfer. Use when you need to segment any object in images using points, boxes, or masks as prompts, or automatically generate all object masks in an image.
|
|
4
|
+
version: 1.0.0
|
|
5
|
+
author: Synthetic Sciences
|
|
6
|
+
license: MIT
|
|
7
|
+
tags: [Multimodal, Image Segmentation, Computer Vision, SAM, Zero-Shot]
|
|
8
|
+
dependencies: [segment-anything, transformers>=4.30.0, torch>=1.7.0]
|
|
9
|
+
---
|
|
10
|
+
|
|
11
|
+
# Segment Anything Model (SAM)
|
|
12
|
+
|
|
13
|
+
Comprehensive guide to using Meta AI's Segment Anything Model for zero-shot image segmentation.
|
|
14
|
+
|
|
15
|
+
## When to use SAM
|
|
16
|
+
|
|
17
|
+
**Use SAM when:**
|
|
18
|
+
- Need to segment any object in images without task-specific training
|
|
19
|
+
- Building interactive annotation tools with point/box prompts
|
|
20
|
+
- Generating training data for other vision models
|
|
21
|
+
- Need zero-shot transfer to new image domains
|
|
22
|
+
- Building object detection/segmentation pipelines
|
|
23
|
+
- Processing medical, satellite, or domain-specific images
|
|
24
|
+
|
|
25
|
+
**Key features:**
|
|
26
|
+
- **Zero-shot segmentation**: Works on any image domain without fine-tuning
|
|
27
|
+
- **Flexible prompts**: Points, bounding boxes, or previous masks
|
|
28
|
+
- **Automatic segmentation**: Generate all object masks automatically
|
|
29
|
+
- **High quality**: Trained on 1.1 billion masks from 11 million images
|
|
30
|
+
- **Multiple model sizes**: ViT-B (fastest), ViT-L, ViT-H (most accurate)
|
|
31
|
+
- **ONNX export**: Deploy in browsers and edge devices
|
|
32
|
+
|
|
33
|
+
**Use alternatives instead:**
|
|
34
|
+
- **YOLO/Detectron2**: For real-time object detection with classes
|
|
35
|
+
- **Mask2Former**: For semantic/panoptic segmentation with categories
|
|
36
|
+
- **GroundingDINO + SAM**: For text-prompted segmentation
|
|
37
|
+
- **SAM 2**: For video segmentation tasks
|
|
38
|
+
|
|
39
|
+
## Quick start
|
|
40
|
+
|
|
41
|
+
### Installation
|
|
42
|
+
|
|
43
|
+
```bash
|
|
44
|
+
# From GitHub
|
|
45
|
+
pip install git+https://github.com/facebookresearch/segment-anything.git
|
|
46
|
+
|
|
47
|
+
# Optional dependencies
|
|
48
|
+
pip install opencv-python pycocotools matplotlib
|
|
49
|
+
|
|
50
|
+
# Or use HuggingFace transformers
|
|
51
|
+
pip install transformers
|
|
52
|
+
```
|
|
53
|
+
|
|
54
|
+
### Download checkpoints
|
|
55
|
+
|
|
56
|
+
```bash
|
|
57
|
+
# ViT-H (largest, most accurate) - 2.4GB
|
|
58
|
+
wget https://dl.fbaipublicfiles.com/segment_anything/sam_vit_h_4b8939.pth
|
|
59
|
+
|
|
60
|
+
# ViT-L (medium) - 1.2GB
|
|
61
|
+
wget https://dl.fbaipublicfiles.com/segment_anything/sam_vit_l_0b3195.pth
|
|
62
|
+
|
|
63
|
+
# ViT-B (smallest, fastest) - 375MB
|
|
64
|
+
wget https://dl.fbaipublicfiles.com/segment_anything/sam_vit_b_01ec64.pth
|
|
65
|
+
```
|
|
66
|
+
|
|
67
|
+
### Basic usage with SamPredictor
|
|
68
|
+
|
|
69
|
+
```python
|
|
70
|
+
import numpy as np
|
|
71
|
+
from segment_anything import sam_model_registry, SamPredictor
|
|
72
|
+
|
|
73
|
+
# Load model
|
|
74
|
+
sam = sam_model_registry["vit_h"](checkpoint="sam_vit_h_4b8939.pth")
|
|
75
|
+
sam.to(device="cuda")
|
|
76
|
+
|
|
77
|
+
# Create predictor
|
|
78
|
+
predictor = SamPredictor(sam)
|
|
79
|
+
|
|
80
|
+
# Set image (computes embeddings once)
|
|
81
|
+
image = cv2.imread("image.jpg")
|
|
82
|
+
image = cv2.cvtColor(image, cv2.COLOR_BGR2RGB)
|
|
83
|
+
predictor.set_image(image)
|
|
84
|
+
|
|
85
|
+
# Predict with point prompts
|
|
86
|
+
input_point = np.array([[500, 375]]) # (x, y) coordinates
|
|
87
|
+
input_label = np.array([1]) # 1 = foreground, 0 = background
|
|
88
|
+
|
|
89
|
+
masks, scores, logits = predictor.predict(
|
|
90
|
+
point_coords=input_point,
|
|
91
|
+
point_labels=input_label,
|
|
92
|
+
multimask_output=True # Returns 3 mask options
|
|
93
|
+
)
|
|
94
|
+
|
|
95
|
+
# Select best mask
|
|
96
|
+
best_mask = masks[np.argmax(scores)]
|
|
97
|
+
```
|
|
98
|
+
|
|
99
|
+
### HuggingFace Transformers
|
|
100
|
+
|
|
101
|
+
```python
|
|
102
|
+
import torch
|
|
103
|
+
from PIL import Image
|
|
104
|
+
from transformers import SamModel, SamProcessor
|
|
105
|
+
|
|
106
|
+
# Load model and processor
|
|
107
|
+
model = SamModel.from_pretrained("facebook/sam-vit-huge")
|
|
108
|
+
processor = SamProcessor.from_pretrained("facebook/sam-vit-huge")
|
|
109
|
+
model.to("cuda")
|
|
110
|
+
|
|
111
|
+
# Process image with point prompt
|
|
112
|
+
image = Image.open("image.jpg")
|
|
113
|
+
input_points = [[[450, 600]]] # Batch of points
|
|
114
|
+
|
|
115
|
+
inputs = processor(image, input_points=input_points, return_tensors="pt")
|
|
116
|
+
inputs = {k: v.to("cuda") for k, v in inputs.items()}
|
|
117
|
+
|
|
118
|
+
# Generate masks
|
|
119
|
+
with torch.no_grad():
|
|
120
|
+
outputs = model(**inputs)
|
|
121
|
+
|
|
122
|
+
# Post-process masks to original size
|
|
123
|
+
masks = processor.image_processor.post_process_masks(
|
|
124
|
+
outputs.pred_masks.cpu(),
|
|
125
|
+
inputs["original_sizes"].cpu(),
|
|
126
|
+
inputs["reshaped_input_sizes"].cpu()
|
|
127
|
+
)
|
|
128
|
+
```
|
|
129
|
+
|
|
130
|
+
## Core concepts
|
|
131
|
+
|
|
132
|
+
### Model architecture
|
|
133
|
+
|
|
134
|
+
```
|
|
135
|
+
SAM Architecture:
|
|
136
|
+
┌─────────────────┐ ┌─────────────────┐ ┌─────────────────┐
|
|
137
|
+
│ Image Encoder │────▶│ Prompt Encoder │────▶│ Mask Decoder │
|
|
138
|
+
│ (ViT) │ │ (Points/Boxes) │ │ (Transformer) │
|
|
139
|
+
└─────────────────┘ └─────────────────┘ └─────────────────┘
|
|
140
|
+
│ │ │
|
|
141
|
+
Image Embeddings Prompt Embeddings Masks + IoU
|
|
142
|
+
(computed once) (per prompt) predictions
|
|
143
|
+
```
|
|
144
|
+
|
|
145
|
+
### Model variants
|
|
146
|
+
|
|
147
|
+
| Model | Checkpoint | Size | Speed | Accuracy |
|
|
148
|
+
|-------|------------|------|-------|----------|
|
|
149
|
+
| ViT-H | `vit_h` | 2.4 GB | Slowest | Best |
|
|
150
|
+
| ViT-L | `vit_l` | 1.2 GB | Medium | Good |
|
|
151
|
+
| ViT-B | `vit_b` | 375 MB | Fastest | Good |
|
|
152
|
+
|
|
153
|
+
### Prompt types
|
|
154
|
+
|
|
155
|
+
| Prompt | Description | Use Case |
|
|
156
|
+
|--------|-------------|----------|
|
|
157
|
+
| Point (foreground) | Click on object | Single object selection |
|
|
158
|
+
| Point (background) | Click outside object | Exclude regions |
|
|
159
|
+
| Bounding box | Rectangle around object | Larger objects |
|
|
160
|
+
| Previous mask | Low-res mask input | Iterative refinement |
|
|
161
|
+
|
|
162
|
+
## Interactive segmentation
|
|
163
|
+
|
|
164
|
+
### Point prompts
|
|
165
|
+
|
|
166
|
+
```python
|
|
167
|
+
# Single foreground point
|
|
168
|
+
input_point = np.array([[500, 375]])
|
|
169
|
+
input_label = np.array([1])
|
|
170
|
+
|
|
171
|
+
masks, scores, logits = predictor.predict(
|
|
172
|
+
point_coords=input_point,
|
|
173
|
+
point_labels=input_label,
|
|
174
|
+
multimask_output=True
|
|
175
|
+
)
|
|
176
|
+
|
|
177
|
+
# Multiple points (foreground + background)
|
|
178
|
+
input_points = np.array([[500, 375], [600, 400], [450, 300]])
|
|
179
|
+
input_labels = np.array([1, 1, 0]) # 2 foreground, 1 background
|
|
180
|
+
|
|
181
|
+
masks, scores, logits = predictor.predict(
|
|
182
|
+
point_coords=input_points,
|
|
183
|
+
point_labels=input_labels,
|
|
184
|
+
multimask_output=False # Single mask when prompts are clear
|
|
185
|
+
)
|
|
186
|
+
```
|
|
187
|
+
|
|
188
|
+
### Box prompts
|
|
189
|
+
|
|
190
|
+
```python
|
|
191
|
+
# Bounding box [x1, y1, x2, y2]
|
|
192
|
+
input_box = np.array([425, 600, 700, 875])
|
|
193
|
+
|
|
194
|
+
masks, scores, logits = predictor.predict(
|
|
195
|
+
box=input_box,
|
|
196
|
+
multimask_output=False
|
|
197
|
+
)
|
|
198
|
+
```
|
|
199
|
+
|
|
200
|
+
### Combined prompts
|
|
201
|
+
|
|
202
|
+
```python
|
|
203
|
+
# Box + points for precise control
|
|
204
|
+
masks, scores, logits = predictor.predict(
|
|
205
|
+
point_coords=np.array([[500, 375]]),
|
|
206
|
+
point_labels=np.array([1]),
|
|
207
|
+
box=np.array([400, 300, 700, 600]),
|
|
208
|
+
multimask_output=False
|
|
209
|
+
)
|
|
210
|
+
```
|
|
211
|
+
|
|
212
|
+
### Iterative refinement
|
|
213
|
+
|
|
214
|
+
```python
|
|
215
|
+
# Initial prediction
|
|
216
|
+
masks, scores, logits = predictor.predict(
|
|
217
|
+
point_coords=np.array([[500, 375]]),
|
|
218
|
+
point_labels=np.array([1]),
|
|
219
|
+
multimask_output=True
|
|
220
|
+
)
|
|
221
|
+
|
|
222
|
+
# Refine with additional point using previous mask
|
|
223
|
+
masks, scores, logits = predictor.predict(
|
|
224
|
+
point_coords=np.array([[500, 375], [550, 400]]),
|
|
225
|
+
point_labels=np.array([1, 0]), # Add background point
|
|
226
|
+
mask_input=logits[np.argmax(scores)][None, :, :], # Use best mask
|
|
227
|
+
multimask_output=False
|
|
228
|
+
)
|
|
229
|
+
```
|
|
230
|
+
|
|
231
|
+
## Automatic mask generation
|
|
232
|
+
|
|
233
|
+
### Basic automatic segmentation
|
|
234
|
+
|
|
235
|
+
```python
|
|
236
|
+
from segment_anything import SamAutomaticMaskGenerator
|
|
237
|
+
|
|
238
|
+
# Create generator
|
|
239
|
+
mask_generator = SamAutomaticMaskGenerator(sam)
|
|
240
|
+
|
|
241
|
+
# Generate all masks
|
|
242
|
+
masks = mask_generator.generate(image)
|
|
243
|
+
|
|
244
|
+
# Each mask contains:
|
|
245
|
+
# - segmentation: binary mask
|
|
246
|
+
# - bbox: [x, y, w, h]
|
|
247
|
+
# - area: pixel count
|
|
248
|
+
# - predicted_iou: quality score
|
|
249
|
+
# - stability_score: robustness score
|
|
250
|
+
# - point_coords: generating point
|
|
251
|
+
```
|
|
252
|
+
|
|
253
|
+
### Customized generation
|
|
254
|
+
|
|
255
|
+
```python
|
|
256
|
+
mask_generator = SamAutomaticMaskGenerator(
|
|
257
|
+
model=sam,
|
|
258
|
+
points_per_side=32, # Grid density (more = more masks)
|
|
259
|
+
pred_iou_thresh=0.88, # Quality threshold
|
|
260
|
+
stability_score_thresh=0.95, # Stability threshold
|
|
261
|
+
crop_n_layers=1, # Multi-scale crops
|
|
262
|
+
crop_n_points_downscale_factor=2,
|
|
263
|
+
min_mask_region_area=100, # Remove tiny masks
|
|
264
|
+
)
|
|
265
|
+
|
|
266
|
+
masks = mask_generator.generate(image)
|
|
267
|
+
```
|
|
268
|
+
|
|
269
|
+
### Filtering masks
|
|
270
|
+
|
|
271
|
+
```python
|
|
272
|
+
# Sort by area (largest first)
|
|
273
|
+
masks = sorted(masks, key=lambda x: x['area'], reverse=True)
|
|
274
|
+
|
|
275
|
+
# Filter by predicted IoU
|
|
276
|
+
high_quality = [m for m in masks if m['predicted_iou'] > 0.9]
|
|
277
|
+
|
|
278
|
+
# Filter by stability score
|
|
279
|
+
stable_masks = [m for m in masks if m['stability_score'] > 0.95]
|
|
280
|
+
```
|
|
281
|
+
|
|
282
|
+
## Batched inference
|
|
283
|
+
|
|
284
|
+
### Multiple images
|
|
285
|
+
|
|
286
|
+
```python
|
|
287
|
+
# Process multiple images efficiently
|
|
288
|
+
images = [cv2.imread(f"image_{i}.jpg") for i in range(10)]
|
|
289
|
+
|
|
290
|
+
all_masks = []
|
|
291
|
+
for image in images:
|
|
292
|
+
predictor.set_image(image)
|
|
293
|
+
masks, _, _ = predictor.predict(
|
|
294
|
+
point_coords=np.array([[500, 375]]),
|
|
295
|
+
point_labels=np.array([1]),
|
|
296
|
+
multimask_output=True
|
|
297
|
+
)
|
|
298
|
+
all_masks.append(masks)
|
|
299
|
+
```
|
|
300
|
+
|
|
301
|
+
### Multiple prompts per image
|
|
302
|
+
|
|
303
|
+
```python
|
|
304
|
+
# Process multiple prompts efficiently (one image encoding)
|
|
305
|
+
predictor.set_image(image)
|
|
306
|
+
|
|
307
|
+
# Batch of point prompts
|
|
308
|
+
points = [
|
|
309
|
+
np.array([[100, 100]]),
|
|
310
|
+
np.array([[200, 200]]),
|
|
311
|
+
np.array([[300, 300]])
|
|
312
|
+
]
|
|
313
|
+
|
|
314
|
+
all_masks = []
|
|
315
|
+
for point in points:
|
|
316
|
+
masks, scores, _ = predictor.predict(
|
|
317
|
+
point_coords=point,
|
|
318
|
+
point_labels=np.array([1]),
|
|
319
|
+
multimask_output=True
|
|
320
|
+
)
|
|
321
|
+
all_masks.append(masks[np.argmax(scores)])
|
|
322
|
+
```
|
|
323
|
+
|
|
324
|
+
## ONNX deployment
|
|
325
|
+
|
|
326
|
+
### Export model
|
|
327
|
+
|
|
328
|
+
```bash
|
|
329
|
+
python scripts/export_onnx_model.py \
|
|
330
|
+
--checkpoint sam_vit_h_4b8939.pth \
|
|
331
|
+
--model-type vit_h \
|
|
332
|
+
--output sam_onnx.onnx \
|
|
333
|
+
--return-single-mask
|
|
334
|
+
```
|
|
335
|
+
|
|
336
|
+
### Use ONNX model
|
|
337
|
+
|
|
338
|
+
```python
|
|
339
|
+
import onnxruntime
|
|
340
|
+
|
|
341
|
+
# Load ONNX model
|
|
342
|
+
ort_session = onnxruntime.InferenceSession("sam_onnx.onnx")
|
|
343
|
+
|
|
344
|
+
# Run inference (image embeddings computed separately)
|
|
345
|
+
masks = ort_session.run(
|
|
346
|
+
None,
|
|
347
|
+
{
|
|
348
|
+
"image_embeddings": image_embeddings,
|
|
349
|
+
"point_coords": point_coords,
|
|
350
|
+
"point_labels": point_labels,
|
|
351
|
+
"mask_input": np.zeros((1, 1, 256, 256), dtype=np.float32),
|
|
352
|
+
"has_mask_input": np.array([0], dtype=np.float32),
|
|
353
|
+
"orig_im_size": np.array([h, w], dtype=np.float32)
|
|
354
|
+
}
|
|
355
|
+
)
|
|
356
|
+
```
|
|
357
|
+
|
|
358
|
+
## Common workflows
|
|
359
|
+
|
|
360
|
+
### Workflow 1: Annotation tool
|
|
361
|
+
|
|
362
|
+
```python
|
|
363
|
+
import cv2
|
|
364
|
+
|
|
365
|
+
# Load model
|
|
366
|
+
predictor = SamPredictor(sam)
|
|
367
|
+
predictor.set_image(image)
|
|
368
|
+
|
|
369
|
+
def on_click(event, x, y, flags, param):
|
|
370
|
+
if event == cv2.EVENT_LBUTTONDOWN:
|
|
371
|
+
# Foreground point
|
|
372
|
+
masks, scores, _ = predictor.predict(
|
|
373
|
+
point_coords=np.array([[x, y]]),
|
|
374
|
+
point_labels=np.array([1]),
|
|
375
|
+
multimask_output=True
|
|
376
|
+
)
|
|
377
|
+
# Display best mask
|
|
378
|
+
display_mask(masks[np.argmax(scores)])
|
|
379
|
+
```
|
|
380
|
+
|
|
381
|
+
### Workflow 2: Object extraction
|
|
382
|
+
|
|
383
|
+
```python
|
|
384
|
+
def extract_object(image, point):
|
|
385
|
+
"""Extract object at point with transparent background."""
|
|
386
|
+
predictor.set_image(image)
|
|
387
|
+
|
|
388
|
+
masks, scores, _ = predictor.predict(
|
|
389
|
+
point_coords=np.array([point]),
|
|
390
|
+
point_labels=np.array([1]),
|
|
391
|
+
multimask_output=True
|
|
392
|
+
)
|
|
393
|
+
|
|
394
|
+
best_mask = masks[np.argmax(scores)]
|
|
395
|
+
|
|
396
|
+
# Create RGBA output
|
|
397
|
+
rgba = np.zeros((image.shape[0], image.shape[1], 4), dtype=np.uint8)
|
|
398
|
+
rgba[:, :, :3] = image
|
|
399
|
+
rgba[:, :, 3] = best_mask * 255
|
|
400
|
+
|
|
401
|
+
return rgba
|
|
402
|
+
```
|
|
403
|
+
|
|
404
|
+
### Workflow 3: Medical image segmentation
|
|
405
|
+
|
|
406
|
+
```python
|
|
407
|
+
# Process medical images (grayscale to RGB)
|
|
408
|
+
medical_image = cv2.imread("scan.png", cv2.IMREAD_GRAYSCALE)
|
|
409
|
+
rgb_image = cv2.cvtColor(medical_image, cv2.COLOR_GRAY2RGB)
|
|
410
|
+
|
|
411
|
+
predictor.set_image(rgb_image)
|
|
412
|
+
|
|
413
|
+
# Segment region of interest
|
|
414
|
+
masks, scores, _ = predictor.predict(
|
|
415
|
+
box=np.array([x1, y1, x2, y2]), # ROI bounding box
|
|
416
|
+
multimask_output=True
|
|
417
|
+
)
|
|
418
|
+
```
|
|
419
|
+
|
|
420
|
+
## Output format
|
|
421
|
+
|
|
422
|
+
### Mask data structure
|
|
423
|
+
|
|
424
|
+
```python
|
|
425
|
+
# SamAutomaticMaskGenerator output
|
|
426
|
+
{
|
|
427
|
+
"segmentation": np.ndarray, # H×W binary mask
|
|
428
|
+
"bbox": [x, y, w, h], # Bounding box
|
|
429
|
+
"area": int, # Pixel count
|
|
430
|
+
"predicted_iou": float, # 0-1 quality score
|
|
431
|
+
"stability_score": float, # 0-1 robustness score
|
|
432
|
+
"crop_box": [x, y, w, h], # Generation crop region
|
|
433
|
+
"point_coords": [[x, y]], # Input point
|
|
434
|
+
}
|
|
435
|
+
```
|
|
436
|
+
|
|
437
|
+
### COCO RLE format
|
|
438
|
+
|
|
439
|
+
```python
|
|
440
|
+
from pycocotools import mask as mask_utils
|
|
441
|
+
|
|
442
|
+
# Encode mask to RLE
|
|
443
|
+
rle = mask_utils.encode(np.asfortranarray(mask.astype(np.uint8)))
|
|
444
|
+
rle["counts"] = rle["counts"].decode("utf-8")
|
|
445
|
+
|
|
446
|
+
# Decode RLE to mask
|
|
447
|
+
decoded_mask = mask_utils.decode(rle)
|
|
448
|
+
```
|
|
449
|
+
|
|
450
|
+
## Performance optimization
|
|
451
|
+
|
|
452
|
+
### GPU memory
|
|
453
|
+
|
|
454
|
+
```python
|
|
455
|
+
# Use smaller model for limited VRAM
|
|
456
|
+
sam = sam_model_registry["vit_b"](checkpoint="sam_vit_b_01ec64.pth")
|
|
457
|
+
|
|
458
|
+
# Process images in batches
|
|
459
|
+
# Clear CUDA cache between large batches
|
|
460
|
+
torch.cuda.empty_cache()
|
|
461
|
+
```
|
|
462
|
+
|
|
463
|
+
### Speed optimization
|
|
464
|
+
|
|
465
|
+
```python
|
|
466
|
+
# Use half precision
|
|
467
|
+
sam = sam.half()
|
|
468
|
+
|
|
469
|
+
# Reduce points for automatic generation
|
|
470
|
+
mask_generator = SamAutomaticMaskGenerator(
|
|
471
|
+
model=sam,
|
|
472
|
+
points_per_side=16, # Default is 32
|
|
473
|
+
)
|
|
474
|
+
|
|
475
|
+
# Use ONNX for deployment
|
|
476
|
+
# Export with --return-single-mask for faster inference
|
|
477
|
+
```
|
|
478
|
+
|
|
479
|
+
## Common issues
|
|
480
|
+
|
|
481
|
+
| Issue | Solution |
|
|
482
|
+
|-------|----------|
|
|
483
|
+
| Out of memory | Use ViT-B model, reduce image size |
|
|
484
|
+
| Slow inference | Use ViT-B, reduce points_per_side |
|
|
485
|
+
| Poor mask quality | Try different prompts, use box + points |
|
|
486
|
+
| Edge artifacts | Use stability_score filtering |
|
|
487
|
+
| Small objects missed | Increase points_per_side |
|
|
488
|
+
|
|
489
|
+
## References
|
|
490
|
+
|
|
491
|
+
- **[Advanced Usage](references/advanced-usage.md)** - Batching, fine-tuning, integration
|
|
492
|
+
- **[Troubleshooting](references/troubleshooting.md)** - Common issues and solutions
|
|
493
|
+
|
|
494
|
+
## Resources
|
|
495
|
+
|
|
496
|
+
- **GitHub**: https://github.com/facebookresearch/segment-anything
|
|
497
|
+
- **Paper**: https://arxiv.org/abs/2304.02643
|
|
498
|
+
- **Demo**: https://segment-anything.com
|
|
499
|
+
- **SAM 2 (Video)**: https://github.com/facebookresearch/segment-anything-2
|
|
500
|
+
- **HuggingFace**: https://huggingface.co/facebook/sam-vit-huge
|