@synsci/cli-darwin-x64 1.1.49
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/bin/skills/accelerate/SKILL.md +332 -0
- package/bin/skills/accelerate/references/custom-plugins.md +453 -0
- package/bin/skills/accelerate/references/megatron-integration.md +489 -0
- package/bin/skills/accelerate/references/performance.md +525 -0
- package/bin/skills/audiocraft/SKILL.md +564 -0
- package/bin/skills/audiocraft/references/advanced-usage.md +666 -0
- package/bin/skills/audiocraft/references/troubleshooting.md +504 -0
- package/bin/skills/autogpt/SKILL.md +403 -0
- package/bin/skills/autogpt/references/advanced-usage.md +535 -0
- package/bin/skills/autogpt/references/troubleshooting.md +420 -0
- package/bin/skills/awq/SKILL.md +310 -0
- package/bin/skills/awq/references/advanced-usage.md +324 -0
- package/bin/skills/awq/references/troubleshooting.md +344 -0
- package/bin/skills/axolotl/SKILL.md +158 -0
- package/bin/skills/axolotl/references/api.md +5548 -0
- package/bin/skills/axolotl/references/dataset-formats.md +1029 -0
- package/bin/skills/axolotl/references/index.md +15 -0
- package/bin/skills/axolotl/references/other.md +3563 -0
- package/bin/skills/bigcode-evaluation-harness/SKILL.md +405 -0
- package/bin/skills/bigcode-evaluation-harness/references/benchmarks.md +393 -0
- package/bin/skills/bigcode-evaluation-harness/references/custom-tasks.md +424 -0
- package/bin/skills/bigcode-evaluation-harness/references/issues.md +394 -0
- package/bin/skills/bitsandbytes/SKILL.md +411 -0
- package/bin/skills/bitsandbytes/references/memory-optimization.md +521 -0
- package/bin/skills/bitsandbytes/references/qlora-training.md +521 -0
- package/bin/skills/bitsandbytes/references/quantization-formats.md +447 -0
- package/bin/skills/blip-2/SKILL.md +564 -0
- package/bin/skills/blip-2/references/advanced-usage.md +680 -0
- package/bin/skills/blip-2/references/troubleshooting.md +526 -0
- package/bin/skills/chroma/SKILL.md +406 -0
- package/bin/skills/chroma/references/integration.md +38 -0
- package/bin/skills/clip/SKILL.md +253 -0
- package/bin/skills/clip/references/applications.md +207 -0
- package/bin/skills/constitutional-ai/SKILL.md +290 -0
- package/bin/skills/crewai/SKILL.md +498 -0
- package/bin/skills/crewai/references/flows.md +438 -0
- package/bin/skills/crewai/references/tools.md +429 -0
- package/bin/skills/crewai/references/troubleshooting.md +480 -0
- package/bin/skills/deepspeed/SKILL.md +141 -0
- package/bin/skills/deepspeed/references/08.md +17 -0
- package/bin/skills/deepspeed/references/09.md +173 -0
- package/bin/skills/deepspeed/references/2020.md +378 -0
- package/bin/skills/deepspeed/references/2023.md +279 -0
- package/bin/skills/deepspeed/references/assets.md +179 -0
- package/bin/skills/deepspeed/references/index.md +35 -0
- package/bin/skills/deepspeed/references/mii.md +118 -0
- package/bin/skills/deepspeed/references/other.md +1191 -0
- package/bin/skills/deepspeed/references/tutorials.md +6554 -0
- package/bin/skills/dspy/SKILL.md +590 -0
- package/bin/skills/dspy/references/examples.md +663 -0
- package/bin/skills/dspy/references/modules.md +475 -0
- package/bin/skills/dspy/references/optimizers.md +566 -0
- package/bin/skills/faiss/SKILL.md +221 -0
- package/bin/skills/faiss/references/index_types.md +280 -0
- package/bin/skills/flash-attention/SKILL.md +367 -0
- package/bin/skills/flash-attention/references/benchmarks.md +215 -0
- package/bin/skills/flash-attention/references/transformers-integration.md +293 -0
- package/bin/skills/gguf/SKILL.md +427 -0
- package/bin/skills/gguf/references/advanced-usage.md +504 -0
- package/bin/skills/gguf/references/troubleshooting.md +442 -0
- package/bin/skills/gptq/SKILL.md +450 -0
- package/bin/skills/gptq/references/calibration.md +337 -0
- package/bin/skills/gptq/references/integration.md +129 -0
- package/bin/skills/gptq/references/troubleshooting.md +95 -0
- package/bin/skills/grpo-rl-training/README.md +97 -0
- package/bin/skills/grpo-rl-training/SKILL.md +572 -0
- package/bin/skills/grpo-rl-training/examples/reward_functions_library.py +393 -0
- package/bin/skills/grpo-rl-training/templates/basic_grpo_training.py +228 -0
- package/bin/skills/guidance/SKILL.md +572 -0
- package/bin/skills/guidance/references/backends.md +554 -0
- package/bin/skills/guidance/references/constraints.md +674 -0
- package/bin/skills/guidance/references/examples.md +767 -0
- package/bin/skills/hqq/SKILL.md +445 -0
- package/bin/skills/hqq/references/advanced-usage.md +528 -0
- package/bin/skills/hqq/references/troubleshooting.md +503 -0
- package/bin/skills/hugging-face-cli/SKILL.md +191 -0
- package/bin/skills/hugging-face-cli/references/commands.md +954 -0
- package/bin/skills/hugging-face-cli/references/examples.md +374 -0
- package/bin/skills/hugging-face-datasets/SKILL.md +547 -0
- package/bin/skills/hugging-face-datasets/examples/diverse_training_examples.json +239 -0
- package/bin/skills/hugging-face-datasets/examples/system_prompt_template.txt +196 -0
- package/bin/skills/hugging-face-datasets/examples/training_examples.json +176 -0
- package/bin/skills/hugging-face-datasets/scripts/dataset_manager.py +522 -0
- package/bin/skills/hugging-face-datasets/scripts/sql_manager.py +844 -0
- package/bin/skills/hugging-face-datasets/templates/chat.json +55 -0
- package/bin/skills/hugging-face-datasets/templates/classification.json +62 -0
- package/bin/skills/hugging-face-datasets/templates/completion.json +51 -0
- package/bin/skills/hugging-face-datasets/templates/custom.json +75 -0
- package/bin/skills/hugging-face-datasets/templates/qa.json +54 -0
- package/bin/skills/hugging-face-datasets/templates/tabular.json +81 -0
- package/bin/skills/hugging-face-evaluation/SKILL.md +656 -0
- package/bin/skills/hugging-face-evaluation/examples/USAGE_EXAMPLES.md +382 -0
- package/bin/skills/hugging-face-evaluation/examples/artificial_analysis_to_hub.py +141 -0
- package/bin/skills/hugging-face-evaluation/examples/example_readme_tables.md +135 -0
- package/bin/skills/hugging-face-evaluation/examples/metric_mapping.json +50 -0
- package/bin/skills/hugging-face-evaluation/requirements.txt +20 -0
- package/bin/skills/hugging-face-evaluation/scripts/evaluation_manager.py +1374 -0
- package/bin/skills/hugging-face-evaluation/scripts/inspect_eval_uv.py +104 -0
- package/bin/skills/hugging-face-evaluation/scripts/inspect_vllm_uv.py +317 -0
- package/bin/skills/hugging-face-evaluation/scripts/lighteval_vllm_uv.py +303 -0
- package/bin/skills/hugging-face-evaluation/scripts/run_eval_job.py +98 -0
- package/bin/skills/hugging-face-evaluation/scripts/run_vllm_eval_job.py +331 -0
- package/bin/skills/hugging-face-evaluation/scripts/test_extraction.py +206 -0
- package/bin/skills/hugging-face-jobs/SKILL.md +1041 -0
- package/bin/skills/hugging-face-jobs/index.html +216 -0
- package/bin/skills/hugging-face-jobs/references/hardware_guide.md +336 -0
- package/bin/skills/hugging-face-jobs/references/hub_saving.md +352 -0
- package/bin/skills/hugging-face-jobs/references/token_usage.md +546 -0
- package/bin/skills/hugging-face-jobs/references/troubleshooting.md +475 -0
- package/bin/skills/hugging-face-jobs/scripts/cot-self-instruct.py +718 -0
- package/bin/skills/hugging-face-jobs/scripts/finepdfs-stats.py +546 -0
- package/bin/skills/hugging-face-jobs/scripts/generate-responses.py +587 -0
- package/bin/skills/hugging-face-model-trainer/SKILL.md +711 -0
- package/bin/skills/hugging-face-model-trainer/references/gguf_conversion.md +296 -0
- package/bin/skills/hugging-face-model-trainer/references/hardware_guide.md +283 -0
- package/bin/skills/hugging-face-model-trainer/references/hub_saving.md +364 -0
- package/bin/skills/hugging-face-model-trainer/references/reliability_principles.md +371 -0
- package/bin/skills/hugging-face-model-trainer/references/trackio_guide.md +189 -0
- package/bin/skills/hugging-face-model-trainer/references/training_methods.md +150 -0
- package/bin/skills/hugging-face-model-trainer/references/training_patterns.md +203 -0
- package/bin/skills/hugging-face-model-trainer/references/troubleshooting.md +282 -0
- package/bin/skills/hugging-face-model-trainer/scripts/convert_to_gguf.py +424 -0
- package/bin/skills/hugging-face-model-trainer/scripts/dataset_inspector.py +417 -0
- package/bin/skills/hugging-face-model-trainer/scripts/estimate_cost.py +150 -0
- package/bin/skills/hugging-face-model-trainer/scripts/train_dpo_example.py +106 -0
- package/bin/skills/hugging-face-model-trainer/scripts/train_grpo_example.py +89 -0
- package/bin/skills/hugging-face-model-trainer/scripts/train_sft_example.py +122 -0
- package/bin/skills/hugging-face-paper-publisher/SKILL.md +627 -0
- package/bin/skills/hugging-face-paper-publisher/examples/example_usage.md +327 -0
- package/bin/skills/hugging-face-paper-publisher/references/quick_reference.md +216 -0
- package/bin/skills/hugging-face-paper-publisher/scripts/paper_manager.py +508 -0
- package/bin/skills/hugging-face-paper-publisher/templates/arxiv.md +299 -0
- package/bin/skills/hugging-face-paper-publisher/templates/ml-report.md +358 -0
- package/bin/skills/hugging-face-paper-publisher/templates/modern.md +319 -0
- package/bin/skills/hugging-face-paper-publisher/templates/standard.md +201 -0
- package/bin/skills/hugging-face-tool-builder/SKILL.md +115 -0
- package/bin/skills/hugging-face-tool-builder/references/baseline_hf_api.py +57 -0
- package/bin/skills/hugging-face-tool-builder/references/baseline_hf_api.sh +40 -0
- package/bin/skills/hugging-face-tool-builder/references/baseline_hf_api.tsx +57 -0
- package/bin/skills/hugging-face-tool-builder/references/find_models_by_paper.sh +230 -0
- package/bin/skills/hugging-face-tool-builder/references/hf_enrich_models.sh +96 -0
- package/bin/skills/hugging-face-tool-builder/references/hf_model_card_frontmatter.sh +188 -0
- package/bin/skills/hugging-face-tool-builder/references/hf_model_papers_auth.sh +171 -0
- package/bin/skills/hugging-face-trackio/SKILL.md +65 -0
- package/bin/skills/hugging-face-trackio/references/logging_metrics.md +206 -0
- package/bin/skills/hugging-face-trackio/references/retrieving_metrics.md +223 -0
- package/bin/skills/huggingface-tokenizers/SKILL.md +516 -0
- package/bin/skills/huggingface-tokenizers/references/algorithms.md +653 -0
- package/bin/skills/huggingface-tokenizers/references/integration.md +637 -0
- package/bin/skills/huggingface-tokenizers/references/pipeline.md +723 -0
- package/bin/skills/huggingface-tokenizers/references/training.md +565 -0
- package/bin/skills/instructor/SKILL.md +740 -0
- package/bin/skills/instructor/references/examples.md +107 -0
- package/bin/skills/instructor/references/providers.md +70 -0
- package/bin/skills/instructor/references/validation.md +606 -0
- package/bin/skills/knowledge-distillation/SKILL.md +458 -0
- package/bin/skills/knowledge-distillation/references/minillm.md +334 -0
- package/bin/skills/lambda-labs/SKILL.md +545 -0
- package/bin/skills/lambda-labs/references/advanced-usage.md +611 -0
- package/bin/skills/lambda-labs/references/troubleshooting.md +530 -0
- package/bin/skills/langchain/SKILL.md +480 -0
- package/bin/skills/langchain/references/agents.md +499 -0
- package/bin/skills/langchain/references/integration.md +562 -0
- package/bin/skills/langchain/references/rag.md +600 -0
- package/bin/skills/langsmith/SKILL.md +422 -0
- package/bin/skills/langsmith/references/advanced-usage.md +548 -0
- package/bin/skills/langsmith/references/troubleshooting.md +537 -0
- package/bin/skills/litgpt/SKILL.md +469 -0
- package/bin/skills/litgpt/references/custom-models.md +568 -0
- package/bin/skills/litgpt/references/distributed-training.md +451 -0
- package/bin/skills/litgpt/references/supported-models.md +336 -0
- package/bin/skills/litgpt/references/training-recipes.md +619 -0
- package/bin/skills/llama-cpp/SKILL.md +258 -0
- package/bin/skills/llama-cpp/references/optimization.md +89 -0
- package/bin/skills/llama-cpp/references/quantization.md +213 -0
- package/bin/skills/llama-cpp/references/server.md +125 -0
- package/bin/skills/llama-factory/SKILL.md +80 -0
- package/bin/skills/llama-factory/references/_images.md +23 -0
- package/bin/skills/llama-factory/references/advanced.md +1055 -0
- package/bin/skills/llama-factory/references/getting_started.md +349 -0
- package/bin/skills/llama-factory/references/index.md +19 -0
- package/bin/skills/llama-factory/references/other.md +31 -0
- package/bin/skills/llamaguard/SKILL.md +337 -0
- package/bin/skills/llamaindex/SKILL.md +569 -0
- package/bin/skills/llamaindex/references/agents.md +83 -0
- package/bin/skills/llamaindex/references/data_connectors.md +108 -0
- package/bin/skills/llamaindex/references/query_engines.md +406 -0
- package/bin/skills/llava/SKILL.md +304 -0
- package/bin/skills/llava/references/training.md +197 -0
- package/bin/skills/lm-evaluation-harness/SKILL.md +490 -0
- package/bin/skills/lm-evaluation-harness/references/api-evaluation.md +490 -0
- package/bin/skills/lm-evaluation-harness/references/benchmark-guide.md +488 -0
- package/bin/skills/lm-evaluation-harness/references/custom-tasks.md +602 -0
- package/bin/skills/lm-evaluation-harness/references/distributed-eval.md +519 -0
- package/bin/skills/long-context/SKILL.md +536 -0
- package/bin/skills/long-context/references/extension_methods.md +468 -0
- package/bin/skills/long-context/references/fine_tuning.md +611 -0
- package/bin/skills/long-context/references/rope.md +402 -0
- package/bin/skills/mamba/SKILL.md +260 -0
- package/bin/skills/mamba/references/architecture-details.md +206 -0
- package/bin/skills/mamba/references/benchmarks.md +255 -0
- package/bin/skills/mamba/references/training-guide.md +388 -0
- package/bin/skills/megatron-core/SKILL.md +366 -0
- package/bin/skills/megatron-core/references/benchmarks.md +249 -0
- package/bin/skills/megatron-core/references/parallelism-guide.md +404 -0
- package/bin/skills/megatron-core/references/production-examples.md +473 -0
- package/bin/skills/megatron-core/references/training-recipes.md +547 -0
- package/bin/skills/miles/SKILL.md +315 -0
- package/bin/skills/miles/references/api-reference.md +141 -0
- package/bin/skills/miles/references/troubleshooting.md +352 -0
- package/bin/skills/mlflow/SKILL.md +704 -0
- package/bin/skills/mlflow/references/deployment.md +744 -0
- package/bin/skills/mlflow/references/model-registry.md +770 -0
- package/bin/skills/mlflow/references/tracking.md +680 -0
- package/bin/skills/modal/SKILL.md +341 -0
- package/bin/skills/modal/references/advanced-usage.md +503 -0
- package/bin/skills/modal/references/troubleshooting.md +494 -0
- package/bin/skills/model-merging/SKILL.md +539 -0
- package/bin/skills/model-merging/references/evaluation.md +462 -0
- package/bin/skills/model-merging/references/examples.md +428 -0
- package/bin/skills/model-merging/references/methods.md +352 -0
- package/bin/skills/model-pruning/SKILL.md +495 -0
- package/bin/skills/model-pruning/references/wanda.md +347 -0
- package/bin/skills/moe-training/SKILL.md +526 -0
- package/bin/skills/moe-training/references/architectures.md +432 -0
- package/bin/skills/moe-training/references/inference.md +348 -0
- package/bin/skills/moe-training/references/training.md +425 -0
- package/bin/skills/nanogpt/SKILL.md +290 -0
- package/bin/skills/nanogpt/references/architecture.md +382 -0
- package/bin/skills/nanogpt/references/data.md +476 -0
- package/bin/skills/nanogpt/references/training.md +564 -0
- package/bin/skills/nemo-curator/SKILL.md +383 -0
- package/bin/skills/nemo-curator/references/deduplication.md +87 -0
- package/bin/skills/nemo-curator/references/filtering.md +102 -0
- package/bin/skills/nemo-evaluator/SKILL.md +494 -0
- package/bin/skills/nemo-evaluator/references/adapter-system.md +340 -0
- package/bin/skills/nemo-evaluator/references/configuration.md +447 -0
- package/bin/skills/nemo-evaluator/references/custom-benchmarks.md +315 -0
- package/bin/skills/nemo-evaluator/references/execution-backends.md +361 -0
- package/bin/skills/nemo-guardrails/SKILL.md +297 -0
- package/bin/skills/nnsight/SKILL.md +436 -0
- package/bin/skills/nnsight/references/README.md +78 -0
- package/bin/skills/nnsight/references/api.md +344 -0
- package/bin/skills/nnsight/references/tutorials.md +300 -0
- package/bin/skills/openrlhf/SKILL.md +249 -0
- package/bin/skills/openrlhf/references/algorithm-comparison.md +404 -0
- package/bin/skills/openrlhf/references/custom-rewards.md +530 -0
- package/bin/skills/openrlhf/references/hybrid-engine.md +287 -0
- package/bin/skills/openrlhf/references/multi-node-training.md +454 -0
- package/bin/skills/outlines/SKILL.md +652 -0
- package/bin/skills/outlines/references/backends.md +615 -0
- package/bin/skills/outlines/references/examples.md +773 -0
- package/bin/skills/outlines/references/json_generation.md +652 -0
- package/bin/skills/peft/SKILL.md +431 -0
- package/bin/skills/peft/references/advanced-usage.md +514 -0
- package/bin/skills/peft/references/troubleshooting.md +480 -0
- package/bin/skills/phoenix/SKILL.md +475 -0
- package/bin/skills/phoenix/references/advanced-usage.md +619 -0
- package/bin/skills/phoenix/references/troubleshooting.md +538 -0
- package/bin/skills/pinecone/SKILL.md +358 -0
- package/bin/skills/pinecone/references/deployment.md +181 -0
- package/bin/skills/pytorch-fsdp/SKILL.md +126 -0
- package/bin/skills/pytorch-fsdp/references/index.md +7 -0
- package/bin/skills/pytorch-fsdp/references/other.md +4249 -0
- package/bin/skills/pytorch-lightning/SKILL.md +346 -0
- package/bin/skills/pytorch-lightning/references/callbacks.md +436 -0
- package/bin/skills/pytorch-lightning/references/distributed.md +490 -0
- package/bin/skills/pytorch-lightning/references/hyperparameter-tuning.md +556 -0
- package/bin/skills/pyvene/SKILL.md +473 -0
- package/bin/skills/pyvene/references/README.md +73 -0
- package/bin/skills/pyvene/references/api.md +383 -0
- package/bin/skills/pyvene/references/tutorials.md +376 -0
- package/bin/skills/qdrant/SKILL.md +493 -0
- package/bin/skills/qdrant/references/advanced-usage.md +648 -0
- package/bin/skills/qdrant/references/troubleshooting.md +631 -0
- package/bin/skills/ray-data/SKILL.md +326 -0
- package/bin/skills/ray-data/references/integration.md +82 -0
- package/bin/skills/ray-data/references/transformations.md +83 -0
- package/bin/skills/ray-train/SKILL.md +406 -0
- package/bin/skills/ray-train/references/multi-node.md +628 -0
- package/bin/skills/rwkv/SKILL.md +260 -0
- package/bin/skills/rwkv/references/architecture-details.md +344 -0
- package/bin/skills/rwkv/references/rwkv7.md +386 -0
- package/bin/skills/rwkv/references/state-management.md +369 -0
- package/bin/skills/saelens/SKILL.md +386 -0
- package/bin/skills/saelens/references/README.md +70 -0
- package/bin/skills/saelens/references/api.md +333 -0
- package/bin/skills/saelens/references/tutorials.md +318 -0
- package/bin/skills/segment-anything/SKILL.md +500 -0
- package/bin/skills/segment-anything/references/advanced-usage.md +589 -0
- package/bin/skills/segment-anything/references/troubleshooting.md +484 -0
- package/bin/skills/sentence-transformers/SKILL.md +255 -0
- package/bin/skills/sentence-transformers/references/models.md +123 -0
- package/bin/skills/sentencepiece/SKILL.md +235 -0
- package/bin/skills/sentencepiece/references/algorithms.md +200 -0
- package/bin/skills/sentencepiece/references/training.md +304 -0
- package/bin/skills/sglang/SKILL.md +442 -0
- package/bin/skills/sglang/references/deployment.md +490 -0
- package/bin/skills/sglang/references/radix-attention.md +413 -0
- package/bin/skills/sglang/references/structured-generation.md +541 -0
- package/bin/skills/simpo/SKILL.md +219 -0
- package/bin/skills/simpo/references/datasets.md +478 -0
- package/bin/skills/simpo/references/hyperparameters.md +452 -0
- package/bin/skills/simpo/references/loss-functions.md +350 -0
- package/bin/skills/skypilot/SKILL.md +509 -0
- package/bin/skills/skypilot/references/advanced-usage.md +491 -0
- package/bin/skills/skypilot/references/troubleshooting.md +570 -0
- package/bin/skills/slime/SKILL.md +464 -0
- package/bin/skills/slime/references/api-reference.md +392 -0
- package/bin/skills/slime/references/troubleshooting.md +386 -0
- package/bin/skills/speculative-decoding/SKILL.md +467 -0
- package/bin/skills/speculative-decoding/references/lookahead.md +309 -0
- package/bin/skills/speculative-decoding/references/medusa.md +350 -0
- package/bin/skills/stable-diffusion/SKILL.md +519 -0
- package/bin/skills/stable-diffusion/references/advanced-usage.md +716 -0
- package/bin/skills/stable-diffusion/references/troubleshooting.md +555 -0
- package/bin/skills/tensorboard/SKILL.md +629 -0
- package/bin/skills/tensorboard/references/integrations.md +638 -0
- package/bin/skills/tensorboard/references/profiling.md +545 -0
- package/bin/skills/tensorboard/references/visualization.md +620 -0
- package/bin/skills/tensorrt-llm/SKILL.md +187 -0
- package/bin/skills/tensorrt-llm/references/multi-gpu.md +298 -0
- package/bin/skills/tensorrt-llm/references/optimization.md +242 -0
- package/bin/skills/tensorrt-llm/references/serving.md +470 -0
- package/bin/skills/tinker/SKILL.md +362 -0
- package/bin/skills/tinker/references/api-reference.md +168 -0
- package/bin/skills/tinker/references/getting-started.md +157 -0
- package/bin/skills/tinker/references/loss-functions.md +163 -0
- package/bin/skills/tinker/references/models-and-lora.md +139 -0
- package/bin/skills/tinker/references/recipes.md +280 -0
- package/bin/skills/tinker/references/reinforcement-learning.md +212 -0
- package/bin/skills/tinker/references/rendering.md +243 -0
- package/bin/skills/tinker/references/supervised-learning.md +232 -0
- package/bin/skills/tinker-training-cost/SKILL.md +187 -0
- package/bin/skills/tinker-training-cost/scripts/calculate_cost.py +123 -0
- package/bin/skills/torchforge/SKILL.md +433 -0
- package/bin/skills/torchforge/references/api-reference.md +327 -0
- package/bin/skills/torchforge/references/troubleshooting.md +409 -0
- package/bin/skills/torchtitan/SKILL.md +358 -0
- package/bin/skills/torchtitan/references/checkpoint.md +181 -0
- package/bin/skills/torchtitan/references/custom-models.md +258 -0
- package/bin/skills/torchtitan/references/float8.md +133 -0
- package/bin/skills/torchtitan/references/fsdp.md +126 -0
- package/bin/skills/transformer-lens/SKILL.md +346 -0
- package/bin/skills/transformer-lens/references/README.md +54 -0
- package/bin/skills/transformer-lens/references/api.md +362 -0
- package/bin/skills/transformer-lens/references/tutorials.md +339 -0
- package/bin/skills/trl-fine-tuning/SKILL.md +455 -0
- package/bin/skills/trl-fine-tuning/references/dpo-variants.md +227 -0
- package/bin/skills/trl-fine-tuning/references/online-rl.md +82 -0
- package/bin/skills/trl-fine-tuning/references/reward-modeling.md +122 -0
- package/bin/skills/trl-fine-tuning/references/sft-training.md +168 -0
- package/bin/skills/unsloth/SKILL.md +80 -0
- package/bin/skills/unsloth/references/index.md +7 -0
- package/bin/skills/unsloth/references/llms-full.md +16799 -0
- package/bin/skills/unsloth/references/llms-txt.md +12044 -0
- package/bin/skills/unsloth/references/llms.md +82 -0
- package/bin/skills/verl/SKILL.md +391 -0
- package/bin/skills/verl/references/api-reference.md +301 -0
- package/bin/skills/verl/references/troubleshooting.md +391 -0
- package/bin/skills/vllm/SKILL.md +364 -0
- package/bin/skills/vllm/references/optimization.md +226 -0
- package/bin/skills/vllm/references/quantization.md +284 -0
- package/bin/skills/vllm/references/server-deployment.md +255 -0
- package/bin/skills/vllm/references/troubleshooting.md +447 -0
- package/bin/skills/weights-and-biases/SKILL.md +590 -0
- package/bin/skills/weights-and-biases/references/artifacts.md +584 -0
- package/bin/skills/weights-and-biases/references/integrations.md +700 -0
- package/bin/skills/weights-and-biases/references/sweeps.md +847 -0
- package/bin/skills/whisper/SKILL.md +317 -0
- package/bin/skills/whisper/references/languages.md +189 -0
- package/bin/synsc +0 -0
- package/package.json +10 -0
|
@@ -0,0 +1,424 @@
|
|
|
1
|
+
#!/usr/bin/env python3
|
|
2
|
+
# /// script
|
|
3
|
+
# requires-python = ">=3.10"
|
|
4
|
+
# dependencies = [
|
|
5
|
+
# "transformers>=4.36.0",
|
|
6
|
+
# "peft>=0.7.0",
|
|
7
|
+
# "torch>=2.0.0",
|
|
8
|
+
# "accelerate>=0.24.0",
|
|
9
|
+
# "huggingface_hub>=0.20.0",
|
|
10
|
+
# "sentencepiece>=0.1.99",
|
|
11
|
+
# "protobuf>=3.20.0",
|
|
12
|
+
# "numpy",
|
|
13
|
+
# "gguf",
|
|
14
|
+
# ]
|
|
15
|
+
# ///
|
|
16
|
+
|
|
17
|
+
"""
|
|
18
|
+
GGUF Conversion Script - Production Ready
|
|
19
|
+
|
|
20
|
+
This script converts a LoRA fine-tuned model to GGUF format for use with:
|
|
21
|
+
- llama.cpp
|
|
22
|
+
- Ollama
|
|
23
|
+
- LM Studio
|
|
24
|
+
- Other GGUF-compatible tools
|
|
25
|
+
|
|
26
|
+
PREREQUISITES (install these FIRST):
|
|
27
|
+
- Ubuntu/Debian: sudo apt-get update && sudo apt-get install -y build-essential cmake
|
|
28
|
+
- RHEL/CentOS: sudo yum groupinstall -y "Development Tools" && sudo yum install -y cmake
|
|
29
|
+
- macOS: xcode-select --install && brew install cmake
|
|
30
|
+
|
|
31
|
+
Usage:
|
|
32
|
+
Set environment variables:
|
|
33
|
+
- ADAPTER_MODEL: Your fine-tuned model (e.g., "username/my-finetuned-model")
|
|
34
|
+
- BASE_MODEL: Base model used for fine-tuning (e.g., "Qwen/Qwen2.5-0.5B")
|
|
35
|
+
- OUTPUT_REPO: Where to upload GGUF files (e.g., "username/my-model-gguf")
|
|
36
|
+
- HF_USERNAME: Your Hugging Face username (optional, for README)
|
|
37
|
+
|
|
38
|
+
Dependencies: All required packages are declared in PEP 723 header above.
|
|
39
|
+
"""
|
|
40
|
+
|
|
41
|
+
import os
|
|
42
|
+
import sys
|
|
43
|
+
import torch
|
|
44
|
+
from transformers import AutoModelForCausalLM, AutoTokenizer
|
|
45
|
+
from peft import PeftModel
|
|
46
|
+
from huggingface_hub import HfApi
|
|
47
|
+
import subprocess
|
|
48
|
+
|
|
49
|
+
|
|
50
|
+
def check_system_dependencies():
|
|
51
|
+
"""Check if required system packages are available."""
|
|
52
|
+
print("🔍 Checking system dependencies...")
|
|
53
|
+
|
|
54
|
+
# Check for git
|
|
55
|
+
if subprocess.run(["which", "git"], capture_output=True).returncode != 0:
|
|
56
|
+
print(" ❌ git is not installed. Please install it:")
|
|
57
|
+
print(" Ubuntu/Debian: sudo apt-get install git")
|
|
58
|
+
print(" RHEL/CentOS: sudo yum install git")
|
|
59
|
+
print(" macOS: brew install git")
|
|
60
|
+
return False
|
|
61
|
+
|
|
62
|
+
# Check for make or cmake
|
|
63
|
+
has_make = subprocess.run(["which", "make"], capture_output=True).returncode == 0
|
|
64
|
+
has_cmake = subprocess.run(["which", "cmake"], capture_output=True).returncode == 0
|
|
65
|
+
|
|
66
|
+
if not has_make and not has_cmake:
|
|
67
|
+
print(" ❌ Neither make nor cmake found. Please install build tools:")
|
|
68
|
+
print(" Ubuntu/Debian: sudo apt-get install build-essential cmake")
|
|
69
|
+
print(" RHEL/CentOS: sudo yum groupinstall 'Development Tools' && sudo yum install cmake")
|
|
70
|
+
print(" macOS: xcode-select --install && brew install cmake")
|
|
71
|
+
return False
|
|
72
|
+
|
|
73
|
+
print(" ✅ System dependencies found")
|
|
74
|
+
return True
|
|
75
|
+
|
|
76
|
+
|
|
77
|
+
def run_command(cmd, description):
|
|
78
|
+
"""Run a command with error handling."""
|
|
79
|
+
print(f" {description}...")
|
|
80
|
+
try:
|
|
81
|
+
result = subprocess.run(
|
|
82
|
+
cmd,
|
|
83
|
+
check=True,
|
|
84
|
+
capture_output=True,
|
|
85
|
+
text=True
|
|
86
|
+
)
|
|
87
|
+
if result.stdout:
|
|
88
|
+
print(f" {result.stdout[:200]}") # Show first 200 chars
|
|
89
|
+
return True
|
|
90
|
+
except subprocess.CalledProcessError as e:
|
|
91
|
+
print(f" ❌ Command failed: {' '.join(cmd)}")
|
|
92
|
+
if e.stdout:
|
|
93
|
+
print(f" STDOUT: {e.stdout[:500]}")
|
|
94
|
+
if e.stderr:
|
|
95
|
+
print(f" STDERR: {e.stderr[:500]}")
|
|
96
|
+
return False
|
|
97
|
+
except FileNotFoundError:
|
|
98
|
+
print(f" ❌ Command not found: {cmd[0]}")
|
|
99
|
+
return False
|
|
100
|
+
|
|
101
|
+
|
|
102
|
+
print("🔄 GGUF Conversion Script")
|
|
103
|
+
print("=" * 60)
|
|
104
|
+
|
|
105
|
+
# Check system dependencies first
|
|
106
|
+
if not check_system_dependencies():
|
|
107
|
+
print("\n❌ Please install the missing system dependencies and try again.")
|
|
108
|
+
sys.exit(1)
|
|
109
|
+
|
|
110
|
+
# Configuration from environment variables
|
|
111
|
+
ADAPTER_MODEL = os.environ.get("ADAPTER_MODEL", "evalstate/qwen-capybara-medium")
|
|
112
|
+
BASE_MODEL = os.environ.get("BASE_MODEL", "Qwen/Qwen2.5-0.5B")
|
|
113
|
+
OUTPUT_REPO = os.environ.get("OUTPUT_REPO", "evalstate/qwen-capybara-medium-gguf")
|
|
114
|
+
username = os.environ.get("HF_USERNAME", ADAPTER_MODEL.split('/')[0])
|
|
115
|
+
|
|
116
|
+
print(f"\n📦 Configuration:")
|
|
117
|
+
print(f" Base model: {BASE_MODEL}")
|
|
118
|
+
print(f" Adapter model: {ADAPTER_MODEL}")
|
|
119
|
+
print(f" Output repo: {OUTPUT_REPO}")
|
|
120
|
+
|
|
121
|
+
# Step 1: Load base model and adapter
|
|
122
|
+
print("\n🔧 Step 1: Loading base model and LoRA adapter...")
|
|
123
|
+
print(" (This may take a few minutes)")
|
|
124
|
+
|
|
125
|
+
try:
|
|
126
|
+
base_model = AutoModelForCausalLM.from_pretrained(
|
|
127
|
+
BASE_MODEL,
|
|
128
|
+
dtype=torch.float16,
|
|
129
|
+
device_map="auto",
|
|
130
|
+
trust_remote_code=True,
|
|
131
|
+
)
|
|
132
|
+
print(" ✅ Base model loaded")
|
|
133
|
+
except Exception as e:
|
|
134
|
+
print(f" ❌ Failed to load base model: {e}")
|
|
135
|
+
sys.exit(1)
|
|
136
|
+
|
|
137
|
+
try:
|
|
138
|
+
# Load and merge adapter
|
|
139
|
+
print(" Loading LoRA adapter...")
|
|
140
|
+
model = PeftModel.from_pretrained(base_model, ADAPTER_MODEL)
|
|
141
|
+
print(" ✅ Adapter loaded")
|
|
142
|
+
|
|
143
|
+
print(" Merging adapter with base model...")
|
|
144
|
+
merged_model = model.merge_and_unload()
|
|
145
|
+
print(" ✅ Models merged!")
|
|
146
|
+
except Exception as e:
|
|
147
|
+
print(f" ❌ Failed to merge models: {e}")
|
|
148
|
+
sys.exit(1)
|
|
149
|
+
|
|
150
|
+
try:
|
|
151
|
+
# Load tokenizer
|
|
152
|
+
tokenizer = AutoTokenizer.from_pretrained(ADAPTER_MODEL, trust_remote_code=True)
|
|
153
|
+
print(" ✅ Tokenizer loaded")
|
|
154
|
+
except Exception as e:
|
|
155
|
+
print(f" ❌ Failed to load tokenizer: {e}")
|
|
156
|
+
sys.exit(1)
|
|
157
|
+
|
|
158
|
+
# Step 2: Save merged model temporarily
|
|
159
|
+
print("\n💾 Step 2: Saving merged model...")
|
|
160
|
+
merged_dir = "/tmp/merged_model"
|
|
161
|
+
try:
|
|
162
|
+
merged_model.save_pretrained(merged_dir, safe_serialization=True)
|
|
163
|
+
tokenizer.save_pretrained(merged_dir)
|
|
164
|
+
print(f" ✅ Merged model saved to {merged_dir}")
|
|
165
|
+
except Exception as e:
|
|
166
|
+
print(f" ❌ Failed to save merged model: {e}")
|
|
167
|
+
sys.exit(1)
|
|
168
|
+
|
|
169
|
+
# Step 3: Install llama.cpp for conversion
|
|
170
|
+
print("\n📥 Step 3: Setting up llama.cpp for GGUF conversion...")
|
|
171
|
+
|
|
172
|
+
# Clone llama.cpp repository
|
|
173
|
+
if not run_command(
|
|
174
|
+
["git", "clone", "https://github.com/ggerganov/llama.cpp.git", "/tmp/llama.cpp"],
|
|
175
|
+
"Cloning llama.cpp repository"
|
|
176
|
+
):
|
|
177
|
+
print(" Trying alternative clone method...")
|
|
178
|
+
# Try shallow clone
|
|
179
|
+
if not run_command(
|
|
180
|
+
["git", "clone", "--depth", "1", "https://github.com/ggerganov/llama.cpp.git", "/tmp/llama.cpp"],
|
|
181
|
+
"Cloning llama.cpp (shallow)"
|
|
182
|
+
):
|
|
183
|
+
sys.exit(1)
|
|
184
|
+
|
|
185
|
+
# Install Python dependencies
|
|
186
|
+
print(" Installing Python dependencies...")
|
|
187
|
+
if not run_command(
|
|
188
|
+
["pip", "install", "-r", "/tmp/llama.cpp/requirements.txt"],
|
|
189
|
+
"Installing llama.cpp requirements"
|
|
190
|
+
):
|
|
191
|
+
print(" ⚠️ Some requirements may already be installed")
|
|
192
|
+
|
|
193
|
+
if not run_command(
|
|
194
|
+
["pip", "install", "sentencepiece", "protobuf"],
|
|
195
|
+
"Installing tokenizer dependencies"
|
|
196
|
+
):
|
|
197
|
+
print(" ⚠️ Tokenizer dependencies may already be installed")
|
|
198
|
+
|
|
199
|
+
# Step 4: Convert to GGUF (FP16)
|
|
200
|
+
print("\n🔄 Step 4: Converting to GGUF format (FP16)...")
|
|
201
|
+
gguf_output_dir = "/tmp/gguf_output"
|
|
202
|
+
os.makedirs(gguf_output_dir, exist_ok=True)
|
|
203
|
+
|
|
204
|
+
convert_script = "/tmp/llama.cpp/convert_hf_to_gguf.py"
|
|
205
|
+
model_name = ADAPTER_MODEL.split('/')[-1]
|
|
206
|
+
gguf_file = f"{gguf_output_dir}/{model_name}-f16.gguf"
|
|
207
|
+
|
|
208
|
+
print(f" Running conversion...")
|
|
209
|
+
if not run_command(
|
|
210
|
+
[
|
|
211
|
+
sys.executable, convert_script,
|
|
212
|
+
merged_dir,
|
|
213
|
+
"--outfile", gguf_file,
|
|
214
|
+
"--outtype", "f16"
|
|
215
|
+
],
|
|
216
|
+
f"Converting to FP16"
|
|
217
|
+
):
|
|
218
|
+
print(" ❌ Conversion failed!")
|
|
219
|
+
sys.exit(1)
|
|
220
|
+
|
|
221
|
+
print(f" ✅ FP16 GGUF created: {gguf_file}")
|
|
222
|
+
|
|
223
|
+
# Step 5: Quantize to different formats
|
|
224
|
+
print("\n⚙️ Step 5: Creating quantized versions...")
|
|
225
|
+
|
|
226
|
+
# Build quantize tool using CMake (more reliable than make)
|
|
227
|
+
print(" Building quantize tool with CMake...")
|
|
228
|
+
os.makedirs("/tmp/llama.cpp/build", exist_ok=True)
|
|
229
|
+
|
|
230
|
+
# Configure with CMake
|
|
231
|
+
if not run_command(
|
|
232
|
+
["cmake", "-B", "/tmp/llama.cpp/build", "-S", "/tmp/llama.cpp",
|
|
233
|
+
"-DGGML_CUDA=OFF"],
|
|
234
|
+
"Configuring with CMake"
|
|
235
|
+
):
|
|
236
|
+
print(" ❌ CMake configuration failed")
|
|
237
|
+
sys.exit(1)
|
|
238
|
+
|
|
239
|
+
# Build just the quantize tool
|
|
240
|
+
if not run_command(
|
|
241
|
+
["cmake", "--build", "/tmp/llama.cpp/build", "--target", "llama-quantize", "-j", "4"],
|
|
242
|
+
"Building llama-quantize"
|
|
243
|
+
):
|
|
244
|
+
print(" ❌ Build failed!")
|
|
245
|
+
sys.exit(1)
|
|
246
|
+
|
|
247
|
+
print(" ✅ Quantize tool built")
|
|
248
|
+
|
|
249
|
+
# Use the CMake build output path
|
|
250
|
+
quantize_bin = "/tmp/llama.cpp/build/bin/llama-quantize"
|
|
251
|
+
|
|
252
|
+
# Common quantization formats
|
|
253
|
+
quant_formats = [
|
|
254
|
+
("Q4_K_M", "4-bit, medium quality (recommended)"),
|
|
255
|
+
("Q5_K_M", "5-bit, higher quality"),
|
|
256
|
+
("Q8_0", "8-bit, very high quality"),
|
|
257
|
+
]
|
|
258
|
+
|
|
259
|
+
quantized_files = []
|
|
260
|
+
for quant_type, description in quant_formats:
|
|
261
|
+
print(f" Creating {quant_type} quantization ({description})...")
|
|
262
|
+
quant_file = f"{gguf_output_dir}/{model_name}-{quant_type.lower()}.gguf"
|
|
263
|
+
|
|
264
|
+
if not run_command(
|
|
265
|
+
[quantize_bin, gguf_file, quant_file, quant_type],
|
|
266
|
+
f"Quantizing to {quant_type}"
|
|
267
|
+
):
|
|
268
|
+
print(f" ⚠️ Skipping {quant_type} due to error")
|
|
269
|
+
continue
|
|
270
|
+
|
|
271
|
+
quantized_files.append((quant_file, quant_type))
|
|
272
|
+
|
|
273
|
+
# Get file size
|
|
274
|
+
size_mb = os.path.getsize(quant_file) / (1024 * 1024)
|
|
275
|
+
print(f" ✅ {quant_type}: {size_mb:.1f} MB")
|
|
276
|
+
|
|
277
|
+
if not quantized_files:
|
|
278
|
+
print(" ❌ No quantized versions were created successfully")
|
|
279
|
+
sys.exit(1)
|
|
280
|
+
|
|
281
|
+
# Step 6: Upload to Hub
|
|
282
|
+
print("\n☁️ Step 6: Uploading to Hugging Face Hub...")
|
|
283
|
+
api = HfApi()
|
|
284
|
+
|
|
285
|
+
# Create repo
|
|
286
|
+
print(f" Creating repository: {OUTPUT_REPO}")
|
|
287
|
+
try:
|
|
288
|
+
api.create_repo(repo_id=OUTPUT_REPO, repo_type="model", exist_ok=True)
|
|
289
|
+
print(" ✅ Repository ready")
|
|
290
|
+
except Exception as e:
|
|
291
|
+
print(f" ℹ️ Repository may already exist: {e}")
|
|
292
|
+
|
|
293
|
+
# Upload FP16 version
|
|
294
|
+
print(" Uploading FP16 GGUF...")
|
|
295
|
+
try:
|
|
296
|
+
api.upload_file(
|
|
297
|
+
path_or_fileobj=gguf_file,
|
|
298
|
+
path_in_repo=f"{model_name}-f16.gguf",
|
|
299
|
+
repo_id=OUTPUT_REPO,
|
|
300
|
+
)
|
|
301
|
+
print(" ✅ FP16 uploaded")
|
|
302
|
+
except Exception as e:
|
|
303
|
+
print(f" ❌ Upload failed: {e}")
|
|
304
|
+
sys.exit(1)
|
|
305
|
+
|
|
306
|
+
# Upload quantized versions
|
|
307
|
+
for quant_file, quant_type in quantized_files:
|
|
308
|
+
print(f" Uploading {quant_type}...")
|
|
309
|
+
try:
|
|
310
|
+
api.upload_file(
|
|
311
|
+
path_or_fileobj=quant_file,
|
|
312
|
+
path_in_repo=f"{model_name}-{quant_type.lower()}.gguf",
|
|
313
|
+
repo_id=OUTPUT_REPO,
|
|
314
|
+
)
|
|
315
|
+
print(f" ✅ {quant_type} uploaded")
|
|
316
|
+
except Exception as e:
|
|
317
|
+
print(f" ❌ Upload failed for {quant_type}: {e}")
|
|
318
|
+
continue
|
|
319
|
+
|
|
320
|
+
# Create README
|
|
321
|
+
print("\n📝 Creating README...")
|
|
322
|
+
readme_content = f"""---
|
|
323
|
+
base_model: {BASE_MODEL}
|
|
324
|
+
tags:
|
|
325
|
+
- gguf
|
|
326
|
+
- llama.cpp
|
|
327
|
+
- quantized
|
|
328
|
+
- trl
|
|
329
|
+
- sft
|
|
330
|
+
---
|
|
331
|
+
|
|
332
|
+
# {OUTPUT_REPO.split('/')[-1]}
|
|
333
|
+
|
|
334
|
+
This is a GGUF conversion of [{ADAPTER_MODEL}](https://huggingface.co/{ADAPTER_MODEL}), which is a LoRA fine-tuned version of [{BASE_MODEL}](https://huggingface.co/{BASE_MODEL}).
|
|
335
|
+
|
|
336
|
+
## Model Details
|
|
337
|
+
|
|
338
|
+
- **Base Model:** {BASE_MODEL}
|
|
339
|
+
- **Fine-tuned Model:** {ADAPTER_MODEL}
|
|
340
|
+
- **Training:** Supervised Fine-Tuning (SFT) with TRL
|
|
341
|
+
- **Format:** GGUF (for llama.cpp, Ollama, LM Studio, etc.)
|
|
342
|
+
|
|
343
|
+
## Available Quantizations
|
|
344
|
+
|
|
345
|
+
| File | Quant | Size | Description | Use Case |
|
|
346
|
+
|------|-------|------|-------------|----------|
|
|
347
|
+
| {model_name}-f16.gguf | F16 | ~1GB | Full precision | Best quality, slower |
|
|
348
|
+
| {model_name}-q8_0.gguf | Q8_0 | ~500MB | 8-bit | High quality |
|
|
349
|
+
| {model_name}-q5_k_m.gguf | Q5_K_M | ~350MB | 5-bit medium | Good quality, smaller |
|
|
350
|
+
| {model_name}-q4_k_m.gguf | Q4_K_M | ~300MB | 4-bit medium | Recommended - good balance |
|
|
351
|
+
|
|
352
|
+
## Usage
|
|
353
|
+
|
|
354
|
+
### With llama.cpp
|
|
355
|
+
|
|
356
|
+
```bash
|
|
357
|
+
# Download model
|
|
358
|
+
huggingface-cli download {OUTPUT_REPO} {model_name}-q4_k_m.gguf
|
|
359
|
+
|
|
360
|
+
# Run with llama.cpp
|
|
361
|
+
./llama-cli -m {model_name}-q4_k_m.gguf -p "Your prompt here"
|
|
362
|
+
```
|
|
363
|
+
|
|
364
|
+
### With Ollama
|
|
365
|
+
|
|
366
|
+
1. Create a `Modelfile`:
|
|
367
|
+
```
|
|
368
|
+
FROM ./{model_name}-q4_k_m.gguf
|
|
369
|
+
```
|
|
370
|
+
|
|
371
|
+
2. Create the model:
|
|
372
|
+
```bash
|
|
373
|
+
ollama create my-model -f Modelfile
|
|
374
|
+
ollama run my-model
|
|
375
|
+
```
|
|
376
|
+
|
|
377
|
+
### With LM Studio
|
|
378
|
+
|
|
379
|
+
1. Download the `.gguf` file
|
|
380
|
+
2. Import into LM Studio
|
|
381
|
+
3. Start chatting!
|
|
382
|
+
|
|
383
|
+
## License
|
|
384
|
+
|
|
385
|
+
Inherits the license from the base model: {BASE_MODEL}
|
|
386
|
+
|
|
387
|
+
## Citation
|
|
388
|
+
|
|
389
|
+
```bibtex
|
|
390
|
+
@misc{{{OUTPUT_REPO.split('/')[-1].replace('-', '_')},
|
|
391
|
+
author = {{{username}}},
|
|
392
|
+
title = {{{OUTPUT_REPO.split('/')[-1]}}},
|
|
393
|
+
year = {{2025}},
|
|
394
|
+
publisher = {{Hugging Face}},
|
|
395
|
+
url = {{https://huggingface.co/{OUTPUT_REPO}}}
|
|
396
|
+
}}
|
|
397
|
+
```
|
|
398
|
+
|
|
399
|
+
---
|
|
400
|
+
|
|
401
|
+
*Converted to GGUF format using llama.cpp*
|
|
402
|
+
"""
|
|
403
|
+
|
|
404
|
+
try:
|
|
405
|
+
api.upload_file(
|
|
406
|
+
path_or_fileobj=readme_content.encode(),
|
|
407
|
+
path_in_repo="README.md",
|
|
408
|
+
repo_id=OUTPUT_REPO,
|
|
409
|
+
)
|
|
410
|
+
print(" ✅ README uploaded")
|
|
411
|
+
except Exception as e:
|
|
412
|
+
print(f" ❌ README upload failed: {e}")
|
|
413
|
+
|
|
414
|
+
print("\n" + "=" * 60)
|
|
415
|
+
print("✅ GGUF Conversion Complete!")
|
|
416
|
+
print(f"📦 Repository: https://huggingface.co/{OUTPUT_REPO}")
|
|
417
|
+
print(f"\n📥 Download with:")
|
|
418
|
+
print(f" huggingface-cli download {OUTPUT_REPO} {model_name}-q4_k_m.gguf")
|
|
419
|
+
print(f"\n🚀 Use with Ollama:")
|
|
420
|
+
print(" 1. Download the GGUF file")
|
|
421
|
+
print(f" 2. Create Modelfile: FROM ./{model_name}-q4_k_m.gguf")
|
|
422
|
+
print(" 3. ollama create my-model -f Modelfile")
|
|
423
|
+
print(" 4. ollama run my-model")
|
|
424
|
+
print("=" * 60)
|