@synsci/cli-darwin-x64 1.1.49

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (373) hide show
  1. package/bin/skills/accelerate/SKILL.md +332 -0
  2. package/bin/skills/accelerate/references/custom-plugins.md +453 -0
  3. package/bin/skills/accelerate/references/megatron-integration.md +489 -0
  4. package/bin/skills/accelerate/references/performance.md +525 -0
  5. package/bin/skills/audiocraft/SKILL.md +564 -0
  6. package/bin/skills/audiocraft/references/advanced-usage.md +666 -0
  7. package/bin/skills/audiocraft/references/troubleshooting.md +504 -0
  8. package/bin/skills/autogpt/SKILL.md +403 -0
  9. package/bin/skills/autogpt/references/advanced-usage.md +535 -0
  10. package/bin/skills/autogpt/references/troubleshooting.md +420 -0
  11. package/bin/skills/awq/SKILL.md +310 -0
  12. package/bin/skills/awq/references/advanced-usage.md +324 -0
  13. package/bin/skills/awq/references/troubleshooting.md +344 -0
  14. package/bin/skills/axolotl/SKILL.md +158 -0
  15. package/bin/skills/axolotl/references/api.md +5548 -0
  16. package/bin/skills/axolotl/references/dataset-formats.md +1029 -0
  17. package/bin/skills/axolotl/references/index.md +15 -0
  18. package/bin/skills/axolotl/references/other.md +3563 -0
  19. package/bin/skills/bigcode-evaluation-harness/SKILL.md +405 -0
  20. package/bin/skills/bigcode-evaluation-harness/references/benchmarks.md +393 -0
  21. package/bin/skills/bigcode-evaluation-harness/references/custom-tasks.md +424 -0
  22. package/bin/skills/bigcode-evaluation-harness/references/issues.md +394 -0
  23. package/bin/skills/bitsandbytes/SKILL.md +411 -0
  24. package/bin/skills/bitsandbytes/references/memory-optimization.md +521 -0
  25. package/bin/skills/bitsandbytes/references/qlora-training.md +521 -0
  26. package/bin/skills/bitsandbytes/references/quantization-formats.md +447 -0
  27. package/bin/skills/blip-2/SKILL.md +564 -0
  28. package/bin/skills/blip-2/references/advanced-usage.md +680 -0
  29. package/bin/skills/blip-2/references/troubleshooting.md +526 -0
  30. package/bin/skills/chroma/SKILL.md +406 -0
  31. package/bin/skills/chroma/references/integration.md +38 -0
  32. package/bin/skills/clip/SKILL.md +253 -0
  33. package/bin/skills/clip/references/applications.md +207 -0
  34. package/bin/skills/constitutional-ai/SKILL.md +290 -0
  35. package/bin/skills/crewai/SKILL.md +498 -0
  36. package/bin/skills/crewai/references/flows.md +438 -0
  37. package/bin/skills/crewai/references/tools.md +429 -0
  38. package/bin/skills/crewai/references/troubleshooting.md +480 -0
  39. package/bin/skills/deepspeed/SKILL.md +141 -0
  40. package/bin/skills/deepspeed/references/08.md +17 -0
  41. package/bin/skills/deepspeed/references/09.md +173 -0
  42. package/bin/skills/deepspeed/references/2020.md +378 -0
  43. package/bin/skills/deepspeed/references/2023.md +279 -0
  44. package/bin/skills/deepspeed/references/assets.md +179 -0
  45. package/bin/skills/deepspeed/references/index.md +35 -0
  46. package/bin/skills/deepspeed/references/mii.md +118 -0
  47. package/bin/skills/deepspeed/references/other.md +1191 -0
  48. package/bin/skills/deepspeed/references/tutorials.md +6554 -0
  49. package/bin/skills/dspy/SKILL.md +590 -0
  50. package/bin/skills/dspy/references/examples.md +663 -0
  51. package/bin/skills/dspy/references/modules.md +475 -0
  52. package/bin/skills/dspy/references/optimizers.md +566 -0
  53. package/bin/skills/faiss/SKILL.md +221 -0
  54. package/bin/skills/faiss/references/index_types.md +280 -0
  55. package/bin/skills/flash-attention/SKILL.md +367 -0
  56. package/bin/skills/flash-attention/references/benchmarks.md +215 -0
  57. package/bin/skills/flash-attention/references/transformers-integration.md +293 -0
  58. package/bin/skills/gguf/SKILL.md +427 -0
  59. package/bin/skills/gguf/references/advanced-usage.md +504 -0
  60. package/bin/skills/gguf/references/troubleshooting.md +442 -0
  61. package/bin/skills/gptq/SKILL.md +450 -0
  62. package/bin/skills/gptq/references/calibration.md +337 -0
  63. package/bin/skills/gptq/references/integration.md +129 -0
  64. package/bin/skills/gptq/references/troubleshooting.md +95 -0
  65. package/bin/skills/grpo-rl-training/README.md +97 -0
  66. package/bin/skills/grpo-rl-training/SKILL.md +572 -0
  67. package/bin/skills/grpo-rl-training/examples/reward_functions_library.py +393 -0
  68. package/bin/skills/grpo-rl-training/templates/basic_grpo_training.py +228 -0
  69. package/bin/skills/guidance/SKILL.md +572 -0
  70. package/bin/skills/guidance/references/backends.md +554 -0
  71. package/bin/skills/guidance/references/constraints.md +674 -0
  72. package/bin/skills/guidance/references/examples.md +767 -0
  73. package/bin/skills/hqq/SKILL.md +445 -0
  74. package/bin/skills/hqq/references/advanced-usage.md +528 -0
  75. package/bin/skills/hqq/references/troubleshooting.md +503 -0
  76. package/bin/skills/hugging-face-cli/SKILL.md +191 -0
  77. package/bin/skills/hugging-face-cli/references/commands.md +954 -0
  78. package/bin/skills/hugging-face-cli/references/examples.md +374 -0
  79. package/bin/skills/hugging-face-datasets/SKILL.md +547 -0
  80. package/bin/skills/hugging-face-datasets/examples/diverse_training_examples.json +239 -0
  81. package/bin/skills/hugging-face-datasets/examples/system_prompt_template.txt +196 -0
  82. package/bin/skills/hugging-face-datasets/examples/training_examples.json +176 -0
  83. package/bin/skills/hugging-face-datasets/scripts/dataset_manager.py +522 -0
  84. package/bin/skills/hugging-face-datasets/scripts/sql_manager.py +844 -0
  85. package/bin/skills/hugging-face-datasets/templates/chat.json +55 -0
  86. package/bin/skills/hugging-face-datasets/templates/classification.json +62 -0
  87. package/bin/skills/hugging-face-datasets/templates/completion.json +51 -0
  88. package/bin/skills/hugging-face-datasets/templates/custom.json +75 -0
  89. package/bin/skills/hugging-face-datasets/templates/qa.json +54 -0
  90. package/bin/skills/hugging-face-datasets/templates/tabular.json +81 -0
  91. package/bin/skills/hugging-face-evaluation/SKILL.md +656 -0
  92. package/bin/skills/hugging-face-evaluation/examples/USAGE_EXAMPLES.md +382 -0
  93. package/bin/skills/hugging-face-evaluation/examples/artificial_analysis_to_hub.py +141 -0
  94. package/bin/skills/hugging-face-evaluation/examples/example_readme_tables.md +135 -0
  95. package/bin/skills/hugging-face-evaluation/examples/metric_mapping.json +50 -0
  96. package/bin/skills/hugging-face-evaluation/requirements.txt +20 -0
  97. package/bin/skills/hugging-face-evaluation/scripts/evaluation_manager.py +1374 -0
  98. package/bin/skills/hugging-face-evaluation/scripts/inspect_eval_uv.py +104 -0
  99. package/bin/skills/hugging-face-evaluation/scripts/inspect_vllm_uv.py +317 -0
  100. package/bin/skills/hugging-face-evaluation/scripts/lighteval_vllm_uv.py +303 -0
  101. package/bin/skills/hugging-face-evaluation/scripts/run_eval_job.py +98 -0
  102. package/bin/skills/hugging-face-evaluation/scripts/run_vllm_eval_job.py +331 -0
  103. package/bin/skills/hugging-face-evaluation/scripts/test_extraction.py +206 -0
  104. package/bin/skills/hugging-face-jobs/SKILL.md +1041 -0
  105. package/bin/skills/hugging-face-jobs/index.html +216 -0
  106. package/bin/skills/hugging-face-jobs/references/hardware_guide.md +336 -0
  107. package/bin/skills/hugging-face-jobs/references/hub_saving.md +352 -0
  108. package/bin/skills/hugging-face-jobs/references/token_usage.md +546 -0
  109. package/bin/skills/hugging-face-jobs/references/troubleshooting.md +475 -0
  110. package/bin/skills/hugging-face-jobs/scripts/cot-self-instruct.py +718 -0
  111. package/bin/skills/hugging-face-jobs/scripts/finepdfs-stats.py +546 -0
  112. package/bin/skills/hugging-face-jobs/scripts/generate-responses.py +587 -0
  113. package/bin/skills/hugging-face-model-trainer/SKILL.md +711 -0
  114. package/bin/skills/hugging-face-model-trainer/references/gguf_conversion.md +296 -0
  115. package/bin/skills/hugging-face-model-trainer/references/hardware_guide.md +283 -0
  116. package/bin/skills/hugging-face-model-trainer/references/hub_saving.md +364 -0
  117. package/bin/skills/hugging-face-model-trainer/references/reliability_principles.md +371 -0
  118. package/bin/skills/hugging-face-model-trainer/references/trackio_guide.md +189 -0
  119. package/bin/skills/hugging-face-model-trainer/references/training_methods.md +150 -0
  120. package/bin/skills/hugging-face-model-trainer/references/training_patterns.md +203 -0
  121. package/bin/skills/hugging-face-model-trainer/references/troubleshooting.md +282 -0
  122. package/bin/skills/hugging-face-model-trainer/scripts/convert_to_gguf.py +424 -0
  123. package/bin/skills/hugging-face-model-trainer/scripts/dataset_inspector.py +417 -0
  124. package/bin/skills/hugging-face-model-trainer/scripts/estimate_cost.py +150 -0
  125. package/bin/skills/hugging-face-model-trainer/scripts/train_dpo_example.py +106 -0
  126. package/bin/skills/hugging-face-model-trainer/scripts/train_grpo_example.py +89 -0
  127. package/bin/skills/hugging-face-model-trainer/scripts/train_sft_example.py +122 -0
  128. package/bin/skills/hugging-face-paper-publisher/SKILL.md +627 -0
  129. package/bin/skills/hugging-face-paper-publisher/examples/example_usage.md +327 -0
  130. package/bin/skills/hugging-face-paper-publisher/references/quick_reference.md +216 -0
  131. package/bin/skills/hugging-face-paper-publisher/scripts/paper_manager.py +508 -0
  132. package/bin/skills/hugging-face-paper-publisher/templates/arxiv.md +299 -0
  133. package/bin/skills/hugging-face-paper-publisher/templates/ml-report.md +358 -0
  134. package/bin/skills/hugging-face-paper-publisher/templates/modern.md +319 -0
  135. package/bin/skills/hugging-face-paper-publisher/templates/standard.md +201 -0
  136. package/bin/skills/hugging-face-tool-builder/SKILL.md +115 -0
  137. package/bin/skills/hugging-face-tool-builder/references/baseline_hf_api.py +57 -0
  138. package/bin/skills/hugging-face-tool-builder/references/baseline_hf_api.sh +40 -0
  139. package/bin/skills/hugging-face-tool-builder/references/baseline_hf_api.tsx +57 -0
  140. package/bin/skills/hugging-face-tool-builder/references/find_models_by_paper.sh +230 -0
  141. package/bin/skills/hugging-face-tool-builder/references/hf_enrich_models.sh +96 -0
  142. package/bin/skills/hugging-face-tool-builder/references/hf_model_card_frontmatter.sh +188 -0
  143. package/bin/skills/hugging-face-tool-builder/references/hf_model_papers_auth.sh +171 -0
  144. package/bin/skills/hugging-face-trackio/SKILL.md +65 -0
  145. package/bin/skills/hugging-face-trackio/references/logging_metrics.md +206 -0
  146. package/bin/skills/hugging-face-trackio/references/retrieving_metrics.md +223 -0
  147. package/bin/skills/huggingface-tokenizers/SKILL.md +516 -0
  148. package/bin/skills/huggingface-tokenizers/references/algorithms.md +653 -0
  149. package/bin/skills/huggingface-tokenizers/references/integration.md +637 -0
  150. package/bin/skills/huggingface-tokenizers/references/pipeline.md +723 -0
  151. package/bin/skills/huggingface-tokenizers/references/training.md +565 -0
  152. package/bin/skills/instructor/SKILL.md +740 -0
  153. package/bin/skills/instructor/references/examples.md +107 -0
  154. package/bin/skills/instructor/references/providers.md +70 -0
  155. package/bin/skills/instructor/references/validation.md +606 -0
  156. package/bin/skills/knowledge-distillation/SKILL.md +458 -0
  157. package/bin/skills/knowledge-distillation/references/minillm.md +334 -0
  158. package/bin/skills/lambda-labs/SKILL.md +545 -0
  159. package/bin/skills/lambda-labs/references/advanced-usage.md +611 -0
  160. package/bin/skills/lambda-labs/references/troubleshooting.md +530 -0
  161. package/bin/skills/langchain/SKILL.md +480 -0
  162. package/bin/skills/langchain/references/agents.md +499 -0
  163. package/bin/skills/langchain/references/integration.md +562 -0
  164. package/bin/skills/langchain/references/rag.md +600 -0
  165. package/bin/skills/langsmith/SKILL.md +422 -0
  166. package/bin/skills/langsmith/references/advanced-usage.md +548 -0
  167. package/bin/skills/langsmith/references/troubleshooting.md +537 -0
  168. package/bin/skills/litgpt/SKILL.md +469 -0
  169. package/bin/skills/litgpt/references/custom-models.md +568 -0
  170. package/bin/skills/litgpt/references/distributed-training.md +451 -0
  171. package/bin/skills/litgpt/references/supported-models.md +336 -0
  172. package/bin/skills/litgpt/references/training-recipes.md +619 -0
  173. package/bin/skills/llama-cpp/SKILL.md +258 -0
  174. package/bin/skills/llama-cpp/references/optimization.md +89 -0
  175. package/bin/skills/llama-cpp/references/quantization.md +213 -0
  176. package/bin/skills/llama-cpp/references/server.md +125 -0
  177. package/bin/skills/llama-factory/SKILL.md +80 -0
  178. package/bin/skills/llama-factory/references/_images.md +23 -0
  179. package/bin/skills/llama-factory/references/advanced.md +1055 -0
  180. package/bin/skills/llama-factory/references/getting_started.md +349 -0
  181. package/bin/skills/llama-factory/references/index.md +19 -0
  182. package/bin/skills/llama-factory/references/other.md +31 -0
  183. package/bin/skills/llamaguard/SKILL.md +337 -0
  184. package/bin/skills/llamaindex/SKILL.md +569 -0
  185. package/bin/skills/llamaindex/references/agents.md +83 -0
  186. package/bin/skills/llamaindex/references/data_connectors.md +108 -0
  187. package/bin/skills/llamaindex/references/query_engines.md +406 -0
  188. package/bin/skills/llava/SKILL.md +304 -0
  189. package/bin/skills/llava/references/training.md +197 -0
  190. package/bin/skills/lm-evaluation-harness/SKILL.md +490 -0
  191. package/bin/skills/lm-evaluation-harness/references/api-evaluation.md +490 -0
  192. package/bin/skills/lm-evaluation-harness/references/benchmark-guide.md +488 -0
  193. package/bin/skills/lm-evaluation-harness/references/custom-tasks.md +602 -0
  194. package/bin/skills/lm-evaluation-harness/references/distributed-eval.md +519 -0
  195. package/bin/skills/long-context/SKILL.md +536 -0
  196. package/bin/skills/long-context/references/extension_methods.md +468 -0
  197. package/bin/skills/long-context/references/fine_tuning.md +611 -0
  198. package/bin/skills/long-context/references/rope.md +402 -0
  199. package/bin/skills/mamba/SKILL.md +260 -0
  200. package/bin/skills/mamba/references/architecture-details.md +206 -0
  201. package/bin/skills/mamba/references/benchmarks.md +255 -0
  202. package/bin/skills/mamba/references/training-guide.md +388 -0
  203. package/bin/skills/megatron-core/SKILL.md +366 -0
  204. package/bin/skills/megatron-core/references/benchmarks.md +249 -0
  205. package/bin/skills/megatron-core/references/parallelism-guide.md +404 -0
  206. package/bin/skills/megatron-core/references/production-examples.md +473 -0
  207. package/bin/skills/megatron-core/references/training-recipes.md +547 -0
  208. package/bin/skills/miles/SKILL.md +315 -0
  209. package/bin/skills/miles/references/api-reference.md +141 -0
  210. package/bin/skills/miles/references/troubleshooting.md +352 -0
  211. package/bin/skills/mlflow/SKILL.md +704 -0
  212. package/bin/skills/mlflow/references/deployment.md +744 -0
  213. package/bin/skills/mlflow/references/model-registry.md +770 -0
  214. package/bin/skills/mlflow/references/tracking.md +680 -0
  215. package/bin/skills/modal/SKILL.md +341 -0
  216. package/bin/skills/modal/references/advanced-usage.md +503 -0
  217. package/bin/skills/modal/references/troubleshooting.md +494 -0
  218. package/bin/skills/model-merging/SKILL.md +539 -0
  219. package/bin/skills/model-merging/references/evaluation.md +462 -0
  220. package/bin/skills/model-merging/references/examples.md +428 -0
  221. package/bin/skills/model-merging/references/methods.md +352 -0
  222. package/bin/skills/model-pruning/SKILL.md +495 -0
  223. package/bin/skills/model-pruning/references/wanda.md +347 -0
  224. package/bin/skills/moe-training/SKILL.md +526 -0
  225. package/bin/skills/moe-training/references/architectures.md +432 -0
  226. package/bin/skills/moe-training/references/inference.md +348 -0
  227. package/bin/skills/moe-training/references/training.md +425 -0
  228. package/bin/skills/nanogpt/SKILL.md +290 -0
  229. package/bin/skills/nanogpt/references/architecture.md +382 -0
  230. package/bin/skills/nanogpt/references/data.md +476 -0
  231. package/bin/skills/nanogpt/references/training.md +564 -0
  232. package/bin/skills/nemo-curator/SKILL.md +383 -0
  233. package/bin/skills/nemo-curator/references/deduplication.md +87 -0
  234. package/bin/skills/nemo-curator/references/filtering.md +102 -0
  235. package/bin/skills/nemo-evaluator/SKILL.md +494 -0
  236. package/bin/skills/nemo-evaluator/references/adapter-system.md +340 -0
  237. package/bin/skills/nemo-evaluator/references/configuration.md +447 -0
  238. package/bin/skills/nemo-evaluator/references/custom-benchmarks.md +315 -0
  239. package/bin/skills/nemo-evaluator/references/execution-backends.md +361 -0
  240. package/bin/skills/nemo-guardrails/SKILL.md +297 -0
  241. package/bin/skills/nnsight/SKILL.md +436 -0
  242. package/bin/skills/nnsight/references/README.md +78 -0
  243. package/bin/skills/nnsight/references/api.md +344 -0
  244. package/bin/skills/nnsight/references/tutorials.md +300 -0
  245. package/bin/skills/openrlhf/SKILL.md +249 -0
  246. package/bin/skills/openrlhf/references/algorithm-comparison.md +404 -0
  247. package/bin/skills/openrlhf/references/custom-rewards.md +530 -0
  248. package/bin/skills/openrlhf/references/hybrid-engine.md +287 -0
  249. package/bin/skills/openrlhf/references/multi-node-training.md +454 -0
  250. package/bin/skills/outlines/SKILL.md +652 -0
  251. package/bin/skills/outlines/references/backends.md +615 -0
  252. package/bin/skills/outlines/references/examples.md +773 -0
  253. package/bin/skills/outlines/references/json_generation.md +652 -0
  254. package/bin/skills/peft/SKILL.md +431 -0
  255. package/bin/skills/peft/references/advanced-usage.md +514 -0
  256. package/bin/skills/peft/references/troubleshooting.md +480 -0
  257. package/bin/skills/phoenix/SKILL.md +475 -0
  258. package/bin/skills/phoenix/references/advanced-usage.md +619 -0
  259. package/bin/skills/phoenix/references/troubleshooting.md +538 -0
  260. package/bin/skills/pinecone/SKILL.md +358 -0
  261. package/bin/skills/pinecone/references/deployment.md +181 -0
  262. package/bin/skills/pytorch-fsdp/SKILL.md +126 -0
  263. package/bin/skills/pytorch-fsdp/references/index.md +7 -0
  264. package/bin/skills/pytorch-fsdp/references/other.md +4249 -0
  265. package/bin/skills/pytorch-lightning/SKILL.md +346 -0
  266. package/bin/skills/pytorch-lightning/references/callbacks.md +436 -0
  267. package/bin/skills/pytorch-lightning/references/distributed.md +490 -0
  268. package/bin/skills/pytorch-lightning/references/hyperparameter-tuning.md +556 -0
  269. package/bin/skills/pyvene/SKILL.md +473 -0
  270. package/bin/skills/pyvene/references/README.md +73 -0
  271. package/bin/skills/pyvene/references/api.md +383 -0
  272. package/bin/skills/pyvene/references/tutorials.md +376 -0
  273. package/bin/skills/qdrant/SKILL.md +493 -0
  274. package/bin/skills/qdrant/references/advanced-usage.md +648 -0
  275. package/bin/skills/qdrant/references/troubleshooting.md +631 -0
  276. package/bin/skills/ray-data/SKILL.md +326 -0
  277. package/bin/skills/ray-data/references/integration.md +82 -0
  278. package/bin/skills/ray-data/references/transformations.md +83 -0
  279. package/bin/skills/ray-train/SKILL.md +406 -0
  280. package/bin/skills/ray-train/references/multi-node.md +628 -0
  281. package/bin/skills/rwkv/SKILL.md +260 -0
  282. package/bin/skills/rwkv/references/architecture-details.md +344 -0
  283. package/bin/skills/rwkv/references/rwkv7.md +386 -0
  284. package/bin/skills/rwkv/references/state-management.md +369 -0
  285. package/bin/skills/saelens/SKILL.md +386 -0
  286. package/bin/skills/saelens/references/README.md +70 -0
  287. package/bin/skills/saelens/references/api.md +333 -0
  288. package/bin/skills/saelens/references/tutorials.md +318 -0
  289. package/bin/skills/segment-anything/SKILL.md +500 -0
  290. package/bin/skills/segment-anything/references/advanced-usage.md +589 -0
  291. package/bin/skills/segment-anything/references/troubleshooting.md +484 -0
  292. package/bin/skills/sentence-transformers/SKILL.md +255 -0
  293. package/bin/skills/sentence-transformers/references/models.md +123 -0
  294. package/bin/skills/sentencepiece/SKILL.md +235 -0
  295. package/bin/skills/sentencepiece/references/algorithms.md +200 -0
  296. package/bin/skills/sentencepiece/references/training.md +304 -0
  297. package/bin/skills/sglang/SKILL.md +442 -0
  298. package/bin/skills/sglang/references/deployment.md +490 -0
  299. package/bin/skills/sglang/references/radix-attention.md +413 -0
  300. package/bin/skills/sglang/references/structured-generation.md +541 -0
  301. package/bin/skills/simpo/SKILL.md +219 -0
  302. package/bin/skills/simpo/references/datasets.md +478 -0
  303. package/bin/skills/simpo/references/hyperparameters.md +452 -0
  304. package/bin/skills/simpo/references/loss-functions.md +350 -0
  305. package/bin/skills/skypilot/SKILL.md +509 -0
  306. package/bin/skills/skypilot/references/advanced-usage.md +491 -0
  307. package/bin/skills/skypilot/references/troubleshooting.md +570 -0
  308. package/bin/skills/slime/SKILL.md +464 -0
  309. package/bin/skills/slime/references/api-reference.md +392 -0
  310. package/bin/skills/slime/references/troubleshooting.md +386 -0
  311. package/bin/skills/speculative-decoding/SKILL.md +467 -0
  312. package/bin/skills/speculative-decoding/references/lookahead.md +309 -0
  313. package/bin/skills/speculative-decoding/references/medusa.md +350 -0
  314. package/bin/skills/stable-diffusion/SKILL.md +519 -0
  315. package/bin/skills/stable-diffusion/references/advanced-usage.md +716 -0
  316. package/bin/skills/stable-diffusion/references/troubleshooting.md +555 -0
  317. package/bin/skills/tensorboard/SKILL.md +629 -0
  318. package/bin/skills/tensorboard/references/integrations.md +638 -0
  319. package/bin/skills/tensorboard/references/profiling.md +545 -0
  320. package/bin/skills/tensorboard/references/visualization.md +620 -0
  321. package/bin/skills/tensorrt-llm/SKILL.md +187 -0
  322. package/bin/skills/tensorrt-llm/references/multi-gpu.md +298 -0
  323. package/bin/skills/tensorrt-llm/references/optimization.md +242 -0
  324. package/bin/skills/tensorrt-llm/references/serving.md +470 -0
  325. package/bin/skills/tinker/SKILL.md +362 -0
  326. package/bin/skills/tinker/references/api-reference.md +168 -0
  327. package/bin/skills/tinker/references/getting-started.md +157 -0
  328. package/bin/skills/tinker/references/loss-functions.md +163 -0
  329. package/bin/skills/tinker/references/models-and-lora.md +139 -0
  330. package/bin/skills/tinker/references/recipes.md +280 -0
  331. package/bin/skills/tinker/references/reinforcement-learning.md +212 -0
  332. package/bin/skills/tinker/references/rendering.md +243 -0
  333. package/bin/skills/tinker/references/supervised-learning.md +232 -0
  334. package/bin/skills/tinker-training-cost/SKILL.md +187 -0
  335. package/bin/skills/tinker-training-cost/scripts/calculate_cost.py +123 -0
  336. package/bin/skills/torchforge/SKILL.md +433 -0
  337. package/bin/skills/torchforge/references/api-reference.md +327 -0
  338. package/bin/skills/torchforge/references/troubleshooting.md +409 -0
  339. package/bin/skills/torchtitan/SKILL.md +358 -0
  340. package/bin/skills/torchtitan/references/checkpoint.md +181 -0
  341. package/bin/skills/torchtitan/references/custom-models.md +258 -0
  342. package/bin/skills/torchtitan/references/float8.md +133 -0
  343. package/bin/skills/torchtitan/references/fsdp.md +126 -0
  344. package/bin/skills/transformer-lens/SKILL.md +346 -0
  345. package/bin/skills/transformer-lens/references/README.md +54 -0
  346. package/bin/skills/transformer-lens/references/api.md +362 -0
  347. package/bin/skills/transformer-lens/references/tutorials.md +339 -0
  348. package/bin/skills/trl-fine-tuning/SKILL.md +455 -0
  349. package/bin/skills/trl-fine-tuning/references/dpo-variants.md +227 -0
  350. package/bin/skills/trl-fine-tuning/references/online-rl.md +82 -0
  351. package/bin/skills/trl-fine-tuning/references/reward-modeling.md +122 -0
  352. package/bin/skills/trl-fine-tuning/references/sft-training.md +168 -0
  353. package/bin/skills/unsloth/SKILL.md +80 -0
  354. package/bin/skills/unsloth/references/index.md +7 -0
  355. package/bin/skills/unsloth/references/llms-full.md +16799 -0
  356. package/bin/skills/unsloth/references/llms-txt.md +12044 -0
  357. package/bin/skills/unsloth/references/llms.md +82 -0
  358. package/bin/skills/verl/SKILL.md +391 -0
  359. package/bin/skills/verl/references/api-reference.md +301 -0
  360. package/bin/skills/verl/references/troubleshooting.md +391 -0
  361. package/bin/skills/vllm/SKILL.md +364 -0
  362. package/bin/skills/vllm/references/optimization.md +226 -0
  363. package/bin/skills/vllm/references/quantization.md +284 -0
  364. package/bin/skills/vllm/references/server-deployment.md +255 -0
  365. package/bin/skills/vllm/references/troubleshooting.md +447 -0
  366. package/bin/skills/weights-and-biases/SKILL.md +590 -0
  367. package/bin/skills/weights-and-biases/references/artifacts.md +584 -0
  368. package/bin/skills/weights-and-biases/references/integrations.md +700 -0
  369. package/bin/skills/weights-and-biases/references/sweeps.md +847 -0
  370. package/bin/skills/whisper/SKILL.md +317 -0
  371. package/bin/skills/whisper/references/languages.md +189 -0
  372. package/bin/synsc +0 -0
  373. package/package.json +10 -0
@@ -0,0 +1,304 @@
1
+ ---
2
+ name: llava
3
+ description: Large Language and Vision Assistant. Enables visual instruction tuning and image-based conversations. Combines CLIP vision encoder with Vicuna/LLaMA language models. Supports multi-turn image chat, visual question answering, and instruction following. Use for vision-language chatbots or image understanding tasks. Best for conversational image analysis.
4
+ version: 1.0.0
5
+ author: Synthetic Sciences
6
+ license: MIT
7
+ tags: [LLaVA, Vision-Language, Multimodal, Visual Question Answering, Image Chat, CLIP, Vicuna, Conversational AI, Instruction Tuning, VQA]
8
+ dependencies: [transformers, torch, pillow]
9
+ ---
10
+
11
+ # LLaVA - Large Language and Vision Assistant
12
+
13
+ Open-source vision-language model for conversational image understanding.
14
+
15
+ ## When to use LLaVA
16
+
17
+ **Use when:**
18
+ - Building vision-language chatbots
19
+ - Visual question answering (VQA)
20
+ - Image description and captioning
21
+ - Multi-turn image conversations
22
+ - Visual instruction following
23
+ - Document understanding with images
24
+
25
+ **Metrics**:
26
+ - **23,000+ GitHub stars**
27
+ - GPT-4V level capabilities (targeted)
28
+ - Apache 2.0 License
29
+ - Multiple model sizes (7B-34B params)
30
+
31
+ **Use alternatives instead**:
32
+ - **GPT-4V**: Highest quality, API-based
33
+ - **CLIP**: Simple zero-shot classification
34
+ - **BLIP-2**: Better for captioning only
35
+ - **Flamingo**: Research, not open-source
36
+
37
+ ## Quick start
38
+
39
+ ### Installation
40
+
41
+ ```bash
42
+ # Clone repository
43
+ git clone https://github.com/haotian-liu/LLaVA
44
+ cd LLaVA
45
+
46
+ # Install
47
+ pip install -e .
48
+ ```
49
+
50
+ ### Basic usage
51
+
52
+ ```python
53
+ from llava.model.builder import load_pretrained_model
54
+ from llava.mm_utils import get_model_name_from_path, process_images, tokenizer_image_token
55
+ from llava.constants import IMAGE_TOKEN_INDEX, DEFAULT_IMAGE_TOKEN
56
+ from llava.conversation import conv_templates
57
+ from PIL import Image
58
+ import torch
59
+
60
+ # Load model
61
+ model_path = "liuhaotian/llava-v1.5-7b"
62
+ tokenizer, model, image_processor, context_len = load_pretrained_model(
63
+ model_path=model_path,
64
+ model_base=None,
65
+ model_name=get_model_name_from_path(model_path)
66
+ )
67
+
68
+ # Load image
69
+ image = Image.open("image.jpg")
70
+ image_tensor = process_images([image], image_processor, model.config)
71
+ image_tensor = image_tensor.to(model.device, dtype=torch.float16)
72
+
73
+ # Create conversation
74
+ conv = conv_templates["llava_v1"].copy()
75
+ conv.append_message(conv.roles[0], DEFAULT_IMAGE_TOKEN + "\nWhat is in this image?")
76
+ conv.append_message(conv.roles[1], None)
77
+ prompt = conv.get_prompt()
78
+
79
+ # Generate response
80
+ input_ids = tokenizer_image_token(prompt, tokenizer, IMAGE_TOKEN_INDEX, return_tensors='pt').unsqueeze(0).to(model.device)
81
+
82
+ with torch.inference_mode():
83
+ output_ids = model.generate(
84
+ input_ids,
85
+ images=image_tensor,
86
+ do_sample=True,
87
+ temperature=0.2,
88
+ max_new_tokens=512
89
+ )
90
+
91
+ response = tokenizer.decode(output_ids[0], skip_special_tokens=True).strip()
92
+ print(response)
93
+ ```
94
+
95
+ ## Available models
96
+
97
+ | Model | Parameters | VRAM | Quality |
98
+ |-------|------------|------|---------|
99
+ | LLaVA-v1.5-7B | 7B | ~14 GB | Good |
100
+ | LLaVA-v1.5-13B | 13B | ~28 GB | Better |
101
+ | LLaVA-v1.6-34B | 34B | ~70 GB | Best |
102
+
103
+ ```python
104
+ # Load different models
105
+ model_7b = "liuhaotian/llava-v1.5-7b"
106
+ model_13b = "liuhaotian/llava-v1.5-13b"
107
+ model_34b = "liuhaotian/llava-v1.6-34b"
108
+
109
+ # 4-bit quantization for lower VRAM
110
+ load_4bit = True # Reduces VRAM by ~4×
111
+ ```
112
+
113
+ ## CLI usage
114
+
115
+ ```bash
116
+ # Single image query
117
+ python -m llava.serve.cli \
118
+ --model-path liuhaotian/llava-v1.5-7b \
119
+ --image-file image.jpg \
120
+ --query "What is in this image?"
121
+
122
+ # Multi-turn conversation
123
+ python -m llava.serve.cli \
124
+ --model-path liuhaotian/llava-v1.5-7b \
125
+ --image-file image.jpg
126
+ # Then type questions interactively
127
+ ```
128
+
129
+ ## Web UI (Gradio)
130
+
131
+ ```bash
132
+ # Launch Gradio interface
133
+ python -m llava.serve.gradio_web_server \
134
+ --model-path liuhaotian/llava-v1.5-7b \
135
+ --load-4bit # Optional: reduce VRAM
136
+
137
+ # Access at http://localhost:7860
138
+ ```
139
+
140
+ ## Multi-turn conversations
141
+
142
+ ```python
143
+ # Initialize conversation
144
+ conv = conv_templates["llava_v1"].copy()
145
+
146
+ # Turn 1
147
+ conv.append_message(conv.roles[0], DEFAULT_IMAGE_TOKEN + "\nWhat is in this image?")
148
+ conv.append_message(conv.roles[1], None)
149
+ response1 = generate(conv, model, image) # "A dog playing in a park"
150
+
151
+ # Turn 2
152
+ conv.messages[-1][1] = response1 # Add previous response
153
+ conv.append_message(conv.roles[0], "What breed is the dog?")
154
+ conv.append_message(conv.roles[1], None)
155
+ response2 = generate(conv, model, image) # "Golden Retriever"
156
+
157
+ # Turn 3
158
+ conv.messages[-1][1] = response2
159
+ conv.append_message(conv.roles[0], "What time of day is it?")
160
+ conv.append_message(conv.roles[1], None)
161
+ response3 = generate(conv, model, image)
162
+ ```
163
+
164
+ ## Common tasks
165
+
166
+ ### Image captioning
167
+
168
+ ```python
169
+ question = "Describe this image in detail."
170
+ response = ask(model, image, question)
171
+ ```
172
+
173
+ ### Visual question answering
174
+
175
+ ```python
176
+ question = "How many people are in the image?"
177
+ response = ask(model, image, question)
178
+ ```
179
+
180
+ ### Object detection (textual)
181
+
182
+ ```python
183
+ question = "List all the objects you can see in this image."
184
+ response = ask(model, image, question)
185
+ ```
186
+
187
+ ### Scene understanding
188
+
189
+ ```python
190
+ question = "What is happening in this scene?"
191
+ response = ask(model, image, question)
192
+ ```
193
+
194
+ ### Document understanding
195
+
196
+ ```python
197
+ question = "What is the main topic of this document?"
198
+ response = ask(model, document_image, question)
199
+ ```
200
+
201
+ ## Training custom model
202
+
203
+ ```bash
204
+ # Stage 1: Feature alignment (558K image-caption pairs)
205
+ bash scripts/v1_5/pretrain.sh
206
+
207
+ # Stage 2: Visual instruction tuning (150K instruction data)
208
+ bash scripts/v1_5/finetune.sh
209
+ ```
210
+
211
+ ## Quantization (reduce VRAM)
212
+
213
+ ```python
214
+ # 4-bit quantization
215
+ tokenizer, model, image_processor, context_len = load_pretrained_model(
216
+ model_path="liuhaotian/llava-v1.5-13b",
217
+ model_base=None,
218
+ model_name=get_model_name_from_path("liuhaotian/llava-v1.5-13b"),
219
+ load_4bit=True # Reduces VRAM ~4×
220
+ )
221
+
222
+ # 8-bit quantization
223
+ load_8bit=True # Reduces VRAM ~2×
224
+ ```
225
+
226
+ ## Best practices
227
+
228
+ 1. **Start with 7B model** - Good quality, manageable VRAM
229
+ 2. **Use 4-bit quantization** - Reduces VRAM significantly
230
+ 3. **GPU required** - CPU inference extremely slow
231
+ 4. **Clear prompts** - Specific questions get better answers
232
+ 5. **Multi-turn conversations** - Maintain conversation context
233
+ 6. **Temperature 0.2-0.7** - Balance creativity/consistency
234
+ 7. **max_new_tokens 512-1024** - For detailed responses
235
+ 8. **Batch processing** - Process multiple images sequentially
236
+
237
+ ## Performance
238
+
239
+ | Model | VRAM (FP16) | VRAM (4-bit) | Speed (tokens/s) |
240
+ |-------|-------------|--------------|------------------|
241
+ | 7B | ~14 GB | ~4 GB | ~20 |
242
+ | 13B | ~28 GB | ~8 GB | ~12 |
243
+ | 34B | ~70 GB | ~18 GB | ~5 |
244
+
245
+ *On A100 GPU*
246
+
247
+ ## Benchmarks
248
+
249
+ LLaVA achieves competitive scores on:
250
+ - **VQAv2**: 78.5%
251
+ - **GQA**: 62.0%
252
+ - **MM-Vet**: 35.4%
253
+ - **MMBench**: 64.3%
254
+
255
+ ## Limitations
256
+
257
+ 1. **Hallucinations** - May describe things not in image
258
+ 2. **Spatial reasoning** - Struggles with precise locations
259
+ 3. **Small text** - Difficulty reading fine print
260
+ 4. **Object counting** - Imprecise for many objects
261
+ 5. **VRAM requirements** - Need powerful GPU
262
+ 6. **Inference speed** - Slower than CLIP
263
+
264
+ ## Integration with frameworks
265
+
266
+ ### LangChain
267
+
268
+ ```python
269
+ from langchain.llms.base import LLM
270
+
271
+ class LLaVALLM(LLM):
272
+ def _call(self, prompt, stop=None):
273
+ # Custom LLaVA inference
274
+ return response
275
+
276
+ llm = LLaVALLM()
277
+ ```
278
+
279
+ ### Gradio App
280
+
281
+ ```python
282
+ import gradio as gr
283
+
284
+ def chat(image, text, history):
285
+ response = ask_llava(model, image, text)
286
+ return response
287
+
288
+ demo = gr.ChatInterface(
289
+ chat,
290
+ additional_inputs=[gr.Image(type="pil")],
291
+ title="LLaVA Chat"
292
+ )
293
+ demo.launch()
294
+ ```
295
+
296
+ ## Resources
297
+
298
+ - **GitHub**: https://github.com/haotian-liu/LLaVA ⭐ 23,000+
299
+ - **Paper**: https://arxiv.org/abs/2304.08485
300
+ - **Demo**: https://llava.hliu.cc
301
+ - **Models**: https://huggingface.co/liuhaotian
302
+ - **License**: Apache 2.0
303
+
304
+
@@ -0,0 +1,197 @@
1
+ # LLaVA Training Guide
2
+
3
+ Guide to training and fine-tuning LLaVA models.
4
+
5
+ ## Training stages
6
+
7
+ ### Stage 1: Feature alignment (Pretraining)
8
+
9
+ **Purpose**: Align vision encoder with language model
10
+
11
+ **Data**: 558K image-caption pairs (CC3M subset)
12
+
13
+ ```bash
14
+ # Download pretrained projector or train from scratch
15
+ bash scripts/v1_5/pretrain.sh
16
+ ```
17
+
18
+ **Configuration:**
19
+ - Base model: Vicuna-7B or LLaMA-2-7B
20
+ - Vision encoder: CLIP ViT-L/14
21
+ - Training time: ~20 hours on 8× A100
22
+
23
+ ### Stage 2: Visual instruction tuning
24
+
25
+ **Purpose**: Teach model to follow visual instructions
26
+
27
+ **Data**: 150K GPT-generated multimodal instruction data
28
+
29
+ ```bash
30
+ # Fine-tune with instruction data
31
+ bash scripts/v1_5/finetune.sh
32
+ ```
33
+
34
+ **Configuration:**
35
+ - Epochs: 1
36
+ - Batch size: 128 (across 8 GPUs)
37
+ - Learning rate: 2e-5
38
+ - Training time: ~24 hours on 8× A100
39
+
40
+ ## Data format
41
+
42
+ ### Instruction data format
43
+
44
+ ```json
45
+ [
46
+ {
47
+ "id": "001",
48
+ "image": "path/to/image.jpg",
49
+ "conversations": [
50
+ {
51
+ "from": "human",
52
+ "value": "<image>\nWhat is in this image?"
53
+ },
54
+ {
55
+ "from": "gpt",
56
+ "value": "The image shows a dog playing in a park."
57
+ },
58
+ {
59
+ "from": "human",
60
+ "value": "What breed is the dog?"
61
+ },
62
+ {
63
+ "from": "gpt",
64
+ "value": "It appears to be a Golden Retriever."
65
+ }
66
+ ]
67
+ }
68
+ ]
69
+ ```
70
+
71
+ ## Fine-tuning on custom data
72
+
73
+ ### Prepare your data
74
+
75
+ ```python
76
+ import json
77
+
78
+ # Create instruction data
79
+ data = []
80
+ for image_path, qa_pairs in your_dataset:
81
+ conversations = []
82
+ for q, a in qa_pairs:
83
+ conversations.append({"from": "human", "value": f"<image>\n{q}"})
84
+ conversations.append({"from": "gpt", "value": a})
85
+
86
+ data.append({
87
+ "id": str(len(data)),
88
+ "image": image_path,
89
+ "conversations": conversations
90
+ })
91
+
92
+ # Save
93
+ with open("custom_data.json", "w") as f:
94
+ json.dump(data, f, indent=2)
95
+ ```
96
+
97
+ ### Fine-tune script
98
+
99
+ ```bash
100
+ #!/bin/bash
101
+
102
+ # Set paths
103
+ DATA_PATH="custom_data.json"
104
+ IMAGE_FOLDER="path/to/images"
105
+ MODEL_PATH="liuhaotian/llava-v1.5-7b"
106
+ OUTPUT_DIR="./checkpoints/llava-custom"
107
+
108
+ # Fine-tune
109
+ deepspeed llava/train/train_mem.py \
110
+ --deepspeed ./scripts/zero2.json \
111
+ --model_name_or_path $MODEL_PATH \
112
+ --version v1 \
113
+ --data_path $DATA_PATH \
114
+ --image_folder $IMAGE_FOLDER \
115
+ --vision_tower openai/clip-vit-large-patch14-336 \
116
+ --mm_projector_type mlp2x_gelu \
117
+ --mm_vision_select_layer -2 \
118
+ --mm_use_im_start_end False \
119
+ --mm_use_im_patch_token False \
120
+ --image_aspect_ratio pad \
121
+ --group_by_modality_length True \
122
+ --bf16 True \
123
+ --output_dir $OUTPUT_DIR \
124
+ --num_train_epochs 1 \
125
+ --per_device_train_batch_size 16 \
126
+ --per_device_eval_batch_size 4 \
127
+ --gradient_accumulation_steps 1 \
128
+ --evaluation_strategy "no" \
129
+ --save_strategy "steps" \
130
+ --save_steps 50000 \
131
+ --save_total_limit 1 \
132
+ --learning_rate 2e-5 \
133
+ --weight_decay 0. \
134
+ --warmup_ratio 0.03 \
135
+ --lr_scheduler_type "cosine" \
136
+ --logging_steps 1 \
137
+ --tf32 True \
138
+ --model_max_length 2048 \
139
+ --gradient_checkpointing True \
140
+ --dataloader_num_workers 4 \
141
+ --lazy_preprocess True \
142
+ --report_to wandb
143
+ ```
144
+
145
+ ## LoRA fine-tuning (memory efficient)
146
+
147
+ ```python
148
+ from peft import LoraConfig, get_peft_model
149
+
150
+ # LoRA config
151
+ lora_config = LoraConfig(
152
+ r=8, # LoRA rank
153
+ lora_alpha=16,
154
+ target_modules=["q_proj", "v_proj"],
155
+ lora_dropout=0.05,
156
+ bias="none",
157
+ task_type="CAUSAL_LM"
158
+ )
159
+
160
+ # Apply LoRA
161
+ model = get_peft_model(base_model, lora_config)
162
+
163
+ # Train with much lower memory
164
+ ```
165
+
166
+ ## Hardware requirements
167
+
168
+ ### Full fine-tuning
169
+
170
+ - **7B model**: 8× A100 (40GB)
171
+ - **13B model**: 8× A100 (80GB)
172
+ - **Training time**: 20-48 hours
173
+
174
+ ### LoRA fine-tuning
175
+
176
+ - **7B model**: 1× A100 (40GB)
177
+ - **13B model**: 2× A100 (40GB)
178
+ - **Training time**: 10-24 hours
179
+
180
+ ## Best practices
181
+
182
+ 1. **Start with pretrained** - Don't train from scratch
183
+ 2. **Use LoRA for efficiency** - 10× less memory
184
+ 3. **Quality over quantity** - 1K high-quality > 10K low-quality
185
+ 4. **Multi-turn conversations** - More engaging than single Q&A
186
+ 5. **Diverse images** - Cover different scenarios
187
+ 6. **Clear instructions** - Specific questions get better answers
188
+ 7. **Monitor loss** - Should decrease smoothly
189
+ 8. **Save checkpoints** - Training can fail
190
+ 9. **Test regularly** - Validate on held-out set
191
+ 10. **Use DeepSpeed** - For multi-GPU training
192
+
193
+ ## Resources
194
+
195
+ - **Training script**: https://github.com/haotian-liu/LLaVA/tree/main/scripts
196
+ - **Data format**: https://github.com/haotian-liu/LLaVA/blob/main/docs/Data.md
197
+ - **Paper**: https://arxiv.org/abs/2304.08485