@synsci/cli-darwin-x64 1.1.97 → 1.1.99

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (1549) hide show
  1. package/bin/synsc +0 -0
  2. package/package.json +1 -1
  3. package/bin/skills/accelerate/SKILL.md +0 -332
  4. package/bin/skills/accelerate/references/custom-plugins.md +0 -453
  5. package/bin/skills/accelerate/references/megatron-integration.md +0 -489
  6. package/bin/skills/accelerate/references/performance.md +0 -525
  7. package/bin/skills/adaptyv/SKILL.md +0 -114
  8. package/bin/skills/adaptyv/reference/api_reference.md +0 -308
  9. package/bin/skills/adaptyv/reference/examples.md +0 -913
  10. package/bin/skills/adaptyv/reference/experiments.md +0 -360
  11. package/bin/skills/adaptyv/reference/protein_optimization.md +0 -637
  12. package/bin/skills/aeon/SKILL.md +0 -374
  13. package/bin/skills/aeon/references/anomaly_detection.md +0 -154
  14. package/bin/skills/aeon/references/classification.md +0 -144
  15. package/bin/skills/aeon/references/clustering.md +0 -123
  16. package/bin/skills/aeon/references/datasets_benchmarking.md +0 -387
  17. package/bin/skills/aeon/references/distances.md +0 -256
  18. package/bin/skills/aeon/references/forecasting.md +0 -140
  19. package/bin/skills/aeon/references/networks.md +0 -289
  20. package/bin/skills/aeon/references/regression.md +0 -118
  21. package/bin/skills/aeon/references/segmentation.md +0 -163
  22. package/bin/skills/aeon/references/similarity_search.md +0 -187
  23. package/bin/skills/aeon/references/transformations.md +0 -246
  24. package/bin/skills/alphafold-database/SKILL.md +0 -513
  25. package/bin/skills/alphafold-database/references/api_reference.md +0 -423
  26. package/bin/skills/anndata/SKILL.md +0 -400
  27. package/bin/skills/anndata/references/best_practices.md +0 -525
  28. package/bin/skills/anndata/references/concatenation.md +0 -396
  29. package/bin/skills/anndata/references/data_structure.md +0 -314
  30. package/bin/skills/anndata/references/io_operations.md +0 -404
  31. package/bin/skills/anndata/references/manipulation.md +0 -516
  32. package/bin/skills/arboreto/SKILL.md +0 -243
  33. package/bin/skills/arboreto/references/algorithms.md +0 -138
  34. package/bin/skills/arboreto/references/basic_inference.md +0 -151
  35. package/bin/skills/arboreto/references/distributed_computing.md +0 -242
  36. package/bin/skills/arboreto/scripts/basic_grn_inference.py +0 -97
  37. package/bin/skills/astropy/SKILL.md +0 -331
  38. package/bin/skills/astropy/references/coordinates.md +0 -273
  39. package/bin/skills/astropy/references/cosmology.md +0 -307
  40. package/bin/skills/astropy/references/fits.md +0 -396
  41. package/bin/skills/astropy/references/tables.md +0 -489
  42. package/bin/skills/astropy/references/time.md +0 -404
  43. package/bin/skills/astropy/references/units.md +0 -178
  44. package/bin/skills/astropy/references/wcs_and_other_modules.md +0 -373
  45. package/bin/skills/audiocraft/SKILL.md +0 -564
  46. package/bin/skills/audiocraft/references/advanced-usage.md +0 -666
  47. package/bin/skills/audiocraft/references/troubleshooting.md +0 -504
  48. package/bin/skills/autogpt/SKILL.md +0 -403
  49. package/bin/skills/autogpt/references/advanced-usage.md +0 -535
  50. package/bin/skills/autogpt/references/troubleshooting.md +0 -420
  51. package/bin/skills/awq/SKILL.md +0 -310
  52. package/bin/skills/awq/references/advanced-usage.md +0 -324
  53. package/bin/skills/awq/references/troubleshooting.md +0 -344
  54. package/bin/skills/axolotl/SKILL.md +0 -158
  55. package/bin/skills/axolotl/references/api.md +0 -5548
  56. package/bin/skills/axolotl/references/dataset-formats.md +0 -1029
  57. package/bin/skills/axolotl/references/index.md +0 -15
  58. package/bin/skills/axolotl/references/other.md +0 -3563
  59. package/bin/skills/benchling-integration/SKILL.md +0 -480
  60. package/bin/skills/benchling-integration/references/api_endpoints.md +0 -883
  61. package/bin/skills/benchling-integration/references/authentication.md +0 -379
  62. package/bin/skills/benchling-integration/references/sdk_reference.md +0 -774
  63. package/bin/skills/bigcode-evaluation-harness/SKILL.md +0 -405
  64. package/bin/skills/bigcode-evaluation-harness/references/benchmarks.md +0 -393
  65. package/bin/skills/bigcode-evaluation-harness/references/custom-tasks.md +0 -424
  66. package/bin/skills/bigcode-evaluation-harness/references/issues.md +0 -394
  67. package/bin/skills/biopython/SKILL.md +0 -443
  68. package/bin/skills/biopython/references/advanced.md +0 -577
  69. package/bin/skills/biopython/references/alignment.md +0 -362
  70. package/bin/skills/biopython/references/blast.md +0 -455
  71. package/bin/skills/biopython/references/databases.md +0 -484
  72. package/bin/skills/biopython/references/phylogenetics.md +0 -566
  73. package/bin/skills/biopython/references/sequence_io.md +0 -285
  74. package/bin/skills/biopython/references/structure.md +0 -564
  75. package/bin/skills/biorxiv-database/SKILL.md +0 -483
  76. package/bin/skills/biorxiv-database/references/api_reference.md +0 -280
  77. package/bin/skills/biorxiv-database/scripts/biorxiv_search.py +0 -445
  78. package/bin/skills/bioservices/SKILL.md +0 -361
  79. package/bin/skills/bioservices/references/identifier_mapping.md +0 -685
  80. package/bin/skills/bioservices/references/services_reference.md +0 -636
  81. package/bin/skills/bioservices/references/workflow_patterns.md +0 -811
  82. package/bin/skills/bioservices/scripts/batch_id_converter.py +0 -347
  83. package/bin/skills/bioservices/scripts/compound_cross_reference.py +0 -378
  84. package/bin/skills/bioservices/scripts/pathway_analysis.py +0 -309
  85. package/bin/skills/bioservices/scripts/protein_analysis_workflow.py +0 -408
  86. package/bin/skills/bitsandbytes/SKILL.md +0 -411
  87. package/bin/skills/bitsandbytes/references/memory-optimization.md +0 -521
  88. package/bin/skills/bitsandbytes/references/qlora-training.md +0 -521
  89. package/bin/skills/bitsandbytes/references/quantization-formats.md +0 -447
  90. package/bin/skills/blip-2/SKILL.md +0 -564
  91. package/bin/skills/blip-2/references/advanced-usage.md +0 -680
  92. package/bin/skills/blip-2/references/troubleshooting.md +0 -526
  93. package/bin/skills/brenda-database/SKILL.md +0 -719
  94. package/bin/skills/brenda-database/references/api_reference.md +0 -537
  95. package/bin/skills/brenda-database/scripts/brenda_queries.py +0 -844
  96. package/bin/skills/brenda-database/scripts/brenda_visualization.py +0 -772
  97. package/bin/skills/brenda-database/scripts/enzyme_pathway_builder.py +0 -1053
  98. package/bin/skills/cellxgene-census/SKILL.md +0 -511
  99. package/bin/skills/cellxgene-census/references/census_schema.md +0 -182
  100. package/bin/skills/cellxgene-census/references/common_patterns.md +0 -351
  101. package/bin/skills/chembl-database/SKILL.md +0 -389
  102. package/bin/skills/chembl-database/references/api_reference.md +0 -272
  103. package/bin/skills/chembl-database/scripts/example_queries.py +0 -278
  104. package/bin/skills/chroma/SKILL.md +0 -406
  105. package/bin/skills/chroma/references/integration.md +0 -38
  106. package/bin/skills/cirq/SKILL.md +0 -346
  107. package/bin/skills/cirq/references/building.md +0 -307
  108. package/bin/skills/cirq/references/experiments.md +0 -572
  109. package/bin/skills/cirq/references/hardware.md +0 -515
  110. package/bin/skills/cirq/references/noise.md +0 -515
  111. package/bin/skills/cirq/references/simulation.md +0 -350
  112. package/bin/skills/cirq/references/transformation.md +0 -416
  113. package/bin/skills/citation-management/SKILL.md +0 -1109
  114. package/bin/skills/citation-management/assets/bibtex_template.bib +0 -264
  115. package/bin/skills/citation-management/assets/citation_checklist.md +0 -386
  116. package/bin/skills/citation-management/references/bibtex_formatting.md +0 -908
  117. package/bin/skills/citation-management/references/citation_validation.md +0 -794
  118. package/bin/skills/citation-management/references/google_scholar_search.md +0 -725
  119. package/bin/skills/citation-management/references/metadata_extraction.md +0 -870
  120. package/bin/skills/citation-management/references/pubmed_search.md +0 -839
  121. package/bin/skills/citation-management/scripts/doi_to_bibtex.py +0 -182
  122. package/bin/skills/citation-management/scripts/extract_metadata.py +0 -570
  123. package/bin/skills/citation-management/scripts/format_bibtex.py +0 -349
  124. package/bin/skills/citation-management/scripts/search_google_scholar.py +0 -251
  125. package/bin/skills/citation-management/scripts/search_pubmed.py +0 -348
  126. package/bin/skills/citation-management/scripts/validate_citations.py +0 -494
  127. package/bin/skills/clinical-decision-support/README.md +0 -129
  128. package/bin/skills/clinical-decision-support/SKILL.md +0 -506
  129. package/bin/skills/clinical-decision-support/assets/biomarker_report_template.tex +0 -380
  130. package/bin/skills/clinical-decision-support/assets/clinical_pathway_template.tex +0 -222
  131. package/bin/skills/clinical-decision-support/assets/cohort_analysis_template.tex +0 -359
  132. package/bin/skills/clinical-decision-support/assets/color_schemes.tex +0 -149
  133. package/bin/skills/clinical-decision-support/assets/example_gbm_cohort.md +0 -208
  134. package/bin/skills/clinical-decision-support/assets/recommendation_strength_guide.md +0 -328
  135. package/bin/skills/clinical-decision-support/assets/treatment_recommendation_template.tex +0 -529
  136. package/bin/skills/clinical-decision-support/references/biomarker_classification.md +0 -719
  137. package/bin/skills/clinical-decision-support/references/clinical_decision_algorithms.md +0 -604
  138. package/bin/skills/clinical-decision-support/references/evidence_synthesis.md +0 -840
  139. package/bin/skills/clinical-decision-support/references/outcome_analysis.md +0 -640
  140. package/bin/skills/clinical-decision-support/references/patient_cohort_analysis.md +0 -427
  141. package/bin/skills/clinical-decision-support/references/treatment_recommendations.md +0 -521
  142. package/bin/skills/clinical-decision-support/scripts/biomarker_classifier.py +0 -383
  143. package/bin/skills/clinical-decision-support/scripts/build_decision_tree.py +0 -417
  144. package/bin/skills/clinical-decision-support/scripts/create_cohort_tables.py +0 -509
  145. package/bin/skills/clinical-decision-support/scripts/generate_survival_analysis.py +0 -441
  146. package/bin/skills/clinical-decision-support/scripts/validate_cds_document.py +0 -326
  147. package/bin/skills/clinical-reports/IMPLEMENTATION_SUMMARY.md +0 -641
  148. package/bin/skills/clinical-reports/README.md +0 -236
  149. package/bin/skills/clinical-reports/SKILL.md +0 -1127
  150. package/bin/skills/clinical-reports/assets/case_report_template.md +0 -352
  151. package/bin/skills/clinical-reports/assets/clinical_trial_csr_template.md +0 -353
  152. package/bin/skills/clinical-reports/assets/clinical_trial_sae_template.md +0 -359
  153. package/bin/skills/clinical-reports/assets/consult_note_template.md +0 -305
  154. package/bin/skills/clinical-reports/assets/discharge_summary_template.md +0 -453
  155. package/bin/skills/clinical-reports/assets/hipaa_compliance_checklist.md +0 -395
  156. package/bin/skills/clinical-reports/assets/history_physical_template.md +0 -305
  157. package/bin/skills/clinical-reports/assets/lab_report_template.md +0 -309
  158. package/bin/skills/clinical-reports/assets/pathology_report_template.md +0 -249
  159. package/bin/skills/clinical-reports/assets/quality_checklist.md +0 -338
  160. package/bin/skills/clinical-reports/assets/radiology_report_template.md +0 -318
  161. package/bin/skills/clinical-reports/assets/soap_note_template.md +0 -253
  162. package/bin/skills/clinical-reports/references/case_report_guidelines.md +0 -570
  163. package/bin/skills/clinical-reports/references/clinical_trial_reporting.md +0 -693
  164. package/bin/skills/clinical-reports/references/data_presentation.md +0 -530
  165. package/bin/skills/clinical-reports/references/diagnostic_reports_standards.md +0 -629
  166. package/bin/skills/clinical-reports/references/medical_terminology.md +0 -588
  167. package/bin/skills/clinical-reports/references/patient_documentation.md +0 -744
  168. package/bin/skills/clinical-reports/references/peer_review_standards.md +0 -585
  169. package/bin/skills/clinical-reports/references/regulatory_compliance.md +0 -577
  170. package/bin/skills/clinical-reports/scripts/check_deidentification.py +0 -332
  171. package/bin/skills/clinical-reports/scripts/compliance_checker.py +0 -78
  172. package/bin/skills/clinical-reports/scripts/extract_clinical_data.py +0 -97
  173. package/bin/skills/clinical-reports/scripts/format_adverse_events.py +0 -97
  174. package/bin/skills/clinical-reports/scripts/generate_report_template.py +0 -149
  175. package/bin/skills/clinical-reports/scripts/terminology_validator.py +0 -126
  176. package/bin/skills/clinical-reports/scripts/validate_case_report.py +0 -323
  177. package/bin/skills/clinical-reports/scripts/validate_trial_report.py +0 -88
  178. package/bin/skills/clinicaltrials-database/SKILL.md +0 -507
  179. package/bin/skills/clinicaltrials-database/references/api_reference.md +0 -358
  180. package/bin/skills/clinicaltrials-database/scripts/query_clinicaltrials.py +0 -215
  181. package/bin/skills/clinpgx-database/SKILL.md +0 -638
  182. package/bin/skills/clinpgx-database/references/api_reference.md +0 -757
  183. package/bin/skills/clinpgx-database/scripts/query_clinpgx.py +0 -518
  184. package/bin/skills/clinvar-database/SKILL.md +0 -362
  185. package/bin/skills/clinvar-database/references/api_reference.md +0 -227
  186. package/bin/skills/clinvar-database/references/clinical_significance.md +0 -218
  187. package/bin/skills/clinvar-database/references/data_formats.md +0 -358
  188. package/bin/skills/clip/SKILL.md +0 -253
  189. package/bin/skills/clip/references/applications.md +0 -207
  190. package/bin/skills/cobrapy/SKILL.md +0 -463
  191. package/bin/skills/cobrapy/references/api_quick_reference.md +0 -655
  192. package/bin/skills/cobrapy/references/workflows.md +0 -593
  193. package/bin/skills/colab-finetuning/SKILL.md +0 -153
  194. package/bin/skills/colab-finetuning/references/bridge-setup.md +0 -68
  195. package/bin/skills/colab-finetuning/references/gpu-tiers.md +0 -54
  196. package/bin/skills/colab-finetuning/references/troubleshooting.md +0 -79
  197. package/bin/skills/constitutional-ai/SKILL.md +0 -290
  198. package/bin/skills/cosmic-database/SKILL.md +0 -336
  199. package/bin/skills/cosmic-database/references/cosmic_data_reference.md +0 -220
  200. package/bin/skills/cosmic-database/scripts/download_cosmic.py +0 -231
  201. package/bin/skills/crewai/SKILL.md +0 -498
  202. package/bin/skills/crewai/references/flows.md +0 -438
  203. package/bin/skills/crewai/references/tools.md +0 -429
  204. package/bin/skills/crewai/references/troubleshooting.md +0 -480
  205. package/bin/skills/dask/SKILL.md +0 -456
  206. package/bin/skills/dask/references/arrays.md +0 -497
  207. package/bin/skills/dask/references/bags.md +0 -468
  208. package/bin/skills/dask/references/best-practices.md +0 -277
  209. package/bin/skills/dask/references/dataframes.md +0 -368
  210. package/bin/skills/dask/references/futures.md +0 -541
  211. package/bin/skills/dask/references/schedulers.md +0 -504
  212. package/bin/skills/datacommons-client/SKILL.md +0 -255
  213. package/bin/skills/datacommons-client/references/getting_started.md +0 -417
  214. package/bin/skills/datacommons-client/references/node.md +0 -250
  215. package/bin/skills/datacommons-client/references/observation.md +0 -185
  216. package/bin/skills/datacommons-client/references/resolve.md +0 -246
  217. package/bin/skills/datamol/SKILL.md +0 -706
  218. package/bin/skills/datamol/references/conformers_module.md +0 -131
  219. package/bin/skills/datamol/references/core_api.md +0 -130
  220. package/bin/skills/datamol/references/descriptors_viz.md +0 -195
  221. package/bin/skills/datamol/references/fragments_scaffolds.md +0 -174
  222. package/bin/skills/datamol/references/io_module.md +0 -109
  223. package/bin/skills/datamol/references/reactions_data.md +0 -218
  224. package/bin/skills/deepchem/SKILL.md +0 -597
  225. package/bin/skills/deepchem/references/api_reference.md +0 -303
  226. package/bin/skills/deepchem/references/workflows.md +0 -491
  227. package/bin/skills/deepchem/scripts/graph_neural_network.py +0 -338
  228. package/bin/skills/deepchem/scripts/predict_solubility.py +0 -224
  229. package/bin/skills/deepchem/scripts/transfer_learning.py +0 -375
  230. package/bin/skills/deepspeed/SKILL.md +0 -141
  231. package/bin/skills/deepspeed/references/08.md +0 -17
  232. package/bin/skills/deepspeed/references/09.md +0 -173
  233. package/bin/skills/deepspeed/references/2020.md +0 -378
  234. package/bin/skills/deepspeed/references/2023.md +0 -279
  235. package/bin/skills/deepspeed/references/assets.md +0 -179
  236. package/bin/skills/deepspeed/references/index.md +0 -35
  237. package/bin/skills/deepspeed/references/mii.md +0 -118
  238. package/bin/skills/deepspeed/references/other.md +0 -1191
  239. package/bin/skills/deepspeed/references/tutorials.md +0 -6554
  240. package/bin/skills/deeptools/SKILL.md +0 -531
  241. package/bin/skills/deeptools/assets/quick_reference.md +0 -58
  242. package/bin/skills/deeptools/references/effective_genome_sizes.md +0 -116
  243. package/bin/skills/deeptools/references/normalization_methods.md +0 -410
  244. package/bin/skills/deeptools/references/tools_reference.md +0 -533
  245. package/bin/skills/deeptools/references/workflows.md +0 -474
  246. package/bin/skills/deeptools/scripts/validate_files.py +0 -195
  247. package/bin/skills/deeptools/scripts/workflow_generator.py +0 -454
  248. package/bin/skills/denario/SKILL.md +0 -215
  249. package/bin/skills/denario/references/examples.md +0 -494
  250. package/bin/skills/denario/references/installation.md +0 -213
  251. package/bin/skills/denario/references/llm_configuration.md +0 -265
  252. package/bin/skills/denario/references/research_pipeline.md +0 -471
  253. package/bin/skills/diffdock/SKILL.md +0 -483
  254. package/bin/skills/diffdock/assets/batch_template.csv +0 -4
  255. package/bin/skills/diffdock/assets/custom_inference_config.yaml +0 -90
  256. package/bin/skills/diffdock/references/confidence_and_limitations.md +0 -182
  257. package/bin/skills/diffdock/references/parameters_reference.md +0 -163
  258. package/bin/skills/diffdock/references/workflows_examples.md +0 -392
  259. package/bin/skills/diffdock/scripts/analyze_results.py +0 -334
  260. package/bin/skills/diffdock/scripts/prepare_batch_csv.py +0 -254
  261. package/bin/skills/diffdock/scripts/setup_check.py +0 -278
  262. package/bin/skills/dnanexus-integration/SKILL.md +0 -383
  263. package/bin/skills/dnanexus-integration/references/app-development.md +0 -247
  264. package/bin/skills/dnanexus-integration/references/configuration.md +0 -646
  265. package/bin/skills/dnanexus-integration/references/data-operations.md +0 -400
  266. package/bin/skills/dnanexus-integration/references/job-execution.md +0 -412
  267. package/bin/skills/dnanexus-integration/references/python-sdk.md +0 -523
  268. package/bin/skills/document-skills/docx/LICENSE.txt +0 -30
  269. package/bin/skills/document-skills/docx/SKILL.md +0 -233
  270. package/bin/skills/document-skills/docx/docx-js.md +0 -350
  271. package/bin/skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/dml-chart.xsd +0 -1499
  272. package/bin/skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/dml-chartDrawing.xsd +0 -146
  273. package/bin/skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/dml-diagram.xsd +0 -1085
  274. package/bin/skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/dml-lockedCanvas.xsd +0 -11
  275. package/bin/skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/dml-main.xsd +0 -3081
  276. package/bin/skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/dml-picture.xsd +0 -23
  277. package/bin/skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/dml-spreadsheetDrawing.xsd +0 -185
  278. package/bin/skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/dml-wordprocessingDrawing.xsd +0 -287
  279. package/bin/skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/pml.xsd +0 -1676
  280. package/bin/skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-additionalCharacteristics.xsd +0 -28
  281. package/bin/skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-bibliography.xsd +0 -144
  282. package/bin/skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-commonSimpleTypes.xsd +0 -174
  283. package/bin/skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-customXmlDataProperties.xsd +0 -25
  284. package/bin/skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-customXmlSchemaProperties.xsd +0 -18
  285. package/bin/skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-documentPropertiesCustom.xsd +0 -59
  286. package/bin/skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-documentPropertiesExtended.xsd +0 -56
  287. package/bin/skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-documentPropertiesVariantTypes.xsd +0 -195
  288. package/bin/skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-math.xsd +0 -582
  289. package/bin/skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-relationshipReference.xsd +0 -25
  290. package/bin/skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/sml.xsd +0 -4439
  291. package/bin/skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/vml-main.xsd +0 -570
  292. package/bin/skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/vml-officeDrawing.xsd +0 -509
  293. package/bin/skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/vml-presentationDrawing.xsd +0 -12
  294. package/bin/skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/vml-spreadsheetDrawing.xsd +0 -108
  295. package/bin/skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/vml-wordprocessingDrawing.xsd +0 -96
  296. package/bin/skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/wml.xsd +0 -3646
  297. package/bin/skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/xml.xsd +0 -116
  298. package/bin/skills/document-skills/docx/ooxml/schemas/ecma/fouth-edition/opc-contentTypes.xsd +0 -42
  299. package/bin/skills/document-skills/docx/ooxml/schemas/ecma/fouth-edition/opc-coreProperties.xsd +0 -50
  300. package/bin/skills/document-skills/docx/ooxml/schemas/ecma/fouth-edition/opc-digSig.xsd +0 -49
  301. package/bin/skills/document-skills/docx/ooxml/schemas/ecma/fouth-edition/opc-relationships.xsd +0 -33
  302. package/bin/skills/document-skills/docx/ooxml/schemas/mce/mc.xsd +0 -75
  303. package/bin/skills/document-skills/docx/ooxml/schemas/microsoft/wml-2010.xsd +0 -560
  304. package/bin/skills/document-skills/docx/ooxml/schemas/microsoft/wml-2012.xsd +0 -67
  305. package/bin/skills/document-skills/docx/ooxml/schemas/microsoft/wml-2018.xsd +0 -14
  306. package/bin/skills/document-skills/docx/ooxml/schemas/microsoft/wml-cex-2018.xsd +0 -20
  307. package/bin/skills/document-skills/docx/ooxml/schemas/microsoft/wml-cid-2016.xsd +0 -13
  308. package/bin/skills/document-skills/docx/ooxml/schemas/microsoft/wml-sdtdatahash-2020.xsd +0 -4
  309. package/bin/skills/document-skills/docx/ooxml/schemas/microsoft/wml-symex-2015.xsd +0 -8
  310. package/bin/skills/document-skills/docx/ooxml/scripts/pack.py +0 -159
  311. package/bin/skills/document-skills/docx/ooxml/scripts/unpack.py +0 -29
  312. package/bin/skills/document-skills/docx/ooxml/scripts/validate.py +0 -69
  313. package/bin/skills/document-skills/docx/ooxml/scripts/validation/__init__.py +0 -15
  314. package/bin/skills/document-skills/docx/ooxml/scripts/validation/base.py +0 -951
  315. package/bin/skills/document-skills/docx/ooxml/scripts/validation/docx.py +0 -274
  316. package/bin/skills/document-skills/docx/ooxml/scripts/validation/pptx.py +0 -315
  317. package/bin/skills/document-skills/docx/ooxml/scripts/validation/redlining.py +0 -279
  318. package/bin/skills/document-skills/docx/ooxml.md +0 -610
  319. package/bin/skills/document-skills/docx/scripts/__init__.py +0 -1
  320. package/bin/skills/document-skills/docx/scripts/document.py +0 -1276
  321. package/bin/skills/document-skills/docx/scripts/templates/comments.xml +0 -3
  322. package/bin/skills/document-skills/docx/scripts/templates/commentsExtended.xml +0 -3
  323. package/bin/skills/document-skills/docx/scripts/templates/commentsExtensible.xml +0 -3
  324. package/bin/skills/document-skills/docx/scripts/templates/commentsIds.xml +0 -3
  325. package/bin/skills/document-skills/docx/scripts/templates/people.xml +0 -3
  326. package/bin/skills/document-skills/docx/scripts/utilities.py +0 -374
  327. package/bin/skills/document-skills/pdf/LICENSE.txt +0 -30
  328. package/bin/skills/document-skills/pdf/SKILL.md +0 -330
  329. package/bin/skills/document-skills/pdf/forms.md +0 -205
  330. package/bin/skills/document-skills/pdf/reference.md +0 -612
  331. package/bin/skills/document-skills/pdf/scripts/check_bounding_boxes.py +0 -70
  332. package/bin/skills/document-skills/pdf/scripts/check_bounding_boxes_test.py +0 -226
  333. package/bin/skills/document-skills/pdf/scripts/check_fillable_fields.py +0 -12
  334. package/bin/skills/document-skills/pdf/scripts/convert_pdf_to_images.py +0 -35
  335. package/bin/skills/document-skills/pdf/scripts/create_validation_image.py +0 -41
  336. package/bin/skills/document-skills/pdf/scripts/extract_form_field_info.py +0 -152
  337. package/bin/skills/document-skills/pdf/scripts/fill_fillable_fields.py +0 -114
  338. package/bin/skills/document-skills/pdf/scripts/fill_pdf_form_with_annotations.py +0 -108
  339. package/bin/skills/document-skills/pptx/LICENSE.txt +0 -30
  340. package/bin/skills/document-skills/pptx/SKILL.md +0 -520
  341. package/bin/skills/document-skills/pptx/html2pptx.md +0 -625
  342. package/bin/skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/dml-chart.xsd +0 -1499
  343. package/bin/skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/dml-chartDrawing.xsd +0 -146
  344. package/bin/skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/dml-diagram.xsd +0 -1085
  345. package/bin/skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/dml-lockedCanvas.xsd +0 -11
  346. package/bin/skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/dml-main.xsd +0 -3081
  347. package/bin/skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/dml-picture.xsd +0 -23
  348. package/bin/skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/dml-spreadsheetDrawing.xsd +0 -185
  349. package/bin/skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/dml-wordprocessingDrawing.xsd +0 -287
  350. package/bin/skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/pml.xsd +0 -1676
  351. package/bin/skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-additionalCharacteristics.xsd +0 -28
  352. package/bin/skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-bibliography.xsd +0 -144
  353. package/bin/skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-commonSimpleTypes.xsd +0 -174
  354. package/bin/skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-customXmlDataProperties.xsd +0 -25
  355. package/bin/skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-customXmlSchemaProperties.xsd +0 -18
  356. package/bin/skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-documentPropertiesCustom.xsd +0 -59
  357. package/bin/skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-documentPropertiesExtended.xsd +0 -56
  358. package/bin/skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-documentPropertiesVariantTypes.xsd +0 -195
  359. package/bin/skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-math.xsd +0 -582
  360. package/bin/skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-relationshipReference.xsd +0 -25
  361. package/bin/skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/sml.xsd +0 -4439
  362. package/bin/skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/vml-main.xsd +0 -570
  363. package/bin/skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/vml-officeDrawing.xsd +0 -509
  364. package/bin/skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/vml-presentationDrawing.xsd +0 -12
  365. package/bin/skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/vml-spreadsheetDrawing.xsd +0 -108
  366. package/bin/skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/vml-wordprocessingDrawing.xsd +0 -96
  367. package/bin/skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/wml.xsd +0 -3646
  368. package/bin/skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/xml.xsd +0 -116
  369. package/bin/skills/document-skills/pptx/ooxml/schemas/ecma/fouth-edition/opc-contentTypes.xsd +0 -42
  370. package/bin/skills/document-skills/pptx/ooxml/schemas/ecma/fouth-edition/opc-coreProperties.xsd +0 -50
  371. package/bin/skills/document-skills/pptx/ooxml/schemas/ecma/fouth-edition/opc-digSig.xsd +0 -49
  372. package/bin/skills/document-skills/pptx/ooxml/schemas/ecma/fouth-edition/opc-relationships.xsd +0 -33
  373. package/bin/skills/document-skills/pptx/ooxml/schemas/mce/mc.xsd +0 -75
  374. package/bin/skills/document-skills/pptx/ooxml/schemas/microsoft/wml-2010.xsd +0 -560
  375. package/bin/skills/document-skills/pptx/ooxml/schemas/microsoft/wml-2012.xsd +0 -67
  376. package/bin/skills/document-skills/pptx/ooxml/schemas/microsoft/wml-2018.xsd +0 -14
  377. package/bin/skills/document-skills/pptx/ooxml/schemas/microsoft/wml-cex-2018.xsd +0 -20
  378. package/bin/skills/document-skills/pptx/ooxml/schemas/microsoft/wml-cid-2016.xsd +0 -13
  379. package/bin/skills/document-skills/pptx/ooxml/schemas/microsoft/wml-sdtdatahash-2020.xsd +0 -4
  380. package/bin/skills/document-skills/pptx/ooxml/schemas/microsoft/wml-symex-2015.xsd +0 -8
  381. package/bin/skills/document-skills/pptx/ooxml/scripts/pack.py +0 -159
  382. package/bin/skills/document-skills/pptx/ooxml/scripts/unpack.py +0 -29
  383. package/bin/skills/document-skills/pptx/ooxml/scripts/validate.py +0 -69
  384. package/bin/skills/document-skills/pptx/ooxml/scripts/validation/__init__.py +0 -15
  385. package/bin/skills/document-skills/pptx/ooxml/scripts/validation/base.py +0 -951
  386. package/bin/skills/document-skills/pptx/ooxml/scripts/validation/docx.py +0 -274
  387. package/bin/skills/document-skills/pptx/ooxml/scripts/validation/pptx.py +0 -315
  388. package/bin/skills/document-skills/pptx/ooxml/scripts/validation/redlining.py +0 -279
  389. package/bin/skills/document-skills/pptx/ooxml.md +0 -427
  390. package/bin/skills/document-skills/pptx/scripts/html2pptx.js +0 -979
  391. package/bin/skills/document-skills/pptx/scripts/inventory.py +0 -1020
  392. package/bin/skills/document-skills/pptx/scripts/rearrange.py +0 -231
  393. package/bin/skills/document-skills/pptx/scripts/replace.py +0 -385
  394. package/bin/skills/document-skills/pptx/scripts/thumbnail.py +0 -450
  395. package/bin/skills/document-skills/xlsx/LICENSE.txt +0 -30
  396. package/bin/skills/document-skills/xlsx/SKILL.md +0 -325
  397. package/bin/skills/document-skills/xlsx/recalc.py +0 -178
  398. package/bin/skills/drugbank-database/SKILL.md +0 -190
  399. package/bin/skills/drugbank-database/references/chemical-analysis.md +0 -590
  400. package/bin/skills/drugbank-database/references/data-access.md +0 -242
  401. package/bin/skills/drugbank-database/references/drug-queries.md +0 -386
  402. package/bin/skills/drugbank-database/references/interactions.md +0 -425
  403. package/bin/skills/drugbank-database/references/targets-pathways.md +0 -518
  404. package/bin/skills/drugbank-database/scripts/drugbank_helper.py +0 -350
  405. package/bin/skills/dspy/SKILL.md +0 -590
  406. package/bin/skills/dspy/references/examples.md +0 -663
  407. package/bin/skills/dspy/references/modules.md +0 -475
  408. package/bin/skills/dspy/references/optimizers.md +0 -566
  409. package/bin/skills/ena-database/SKILL.md +0 -204
  410. package/bin/skills/ena-database/references/api_reference.md +0 -490
  411. package/bin/skills/ensembl-database/SKILL.md +0 -311
  412. package/bin/skills/ensembl-database/references/api_endpoints.md +0 -346
  413. package/bin/skills/ensembl-database/scripts/ensembl_query.py +0 -427
  414. package/bin/skills/esm/SKILL.md +0 -306
  415. package/bin/skills/esm/references/esm-c-api.md +0 -583
  416. package/bin/skills/esm/references/esm3-api.md +0 -452
  417. package/bin/skills/esm/references/forge-api.md +0 -657
  418. package/bin/skills/esm/references/workflows.md +0 -685
  419. package/bin/skills/etetoolkit/SKILL.md +0 -623
  420. package/bin/skills/etetoolkit/references/api_reference.md +0 -583
  421. package/bin/skills/etetoolkit/references/visualization.md +0 -783
  422. package/bin/skills/etetoolkit/references/workflows.md +0 -774
  423. package/bin/skills/etetoolkit/scripts/quick_visualize.py +0 -214
  424. package/bin/skills/etetoolkit/scripts/tree_operations.py +0 -229
  425. package/bin/skills/exploratory-data-analysis/SKILL.md +0 -446
  426. package/bin/skills/exploratory-data-analysis/assets/report_template.md +0 -196
  427. package/bin/skills/exploratory-data-analysis/references/bioinformatics_genomics_formats.md +0 -664
  428. package/bin/skills/exploratory-data-analysis/references/chemistry_molecular_formats.md +0 -664
  429. package/bin/skills/exploratory-data-analysis/references/general_scientific_formats.md +0 -518
  430. package/bin/skills/exploratory-data-analysis/references/microscopy_imaging_formats.md +0 -620
  431. package/bin/skills/exploratory-data-analysis/references/proteomics_metabolomics_formats.md +0 -517
  432. package/bin/skills/exploratory-data-analysis/references/spectroscopy_analytical_formats.md +0 -633
  433. package/bin/skills/exploratory-data-analysis/scripts/eda_analyzer.py +0 -547
  434. package/bin/skills/faiss/SKILL.md +0 -221
  435. package/bin/skills/faiss/references/index_types.md +0 -280
  436. package/bin/skills/fda-database/SKILL.md +0 -518
  437. package/bin/skills/fda-database/references/animal_veterinary.md +0 -377
  438. package/bin/skills/fda-database/references/api_basics.md +0 -687
  439. package/bin/skills/fda-database/references/devices.md +0 -632
  440. package/bin/skills/fda-database/references/drugs.md +0 -468
  441. package/bin/skills/fda-database/references/foods.md +0 -374
  442. package/bin/skills/fda-database/references/other.md +0 -472
  443. package/bin/skills/fda-database/scripts/fda_examples.py +0 -335
  444. package/bin/skills/fda-database/scripts/fda_query.py +0 -440
  445. package/bin/skills/fireworks-ai/SKILL.md +0 -665
  446. package/bin/skills/flash-attention/SKILL.md +0 -367
  447. package/bin/skills/flash-attention/references/benchmarks.md +0 -215
  448. package/bin/skills/flash-attention/references/transformers-integration.md +0 -293
  449. package/bin/skills/flowio/SKILL.md +0 -608
  450. package/bin/skills/flowio/references/api_reference.md +0 -372
  451. package/bin/skills/fluidsim/SKILL.md +0 -349
  452. package/bin/skills/fluidsim/references/advanced_features.md +0 -398
  453. package/bin/skills/fluidsim/references/installation.md +0 -68
  454. package/bin/skills/fluidsim/references/output_analysis.md +0 -283
  455. package/bin/skills/fluidsim/references/parameters.md +0 -198
  456. package/bin/skills/fluidsim/references/simulation_workflow.md +0 -172
  457. package/bin/skills/fluidsim/references/solvers.md +0 -94
  458. package/bin/skills/fred-economic-data/SKILL.md +0 -433
  459. package/bin/skills/fred-economic-data/references/api_basics.md +0 -212
  460. package/bin/skills/fred-economic-data/references/categories.md +0 -442
  461. package/bin/skills/fred-economic-data/references/geofred.md +0 -588
  462. package/bin/skills/fred-economic-data/references/releases.md +0 -642
  463. package/bin/skills/fred-economic-data/references/series.md +0 -584
  464. package/bin/skills/fred-economic-data/references/sources.md +0 -423
  465. package/bin/skills/fred-economic-data/references/tags.md +0 -485
  466. package/bin/skills/fred-economic-data/scripts/fred_examples.py +0 -354
  467. package/bin/skills/fred-economic-data/scripts/fred_query.py +0 -590
  468. package/bin/skills/gene-database/SKILL.md +0 -179
  469. package/bin/skills/gene-database/references/api_reference.md +0 -404
  470. package/bin/skills/gene-database/references/common_workflows.md +0 -428
  471. package/bin/skills/gene-database/scripts/batch_gene_lookup.py +0 -298
  472. package/bin/skills/gene-database/scripts/fetch_gene_data.py +0 -277
  473. package/bin/skills/gene-database/scripts/query_gene.py +0 -251
  474. package/bin/skills/generate-image/SKILL.md +0 -178
  475. package/bin/skills/generate-image/scripts/generate_image.py +0 -254
  476. package/bin/skills/geniml/SKILL.md +0 -318
  477. package/bin/skills/geniml/references/bedspace.md +0 -127
  478. package/bin/skills/geniml/references/consensus_peaks.md +0 -238
  479. package/bin/skills/geniml/references/region2vec.md +0 -90
  480. package/bin/skills/geniml/references/scembed.md +0 -197
  481. package/bin/skills/geniml/references/utilities.md +0 -385
  482. package/bin/skills/geo-database/SKILL.md +0 -815
  483. package/bin/skills/geo-database/references/geo_reference.md +0 -829
  484. package/bin/skills/geopandas/SKILL.md +0 -251
  485. package/bin/skills/geopandas/references/crs-management.md +0 -243
  486. package/bin/skills/geopandas/references/data-io.md +0 -165
  487. package/bin/skills/geopandas/references/data-structures.md +0 -70
  488. package/bin/skills/geopandas/references/geometric-operations.md +0 -221
  489. package/bin/skills/geopandas/references/spatial-analysis.md +0 -184
  490. package/bin/skills/geopandas/references/visualization.md +0 -243
  491. package/bin/skills/get-available-resources/SKILL.md +0 -277
  492. package/bin/skills/get-available-resources/scripts/detect_resources.py +0 -401
  493. package/bin/skills/gget/SKILL.md +0 -871
  494. package/bin/skills/gget/references/database_info.md +0 -300
  495. package/bin/skills/gget/references/module_reference.md +0 -467
  496. package/bin/skills/gget/references/workflows.md +0 -814
  497. package/bin/skills/gget/scripts/batch_sequence_analysis.py +0 -191
  498. package/bin/skills/gget/scripts/enrichment_pipeline.py +0 -235
  499. package/bin/skills/gget/scripts/gene_analysis.py +0 -161
  500. package/bin/skills/gguf/SKILL.md +0 -427
  501. package/bin/skills/gguf/references/advanced-usage.md +0 -504
  502. package/bin/skills/gguf/references/troubleshooting.md +0 -442
  503. package/bin/skills/gptq/SKILL.md +0 -450
  504. package/bin/skills/gptq/references/calibration.md +0 -337
  505. package/bin/skills/gptq/references/integration.md +0 -129
  506. package/bin/skills/gptq/references/troubleshooting.md +0 -95
  507. package/bin/skills/groq/SKILL.md +0 -347
  508. package/bin/skills/grpo-rl-training/README.md +0 -97
  509. package/bin/skills/grpo-rl-training/SKILL.md +0 -572
  510. package/bin/skills/grpo-rl-training/examples/reward_functions_library.py +0 -393
  511. package/bin/skills/grpo-rl-training/templates/basic_grpo_training.py +0 -228
  512. package/bin/skills/gtars/SKILL.md +0 -285
  513. package/bin/skills/gtars/references/cli.md +0 -222
  514. package/bin/skills/gtars/references/coverage.md +0 -172
  515. package/bin/skills/gtars/references/overlap.md +0 -156
  516. package/bin/skills/gtars/references/python-api.md +0 -211
  517. package/bin/skills/gtars/references/refget.md +0 -147
  518. package/bin/skills/gtars/references/tokenizers.md +0 -103
  519. package/bin/skills/guidance/SKILL.md +0 -572
  520. package/bin/skills/guidance/references/backends.md +0 -554
  521. package/bin/skills/guidance/references/constraints.md +0 -674
  522. package/bin/skills/guidance/references/examples.md +0 -767
  523. package/bin/skills/gwas-database/SKILL.md +0 -608
  524. package/bin/skills/gwas-database/references/api_reference.md +0 -793
  525. package/bin/skills/histolab/SKILL.md +0 -678
  526. package/bin/skills/histolab/references/filters_preprocessing.md +0 -514
  527. package/bin/skills/histolab/references/slide_management.md +0 -172
  528. package/bin/skills/histolab/references/tile_extraction.md +0 -421
  529. package/bin/skills/histolab/references/tissue_masks.md +0 -251
  530. package/bin/skills/histolab/references/visualization.md +0 -547
  531. package/bin/skills/hmdb-database/SKILL.md +0 -196
  532. package/bin/skills/hmdb-database/references/hmdb_data_fields.md +0 -267
  533. package/bin/skills/hqq/SKILL.md +0 -445
  534. package/bin/skills/hqq/references/advanced-usage.md +0 -528
  535. package/bin/skills/hqq/references/troubleshooting.md +0 -503
  536. package/bin/skills/hugging-face-cli/SKILL.md +0 -191
  537. package/bin/skills/hugging-face-cli/references/commands.md +0 -954
  538. package/bin/skills/hugging-face-cli/references/examples.md +0 -374
  539. package/bin/skills/hugging-face-datasets/SKILL.md +0 -547
  540. package/bin/skills/hugging-face-datasets/examples/diverse_training_examples.json +0 -239
  541. package/bin/skills/hugging-face-datasets/examples/system_prompt_template.txt +0 -196
  542. package/bin/skills/hugging-face-datasets/examples/training_examples.json +0 -176
  543. package/bin/skills/hugging-face-datasets/scripts/dataset_manager.py +0 -522
  544. package/bin/skills/hugging-face-datasets/scripts/sql_manager.py +0 -844
  545. package/bin/skills/hugging-face-datasets/templates/chat.json +0 -55
  546. package/bin/skills/hugging-face-datasets/templates/classification.json +0 -62
  547. package/bin/skills/hugging-face-datasets/templates/completion.json +0 -51
  548. package/bin/skills/hugging-face-datasets/templates/custom.json +0 -75
  549. package/bin/skills/hugging-face-datasets/templates/qa.json +0 -54
  550. package/bin/skills/hugging-face-datasets/templates/tabular.json +0 -81
  551. package/bin/skills/hugging-face-evaluation/SKILL.md +0 -656
  552. package/bin/skills/hugging-face-evaluation/examples/.env.example +0 -7
  553. package/bin/skills/hugging-face-evaluation/examples/USAGE_EXAMPLES.md +0 -382
  554. package/bin/skills/hugging-face-evaluation/examples/artificial_analysis_to_hub.py +0 -141
  555. package/bin/skills/hugging-face-evaluation/examples/example_readme_tables.md +0 -135
  556. package/bin/skills/hugging-face-evaluation/examples/metric_mapping.json +0 -50
  557. package/bin/skills/hugging-face-evaluation/requirements.txt +0 -20
  558. package/bin/skills/hugging-face-evaluation/scripts/evaluation_manager.py +0 -1374
  559. package/bin/skills/hugging-face-evaluation/scripts/inspect_eval_uv.py +0 -104
  560. package/bin/skills/hugging-face-evaluation/scripts/inspect_vllm_uv.py +0 -317
  561. package/bin/skills/hugging-face-evaluation/scripts/lighteval_vllm_uv.py +0 -303
  562. package/bin/skills/hugging-face-evaluation/scripts/run_eval_job.py +0 -98
  563. package/bin/skills/hugging-face-evaluation/scripts/run_vllm_eval_job.py +0 -331
  564. package/bin/skills/hugging-face-evaluation/scripts/test_extraction.py +0 -206
  565. package/bin/skills/hugging-face-jobs/SKILL.md +0 -1040
  566. package/bin/skills/hugging-face-jobs/index.html +0 -216
  567. package/bin/skills/hugging-face-jobs/references/hardware_guide.md +0 -336
  568. package/bin/skills/hugging-face-jobs/references/hub_saving.md +0 -352
  569. package/bin/skills/hugging-face-jobs/references/token_usage.md +0 -546
  570. package/bin/skills/hugging-face-jobs/references/troubleshooting.md +0 -475
  571. package/bin/skills/hugging-face-jobs/scripts/cot-self-instruct.py +0 -718
  572. package/bin/skills/hugging-face-jobs/scripts/finepdfs-stats.py +0 -546
  573. package/bin/skills/hugging-face-jobs/scripts/generate-responses.py +0 -587
  574. package/bin/skills/hugging-face-model-trainer/SKILL.md +0 -710
  575. package/bin/skills/hugging-face-model-trainer/references/gguf_conversion.md +0 -296
  576. package/bin/skills/hugging-face-model-trainer/references/hardware_guide.md +0 -283
  577. package/bin/skills/hugging-face-model-trainer/references/hub_saving.md +0 -364
  578. package/bin/skills/hugging-face-model-trainer/references/reliability_principles.md +0 -371
  579. package/bin/skills/hugging-face-model-trainer/references/trackio_guide.md +0 -189
  580. package/bin/skills/hugging-face-model-trainer/references/training_methods.md +0 -150
  581. package/bin/skills/hugging-face-model-trainer/references/training_patterns.md +0 -203
  582. package/bin/skills/hugging-face-model-trainer/references/troubleshooting.md +0 -282
  583. package/bin/skills/hugging-face-model-trainer/scripts/convert_to_gguf.py +0 -424
  584. package/bin/skills/hugging-face-model-trainer/scripts/dataset_inspector.py +0 -417
  585. package/bin/skills/hugging-face-model-trainer/scripts/estimate_cost.py +0 -150
  586. package/bin/skills/hugging-face-model-trainer/scripts/train_dpo_example.py +0 -106
  587. package/bin/skills/hugging-face-model-trainer/scripts/train_grpo_example.py +0 -89
  588. package/bin/skills/hugging-face-model-trainer/scripts/train_sft_example.py +0 -122
  589. package/bin/skills/hugging-face-paper-publisher/SKILL.md +0 -627
  590. package/bin/skills/hugging-face-paper-publisher/examples/example_usage.md +0 -327
  591. package/bin/skills/hugging-face-paper-publisher/references/quick_reference.md +0 -216
  592. package/bin/skills/hugging-face-paper-publisher/scripts/paper_manager.py +0 -508
  593. package/bin/skills/hugging-face-paper-publisher/templates/arxiv.md +0 -299
  594. package/bin/skills/hugging-face-paper-publisher/templates/ml-report.md +0 -358
  595. package/bin/skills/hugging-face-paper-publisher/templates/modern.md +0 -319
  596. package/bin/skills/hugging-face-paper-publisher/templates/standard.md +0 -201
  597. package/bin/skills/hugging-face-tool-builder/SKILL.md +0 -115
  598. package/bin/skills/hugging-face-tool-builder/references/baseline_hf_api.py +0 -57
  599. package/bin/skills/hugging-face-tool-builder/references/baseline_hf_api.sh +0 -40
  600. package/bin/skills/hugging-face-tool-builder/references/baseline_hf_api.tsx +0 -57
  601. package/bin/skills/hugging-face-tool-builder/references/find_models_by_paper.sh +0 -230
  602. package/bin/skills/hugging-face-tool-builder/references/hf_enrich_models.sh +0 -96
  603. package/bin/skills/hugging-face-tool-builder/references/hf_model_card_frontmatter.sh +0 -188
  604. package/bin/skills/hugging-face-tool-builder/references/hf_model_papers_auth.sh +0 -171
  605. package/bin/skills/hugging-face-trackio/.claude-plugin/plugin.json +0 -19
  606. package/bin/skills/hugging-face-trackio/SKILL.md +0 -65
  607. package/bin/skills/hugging-face-trackio/references/logging_metrics.md +0 -206
  608. package/bin/skills/hugging-face-trackio/references/retrieving_metrics.md +0 -223
  609. package/bin/skills/huggingface-tokenizers/SKILL.md +0 -516
  610. package/bin/skills/huggingface-tokenizers/references/algorithms.md +0 -653
  611. package/bin/skills/huggingface-tokenizers/references/integration.md +0 -637
  612. package/bin/skills/huggingface-tokenizers/references/pipeline.md +0 -723
  613. package/bin/skills/huggingface-tokenizers/references/training.md +0 -565
  614. package/bin/skills/hypogenic/SKILL.md +0 -655
  615. package/bin/skills/hypogenic/references/config_template.yaml +0 -150
  616. package/bin/skills/hypothesis-generation/SKILL.md +0 -293
  617. package/bin/skills/hypothesis-generation/assets/FORMATTING_GUIDE.md +0 -672
  618. package/bin/skills/hypothesis-generation/assets/hypothesis_generation.sty +0 -307
  619. package/bin/skills/hypothesis-generation/assets/hypothesis_report_template.tex +0 -572
  620. package/bin/skills/hypothesis-generation/references/experimental_design_patterns.md +0 -329
  621. package/bin/skills/hypothesis-generation/references/hypothesis_quality_criteria.md +0 -198
  622. package/bin/skills/hypothesis-generation/references/literature_search_strategies.md +0 -622
  623. package/bin/skills/imaging-data-commons/SKILL.md +0 -1182
  624. package/bin/skills/imaging-data-commons/references/bigquery_guide.md +0 -556
  625. package/bin/skills/imaging-data-commons/references/cli_guide.md +0 -272
  626. package/bin/skills/imaging-data-commons/references/cloud_storage_guide.md +0 -333
  627. package/bin/skills/imaging-data-commons/references/dicomweb_guide.md +0 -399
  628. package/bin/skills/infographics/SKILL.md +0 -563
  629. package/bin/skills/infographics/references/color_palettes.md +0 -496
  630. package/bin/skills/infographics/references/design_principles.md +0 -636
  631. package/bin/skills/infographics/references/infographic_types.md +0 -907
  632. package/bin/skills/infographics/scripts/generate_infographic.py +0 -234
  633. package/bin/skills/infographics/scripts/generate_infographic_ai.py +0 -1290
  634. package/bin/skills/instructor/SKILL.md +0 -740
  635. package/bin/skills/instructor/references/examples.md +0 -107
  636. package/bin/skills/instructor/references/providers.md +0 -70
  637. package/bin/skills/instructor/references/validation.md +0 -606
  638. package/bin/skills/iso-13485-certification/SKILL.md +0 -680
  639. package/bin/skills/iso-13485-certification/assets/templates/procedures/CAPA-procedure-template.md +0 -453
  640. package/bin/skills/iso-13485-certification/assets/templates/procedures/document-control-procedure-template.md +0 -567
  641. package/bin/skills/iso-13485-certification/assets/templates/quality-manual-template.md +0 -521
  642. package/bin/skills/iso-13485-certification/references/gap-analysis-checklist.md +0 -568
  643. package/bin/skills/iso-13485-certification/references/iso-13485-requirements.md +0 -610
  644. package/bin/skills/iso-13485-certification/references/mandatory-documents.md +0 -606
  645. package/bin/skills/iso-13485-certification/references/quality-manual-guide.md +0 -688
  646. package/bin/skills/iso-13485-certification/scripts/gap_analyzer.py +0 -440
  647. package/bin/skills/kegg-database/SKILL.md +0 -377
  648. package/bin/skills/kegg-database/references/kegg_reference.md +0 -326
  649. package/bin/skills/kegg-database/scripts/kegg_api.py +0 -251
  650. package/bin/skills/knowledge-distillation/SKILL.md +0 -458
  651. package/bin/skills/knowledge-distillation/references/minillm.md +0 -334
  652. package/bin/skills/labarchive-integration/SKILL.md +0 -268
  653. package/bin/skills/labarchive-integration/references/api_reference.md +0 -342
  654. package/bin/skills/labarchive-integration/references/authentication_guide.md +0 -357
  655. package/bin/skills/labarchive-integration/references/integrations.md +0 -425
  656. package/bin/skills/labarchive-integration/scripts/entry_operations.py +0 -334
  657. package/bin/skills/labarchive-integration/scripts/notebook_operations.py +0 -269
  658. package/bin/skills/labarchive-integration/scripts/setup_config.py +0 -205
  659. package/bin/skills/lambda-labs/SKILL.md +0 -545
  660. package/bin/skills/lambda-labs/references/advanced-usage.md +0 -611
  661. package/bin/skills/lambda-labs/references/troubleshooting.md +0 -530
  662. package/bin/skills/lamindb/SKILL.md +0 -390
  663. package/bin/skills/lamindb/references/annotation-validation.md +0 -513
  664. package/bin/skills/lamindb/references/core-concepts.md +0 -380
  665. package/bin/skills/lamindb/references/data-management.md +0 -433
  666. package/bin/skills/lamindb/references/integrations.md +0 -642
  667. package/bin/skills/lamindb/references/ontologies.md +0 -497
  668. package/bin/skills/lamindb/references/setup-deployment.md +0 -733
  669. package/bin/skills/langchain/SKILL.md +0 -480
  670. package/bin/skills/langchain/references/agents.md +0 -499
  671. package/bin/skills/langchain/references/integration.md +0 -562
  672. package/bin/skills/langchain/references/rag.md +0 -600
  673. package/bin/skills/langsmith/SKILL.md +0 -422
  674. package/bin/skills/langsmith/references/advanced-usage.md +0 -548
  675. package/bin/skills/langsmith/references/troubleshooting.md +0 -537
  676. package/bin/skills/latchbio-integration/SKILL.md +0 -353
  677. package/bin/skills/latchbio-integration/references/data-management.md +0 -427
  678. package/bin/skills/latchbio-integration/references/resource-configuration.md +0 -429
  679. package/bin/skills/latchbio-integration/references/verified-workflows.md +0 -487
  680. package/bin/skills/latchbio-integration/references/workflow-creation.md +0 -254
  681. package/bin/skills/latex-posters/README.md +0 -417
  682. package/bin/skills/latex-posters/SKILL.md +0 -1602
  683. package/bin/skills/latex-posters/assets/baposter_template.tex +0 -257
  684. package/bin/skills/latex-posters/assets/beamerposter_template.tex +0 -244
  685. package/bin/skills/latex-posters/assets/poster_quality_checklist.md +0 -358
  686. package/bin/skills/latex-posters/assets/tikzposter_template.tex +0 -251
  687. package/bin/skills/latex-posters/references/latex_poster_packages.md +0 -745
  688. package/bin/skills/latex-posters/references/poster_content_guide.md +0 -748
  689. package/bin/skills/latex-posters/references/poster_design_principles.md +0 -806
  690. package/bin/skills/latex-posters/references/poster_layout_design.md +0 -900
  691. package/bin/skills/latex-posters/scripts/review_poster.sh +0 -214
  692. package/bin/skills/literature-review/SKILL.md +0 -641
  693. package/bin/skills/literature-review/assets/review_template.md +0 -412
  694. package/bin/skills/literature-review/references/citation_styles.md +0 -166
  695. package/bin/skills/literature-review/references/database_strategies.md +0 -455
  696. package/bin/skills/literature-review/scripts/generate_pdf.py +0 -184
  697. package/bin/skills/literature-review/scripts/search_databases.py +0 -310
  698. package/bin/skills/literature-review/scripts/verify_citations.py +0 -218
  699. package/bin/skills/litgpt/SKILL.md +0 -469
  700. package/bin/skills/litgpt/references/custom-models.md +0 -568
  701. package/bin/skills/litgpt/references/distributed-training.md +0 -451
  702. package/bin/skills/litgpt/references/supported-models.md +0 -336
  703. package/bin/skills/litgpt/references/training-recipes.md +0 -619
  704. package/bin/skills/llama-cpp/SKILL.md +0 -258
  705. package/bin/skills/llama-cpp/references/optimization.md +0 -89
  706. package/bin/skills/llama-cpp/references/quantization.md +0 -213
  707. package/bin/skills/llama-cpp/references/server.md +0 -125
  708. package/bin/skills/llama-factory/SKILL.md +0 -80
  709. package/bin/skills/llama-factory/references/_images.md +0 -23
  710. package/bin/skills/llama-factory/references/advanced.md +0 -1055
  711. package/bin/skills/llama-factory/references/getting_started.md +0 -349
  712. package/bin/skills/llama-factory/references/index.md +0 -19
  713. package/bin/skills/llama-factory/references/other.md +0 -31
  714. package/bin/skills/llamaguard/SKILL.md +0 -337
  715. package/bin/skills/llamaindex/SKILL.md +0 -569
  716. package/bin/skills/llamaindex/references/agents.md +0 -83
  717. package/bin/skills/llamaindex/references/data_connectors.md +0 -108
  718. package/bin/skills/llamaindex/references/query_engines.md +0 -406
  719. package/bin/skills/llava/SKILL.md +0 -304
  720. package/bin/skills/llava/references/training.md +0 -197
  721. package/bin/skills/llm-as-judge-evaluation/SKILL.md +0 -385
  722. package/bin/skills/llm-as-judge-evaluation/references/pairwise-comparison.md +0 -95
  723. package/bin/skills/llm-as-judge-evaluation/references/scoring-rubrics.md +0 -169
  724. package/bin/skills/lm-evaluation-harness/SKILL.md +0 -490
  725. package/bin/skills/lm-evaluation-harness/references/api-evaluation.md +0 -490
  726. package/bin/skills/lm-evaluation-harness/references/benchmark-guide.md +0 -488
  727. package/bin/skills/lm-evaluation-harness/references/custom-tasks.md +0 -602
  728. package/bin/skills/lm-evaluation-harness/references/distributed-eval.md +0 -519
  729. package/bin/skills/long-context/SKILL.md +0 -536
  730. package/bin/skills/long-context/references/extension_methods.md +0 -468
  731. package/bin/skills/long-context/references/fine_tuning.md +0 -611
  732. package/bin/skills/long-context/references/rope.md +0 -402
  733. package/bin/skills/mamba/SKILL.md +0 -260
  734. package/bin/skills/mamba/references/architecture-details.md +0 -206
  735. package/bin/skills/mamba/references/benchmarks.md +0 -255
  736. package/bin/skills/mamba/references/training-guide.md +0 -388
  737. package/bin/skills/market-research-reports/SKILL.md +0 -904
  738. package/bin/skills/market-research-reports/assets/FORMATTING_GUIDE.md +0 -428
  739. package/bin/skills/market-research-reports/assets/market_report_template.tex +0 -1380
  740. package/bin/skills/market-research-reports/assets/market_research.sty +0 -564
  741. package/bin/skills/market-research-reports/references/data_analysis_patterns.md +0 -548
  742. package/bin/skills/market-research-reports/references/report_structure_guide.md +0 -999
  743. package/bin/skills/market-research-reports/references/visual_generation_guide.md +0 -1077
  744. package/bin/skills/market-research-reports/scripts/generate_market_visuals.py +0 -472
  745. package/bin/skills/markitdown/INSTALLATION_GUIDE.md +0 -318
  746. package/bin/skills/markitdown/LICENSE.txt +0 -22
  747. package/bin/skills/markitdown/OPENROUTER_INTEGRATION.md +0 -359
  748. package/bin/skills/markitdown/QUICK_REFERENCE.md +0 -309
  749. package/bin/skills/markitdown/README.md +0 -184
  750. package/bin/skills/markitdown/SKILL.md +0 -486
  751. package/bin/skills/markitdown/SKILL_SUMMARY.md +0 -307
  752. package/bin/skills/markitdown/assets/example_usage.md +0 -463
  753. package/bin/skills/markitdown/references/api_reference.md +0 -399
  754. package/bin/skills/markitdown/references/file_formats.md +0 -542
  755. package/bin/skills/markitdown/scripts/batch_convert.py +0 -195
  756. package/bin/skills/markitdown/scripts/convert_literature.py +0 -262
  757. package/bin/skills/markitdown/scripts/convert_with_ai.py +0 -224
  758. package/bin/skills/matchms/SKILL.md +0 -203
  759. package/bin/skills/matchms/references/filtering.md +0 -288
  760. package/bin/skills/matchms/references/importing_exporting.md +0 -416
  761. package/bin/skills/matchms/references/similarity.md +0 -380
  762. package/bin/skills/matchms/references/workflows.md +0 -647
  763. package/bin/skills/matlab/SKILL.md +0 -376
  764. package/bin/skills/matlab/references/data-import-export.md +0 -479
  765. package/bin/skills/matlab/references/executing-scripts.md +0 -444
  766. package/bin/skills/matlab/references/graphics-visualization.md +0 -579
  767. package/bin/skills/matlab/references/mathematics.md +0 -553
  768. package/bin/skills/matlab/references/matrices-arrays.md +0 -349
  769. package/bin/skills/matlab/references/octave-compatibility.md +0 -544
  770. package/bin/skills/matlab/references/programming.md +0 -672
  771. package/bin/skills/matlab/references/python-integration.md +0 -433
  772. package/bin/skills/matplotlib/SKILL.md +0 -361
  773. package/bin/skills/matplotlib/references/api_reference.md +0 -412
  774. package/bin/skills/matplotlib/references/common_issues.md +0 -563
  775. package/bin/skills/matplotlib/references/plot_types.md +0 -476
  776. package/bin/skills/matplotlib/references/styling_guide.md +0 -589
  777. package/bin/skills/matplotlib/scripts/plot_template.py +0 -401
  778. package/bin/skills/matplotlib/scripts/style_configurator.py +0 -409
  779. package/bin/skills/medchem/SKILL.md +0 -406
  780. package/bin/skills/medchem/references/api_guide.md +0 -600
  781. package/bin/skills/medchem/references/rules_catalog.md +0 -604
  782. package/bin/skills/medchem/scripts/filter_molecules.py +0 -418
  783. package/bin/skills/megatron-core/SKILL.md +0 -366
  784. package/bin/skills/megatron-core/references/benchmarks.md +0 -249
  785. package/bin/skills/megatron-core/references/parallelism-guide.md +0 -404
  786. package/bin/skills/megatron-core/references/production-examples.md +0 -473
  787. package/bin/skills/megatron-core/references/training-recipes.md +0 -547
  788. package/bin/skills/metabolomics-workbench-database/SKILL.md +0 -259
  789. package/bin/skills/metabolomics-workbench-database/references/api_reference.md +0 -494
  790. package/bin/skills/miles/SKILL.md +0 -315
  791. package/bin/skills/miles/references/api-reference.md +0 -141
  792. package/bin/skills/miles/references/troubleshooting.md +0 -352
  793. package/bin/skills/ml-paper-writing/SKILL.md +0 -937
  794. package/bin/skills/ml-paper-writing/references/checklists.md +0 -361
  795. package/bin/skills/ml-paper-writing/references/citation-workflow.md +0 -562
  796. package/bin/skills/ml-paper-writing/references/reviewer-guidelines.md +0 -367
  797. package/bin/skills/ml-paper-writing/references/sources.md +0 -159
  798. package/bin/skills/ml-paper-writing/references/writing-guide.md +0 -476
  799. package/bin/skills/ml-paper-writing/templates/README.md +0 -251
  800. package/bin/skills/ml-paper-writing/templates/aaai2026/README.md +0 -534
  801. package/bin/skills/ml-paper-writing/templates/aaai2026/aaai2026-unified-supp.tex +0 -144
  802. package/bin/skills/ml-paper-writing/templates/aaai2026/aaai2026-unified-template.tex +0 -952
  803. package/bin/skills/ml-paper-writing/templates/aaai2026/aaai2026.bib +0 -111
  804. package/bin/skills/ml-paper-writing/templates/aaai2026/aaai2026.bst +0 -1493
  805. package/bin/skills/ml-paper-writing/templates/aaai2026/aaai2026.sty +0 -315
  806. package/bin/skills/ml-paper-writing/templates/acl/README.md +0 -50
  807. package/bin/skills/ml-paper-writing/templates/acl/acl.sty +0 -312
  808. package/bin/skills/ml-paper-writing/templates/acl/acl_latex.tex +0 -377
  809. package/bin/skills/ml-paper-writing/templates/acl/acl_lualatex.tex +0 -101
  810. package/bin/skills/ml-paper-writing/templates/acl/acl_natbib.bst +0 -1940
  811. package/bin/skills/ml-paper-writing/templates/acl/anthology.bib.txt +0 -26
  812. package/bin/skills/ml-paper-writing/templates/acl/custom.bib +0 -70
  813. package/bin/skills/ml-paper-writing/templates/acl/formatting.md +0 -326
  814. package/bin/skills/ml-paper-writing/templates/colm2025/README.md +0 -3
  815. package/bin/skills/ml-paper-writing/templates/colm2025/colm2025_conference.bib +0 -11
  816. package/bin/skills/ml-paper-writing/templates/colm2025/colm2025_conference.bst +0 -1440
  817. package/bin/skills/ml-paper-writing/templates/colm2025/colm2025_conference.pdf +0 -0
  818. package/bin/skills/ml-paper-writing/templates/colm2025/colm2025_conference.sty +0 -218
  819. package/bin/skills/ml-paper-writing/templates/colm2025/colm2025_conference.tex +0 -305
  820. package/bin/skills/ml-paper-writing/templates/colm2025/fancyhdr.sty +0 -485
  821. package/bin/skills/ml-paper-writing/templates/colm2025/math_commands.tex +0 -508
  822. package/bin/skills/ml-paper-writing/templates/colm2025/natbib.sty +0 -1246
  823. package/bin/skills/ml-paper-writing/templates/iclr2026/fancyhdr.sty +0 -485
  824. package/bin/skills/ml-paper-writing/templates/iclr2026/iclr2026_conference.bib +0 -24
  825. package/bin/skills/ml-paper-writing/templates/iclr2026/iclr2026_conference.bst +0 -1440
  826. package/bin/skills/ml-paper-writing/templates/iclr2026/iclr2026_conference.pdf +0 -0
  827. package/bin/skills/ml-paper-writing/templates/iclr2026/iclr2026_conference.sty +0 -246
  828. package/bin/skills/ml-paper-writing/templates/iclr2026/iclr2026_conference.tex +0 -414
  829. package/bin/skills/ml-paper-writing/templates/iclr2026/math_commands.tex +0 -508
  830. package/bin/skills/ml-paper-writing/templates/iclr2026/natbib.sty +0 -1246
  831. package/bin/skills/ml-paper-writing/templates/icml2026/algorithm.sty +0 -79
  832. package/bin/skills/ml-paper-writing/templates/icml2026/algorithmic.sty +0 -201
  833. package/bin/skills/ml-paper-writing/templates/icml2026/example_paper.bib +0 -75
  834. package/bin/skills/ml-paper-writing/templates/icml2026/example_paper.pdf +0 -0
  835. package/bin/skills/ml-paper-writing/templates/icml2026/example_paper.tex +0 -662
  836. package/bin/skills/ml-paper-writing/templates/icml2026/fancyhdr.sty +0 -864
  837. package/bin/skills/ml-paper-writing/templates/icml2026/icml2026.bst +0 -1443
  838. package/bin/skills/ml-paper-writing/templates/icml2026/icml2026.sty +0 -767
  839. package/bin/skills/ml-paper-writing/templates/icml2026/icml_numpapers.pdf +0 -0
  840. package/bin/skills/ml-paper-writing/templates/neurips2025/Makefile +0 -36
  841. package/bin/skills/ml-paper-writing/templates/neurips2025/extra_pkgs.tex +0 -53
  842. package/bin/skills/ml-paper-writing/templates/neurips2025/main.tex +0 -38
  843. package/bin/skills/ml-paper-writing/templates/neurips2025/neurips.sty +0 -382
  844. package/bin/skills/mlflow/SKILL.md +0 -704
  845. package/bin/skills/mlflow/references/deployment.md +0 -744
  846. package/bin/skills/mlflow/references/model-registry.md +0 -770
  847. package/bin/skills/mlflow/references/tracking.md +0 -680
  848. package/bin/skills/modal/SKILL.md +0 -418
  849. package/bin/skills/modal/references/advanced-patterns.md +0 -695
  850. package/bin/skills/modal/references/examples-catalog.md +0 -423
  851. package/bin/skills/modal/references/troubleshooting.md +0 -494
  852. package/bin/skills/modal-research-gpu/SKILL.md +0 -238
  853. package/bin/skills/model-economics/SKILL.md +0 -238
  854. package/bin/skills/model-merging/SKILL.md +0 -539
  855. package/bin/skills/model-merging/references/evaluation.md +0 -462
  856. package/bin/skills/model-merging/references/examples.md +0 -428
  857. package/bin/skills/model-merging/references/methods.md +0 -352
  858. package/bin/skills/model-pruning/SKILL.md +0 -495
  859. package/bin/skills/model-pruning/references/wanda.md +0 -347
  860. package/bin/skills/moe-training/SKILL.md +0 -526
  861. package/bin/skills/moe-training/references/architectures.md +0 -432
  862. package/bin/skills/moe-training/references/inference.md +0 -348
  863. package/bin/skills/moe-training/references/training.md +0 -425
  864. package/bin/skills/molfeat/SKILL.md +0 -511
  865. package/bin/skills/molfeat/references/api_reference.md +0 -428
  866. package/bin/skills/molfeat/references/available_featurizers.md +0 -333
  867. package/bin/skills/molfeat/references/examples.md +0 -723
  868. package/bin/skills/nanogpt/SKILL.md +0 -290
  869. package/bin/skills/nanogpt/references/architecture.md +0 -382
  870. package/bin/skills/nanogpt/references/data.md +0 -476
  871. package/bin/skills/nanogpt/references/training.md +0 -564
  872. package/bin/skills/nemo-curator/SKILL.md +0 -383
  873. package/bin/skills/nemo-curator/references/deduplication.md +0 -87
  874. package/bin/skills/nemo-curator/references/filtering.md +0 -102
  875. package/bin/skills/nemo-evaluator/SKILL.md +0 -494
  876. package/bin/skills/nemo-evaluator/references/adapter-system.md +0 -340
  877. package/bin/skills/nemo-evaluator/references/configuration.md +0 -447
  878. package/bin/skills/nemo-evaluator/references/custom-benchmarks.md +0 -315
  879. package/bin/skills/nemo-evaluator/references/execution-backends.md +0 -361
  880. package/bin/skills/nemo-guardrails/SKILL.md +0 -297
  881. package/bin/skills/networkx/SKILL.md +0 -437
  882. package/bin/skills/networkx/references/algorithms.md +0 -383
  883. package/bin/skills/networkx/references/generators.md +0 -378
  884. package/bin/skills/networkx/references/graph-basics.md +0 -283
  885. package/bin/skills/networkx/references/io.md +0 -441
  886. package/bin/skills/networkx/references/visualization.md +0 -529
  887. package/bin/skills/neurokit2/SKILL.md +0 -356
  888. package/bin/skills/neurokit2/references/bio_module.md +0 -417
  889. package/bin/skills/neurokit2/references/complexity.md +0 -715
  890. package/bin/skills/neurokit2/references/ecg_cardiac.md +0 -355
  891. package/bin/skills/neurokit2/references/eda.md +0 -497
  892. package/bin/skills/neurokit2/references/eeg.md +0 -506
  893. package/bin/skills/neurokit2/references/emg.md +0 -408
  894. package/bin/skills/neurokit2/references/eog.md +0 -407
  895. package/bin/skills/neurokit2/references/epochs_events.md +0 -471
  896. package/bin/skills/neurokit2/references/hrv.md +0 -480
  897. package/bin/skills/neurokit2/references/ppg.md +0 -413
  898. package/bin/skills/neurokit2/references/rsp.md +0 -510
  899. package/bin/skills/neurokit2/references/signal_processing.md +0 -648
  900. package/bin/skills/neuropixels-analysis/SKILL.md +0 -350
  901. package/bin/skills/neuropixels-analysis/assets/analysis_template.py +0 -271
  902. package/bin/skills/neuropixels-analysis/references/AI_CURATION.md +0 -345
  903. package/bin/skills/neuropixels-analysis/references/ANALYSIS.md +0 -392
  904. package/bin/skills/neuropixels-analysis/references/AUTOMATED_CURATION.md +0 -358
  905. package/bin/skills/neuropixels-analysis/references/MOTION_CORRECTION.md +0 -323
  906. package/bin/skills/neuropixels-analysis/references/PREPROCESSING.md +0 -273
  907. package/bin/skills/neuropixels-analysis/references/QUALITY_METRICS.md +0 -359
  908. package/bin/skills/neuropixels-analysis/references/SPIKE_SORTING.md +0 -339
  909. package/bin/skills/neuropixels-analysis/references/api_reference.md +0 -415
  910. package/bin/skills/neuropixels-analysis/references/plotting_guide.md +0 -454
  911. package/bin/skills/neuropixels-analysis/references/standard_workflow.md +0 -385
  912. package/bin/skills/neuropixels-analysis/scripts/compute_metrics.py +0 -178
  913. package/bin/skills/neuropixels-analysis/scripts/explore_recording.py +0 -168
  914. package/bin/skills/neuropixels-analysis/scripts/export_to_phy.py +0 -79
  915. package/bin/skills/neuropixels-analysis/scripts/neuropixels_pipeline.py +0 -432
  916. package/bin/skills/neuropixels-analysis/scripts/preprocess_recording.py +0 -122
  917. package/bin/skills/neuropixels-analysis/scripts/run_sorting.py +0 -98
  918. package/bin/skills/nnsight/SKILL.md +0 -436
  919. package/bin/skills/nnsight/references/README.md +0 -78
  920. package/bin/skills/nnsight/references/api.md +0 -344
  921. package/bin/skills/nnsight/references/tutorials.md +0 -300
  922. package/bin/skills/offer-k-dense-web/SKILL.md +0 -21
  923. package/bin/skills/omero-integration/SKILL.md +0 -251
  924. package/bin/skills/omero-integration/references/advanced.md +0 -631
  925. package/bin/skills/omero-integration/references/connection.md +0 -369
  926. package/bin/skills/omero-integration/references/data_access.md +0 -544
  927. package/bin/skills/omero-integration/references/image_processing.md +0 -665
  928. package/bin/skills/omero-integration/references/metadata.md +0 -688
  929. package/bin/skills/omero-integration/references/rois.md +0 -648
  930. package/bin/skills/omero-integration/references/scripts.md +0 -637
  931. package/bin/skills/omero-integration/references/tables.md +0 -532
  932. package/bin/skills/openalex-database/SKILL.md +0 -494
  933. package/bin/skills/openalex-database/references/api_guide.md +0 -371
  934. package/bin/skills/openalex-database/references/common_queries.md +0 -381
  935. package/bin/skills/openalex-database/scripts/openalex_client.py +0 -337
  936. package/bin/skills/openalex-database/scripts/query_helpers.py +0 -306
  937. package/bin/skills/openrlhf/SKILL.md +0 -249
  938. package/bin/skills/openrlhf/references/algorithm-comparison.md +0 -404
  939. package/bin/skills/openrlhf/references/custom-rewards.md +0 -530
  940. package/bin/skills/openrlhf/references/hybrid-engine.md +0 -287
  941. package/bin/skills/openrlhf/references/multi-node-training.md +0 -454
  942. package/bin/skills/opentargets-database/SKILL.md +0 -373
  943. package/bin/skills/opentargets-database/references/api_reference.md +0 -249
  944. package/bin/skills/opentargets-database/references/evidence_types.md +0 -306
  945. package/bin/skills/opentargets-database/references/target_annotations.md +0 -401
  946. package/bin/skills/opentargets-database/scripts/query_opentargets.py +0 -403
  947. package/bin/skills/opentrons-integration/SKILL.md +0 -573
  948. package/bin/skills/opentrons-integration/references/api_reference.md +0 -366
  949. package/bin/skills/opentrons-integration/scripts/basic_protocol_template.py +0 -67
  950. package/bin/skills/opentrons-integration/scripts/pcr_setup_template.py +0 -154
  951. package/bin/skills/opentrons-integration/scripts/serial_dilution_template.py +0 -96
  952. package/bin/skills/outlines/SKILL.md +0 -652
  953. package/bin/skills/outlines/references/backends.md +0 -615
  954. package/bin/skills/outlines/references/examples.md +0 -773
  955. package/bin/skills/outlines/references/json_generation.md +0 -652
  956. package/bin/skills/paper-2-web/SKILL.md +0 -491
  957. package/bin/skills/paper-2-web/references/installation.md +0 -141
  958. package/bin/skills/paper-2-web/references/paper2poster.md +0 -346
  959. package/bin/skills/paper-2-web/references/paper2video.md +0 -305
  960. package/bin/skills/paper-2-web/references/paper2web.md +0 -187
  961. package/bin/skills/paper-2-web/references/usage_examples.md +0 -436
  962. package/bin/skills/pathml/SKILL.md +0 -166
  963. package/bin/skills/pathml/references/data_management.md +0 -742
  964. package/bin/skills/pathml/references/graphs.md +0 -653
  965. package/bin/skills/pathml/references/image_loading.md +0 -448
  966. package/bin/skills/pathml/references/machine_learning.md +0 -725
  967. package/bin/skills/pathml/references/multiparametric.md +0 -686
  968. package/bin/skills/pathml/references/preprocessing.md +0 -722
  969. package/bin/skills/pdb-database/SKILL.md +0 -309
  970. package/bin/skills/pdb-database/references/api_reference.md +0 -617
  971. package/bin/skills/peer-review/SKILL.md +0 -702
  972. package/bin/skills/peer-review/references/calibration_guidelines.md +0 -196
  973. package/bin/skills/peer-review/references/common_issues.md +0 -552
  974. package/bin/skills/peer-review/references/paper_mechanics.md +0 -269
  975. package/bin/skills/peer-review/references/reporting_standards.md +0 -290
  976. package/bin/skills/peer-review/references/scoring_rubric.md +0 -239
  977. package/bin/skills/peft/SKILL.md +0 -431
  978. package/bin/skills/peft/references/advanced-usage.md +0 -514
  979. package/bin/skills/peft/references/troubleshooting.md +0 -480
  980. package/bin/skills/pennylane/SKILL.md +0 -226
  981. package/bin/skills/pennylane/references/advanced_features.md +0 -667
  982. package/bin/skills/pennylane/references/devices_backends.md +0 -596
  983. package/bin/skills/pennylane/references/getting_started.md +0 -227
  984. package/bin/skills/pennylane/references/optimization.md +0 -671
  985. package/bin/skills/pennylane/references/quantum_chemistry.md +0 -567
  986. package/bin/skills/pennylane/references/quantum_circuits.md +0 -437
  987. package/bin/skills/pennylane/references/quantum_ml.md +0 -571
  988. package/bin/skills/perplexity-search/SKILL.md +0 -448
  989. package/bin/skills/perplexity-search/assets/.env.example +0 -16
  990. package/bin/skills/perplexity-search/references/model_comparison.md +0 -386
  991. package/bin/skills/perplexity-search/references/openrouter_setup.md +0 -454
  992. package/bin/skills/perplexity-search/references/search_strategies.md +0 -258
  993. package/bin/skills/perplexity-search/scripts/perplexity_search.py +0 -277
  994. package/bin/skills/perplexity-search/scripts/setup_env.py +0 -171
  995. package/bin/skills/phoenix/SKILL.md +0 -475
  996. package/bin/skills/phoenix/references/advanced-usage.md +0 -619
  997. package/bin/skills/phoenix/references/troubleshooting.md +0 -538
  998. package/bin/skills/pinecone/SKILL.md +0 -358
  999. package/bin/skills/pinecone/references/deployment.md +0 -181
  1000. package/bin/skills/plotly/SKILL.md +0 -267
  1001. package/bin/skills/plotly/references/chart-types.md +0 -488
  1002. package/bin/skills/plotly/references/export-interactivity.md +0 -453
  1003. package/bin/skills/plotly/references/graph-objects.md +0 -302
  1004. package/bin/skills/plotly/references/layouts-styling.md +0 -457
  1005. package/bin/skills/plotly/references/plotly-express.md +0 -213
  1006. package/bin/skills/polars/SKILL.md +0 -387
  1007. package/bin/skills/polars/references/best_practices.md +0 -649
  1008. package/bin/skills/polars/references/core_concepts.md +0 -378
  1009. package/bin/skills/polars/references/io_guide.md +0 -557
  1010. package/bin/skills/polars/references/operations.md +0 -602
  1011. package/bin/skills/polars/references/pandas_migration.md +0 -417
  1012. package/bin/skills/polars/references/transformations.md +0 -549
  1013. package/bin/skills/pptx-posters/SKILL.md +0 -410
  1014. package/bin/skills/pptx-posters/assets/poster_html_template.html +0 -257
  1015. package/bin/skills/pptx-posters/assets/poster_quality_checklist.md +0 -358
  1016. package/bin/skills/pptx-posters/references/poster_content_guide.md +0 -748
  1017. package/bin/skills/pptx-posters/references/poster_design_principles.md +0 -806
  1018. package/bin/skills/pptx-posters/references/poster_layout_design.md +0 -900
  1019. package/bin/skills/prime-intellect-lab/README.md +0 -69
  1020. package/bin/skills/prime-intellect-lab/SKILL.md +0 -598
  1021. package/bin/skills/prime-intellect-lab/templates/basic_rl_training.toml +0 -82
  1022. package/bin/skills/protocolsio-integration/SKILL.md +0 -421
  1023. package/bin/skills/protocolsio-integration/references/additional_features.md +0 -387
  1024. package/bin/skills/protocolsio-integration/references/authentication.md +0 -100
  1025. package/bin/skills/protocolsio-integration/references/discussions.md +0 -225
  1026. package/bin/skills/protocolsio-integration/references/file_manager.md +0 -412
  1027. package/bin/skills/protocolsio-integration/references/protocols_api.md +0 -294
  1028. package/bin/skills/protocolsio-integration/references/workspaces.md +0 -293
  1029. package/bin/skills/pubchem-database/SKILL.md +0 -574
  1030. package/bin/skills/pubchem-database/references/api_reference.md +0 -440
  1031. package/bin/skills/pubchem-database/scripts/bioactivity_query.py +0 -367
  1032. package/bin/skills/pubchem-database/scripts/compound_search.py +0 -297
  1033. package/bin/skills/pubmed-database/SKILL.md +0 -460
  1034. package/bin/skills/pubmed-database/references/api_reference.md +0 -298
  1035. package/bin/skills/pubmed-database/references/common_queries.md +0 -453
  1036. package/bin/skills/pubmed-database/references/search_syntax.md +0 -436
  1037. package/bin/skills/pufferlib/SKILL.md +0 -436
  1038. package/bin/skills/pufferlib/references/environments.md +0 -508
  1039. package/bin/skills/pufferlib/references/integration.md +0 -621
  1040. package/bin/skills/pufferlib/references/policies.md +0 -653
  1041. package/bin/skills/pufferlib/references/training.md +0 -360
  1042. package/bin/skills/pufferlib/references/vectorization.md +0 -557
  1043. package/bin/skills/pufferlib/scripts/env_template.py +0 -340
  1044. package/bin/skills/pufferlib/scripts/train_template.py +0 -239
  1045. package/bin/skills/pydeseq2/SKILL.md +0 -559
  1046. package/bin/skills/pydeseq2/references/api_reference.md +0 -228
  1047. package/bin/skills/pydeseq2/references/workflow_guide.md +0 -582
  1048. package/bin/skills/pydeseq2/scripts/run_deseq2_analysis.py +0 -353
  1049. package/bin/skills/pydicom/SKILL.md +0 -434
  1050. package/bin/skills/pydicom/references/common_tags.md +0 -228
  1051. package/bin/skills/pydicom/references/transfer_syntaxes.md +0 -352
  1052. package/bin/skills/pydicom/scripts/anonymize_dicom.py +0 -137
  1053. package/bin/skills/pydicom/scripts/dicom_to_image.py +0 -172
  1054. package/bin/skills/pydicom/scripts/extract_metadata.py +0 -173
  1055. package/bin/skills/pyhealth/SKILL.md +0 -491
  1056. package/bin/skills/pyhealth/references/datasets.md +0 -178
  1057. package/bin/skills/pyhealth/references/medical_coding.md +0 -284
  1058. package/bin/skills/pyhealth/references/models.md +0 -594
  1059. package/bin/skills/pyhealth/references/preprocessing.md +0 -638
  1060. package/bin/skills/pyhealth/references/tasks.md +0 -379
  1061. package/bin/skills/pyhealth/references/training_evaluation.md +0 -648
  1062. package/bin/skills/pylabrobot/SKILL.md +0 -185
  1063. package/bin/skills/pylabrobot/references/analytical-equipment.md +0 -464
  1064. package/bin/skills/pylabrobot/references/hardware-backends.md +0 -480
  1065. package/bin/skills/pylabrobot/references/liquid-handling.md +0 -403
  1066. package/bin/skills/pylabrobot/references/material-handling.md +0 -620
  1067. package/bin/skills/pylabrobot/references/resources.md +0 -489
  1068. package/bin/skills/pylabrobot/references/visualization.md +0 -532
  1069. package/bin/skills/pymatgen/SKILL.md +0 -691
  1070. package/bin/skills/pymatgen/references/analysis_modules.md +0 -530
  1071. package/bin/skills/pymatgen/references/core_classes.md +0 -318
  1072. package/bin/skills/pymatgen/references/io_formats.md +0 -469
  1073. package/bin/skills/pymatgen/references/materials_project_api.md +0 -517
  1074. package/bin/skills/pymatgen/references/transformations_workflows.md +0 -591
  1075. package/bin/skills/pymatgen/scripts/phase_diagram_generator.py +0 -233
  1076. package/bin/skills/pymatgen/scripts/structure_analyzer.py +0 -266
  1077. package/bin/skills/pymatgen/scripts/structure_converter.py +0 -169
  1078. package/bin/skills/pymc/SKILL.md +0 -572
  1079. package/bin/skills/pymc/assets/hierarchical_model_template.py +0 -333
  1080. package/bin/skills/pymc/assets/linear_regression_template.py +0 -241
  1081. package/bin/skills/pymc/references/distributions.md +0 -320
  1082. package/bin/skills/pymc/references/sampling_inference.md +0 -424
  1083. package/bin/skills/pymc/references/workflows.md +0 -526
  1084. package/bin/skills/pymc/scripts/model_comparison.py +0 -387
  1085. package/bin/skills/pymc/scripts/model_diagnostics.py +0 -350
  1086. package/bin/skills/pymoo/SKILL.md +0 -571
  1087. package/bin/skills/pymoo/references/algorithms.md +0 -180
  1088. package/bin/skills/pymoo/references/constraints_mcdm.md +0 -417
  1089. package/bin/skills/pymoo/references/operators.md +0 -345
  1090. package/bin/skills/pymoo/references/problems.md +0 -265
  1091. package/bin/skills/pymoo/references/visualization.md +0 -353
  1092. package/bin/skills/pymoo/scripts/custom_problem_example.py +0 -181
  1093. package/bin/skills/pymoo/scripts/decision_making_example.py +0 -161
  1094. package/bin/skills/pymoo/scripts/many_objective_example.py +0 -72
  1095. package/bin/skills/pymoo/scripts/multi_objective_example.py +0 -63
  1096. package/bin/skills/pymoo/scripts/single_objective_example.py +0 -59
  1097. package/bin/skills/pyopenms/SKILL.md +0 -217
  1098. package/bin/skills/pyopenms/references/data_structures.md +0 -497
  1099. package/bin/skills/pyopenms/references/feature_detection.md +0 -410
  1100. package/bin/skills/pyopenms/references/file_io.md +0 -349
  1101. package/bin/skills/pyopenms/references/identification.md +0 -422
  1102. package/bin/skills/pyopenms/references/metabolomics.md +0 -482
  1103. package/bin/skills/pyopenms/references/signal_processing.md +0 -433
  1104. package/bin/skills/pysam/SKILL.md +0 -265
  1105. package/bin/skills/pysam/references/alignment_files.md +0 -280
  1106. package/bin/skills/pysam/references/common_workflows.md +0 -520
  1107. package/bin/skills/pysam/references/sequence_files.md +0 -407
  1108. package/bin/skills/pysam/references/variant_files.md +0 -365
  1109. package/bin/skills/pytdc/SKILL.md +0 -460
  1110. package/bin/skills/pytdc/references/datasets.md +0 -246
  1111. package/bin/skills/pytdc/references/oracles.md +0 -400
  1112. package/bin/skills/pytdc/references/utilities.md +0 -684
  1113. package/bin/skills/pytdc/scripts/benchmark_evaluation.py +0 -327
  1114. package/bin/skills/pytdc/scripts/load_and_split_data.py +0 -214
  1115. package/bin/skills/pytdc/scripts/molecular_generation.py +0 -404
  1116. package/bin/skills/pytorch-fsdp/SKILL.md +0 -126
  1117. package/bin/skills/pytorch-fsdp/references/index.md +0 -7
  1118. package/bin/skills/pytorch-fsdp/references/other.md +0 -4249
  1119. package/bin/skills/pytorch-lightning/SKILL.md +0 -346
  1120. package/bin/skills/pytorch-lightning/references/callbacks.md +0 -436
  1121. package/bin/skills/pytorch-lightning/references/distributed.md +0 -490
  1122. package/bin/skills/pytorch-lightning/references/hyperparameter-tuning.md +0 -556
  1123. package/bin/skills/pyvene/SKILL.md +0 -473
  1124. package/bin/skills/pyvene/references/README.md +0 -73
  1125. package/bin/skills/pyvene/references/api.md +0 -383
  1126. package/bin/skills/pyvene/references/tutorials.md +0 -376
  1127. package/bin/skills/qdrant/SKILL.md +0 -493
  1128. package/bin/skills/qdrant/references/advanced-usage.md +0 -648
  1129. package/bin/skills/qdrant/references/troubleshooting.md +0 -631
  1130. package/bin/skills/qiskit/SKILL.md +0 -275
  1131. package/bin/skills/qiskit/references/algorithms.md +0 -607
  1132. package/bin/skills/qiskit/references/backends.md +0 -433
  1133. package/bin/skills/qiskit/references/circuits.md +0 -197
  1134. package/bin/skills/qiskit/references/patterns.md +0 -533
  1135. package/bin/skills/qiskit/references/primitives.md +0 -277
  1136. package/bin/skills/qiskit/references/setup.md +0 -99
  1137. package/bin/skills/qiskit/references/transpilation.md +0 -286
  1138. package/bin/skills/qiskit/references/visualization.md +0 -415
  1139. package/bin/skills/qutip/SKILL.md +0 -318
  1140. package/bin/skills/qutip/references/advanced.md +0 -555
  1141. package/bin/skills/qutip/references/analysis.md +0 -523
  1142. package/bin/skills/qutip/references/core_concepts.md +0 -293
  1143. package/bin/skills/qutip/references/time_evolution.md +0 -348
  1144. package/bin/skills/qutip/references/visualization.md +0 -431
  1145. package/bin/skills/ray-data/SKILL.md +0 -326
  1146. package/bin/skills/ray-data/references/integration.md +0 -82
  1147. package/bin/skills/ray-data/references/transformations.md +0 -83
  1148. package/bin/skills/ray-train/SKILL.md +0 -406
  1149. package/bin/skills/ray-train/references/multi-node.md +0 -628
  1150. package/bin/skills/rdkit/SKILL.md +0 -780
  1151. package/bin/skills/rdkit/references/api_reference.md +0 -432
  1152. package/bin/skills/rdkit/references/descriptors_reference.md +0 -595
  1153. package/bin/skills/rdkit/references/smarts_patterns.md +0 -668
  1154. package/bin/skills/rdkit/scripts/molecular_properties.py +0 -243
  1155. package/bin/skills/rdkit/scripts/similarity_search.py +0 -297
  1156. package/bin/skills/rdkit/scripts/substructure_filter.py +0 -386
  1157. package/bin/skills/reactome-database/SKILL.md +0 -278
  1158. package/bin/skills/reactome-database/references/api_reference.md +0 -465
  1159. package/bin/skills/reactome-database/scripts/reactome_query.py +0 -286
  1160. package/bin/skills/research-grants/README.md +0 -285
  1161. package/bin/skills/research-grants/SKILL.md +0 -938
  1162. package/bin/skills/research-grants/assets/budget_justification_template.md +0 -453
  1163. package/bin/skills/research-grants/assets/nih_specific_aims_template.md +0 -166
  1164. package/bin/skills/research-grants/assets/nsf_project_summary_template.md +0 -92
  1165. package/bin/skills/research-grants/references/broader_impacts.md +0 -392
  1166. package/bin/skills/research-grants/references/darpa_guidelines.md +0 -636
  1167. package/bin/skills/research-grants/references/doe_guidelines.md +0 -586
  1168. package/bin/skills/research-grants/references/nih_guidelines.md +0 -851
  1169. package/bin/skills/research-grants/references/nsf_guidelines.md +0 -570
  1170. package/bin/skills/research-grants/references/specific_aims_guide.md +0 -458
  1171. package/bin/skills/research-lookup/README.md +0 -156
  1172. package/bin/skills/research-lookup/SKILL.md +0 -606
  1173. package/bin/skills/research-lookup/examples.py +0 -174
  1174. package/bin/skills/research-lookup/lookup.py +0 -187
  1175. package/bin/skills/research-lookup/research_lookup.py +0 -483
  1176. package/bin/skills/research-lookup/scripts/research_lookup.py +0 -483
  1177. package/bin/skills/rowan/SKILL.md +0 -427
  1178. package/bin/skills/rowan/references/api_reference.md +0 -413
  1179. package/bin/skills/rowan/references/molecule_handling.md +0 -429
  1180. package/bin/skills/rowan/references/proteins_and_organization.md +0 -499
  1181. package/bin/skills/rowan/references/rdkit_native.md +0 -438
  1182. package/bin/skills/rowan/references/results_interpretation.md +0 -481
  1183. package/bin/skills/rowan/references/workflow_types.md +0 -591
  1184. package/bin/skills/rwkv/SKILL.md +0 -260
  1185. package/bin/skills/rwkv/references/architecture-details.md +0 -344
  1186. package/bin/skills/rwkv/references/rwkv7.md +0 -386
  1187. package/bin/skills/rwkv/references/state-management.md +0 -369
  1188. package/bin/skills/saelens/SKILL.md +0 -386
  1189. package/bin/skills/saelens/references/README.md +0 -70
  1190. package/bin/skills/saelens/references/api.md +0 -333
  1191. package/bin/skills/saelens/references/tutorials.md +0 -318
  1192. package/bin/skills/scanpy/SKILL.md +0 -386
  1193. package/bin/skills/scanpy/assets/analysis_template.py +0 -295
  1194. package/bin/skills/scanpy/references/api_reference.md +0 -251
  1195. package/bin/skills/scanpy/references/plotting_guide.md +0 -352
  1196. package/bin/skills/scanpy/references/standard_workflow.md +0 -206
  1197. package/bin/skills/scanpy/scripts/qc_analysis.py +0 -200
  1198. package/bin/skills/scholar-evaluation/SKILL.md +0 -289
  1199. package/bin/skills/scholar-evaluation/references/evaluation_framework.md +0 -663
  1200. package/bin/skills/scholar-evaluation/scripts/calculate_scores.py +0 -366
  1201. package/bin/skills/scientific-brainstorming/SKILL.md +0 -191
  1202. package/bin/skills/scientific-brainstorming/references/brainstorming_methods.md +0 -326
  1203. package/bin/skills/scientific-critical-thinking/SKILL.md +0 -566
  1204. package/bin/skills/scientific-critical-thinking/references/common_biases.md +0 -364
  1205. package/bin/skills/scientific-critical-thinking/references/evidence_hierarchy.md +0 -484
  1206. package/bin/skills/scientific-critical-thinking/references/experimental_design.md +0 -496
  1207. package/bin/skills/scientific-critical-thinking/references/logical_fallacies.md +0 -478
  1208. package/bin/skills/scientific-critical-thinking/references/scientific_method.md +0 -169
  1209. package/bin/skills/scientific-critical-thinking/references/statistical_pitfalls.md +0 -506
  1210. package/bin/skills/scientific-schematics/QUICK_REFERENCE.md +0 -207
  1211. package/bin/skills/scientific-schematics/README.md +0 -327
  1212. package/bin/skills/scientific-schematics/SKILL.md +0 -615
  1213. package/bin/skills/scientific-schematics/example_usage.sh +0 -89
  1214. package/bin/skills/scientific-schematics/references/best_practices.md +0 -559
  1215. package/bin/skills/scientific-schematics/scripts/generate_schematic.py +0 -135
  1216. package/bin/skills/scientific-schematics/scripts/generate_schematic_ai.py +0 -837
  1217. package/bin/skills/scientific-schematics/test_ai_generation.py +0 -243
  1218. package/bin/skills/scientific-slides/SKILL.md +0 -942
  1219. package/bin/skills/scientific-slides/assets/timing_guidelines.md +0 -597
  1220. package/bin/skills/scientific-slides/references/data_visualization_slides.md +0 -708
  1221. package/bin/skills/scientific-slides/references/presentation_structure.md +0 -642
  1222. package/bin/skills/scientific-slides/references/slide_design_principles.md +0 -849
  1223. package/bin/skills/scientific-slides/references/talk_types_guide.md +0 -687
  1224. package/bin/skills/scientific-slides/references/visual_review_workflow.md +0 -775
  1225. package/bin/skills/scientific-slides/scripts/generate_slide_image.py +0 -143
  1226. package/bin/skills/scientific-slides/scripts/generate_slide_image_ai.py +0 -748
  1227. package/bin/skills/scientific-slides/scripts/pdf_to_images.py +0 -201
  1228. package/bin/skills/scientific-slides/scripts/slides_to_pdf.py +0 -220
  1229. package/bin/skills/scientific-slides/scripts/validate_presentation.py +0 -367
  1230. package/bin/skills/scientific-visualization/SKILL.md +0 -779
  1231. package/bin/skills/scientific-visualization/assets/color_palettes.py +0 -197
  1232. package/bin/skills/scientific-visualization/assets/nature.mplstyle +0 -63
  1233. package/bin/skills/scientific-visualization/assets/presentation.mplstyle +0 -61
  1234. package/bin/skills/scientific-visualization/assets/publication.mplstyle +0 -68
  1235. package/bin/skills/scientific-visualization/references/color_palettes.md +0 -348
  1236. package/bin/skills/scientific-visualization/references/journal_requirements.md +0 -320
  1237. package/bin/skills/scientific-visualization/references/matplotlib_examples.md +0 -620
  1238. package/bin/skills/scientific-visualization/references/publication_guidelines.md +0 -205
  1239. package/bin/skills/scientific-visualization/scripts/figure_export.py +0 -343
  1240. package/bin/skills/scientific-visualization/scripts/style_presets.py +0 -416
  1241. package/bin/skills/scientific-writing/SKILL.md +0 -714
  1242. package/bin/skills/scientific-writing/assets/REPORT_FORMATTING_GUIDE.md +0 -574
  1243. package/bin/skills/scientific-writing/assets/scientific_report.sty +0 -606
  1244. package/bin/skills/scientific-writing/assets/scientific_report_template.tex +0 -449
  1245. package/bin/skills/scientific-writing/references/citation_styles.md +0 -720
  1246. package/bin/skills/scientific-writing/references/figures_tables.md +0 -806
  1247. package/bin/skills/scientific-writing/references/imrad_structure.md +0 -686
  1248. package/bin/skills/scientific-writing/references/professional_report_formatting.md +0 -664
  1249. package/bin/skills/scientific-writing/references/reporting_guidelines.md +0 -748
  1250. package/bin/skills/scientific-writing/references/writing_principles.md +0 -824
  1251. package/bin/skills/scikit-bio/SKILL.md +0 -437
  1252. package/bin/skills/scikit-bio/references/api_reference.md +0 -749
  1253. package/bin/skills/scikit-learn/SKILL.md +0 -521
  1254. package/bin/skills/scikit-learn/references/model_evaluation.md +0 -592
  1255. package/bin/skills/scikit-learn/references/pipelines_and_composition.md +0 -612
  1256. package/bin/skills/scikit-learn/references/preprocessing.md +0 -606
  1257. package/bin/skills/scikit-learn/references/quick_reference.md +0 -433
  1258. package/bin/skills/scikit-learn/references/supervised_learning.md +0 -378
  1259. package/bin/skills/scikit-learn/references/unsupervised_learning.md +0 -505
  1260. package/bin/skills/scikit-learn/scripts/classification_pipeline.py +0 -257
  1261. package/bin/skills/scikit-learn/scripts/clustering_analysis.py +0 -386
  1262. package/bin/skills/scikit-survival/SKILL.md +0 -399
  1263. package/bin/skills/scikit-survival/references/competing-risks.md +0 -397
  1264. package/bin/skills/scikit-survival/references/cox-models.md +0 -182
  1265. package/bin/skills/scikit-survival/references/data-handling.md +0 -494
  1266. package/bin/skills/scikit-survival/references/ensemble-models.md +0 -327
  1267. package/bin/skills/scikit-survival/references/evaluation-metrics.md +0 -378
  1268. package/bin/skills/scikit-survival/references/svm-models.md +0 -411
  1269. package/bin/skills/scvi-tools/SKILL.md +0 -190
  1270. package/bin/skills/scvi-tools/references/differential-expression.md +0 -581
  1271. package/bin/skills/scvi-tools/references/models-atac-seq.md +0 -321
  1272. package/bin/skills/scvi-tools/references/models-multimodal.md +0 -367
  1273. package/bin/skills/scvi-tools/references/models-scrna-seq.md +0 -330
  1274. package/bin/skills/scvi-tools/references/models-spatial.md +0 -438
  1275. package/bin/skills/scvi-tools/references/models-specialized.md +0 -408
  1276. package/bin/skills/scvi-tools/references/theoretical-foundations.md +0 -438
  1277. package/bin/skills/scvi-tools/references/workflows.md +0 -546
  1278. package/bin/skills/seaborn/SKILL.md +0 -673
  1279. package/bin/skills/seaborn/references/examples.md +0 -822
  1280. package/bin/skills/seaborn/references/function_reference.md +0 -770
  1281. package/bin/skills/seaborn/references/objects_interface.md +0 -964
  1282. package/bin/skills/segment-anything/SKILL.md +0 -500
  1283. package/bin/skills/segment-anything/references/advanced-usage.md +0 -589
  1284. package/bin/skills/segment-anything/references/troubleshooting.md +0 -484
  1285. package/bin/skills/sentence-transformers/SKILL.md +0 -255
  1286. package/bin/skills/sentence-transformers/references/models.md +0 -123
  1287. package/bin/skills/sentencepiece/SKILL.md +0 -235
  1288. package/bin/skills/sentencepiece/references/algorithms.md +0 -200
  1289. package/bin/skills/sentencepiece/references/training.md +0 -304
  1290. package/bin/skills/sglang/SKILL.md +0 -442
  1291. package/bin/skills/sglang/references/deployment.md +0 -490
  1292. package/bin/skills/sglang/references/radix-attention.md +0 -413
  1293. package/bin/skills/sglang/references/structured-generation.md +0 -541
  1294. package/bin/skills/shap/SKILL.md +0 -566
  1295. package/bin/skills/shap/references/explainers.md +0 -339
  1296. package/bin/skills/shap/references/plots.md +0 -507
  1297. package/bin/skills/shap/references/theory.md +0 -449
  1298. package/bin/skills/shap/references/workflows.md +0 -605
  1299. package/bin/skills/simpo/SKILL.md +0 -219
  1300. package/bin/skills/simpo/references/datasets.md +0 -478
  1301. package/bin/skills/simpo/references/hyperparameters.md +0 -452
  1302. package/bin/skills/simpo/references/loss-functions.md +0 -350
  1303. package/bin/skills/simpy/SKILL.md +0 -429
  1304. package/bin/skills/simpy/references/events.md +0 -374
  1305. package/bin/skills/simpy/references/monitoring.md +0 -475
  1306. package/bin/skills/simpy/references/process-interaction.md +0 -424
  1307. package/bin/skills/simpy/references/real-time.md +0 -395
  1308. package/bin/skills/simpy/references/resources.md +0 -275
  1309. package/bin/skills/simpy/scripts/basic_simulation_template.py +0 -193
  1310. package/bin/skills/simpy/scripts/resource_monitor.py +0 -345
  1311. package/bin/skills/skypilot/SKILL.md +0 -509
  1312. package/bin/skills/skypilot/references/advanced-usage.md +0 -491
  1313. package/bin/skills/skypilot/references/troubleshooting.md +0 -570
  1314. package/bin/skills/slime/SKILL.md +0 -464
  1315. package/bin/skills/slime/references/api-reference.md +0 -392
  1316. package/bin/skills/slime/references/troubleshooting.md +0 -386
  1317. package/bin/skills/speculative-decoding/SKILL.md +0 -467
  1318. package/bin/skills/speculative-decoding/references/lookahead.md +0 -309
  1319. package/bin/skills/speculative-decoding/references/medusa.md +0 -350
  1320. package/bin/skills/stable-baselines3/SKILL.md +0 -299
  1321. package/bin/skills/stable-baselines3/references/algorithms.md +0 -333
  1322. package/bin/skills/stable-baselines3/references/callbacks.md +0 -556
  1323. package/bin/skills/stable-baselines3/references/custom_environments.md +0 -526
  1324. package/bin/skills/stable-baselines3/references/vectorized_envs.md +0 -568
  1325. package/bin/skills/stable-baselines3/scripts/custom_env_template.py +0 -314
  1326. package/bin/skills/stable-baselines3/scripts/evaluate_agent.py +0 -245
  1327. package/bin/skills/stable-baselines3/scripts/train_rl_agent.py +0 -165
  1328. package/bin/skills/stable-diffusion/SKILL.md +0 -519
  1329. package/bin/skills/stable-diffusion/references/advanced-usage.md +0 -716
  1330. package/bin/skills/stable-diffusion/references/troubleshooting.md +0 -555
  1331. package/bin/skills/statistical-analysis/SKILL.md +0 -632
  1332. package/bin/skills/statistical-analysis/references/assumptions_and_diagnostics.md +0 -369
  1333. package/bin/skills/statistical-analysis/references/bayesian_statistics.md +0 -661
  1334. package/bin/skills/statistical-analysis/references/effect_sizes_and_power.md +0 -581
  1335. package/bin/skills/statistical-analysis/references/reporting_standards.md +0 -469
  1336. package/bin/skills/statistical-analysis/references/test_selection_guide.md +0 -129
  1337. package/bin/skills/statistical-analysis/scripts/assumption_checks.py +0 -539
  1338. package/bin/skills/statsmodels/SKILL.md +0 -614
  1339. package/bin/skills/statsmodels/references/discrete_choice.md +0 -669
  1340. package/bin/skills/statsmodels/references/glm.md +0 -619
  1341. package/bin/skills/statsmodels/references/linear_models.md +0 -447
  1342. package/bin/skills/statsmodels/references/stats_diagnostics.md +0 -859
  1343. package/bin/skills/statsmodels/references/time_series.md +0 -716
  1344. package/bin/skills/string-database/SKILL.md +0 -534
  1345. package/bin/skills/string-database/references/string_reference.md +0 -455
  1346. package/bin/skills/string-database/scripts/string_api.py +0 -369
  1347. package/bin/skills/sympy/SKILL.md +0 -500
  1348. package/bin/skills/sympy/references/advanced-topics.md +0 -635
  1349. package/bin/skills/sympy/references/code-generation-printing.md +0 -599
  1350. package/bin/skills/sympy/references/core-capabilities.md +0 -348
  1351. package/bin/skills/sympy/references/matrices-linear-algebra.md +0 -526
  1352. package/bin/skills/sympy/references/physics-mechanics.md +0 -592
  1353. package/bin/skills/tensorboard/SKILL.md +0 -629
  1354. package/bin/skills/tensorboard/references/integrations.md +0 -638
  1355. package/bin/skills/tensorboard/references/profiling.md +0 -545
  1356. package/bin/skills/tensorboard/references/visualization.md +0 -620
  1357. package/bin/skills/tensorpool/SKILL.md +0 -519
  1358. package/bin/skills/tensorrt-llm/SKILL.md +0 -187
  1359. package/bin/skills/tensorrt-llm/references/multi-gpu.md +0 -298
  1360. package/bin/skills/tensorrt-llm/references/optimization.md +0 -242
  1361. package/bin/skills/tensorrt-llm/references/serving.md +0 -470
  1362. package/bin/skills/tinker/SKILL.md +0 -466
  1363. package/bin/skills/tinker/references/api-reference.md +0 -168
  1364. package/bin/skills/tinker/references/dpo-and-preference.md +0 -174
  1365. package/bin/skills/tinker/references/evaluations.md +0 -183
  1366. package/bin/skills/tinker/references/getting-started.md +0 -157
  1367. package/bin/skills/tinker/references/loss-functions.md +0 -163
  1368. package/bin/skills/tinker/references/models-and-lora.md +0 -148
  1369. package/bin/skills/tinker/references/recipes.md +0 -326
  1370. package/bin/skills/tinker/references/reinforcement-learning.md +0 -357
  1371. package/bin/skills/tinker/references/rendering.md +0 -255
  1372. package/bin/skills/tinker/references/supervised-learning.md +0 -256
  1373. package/bin/skills/tinker-training-cost/SKILL.md +0 -187
  1374. package/bin/skills/tinker-training-cost/scripts/calculate_cost.py +0 -123
  1375. package/bin/skills/together-ai/SKILL.md +0 -722
  1376. package/bin/skills/torch_geometric/SKILL.md +0 -676
  1377. package/bin/skills/torch_geometric/references/datasets_reference.md +0 -574
  1378. package/bin/skills/torch_geometric/references/layers_reference.md +0 -485
  1379. package/bin/skills/torch_geometric/references/transforms_reference.md +0 -679
  1380. package/bin/skills/torch_geometric/scripts/benchmark_model.py +0 -309
  1381. package/bin/skills/torch_geometric/scripts/create_gnn_template.py +0 -529
  1382. package/bin/skills/torch_geometric/scripts/visualize_graph.py +0 -313
  1383. package/bin/skills/torchdrug/SKILL.md +0 -450
  1384. package/bin/skills/torchdrug/references/core_concepts.md +0 -565
  1385. package/bin/skills/torchdrug/references/datasets.md +0 -380
  1386. package/bin/skills/torchdrug/references/knowledge_graphs.md +0 -320
  1387. package/bin/skills/torchdrug/references/models_architectures.md +0 -541
  1388. package/bin/skills/torchdrug/references/molecular_generation.md +0 -352
  1389. package/bin/skills/torchdrug/references/molecular_property_prediction.md +0 -169
  1390. package/bin/skills/torchdrug/references/protein_modeling.md +0 -272
  1391. package/bin/skills/torchdrug/references/retrosynthesis.md +0 -436
  1392. package/bin/skills/torchforge/SKILL.md +0 -433
  1393. package/bin/skills/torchforge/references/api-reference.md +0 -327
  1394. package/bin/skills/torchforge/references/troubleshooting.md +0 -409
  1395. package/bin/skills/torchtitan/SKILL.md +0 -358
  1396. package/bin/skills/torchtitan/references/checkpoint.md +0 -181
  1397. package/bin/skills/torchtitan/references/custom-models.md +0 -258
  1398. package/bin/skills/torchtitan/references/float8.md +0 -133
  1399. package/bin/skills/torchtitan/references/fsdp.md +0 -126
  1400. package/bin/skills/training-data-pipeline/SKILL.md +0 -427
  1401. package/bin/skills/training-data-pipeline/references/data-quality.md +0 -136
  1402. package/bin/skills/training-data-pipeline/references/frontier-distillation.md +0 -129
  1403. package/bin/skills/training-data-pipeline/references/production-data-formatting.md +0 -126
  1404. package/bin/skills/transformer-lens/SKILL.md +0 -346
  1405. package/bin/skills/transformer-lens/references/README.md +0 -54
  1406. package/bin/skills/transformer-lens/references/api.md +0 -362
  1407. package/bin/skills/transformer-lens/references/tutorials.md +0 -339
  1408. package/bin/skills/transformers/SKILL.md +0 -164
  1409. package/bin/skills/transformers/references/generation.md +0 -467
  1410. package/bin/skills/transformers/references/models.md +0 -361
  1411. package/bin/skills/transformers/references/pipelines.md +0 -335
  1412. package/bin/skills/transformers/references/tokenizers.md +0 -447
  1413. package/bin/skills/transformers/references/training.md +0 -500
  1414. package/bin/skills/treatment-plans/README.md +0 -488
  1415. package/bin/skills/treatment-plans/SKILL.md +0 -1579
  1416. package/bin/skills/treatment-plans/assets/STYLING_QUICK_REFERENCE.md +0 -185
  1417. package/bin/skills/treatment-plans/assets/chronic_disease_management_plan.tex +0 -665
  1418. package/bin/skills/treatment-plans/assets/general_medical_treatment_plan.tex +0 -547
  1419. package/bin/skills/treatment-plans/assets/medical_treatment_plan.sty +0 -222
  1420. package/bin/skills/treatment-plans/assets/mental_health_treatment_plan.tex +0 -774
  1421. package/bin/skills/treatment-plans/assets/one_page_treatment_plan.tex +0 -193
  1422. package/bin/skills/treatment-plans/assets/pain_management_plan.tex +0 -799
  1423. package/bin/skills/treatment-plans/assets/perioperative_care_plan.tex +0 -753
  1424. package/bin/skills/treatment-plans/assets/quality_checklist.md +0 -471
  1425. package/bin/skills/treatment-plans/assets/rehabilitation_treatment_plan.tex +0 -756
  1426. package/bin/skills/treatment-plans/references/goal_setting_frameworks.md +0 -411
  1427. package/bin/skills/treatment-plans/references/intervention_guidelines.md +0 -507
  1428. package/bin/skills/treatment-plans/references/regulatory_compliance.md +0 -476
  1429. package/bin/skills/treatment-plans/references/specialty_specific_guidelines.md +0 -655
  1430. package/bin/skills/treatment-plans/references/treatment_plan_standards.md +0 -485
  1431. package/bin/skills/treatment-plans/scripts/check_completeness.py +0 -322
  1432. package/bin/skills/treatment-plans/scripts/generate_template.py +0 -233
  1433. package/bin/skills/treatment-plans/scripts/timeline_generator.py +0 -385
  1434. package/bin/skills/treatment-plans/scripts/validate_treatment_plan.py +0 -369
  1435. package/bin/skills/trl-fine-tuning/SKILL.md +0 -455
  1436. package/bin/skills/trl-fine-tuning/references/dpo-variants.md +0 -227
  1437. package/bin/skills/trl-fine-tuning/references/online-rl.md +0 -82
  1438. package/bin/skills/trl-fine-tuning/references/reward-modeling.md +0 -122
  1439. package/bin/skills/trl-fine-tuning/references/sft-training.md +0 -168
  1440. package/bin/skills/umap-learn/SKILL.md +0 -479
  1441. package/bin/skills/umap-learn/references/api_reference.md +0 -532
  1442. package/bin/skills/uniprot-database/SKILL.md +0 -195
  1443. package/bin/skills/uniprot-database/references/api_examples.md +0 -413
  1444. package/bin/skills/uniprot-database/references/api_fields.md +0 -275
  1445. package/bin/skills/uniprot-database/references/id_mapping_databases.md +0 -285
  1446. package/bin/skills/uniprot-database/references/query_syntax.md +0 -256
  1447. package/bin/skills/uniprot-database/scripts/uniprot_client.py +0 -341
  1448. package/bin/skills/unsloth/SKILL.md +0 -635
  1449. package/bin/skills/unsloth/docs/advanced-rl.md +0 -222
  1450. package/bin/skills/unsloth/docs/chat-templates.md +0 -141
  1451. package/bin/skills/unsloth/docs/datasets.md +0 -489
  1452. package/bin/skills/unsloth/docs/docker-extended.md +0 -99
  1453. package/bin/skills/unsloth/docs/dynamic-ggufs-2.0.md +0 -116
  1454. package/bin/skills/unsloth/docs/dynamic-ggufs-aider.md +0 -118
  1455. package/bin/skills/unsloth/docs/faq.md +0 -91
  1456. package/bin/skills/unsloth/docs/fp16-vs-bf16.md +0 -61
  1457. package/bin/skills/unsloth/docs/fp8-rl.md +0 -224
  1458. package/bin/skills/unsloth/docs/glm-4.7-flash.md +0 -997
  1459. package/bin/skills/unsloth/docs/inference-deployment-overview.md +0 -17
  1460. package/bin/skills/unsloth/docs/inference.md +0 -27
  1461. package/bin/skills/unsloth/docs/installation-docker.md +0 -155
  1462. package/bin/skills/unsloth/docs/installation-pip.md +0 -148
  1463. package/bin/skills/unsloth/docs/kernels-packing.md +0 -190
  1464. package/bin/skills/unsloth/docs/kimi-k2.5.md +0 -634
  1465. package/bin/skills/unsloth/docs/lm-studio.md +0 -235
  1466. package/bin/skills/unsloth/docs/lora-hot-swapping.md +0 -75
  1467. package/bin/skills/unsloth/docs/lora-hyperparameters.md +0 -363
  1468. package/bin/skills/unsloth/docs/memory-efficient-rl.md +0 -267
  1469. package/bin/skills/unsloth/docs/model-selection.md +0 -70
  1470. package/bin/skills/unsloth/docs/models.md +0 -532
  1471. package/bin/skills/unsloth/docs/multi-gpu-ddp.md +0 -90
  1472. package/bin/skills/unsloth/docs/notebooks.md +0 -223
  1473. package/bin/skills/unsloth/docs/overview.md +0 -110
  1474. package/bin/skills/unsloth/docs/qwen3-coder-next-extended.md +0 -900
  1475. package/bin/skills/unsloth/docs/qwen3-coder-next.md +0 -900
  1476. package/bin/skills/unsloth/docs/requirements.md +0 -45
  1477. package/bin/skills/unsloth/docs/reward-hacking.md +0 -25
  1478. package/bin/skills/unsloth/docs/saving-to-gguf.md +0 -138
  1479. package/bin/skills/unsloth/docs/saving-to-ollama.md +0 -46
  1480. package/bin/skills/unsloth/docs/sglang-guide.md +0 -278
  1481. package/bin/skills/unsloth/docs/speculative-decoding.md +0 -70
  1482. package/bin/skills/unsloth/docs/tool-calling.md +0 -334
  1483. package/bin/skills/unsloth/docs/troubleshooting-faq.md +0 -204
  1484. package/bin/skills/unsloth/docs/troubleshooting-inference.md +0 -26
  1485. package/bin/skills/unsloth/docs/tts-fine-tuning.md +0 -149
  1486. package/bin/skills/unsloth/docs/tutorial-grpo.md +0 -273
  1487. package/bin/skills/unsloth/docs/tutorial-llama3-ollama.md +0 -356
  1488. package/bin/skills/unsloth/docs/vision-fine-tuning.md +0 -135
  1489. package/bin/skills/unsloth/docs/vision-rl.md +0 -170
  1490. package/bin/skills/unsloth/docs/vllm-engine-arguments.md +0 -43
  1491. package/bin/skills/unsloth/docs/vllm-guide.md +0 -98
  1492. package/bin/skills/uspto-database/SKILL.md +0 -607
  1493. package/bin/skills/uspto-database/references/additional_apis.md +0 -394
  1494. package/bin/skills/uspto-database/references/patentsearch_api.md +0 -266
  1495. package/bin/skills/uspto-database/references/peds_api.md +0 -212
  1496. package/bin/skills/uspto-database/references/trademark_api.md +0 -358
  1497. package/bin/skills/uspto-database/scripts/patent_search.py +0 -290
  1498. package/bin/skills/uspto-database/scripts/peds_client.py +0 -285
  1499. package/bin/skills/uspto-database/scripts/trademark_client.py +0 -311
  1500. package/bin/skills/vaex/SKILL.md +0 -182
  1501. package/bin/skills/vaex/references/core_dataframes.md +0 -367
  1502. package/bin/skills/vaex/references/data_processing.md +0 -555
  1503. package/bin/skills/vaex/references/io_operations.md +0 -703
  1504. package/bin/skills/vaex/references/machine_learning.md +0 -728
  1505. package/bin/skills/vaex/references/performance.md +0 -571
  1506. package/bin/skills/vaex/references/visualization.md +0 -613
  1507. package/bin/skills/venue-templates/SKILL.md +0 -686
  1508. package/bin/skills/venue-templates/assets/examples/cell_summary_example.md +0 -247
  1509. package/bin/skills/venue-templates/assets/examples/medical_structured_abstract.md +0 -313
  1510. package/bin/skills/venue-templates/assets/examples/nature_abstract_examples.md +0 -213
  1511. package/bin/skills/venue-templates/assets/examples/neurips_introduction_example.md +0 -245
  1512. package/bin/skills/venue-templates/assets/grants/nih_specific_aims.tex +0 -235
  1513. package/bin/skills/venue-templates/assets/grants/nsf_proposal_template.tex +0 -375
  1514. package/bin/skills/venue-templates/assets/journals/nature_article.tex +0 -171
  1515. package/bin/skills/venue-templates/assets/journals/neurips_article.tex +0 -283
  1516. package/bin/skills/venue-templates/assets/journals/plos_one.tex +0 -317
  1517. package/bin/skills/venue-templates/assets/posters/beamerposter_academic.tex +0 -311
  1518. package/bin/skills/venue-templates/references/cell_press_style.md +0 -483
  1519. package/bin/skills/venue-templates/references/conferences_formatting.md +0 -564
  1520. package/bin/skills/venue-templates/references/cs_conference_style.md +0 -463
  1521. package/bin/skills/venue-templates/references/grants_requirements.md +0 -787
  1522. package/bin/skills/venue-templates/references/journals_formatting.md +0 -486
  1523. package/bin/skills/venue-templates/references/medical_journal_styles.md +0 -535
  1524. package/bin/skills/venue-templates/references/ml_conference_style.md +0 -556
  1525. package/bin/skills/venue-templates/references/nature_science_style.md +0 -405
  1526. package/bin/skills/venue-templates/references/posters_guidelines.md +0 -628
  1527. package/bin/skills/venue-templates/references/reviewer_expectations.md +0 -417
  1528. package/bin/skills/venue-templates/references/venue_writing_styles.md +0 -321
  1529. package/bin/skills/venue-templates/scripts/customize_template.py +0 -195
  1530. package/bin/skills/venue-templates/scripts/query_template.py +0 -266
  1531. package/bin/skills/venue-templates/scripts/validate_format.py +0 -250
  1532. package/bin/skills/verl/SKILL.md +0 -391
  1533. package/bin/skills/verl/references/api-reference.md +0 -301
  1534. package/bin/skills/verl/references/troubleshooting.md +0 -391
  1535. package/bin/skills/vllm/SKILL.md +0 -364
  1536. package/bin/skills/vllm/references/optimization.md +0 -226
  1537. package/bin/skills/vllm/references/quantization.md +0 -284
  1538. package/bin/skills/vllm/references/server-deployment.md +0 -255
  1539. package/bin/skills/vllm/references/troubleshooting.md +0 -447
  1540. package/bin/skills/weights-and-biases/SKILL.md +0 -590
  1541. package/bin/skills/weights-and-biases/references/artifacts.md +0 -584
  1542. package/bin/skills/weights-and-biases/references/integrations.md +0 -700
  1543. package/bin/skills/weights-and-biases/references/sweeps.md +0 -847
  1544. package/bin/skills/whisper/SKILL.md +0 -317
  1545. package/bin/skills/whisper/references/languages.md +0 -189
  1546. package/bin/skills/zarr-python/SKILL.md +0 -779
  1547. package/bin/skills/zarr-python/references/api_reference.md +0 -515
  1548. package/bin/skills/zinc-database/SKILL.md +0 -404
  1549. package/bin/skills/zinc-database/references/api_reference.md +0 -692
@@ -1,1191 +0,0 @@
1
- # Deepspeed - Other
2
-
3
- **Pages:** 15
4
-
5
- ---
6
-
7
- ## Training Overview and Features
8
-
9
- **URL:** https://www.deepspeed.ai/training/
10
-
11
- **Contents:**
12
- - Training Overview and Features
13
- - Contents
14
- - Overview
15
- - Distributed, Effective, and Efficient Training with Ease
16
- - Speed
17
- - Memory efficiency
18
- - Scalability
19
- - Communication efficiency
20
- - Data efficiency
21
- - Supporting long sequence length
22
-
23
- Training advanced deep learning models is challenging. Beyond model design, model scientists also need to set up the state-of-the-art training techniques such as distributed training, mixed precision, gradient accumulation, and checkpointing. Yet still, scientists may not achieve the desired system performance and convergence rate. Large model sizes are even more challenging: a large model easily runs out of memory with pure data parallelism and it is difficult to use model parallelism. DeepSpeed addresses these challenges to accelerate model development and training.
24
-
25
- The DeepSpeed API is a lightweight wrapper on PyTorch. This means that you can use everything you love in PyTorch and without learning a new platform. In addition, DeepSpeed manages all of the boilerplate state-of-the-art training techniques, such as distributed training, mixed precision, gradient accumulation, and checkpoints so that you can focus on your model development. Most importantly, you can leverage the distinctive efficiency and effectiveness benefit of DeepSpeed to boost speed and scale with just a few lines of code changes to your PyTorch models.
26
-
27
- DeepSpeed achieves high performance and fast convergence through a combination of efficiency optimizations on compute/communication/memory/IO and effectiveness optimizations on advanced hyperparameter tuning and optimizers. For example:
28
-
29
- DeepSpeed trains BERT-large to parity in 44 mins using 1024 V100 GPUs (64 DGX-2 boxes) and in 2.4 hours using 256 GPUs (16 DGX-2 boxes).
30
-
31
- BERT-large Training Times
32
-
33
- BERT code and tutorials will be available soon.
34
-
35
- DeepSpeed trains GPT2 (1.5 billion parameters) 3.75x faster than state-of-art, NVIDIA Megatron on Azure GPUs.
36
-
37
- Read more: GPT tutorial
38
-
39
- DeepSpeed provides memory-efficient data parallelism and enables training models without model parallelism. For example, DeepSpeed can train models with up to 13 billion parameters on a single GPU. In comparison, existing frameworks (e.g., PyTorch’s Distributed Data Parallel) run out of memory with 1.4 billion parameter models.
40
-
41
- DeepSpeed reduces the training memory footprint through a novel solution called Zero Redundancy Optimizer (ZeRO). Unlike basic data parallelism where memory states are replicated across data-parallel processes, ZeRO partitions model states and gradients to save significant memory. Furthermore, it also reduces activation memory and fragmented memory. The current implementation (ZeRO-2) reduces memory by up to 8x relative to the state-of-art. You can read more about ZeRO in our paper, and in our blog posts related to ZeRO-1 and ZeRO-2.
42
-
43
- With this impressive memory reduction, early adopters of DeepSpeed have already produced a language model (LM) with over 17B parameters called Turing-NLG, establishing a new SOTA in the LM category.
44
-
45
- For model scientists with limited GPU resources, ZeRO-Offload leverages both CPU and GPU memory for training large models. Using a machine with a single GPU, our users can run models of up to 13 billion parameters without running out of memory, 10x bigger than the existing approaches, while obtaining competitive throughput. This feature democratizes multi-billion-parameter model training and opens the window for many deep learning practitioners to explore bigger and better models.
46
-
47
- DeepSpeed supports efficient data parallelism, model parallelism, pipeline parallelism and their combinations, which we call 3D parallelism.
48
-
49
- DeepSpeed can run large models more efficiently, up to 10x faster for models with various sizes spanning 1.5B to hundred billion. More specifically, the data parallelism powered by ZeRO is complementary and can be combined with different types of model parallelism. It allows DeepSpeed to fit models using lower degree of model parallelism and higher batch size, offering significant performance gains compared to using model parallelism alone.
50
-
51
- Read more: ZeRO paper, and GPT tutorial.
52
-
53
- The figure depicts system throughput improvements of DeepSpeed (combining ZeRO-powered data parallelism with model parallelism of NVIDIA Megatron-LM) over using Megatron-LM alone.
54
-
55
- Pipeline parallelism of DeepSpeed reduce communication volume during distributed training, which allows users to train multi-billion-parameter models 2–7x faster on clusters with limited network bandwidth.
56
-
57
- 1-bit Adam, 0/1 Adam and 1-bit LAMB reduce communication volume by up to 26x while achieving similar convergence efficiency to Adam, allowing for scaling to different types of GPU clusters and networks. 1-bit Adam blog post, 1-bit Adam tutorial, 0/1 Adam tutorial, 1-bit LAMB tutorial.
58
-
59
- DeepSpeed Data Efficiency Library provides efficient data sampling via curriculum learning and efficient data routing via random layerwise token dropping. The composed solution enables up to 2x data and 2x time saving during GPT-3/BERT pretraining and GPT/ViT finetuning, or further improve model quality under the same data/time. See more in the tutorial.
60
-
61
- DeepSpeed offers sparse attention kernels—an instrumental technology to support long sequences of model inputs, whether for text, image, or sound. Compared with the classic dense Transformers, it powers an order-of-magnitude longer input sequence and obtains up to 6x faster execution with comparable accuracy. It also outperforms state-of-the-art sparse implementations with 1.5–3x faster execution. Furthermore, our sparse kernels support efficient execution of flexible sparse format and empower users to innovate on their custom sparse structures. Read more here.
62
-
63
- DeepSpeed supports advanced hyperparameter tuning and large batch size optimizers such as LAMB. These improve the effectiveness of model training and reduce the number of samples required to convergence to desired accuracy.
64
-
65
- Read more: Tuning tutorial.
66
-
67
- Only a few lines of code changes are needed to enable a PyTorch model to use DeepSpeed and ZeRO. Compared to current model parallelism libraries, DeepSpeed does not require a code redesign or model refactoring. It also does not put limitations on model dimensions (such as number of attention heads, hidden sizes, and others), batch size, or any other training parameters. For models of up to 13 billion parameters, you can use ZeRO-powered data parallelism conveniently without requiring model parallelism, while in contrast, standard data parallelism will run out of memory for models with more than 1.4 billion parameters. In addition, DeepSpeed conveniently supports flexible combination of ZeRO-powered data parallelism with custom model parallelisms, such as tensor slicing of NVIDIA’s Megatron-LM.
68
-
69
- Below we provide a brief feature list, see our detailed feature overview for descriptions and usage.
70
-
71
- title: “Feature Overview” layout: single permalink: /features/ toc: true toc_label: “Contents” —
72
-
73
- Enable 16-bit (FP16) training by in the deepspeed_config JSON.
74
-
75
- Easily switch between single-GPU, single-node multi-GPU, or multi-node multi-GPU execution by specifying resources with a hostfile.
76
-
77
- The script <client_entry.py> will execute on the resources specified in <hostfile>.
78
-
79
- DeepSpeed provides pipeline parallelism for memory- and communication- efficient training. DeepSpeed supports a hybrid combination of data, model, and pipeline parallelism and has scaled to over one trillion parameters using 3D parallelism. Pipeline parallelism can also improve communication efficiency and has accelerated training by up to 7x on low-bandwidth clusters.
80
-
81
- DeepSpeed supports all forms of model parallelism including tensor slicing based approaches such as the Megatron-LM. It does so by only requiring the model parallelism framework to provide a model parallelism unit (mpu) that implements a few bookkeeping functionalities:
82
-
83
- DeepSpeed is fully compatible with Megatron. Please see the Megatron-LM tutorial for details.
84
-
85
- The Zero Redundancy Optimizer (ZeRO) is at the heart of DeepSpeed and enables large model training at a scale that is simply not possible with model parallelism alone. When enabled, ZeRO allows training models with over 13 billion parameters without any model parallelism, and up to 200 billion parameter models with model parallelism on current generation hardware.
86
-
87
- For more details see the ZeRO paper, GPT tutorial on integration with DeepSpeed.
88
-
89
- Optimizer State and Gradient Partitioning in ZeRO reduces the memory consumption of the model states (optimizer states, gradients and parameters) by 8x compared to standard data parallelism by partitioning these states across data parallel process instead of replicating them.
90
-
91
- Activation Partitioning is a memory optimization in ZeRO that can reduce the memory consumed by activations during model parallel training (MP). In MP certain activations maybe required by all MP processes, resulting in a replication of activations across MP GPUs. Activation Partitioning stores these activations in a partitioned state once they are used for computation in the forward propagation. These activations are allgathered right before they are needed again during the backward propagation. By storing activations in a partitioned state, ZeRO in DeepSpeed can reduce the activation memory footprint proportional to the MP degree.
92
-
93
- CBO enables high network and memory throughput while restricting memory usage to a constant size. For memory- and network-bound operations such as normalization or allreduce collectives, the performance depends on the size of the operand. Simply fusing all operands into a single large operand can enable great throughput at the expense of unnecessary memory overhead. CBO in DeepSpeed fuses smaller operands into approximately a pre-defined sized buffer large enough to achieve great performance without the unnecessary memory overhead.
94
-
95
- CMO reduces memory fragmentation during training, preventing out of memory errors due to lack of contiguous memory. Memory fragmentation is a result of interleaving between short lived and long lived memory objects. During the forward propagation activation checkpoints are long lived but the activations that recomputed are short lived. Similarly, during the backward computation, the activation gradients are short lived while the parameter gradients are long lived. CMO transfers activation checkpoints and parameter gradients to contiguous buffers preventing memory fragmentation.
96
-
97
- ZeRO-Offload pushes the boundary of the maximum model size that can be trained efficiently using minimal GPU resources, by exploiting computational and memory resources on both GPUs and their host CPUs. It allows training up to 13-billion-parameter models on a single NVIDIA V100 GPU, 10x larger than the state-of-the-art, while retaining high training throughput of over 30 teraflops per GPU.
98
-
99
- For more details see the ZeRO-Offload release blog, and tutorial on integration with DeepSpeed.
100
-
101
- Gradient accumulation allows running larger batch size with limited memory by breaking an effective batch into several sequential micro-batches, and averaging the parameter gradients across these micro-batches. Furthermore, instead of averaging the gradients of each micro-batch across all GPUs, the gradients are averaged locally during each step of the sequence, and a single allreduce is done at the end of the sequence to produce the averaged gradients for the effective batch across all GPUs. This strategy significantly reduces the communication involved over the approach of averaging globally for each micro-batch, specially when the number of micro-batches per effective batch is large.
102
-
103
- During back propagation, DeepSpeed can overlap the communication required for averaging parameter gradients that have already been computed with the ongoing gradient computation. This computation-communication overlap allows DeepSpeed to achieve higher throughput even at modest batch sizes.
104
-
105
- The DeepSpeed core API consists of just a handful of methods:
106
-
107
- DeepSpeed supports most of the features described in this document, via the use of these API, along with a deepspeed_config JSON file for enabling and disabling the features. Please see the core API doc for more details.
108
-
109
- DeepSpeed’s Activation Checkpointing API supports activation checkpoint partitioning, cpu checkpointing, and contiguous memory optimizations, while also allowing layerwise profiling. Please see the core API doc for more details.
110
-
111
- DeepSpeed handles gradient clipping under the hood based on the max gradient norm specified by the user. Please see the core API doc for more details.
112
-
113
- DeepSpeed internally handles loss scaling for mixed precision training. The parameters for loss scaling can be specified in the deepspeed_config JSON file. Please see the core API doc for more details.
114
-
115
- DeepSpeed has three communication-efficient optimizers called 1-bit Adam, 0/1 Adam and 1-bit LAMB. They offer the same convergence as Adam/LAMB, incur up to 26x less communication that enables up to 6.6x higher throughput for BERT-Large pretraining and up to 2.7x higher throughput for SQuAD fine-tuning on bandwidth-limited clusters. For more details on usage and performance, please refer to the 1-bit Adam tutorial, 1-bit Adam blog post, 0/1 Adam tutorial and 1-bit LAMB tutorial. For technical details, please refer to the 1-bit Adam paper, 0/1 Adam paper and 1-bit LAMB paper.
116
-
117
- With DeepSpeed, the user can choose to use a high performance implementation of ADAM from NVIDIA, or any training optimizer that extends torch’s torch.optim.Optimizer class.
118
-
119
- We introduce an efficient implementation of Adam optimizer on CPU that improves the parameter-update performance by nearly an order of magnitude. We use the AVX SIMD instructions on Intel-x86 architecture for the CPU-Adam implementation. We support both AVX-512 and AVX-2 instruction sets. DeepSpeed uses AVX-2 by default which can be switched to AVX-512 by setting the build flag, DS_BUILD_AVX512 to 1 when installing DeepSpeed. Using AVX-512, we observe 5.1x to 6.5x speedups considering the model-size between 1 to 10 billion parameters with respect to torch-adam.
120
-
121
- Mixed precision training is handled by the DeepSpeed FP16 Optimizer. This optimizer not only handles FP16 training but is also highly efficient. The performance of weight update is primarily dominated by the memory bandwidth, and the achieved memory bandwidth is dependent on the size of the input operands. The FP16 Optimizer is designed to maximize the achievable memory bandwidth by merging all the parameters of the model into a single large buffer, and applying the weight updates in a single kernel, allowing it to achieve high memory bandwidth.
122
-
123
- DeepSpeed makes it easy to train with large batch sizes by enabling the LAMB Optimizer. For more details on LAMB, see the LAMB paper.
124
-
125
- DeepSpeed can train models with up to 13 billion parameters without model parallelism, and models with up to 200 billion parameters with 16-way model parallelism. This leap in model size is possible through the memory efficiency achieved via the ZeRO Optimizer. For more details see ZeRO paper .
126
-
127
- DeepSpeed can simplify checkpointing for you regardless of whether you are using data parallel training, model parallel training, mixed-precision training, a mix of these three, or using the zero optimizer to enable larger model sizes. Please see the Getting Started guide and the core API doc for more details.
128
-
129
- DeepSpeed supports multiple Learning Rate Schedules to enable faster convergence for large batch scaling.
130
-
131
- Please refer to the Learning Rate Range Test tutorial.
132
-
133
- Please refer to the 1Cycle Learning Rate Schedule tutorial.
134
-
135
- DeepSpeed abstracts away data parallelism and model parallelism from the user when it comes to data loading. Users simply provide a PyTorch dataset, and DeepSpeed data loader can automatically handle batch creation appropriately.
136
-
137
- Please refer to the Data Efficiency tutorial.
138
-
139
- Please refer to the Curriculum Learning tutorial. Note that the Data Efficiency Library above provides more general curriculum learning support. This legacy curriculum learning feature is still supported but we recommend to use the Data Efficiency Library.
140
-
141
- DeepSpeed provides a set of tools for performance analysis and debugging.
142
-
143
- DeepSpeed provides a detailed breakdown of the time spent in different parts of the training. This can be enabled by setting the following in the deepspeed_config file.
144
-
145
- When activation checkpointing is enabled, profiling the forward and backward time of each checkpoint function can be enabled in the deepspeed_config file.
146
-
147
- The DeepSpeed flops profiler measures the time, flops and parameters of a PyTorch model and shows which modules or layers are the bottleneck. When used with the DeepSpeed runtime, the flops profiler can be configured in the deepspeed_config file as follows:
148
-
149
- The flops profiler can also be used as a standalone package. Please refer to the Flops Profiler tutorial for more details.
150
-
151
- The DeepSpeed Autotuner uses model information, system information, and heuristics to efficiently tune Zero stage, micro batch size, and other Zero configurations. Using the autotuning feature requires no code change from DeepSpeed users. While "autotuning": {"enabled": true} is the minimal required to enable autotuning, there are other parameters users can define to configure the autotuning process. Below shows major parameters and their default values in the autotuning configuration. Please refer to the Autotuning tutorial for more details.
152
-
153
- The flops profiler can also be used as a standalone package. Please refer to the Flops Profiler tutorial for more details.
154
-
155
- The DeepSpeed Monitor logs live training metrics to one or more monitoring backends, including PyTorch’s TensorBoard, WandB, or simply to CSV files. The Monitor can be configured with one or more backends in the deepspeed_config file as follows:
156
-
157
- The Monitor can also be added to log custom metrics and client codes. Please refer to the Monitor tutorial for more details.
158
-
159
- DeepSpeed provides logging of all communication operations launched within deepspeed.comm. The communication logger can be configured in the deepspeed_config file as follows:
160
-
161
- Client codes can then print a summary with a call to deepspeed.comm.log_summary(). For more details and example usage, see the Communication Logging tutorial.
162
-
163
- DeepSpeed offers sparse attention to support long sequences. Please refer to the Sparse Attention tutorial.
164
-
165
- To learn more about training Mixture of Experts (MoE) models with DeepSpeed, see our tutorial for more details.
166
-
167
- **Examples:**
168
-
169
- Example 1 (unknown):
170
- ```unknown
171
- "fp16": {
172
- "enabled": true,
173
- "loss_scale": 0,
174
- "loss_scale_window": 1000,
175
- "hysteresis": 2,
176
- "consecutive_hysteresis": false,
177
- "min_loss_scale": 1
178
- }
179
- ```
180
-
181
- Example 2 (unknown):
182
- ```unknown
183
- deepspeed --hostfile=<hostfile> \
184
- <client_entry.py> <client args> \
185
- --deepspeed --deepspeed_config ds_config.json
186
- ```
187
-
188
- Example 3 (unknown):
189
- ```unknown
190
- mpu.get_model_parallel_rank()
191
- mpu.get_model_parallel_group()
192
- mpu.get_model_parallel_world_size()
193
-
194
- mpu.get_data_parallel_rank()
195
- mpu.get_data_parallel_group()
196
- mpu.get_data_parallel_world_size()
197
- ```
198
-
199
- Example 4 (unknown):
200
- ```unknown
201
- {
202
- "gradient_clipping": 1.0
203
- }
204
- ```
205
-
206
- ---
207
-
208
- ## Latest News
209
-
210
- **URL:** https://www.deepspeed.ai/
211
-
212
- **Contents:**
213
- - Latest News
214
- - Contents
215
- - Extreme Speed and Scale for DL Training
216
- - DeepSpeed Adoption
217
- - Contributing
218
- - Contributor License Agreement
219
- - Code of Conduct
220
- - Publications
221
- - Videos
222
-
223
- [2025/10] SuperOffload: Unleashing the Power of Large-Scale LLM Training on Superchips
224
-
225
- [2025/10] Study of ZenFlow and ZeRO offload performance with DeepSpeed CPU core binding
226
-
227
- [2025/08] ZenFlow: Stall-Free Offloading Engine for LLM Training
228
-
229
- [2025/06] Arctic Long Sequence Training (ALST) with DeepSpeed: Scalable And Efficient Training For Multi-Million Token Sequences
230
-
231
- [2025/06] DeepNVMe: Affordable I/O scaling for Deep Learning Applications
232
-
233
- DeepSpeed enabled the world’s most powerful language models (at the time of this writing) such as MT-530B and BLOOM. DeepSpeed offers a confluence of system innovations, that has made large scale DL training effective, and efficient, greatly improved ease of use, and redefined the DL training landscape in terms of scale that is possible. These innovations include ZeRO, 3D-Parallelism, DeepSpeed-MoE, ZeRO-Infinity, etc.
234
-
235
- DeepSpeed has been used to train many different large-scale models. Below is a list of several examples that we are aware of (if you’d like to include your model please submit a PR):
236
-
237
- DeepSpeed has been integrated with several different popular open-source DL frameworks such as:
238
-
239
- DeepSpeed is an integral part of Microsoft’s AI at Scale initiative to enable next-generation AI capabilities at scale.
240
-
241
- DeepSpeed welcomes your contributions! Please see our contributing guide for more details on formatting, testing, etc.
242
-
243
- This project welcomes contributions and suggestions. Most contributions require you to agree to a Contributor License Agreement (CLA) declaring that you have the right to, and actually do, grant us the rights to use your contribution. For details, visit https://cla.opensource.microsoft.com.
244
-
245
- When you submit a pull request, a CLA bot will automatically determine whether you need to provide a CLA and decorate the PR appropriately (e.g., status check, comment). Simply follow the instructions provided by the bot. You will only need to do this once across all repos using our CLA.
246
-
247
- This project has adopted the Microsoft Open Source Code of Conduct. For more information see the Code of Conduct FAQ or contact opencode@microsoft.com with any additional questions or comments.
248
-
249
- Xinyu Lian, Sam Ade Jacobs, Lev Kurilenko, Masahiro Tanaka, Stas Bekman, Olatunji Ruwase, Minjia Zhang. (2024) Universal Checkpointing: Efficient and Flexible Checkpointing for Large Scale Distributed Training arXiv:2406.18820
250
-
251
- ---
252
-
253
- ## Supporting efficient large model training on AMD Instinct GPUs with DeepSpeed
254
-
255
- **URL:** https://www.deepspeed.ai/2022/03/20/amd-support.html
256
-
257
- **Contents:**
258
- - Supporting efficient large model training on AMD Instinct GPUs with DeepSpeed
259
- - Contents
260
-
261
- Updated: March 20, 2022
262
-
263
- ---
264
-
265
- ## DeepSpeed Configuration JSON
266
-
267
- **URL:** https://www.deepspeed.ai/docs/config-json/
268
-
269
- **Contents:**
270
- - DeepSpeed Configuration JSON
271
- - Contents
272
- - Batch Size Related Parameters
273
- - Optimizer Parameters
274
- - Scheduler Parameters
275
- - Communication options
276
- - FP16 training options
277
- - BFLOAT16 training options
278
- - Automatic mixed precision (AMP) training options
279
- - Gradient Clipping
280
-
281
- Note: train_batch_size must be equal to train_micro_batch_size_per_gpu * gradient_accumulation_steps * number of GPUs. For simplicity, you can choose to only specify two of the three parameters, the last one will be inferred automatically by DeepSpeed.
282
-
283
- train_batch_size: [integer]
284
-
285
- train_micro_batch_size_per_gpu: [integer]
286
-
287
- gradient_accumulation_steps: [integer]
288
-
289
- optimizer: [dictionary]
290
-
291
- Example of optimizer with Adam
292
-
293
- The Adam optimizer also supports the following two params keys/values in addition to the standard parameters from torch.optim.Adam:
294
-
295
- Another example of optimizer with 1-bit Adam specific parameters is as follows.
296
-
297
- The 1-bit Adam optimizer supports the following three params keys/values in addition to the standard Adam (learn more in our tutorial):
298
-
299
- A variant optimizer for 1-bit Adam is 0/1 Adam, which further optimizes 1-bit Adam via adaptive variance freezing and 1-bit synchronization over optimizer states.
300
-
301
- 0/1 Adam supports the following params key/values in addition to standard Adam (learn more in our tutorial.)
302
-
303
- Another example of optimizer with 1-bit LAMB
304
-
305
- The 1-bit LAMB optimizer supports the following params keys/values in addition to the standard LAMB (learn more in our tutorial):
306
-
307
- DeepSpeed calls the step() method of the scheduler at every training step when model_engine.step() is executed.
308
-
309
- scheduler: [dictionary]
310
-
311
- communication_data_type: [string]
312
-
313
- prescale_gradients: [boolean]
314
-
315
- gradient_predivide_factor: [float]
316
-
317
- sparse_gradients: [boolean]
318
-
319
- Note: this mode cannot be combined with the amp mode described below.
320
-
321
- fp16:enabled: [boolean]
322
-
323
- fp16:auto_cast: [boolean]
324
-
325
- fp16:loss_scale: [float]
326
-
327
- fp16:initial_scale_power: [integer]
328
-
329
- fp16:loss_scale_window: [integer]
330
-
331
- fp16:hysteresis: [integer]
332
-
333
- fp16:consecutive_hysteresis: [boolean]
334
-
335
- fp16:min_loss_scale: [integer]
336
-
337
- Note: this mode cannot be combined with the amp mode described below.
338
-
339
- Note: this mode cannot be combined with the fp16 mode described above.
340
-
341
- bf16:enabled: [boolean]
342
-
343
- Note: this mode cannot be combined with the fp16 mode described above. In addition this mode is not currently compatible with ZeRO.
344
-
345
- amp:enabled: [boolean]
346
-
347
- amp params: [various]
348
-
349
- gradient_clipping: [float]
350
-
351
- Enabling and configuring ZeRO memory optimizations
352
-
353
- zero_optimization: [dictionary]
354
-
355
- allgather_partitions: [boolean]
356
-
357
- allgather_bucket_size: [integer]
358
-
359
- overlap_comm: [boolean]
360
-
361
- reduce_scatter: [boolean]
362
-
363
- reduce_bucket_size: [integer]
364
-
365
- contiguous_gradients: [boolean]
366
-
367
- load_from_fp32_weights: [boolean]
368
-
369
- grad_hooks: [boolean]
370
-
371
- round_robin_gradients: [boolean]
372
-
373
- offload_param: [dictionary]
374
-
375
- offload_optimizer: [dictionary]
376
-
377
- stage3_max_live_parameters: [integer]
378
-
379
- stage3_max_reuse_distance: [integer]
380
-
381
- stage3_prefetch_bucket_size: [integer]
382
-
383
- stage3_param_persistence_threshold: [integer]
384
-
385
- stage3_gather_16bit_weights_on_model_save: [boolean]
386
-
387
- stage3_module_granularity_threshold: [integer] | Description | Default | |——————————————————————————————————————————————————————————————————————————————————————————–| ——- | | The granularity of a module is determined by the ratio of parameter_count / (1 + descendant_count). ZeRO3 classifies modules with a granularity below the threshold as fine-grained, treating them as integral units during parameter fetching. This reduces host and communication overhead from separate hooks. | 0 |
388
-
389
- zero_hpz_partition_size: [integer]
390
-
391
- zero_quantized_weights: [boolean]
392
-
393
- zero_quantized_gradients: [boolean]
394
-
395
- log_trace_cache_warnings: [boolean]
396
-
397
- cpu_offload: [boolean]
398
-
399
- Deprecated: cpu_offload is deprecated and will be removed in future, please use offload_optimizer instead.
400
-
401
- Enabling and configuring ZeRO optimization of parameter offloading to CPU/NVMe. Available only with ZeRO stage 3. Note that if the value of “device” is not specified or not supported, an assertion will be triggered.
402
-
403
- pin_memory: [boolean]
404
-
405
- buffer_count: [integer]
406
-
407
- buffer_size: [integer]
408
-
409
- max_in_cpu: [integer]
410
-
411
- Enabling and configuring ZeRO optimization of offloading optimizer computation to CPU and state to CPU/NVMe. CPU offloading is available with ZeRO stage 1, 2, 3. NVMe offloading is available only with ZeRO stage 3. Note that if the value of “device” is not specified or not supported, an assertion will be triggered.
412
-
413
- pin_memory: [boolean]
414
-
415
- buffer_count: [integer]
416
-
417
- Configuring the asynchronous I/O module for offloading parameter and optimizer states to persistent (NVMe) storage. This module uses Linux native asynchronous I/O (libaio).
418
-
419
- block_size: [integer]
420
-
421
- queue_depth: [integer]
422
-
423
- thread_count: [integer]
424
-
425
- single_submit: [boolean]
426
-
427
- overlap_events: [boolean]
428
-
429
- ignore_unused_parameters: [boolean]
430
-
431
- steps_per_print: [integer]
432
-
433
- wall_clock_breakdown: [boolean]
434
-
435
- dump_state: [boolean]
436
-
437
- results_dir: [string]
438
-
439
- start_profile_step: [integer]
440
-
441
- end_profile_step: [integer]
442
-
443
- max_train_batch_size: [int]
444
-
445
- num_tuning_micro_batch_sizes: [integer]
446
-
447
- tuner_early_stopping: [integer]
448
-
449
- tuner_num_trials: [integer]
450
-
451
- profile_step: [integer]
452
-
453
- module_depth: [integer]
454
-
455
- top_modules: [integer]
456
-
457
- output_file: [string]
458
-
459
- partition_activations: [boolean]
460
-
461
- cpu_checkpointing: [boolean]
462
-
463
- contiguous_memory_optimization: [boolean]
464
-
465
- number_checkpoints: [integer]
466
-
467
- synchronize_checkpoint_boundary: [boolean]
468
-
469
- sparse_attention: [dictionary]
470
-
471
- Example of sparse_attention
472
-
473
- DeepSpeed Data Efficiency Library includes two techniques: curriculum learning and random layerwise token dropping (random-LTD). Read more about how to use the DeepSpeed Data Efficiency Library in our tutorial.
474
-
475
- data_efficiency: [dictionary]
476
-
477
- data_routing: [dictionary]
478
-
479
- data_sampling: [dictionary]
480
-
481
- random_ltd: [dictionary]
482
-
483
- curriculum_learning: [dictionary]
484
-
485
- Note: On 12/12/2022, we released DeepSpeed Data Efficiency Library which provides a more general curriculum learning support. This legacy curriculum learning feature below is still supported but we recommend to use the Data Efficiency Library.
486
-
487
- curriculum_type: [string]
488
-
489
- min_difficulty: [integer]
490
-
491
- max_difficulty: [integer]
492
-
493
- schedule_type: [string]
494
-
495
- total_curriculum_step: [integer]
496
-
497
- difficulty_step: [integer]
498
-
499
- root_degree: [integer]
500
-
501
- difficulty: [list of integer]
502
-
503
- max_step: [list of integer]
504
-
505
- Note: Deepspeed logs to TensorBoard through PyTorch. Logging to TensorBoard requires that the tensorboard package is installed (read more in the PyTorch documentation).
506
-
507
- Note: Logging to WandB requires that the wandb package is installed (read more in the WandB documentation).
508
-
509
- Note: Logging to Comet requires that the comet_ml package is installed (read more in the Comet documentation).
510
-
511
- Deepspeed’s Monitor module can log training details into a Tensorboard-compatible file, to WandB, to Comet or to simple CSV files. Below is an overview of what DeepSpeed will log automatically.
512
-
513
- tensorboard: [dictionary]
514
-
515
- Example of tensorboard configuration:
516
-
517
- Example of wandb configuration:
518
-
519
- Example of comet configuration:
520
-
521
- csv_monitor: [dictionary]
522
-
523
- Example of csv_monitor configuration:
524
-
525
- DeepSpeed provides a flexible communication logging tool which can automatically detect and record communication operations launched via deepspeed.comm. NOTE: All logging communication calls are synchronized in order to provide accurate timing information. This may hamper performance if your model heavily uses asynchronous communication operations.
526
-
527
- Once the logs are populated, they can be summarized with deepspeed.comm.log_summary(). For more detail and example usage, see the tutorial
528
-
529
- comms_logger: [dictionary]
530
-
531
- Example of recommended comms_logger configuration:
532
-
533
- Example of comms_logger configuration for logging specific operations only:
534
-
535
- Note: Compression has seven different components, including layer reduction, weight quantization, activation quantization, sparse pruning, row pruning, head pruning, and channel pruning. We explain them one by one with simple json examples. Read more about how to use the DeepSpeed Compression library in our tutorial.
536
-
537
- Note: Layer reduction works much better when using knowledage distillation (learn more in our tutorial):
538
-
539
- layer_reduction: [dictionary]
540
-
541
- shared_parameters: [dictionary]
542
-
543
- Shared parameters for all weight quantization groups.
544
-
545
- different_groups: [dictionary]
546
-
547
- Different quantization sets, this is used for different quantization parameters. In this example, we give two different sets. In practice, you can choose the number of sets based on your requirements.
548
-
549
- shared_parameters: [dictionary]
550
-
551
- Shared parameters for all activation quantization groups.
552
-
553
- different_groups: [dictionary]
554
-
555
- Different quantization sets, this is used for different quantization parameters. In this example, we give one set. In practice, you can choose the number of sets based on your requirements.
556
-
557
- shared_parameters: [dictionary]
558
-
559
- Shared parameters for all sparse pruning groups.
560
-
561
- different_groups: [dictionary]
562
-
563
- Different pruning sets, this is used for different pruning parameters. In this example, we give one set. In practice, you can choose the number of sets based on your requirements. Note for snip_momentum method, you can leave it as empty.
564
-
565
- Note: Row Pruning is a feature designed for two back-to-back linear layers (e.g., Feed Forward Network in Transformers). As such, we suggested use row pruning for the first linear layer (i.e., the intermediate.dense layer for BERT). Reducing the row dimension of this matrix can help reducing the column of the follow-up matrix (i.e., layer.\\w+.output.dense layer for BERT). It should also work for other linear layers as well.
566
-
567
- shared_parameters: [dictionary]
568
-
569
- Shared parameters for all row pruning groups.
570
-
571
- different_groups: [dictionary]
572
-
573
- Different pruning sets, this is used for different pruning parameters. In this example, we give one set. In practice, you can choose the number of sets based on your requirements.
574
-
575
- Note: Head Pruning is a feature designed for two attention layers (e.g., Multi Head Attention in Transformers). For now, it can only be applied to output matrix of the Transformer (i.e., attention.output.dense in BERT). Pruning the output matrix can lead to the pruning of Query/Key/Value matrix as well.
576
-
577
- shared_parameters: [dictionary]
578
-
579
- Shared parameters for all head pruning groups.
580
-
581
- different_groups: [dictionary]
582
-
583
- Different pruning sets, this is used for different pruning parameters. In this example, we give one set. In practice, you can choose the number of sets based on your requirements.
584
-
585
- Note: Channel Pruning is a feature designed for two back-to-back CONV2d layers (e.g., residual connection in ResNet). As such, we suggested use channel pruning for the first CONV2d layer. Reducing the number of output channels of this layer can help reducing the number of input channels the follow-up layer. It should also work for other CONV2d layers as well.
586
-
587
- shared_parameters: [dictionary]
588
-
589
- Shared parameters for all channel pruning groups.
590
-
591
- different_groups: [dictionary]
592
-
593
- Different pruning sets, this is used for different pruning parameters. In this example, we give one set. In practice, you can choose the number of sets based on your requirements.
594
-
595
- load_universal: [boolean]
596
-
597
- use_node_local_storage: [boolean]
598
-
599
- pipeline_stage: [boolean]
600
-
601
- **Examples:**
602
-
603
- Example 1 (unknown):
604
- ```unknown
605
- "optimizer": {
606
- "type": "Adam",
607
- "params": {
608
- "lr": 0.001,
609
- "betas": [
610
- 0.8,
611
- 0.999
612
- ],
613
- "eps": 1e-8,
614
- "weight_decay": 3e-7
615
- }
616
- }
617
- ```
618
-
619
- Example 2 (unknown):
620
- ```unknown
621
- "optimizer": {
622
- "type": "OneBitAdam",
623
- "params": {
624
- "lr": 0.001,
625
- "betas": [
626
- 0.8,
627
- 0.999
628
- ],
629
- "eps": 1e-8,
630
- "weight_decay": 3e-7,
631
- "freeze_step": 400,
632
- "cuda_aware": false,
633
- "comm_backend_name": "nccl"
634
- }
635
- }
636
- ```
637
-
638
- Example 3 (unknown):
639
- ```unknown
640
- "optimizer": {
641
- "type": "ZeroOneAdam",
642
- "params": {
643
- "lr": 1e-3,
644
- "weight_decay": 0.01,
645
- "bias_correction": false,
646
- "var_freeze_step": 1000,
647
- "var_update_scaler": 16,
648
- "local_step_scaler": 1000,
649
- "local_step_clipper": 16,
650
- "cuda_aware": false,
651
- "comm_backend_name": "nccl"
652
- }
653
- }
654
- ```
655
-
656
- Example 4 (unknown):
657
- ```unknown
658
- "optimizer": {
659
- "type": "OneBitLamb",
660
- "params": {
661
- "lr": 11e-3,
662
- "weight_decay": 0.01,
663
- "bias_correction": false,
664
- "max_coeff": 0.3,
665
- "min_coeff": 0.01,
666
- "freeze_step": 1000,
667
- "cuda_aware": false,
668
- "comm_backend_name": "nccl",
669
- "coeff_beta": 0.9,
670
- "factor_max": 4.0,
671
- "factor_min": 0.5,
672
- "factor_threshold": 0.1
673
- }
674
- }
675
- ```
676
-
677
- ---
678
-
679
- ## DeepSpeed ZeRO-3 Offload
680
-
681
- **URL:** https://www.deepspeed.ai/2021/03/07/zero3-offload.html
682
-
683
- **Contents:**
684
- - DeepSpeed ZeRO-3 Offload
685
- - Contents
686
- - Overview of ZeRO family of technology
687
- - ZeRO-3 Offload
688
- - Unprecedented model scale
689
- - Ease of supporting very large models
690
- - Excellent training efficiency
691
- - How to use ZeRO-3 Offload
692
-
693
- Today we are announcing the release of ZeRO-3 Offload, a highly efficient and easy to use implementation of ZeRO Stage 3 and ZeRO Offload combined, geared towards our continued goal of democratizing AI by making efficient large-scale DL training available to everyone. The key benefits of ZeRO-3 Offload are:
694
-
695
- The ZeRO Redundancy Optimizer (abbreviated ZeRO) is a family of memory optimization technologies for large-scale distributed deep learning. Unlike data parallelism (that is efficient but can only support a limited model size) or model parallelism (that can support larger model sizes but requires significant code refactoring while adding communication overhead that limits efficiency), ZeRO allows fitting larger models in memory without requiring code refactoring while remaining very efficient. ZeRO does so by eliminating the memory redundancy that is inherent in data parallelism while limiting the communication overhead to a minimum. ZeRO removes the memory redundancies across data-parallel processes by partitioning the three model states (optimizer states, gradients, and parameters) across data-parallel processes instead of replicating them. By doing this, it boosts memory efficiency compared to classic data-parallelism while retaining its computational granularity and communication efficiency. There are three stages in ZeRO corresponding to three model states, as shown in the Figure 1: the first stage (ZeRO-1) partitions only the optimizer states, the second stage (ZeRO-2) partitions both the optimizer states and the gradients and the final stage (ZeRO-3) partitions all three model states (for more details see the ZeRO paper).
696
-
697
- Figure 1. Overview of ZeRO memory savings
698
-
699
- In addition to these three stages, ZeRO family of technology also consists of ZeRO-2 Offload. ZeRO-2 Offload is a heterogeneous DL training technology that works in conjunction with ZeRO-2 to offload partitioned optimizer states and gradients to CPU memory. ZeRO-2 Offload offers the full memory advantage of ZeRO-2 even on a single GPU, while at the same time offering great scalability of ZeRO-2 on multi-GPU setup. DeepSpeed library has been offering ZeRO-2 Offload since Sept 2020. For details, please see below:
700
-
701
- With today’s release of ZeRO-3 Offload, we are adding support for partitioning and offloading parameters in addition to optimizer states and gradients partitioning already supported by ZeRO-2 Offload in DeepSpeed. With parameter partitioning ZeRO-3 Offload implements the full set of features in the three stages of ZeRO, that allows for a linear growth in model size with the number of GPUs. In addition, ZeRO-3 Offload can also optionally offload all these model states to CPU to further reduce GPU memory consumption, leveraging both CPU and GPU to maximize memory and compute efficiency of the entire system.
702
-
703
- We believe ZeRO-3 Offload offers a massive leap for large model training, in three regards:
704
-
705
- i) Unprecedented model scale,
706
-
707
- ii) Ease of supporting very-large models, and
708
-
709
- iii) Achieving excellent training efficiency.
710
-
711
- Unlike ZeRO-2 and ZeRO-Offload where the parameters have to fit in the memory of a single GPU, ZeRO-3 Offload can partition the parameters across GPUs, and offload them to CPU, supporting model sizes that are much larger than the memory on a single GPU. Furthermore, ZeRO-3 Offload goes beyond the state-of-the-art hybrid 3D-parallelism (data, model and pipeline parallelism combined). While 3D Parallelism is limited by the aggregate GPU memory, ZeRO-3 Offload can exploit both GPU and CPU memory, the latter of which is much larger and cheaper compared to GPU memory. This allows ZeRO-3 Offload to train larger model sizes with the given GPU and CPU resources than any other currently available technology.
712
-
713
- Model Scale on Single GPU: ZeRO-3 Offload can train models with over 40B parameters efficiently on a single GPU (e.g., 32GB V100 GPU + 1.5TB CPU memory). This is 3x larger than what is possible with ZeRO-2 Offload, the current state-of-the art.
714
-
715
- Model Scale on Multi-GPUs: With ZeRO-3 Offload you can train a trillion and two trillion parameter models on NVIDIA 32GB V100 DGX-2 cluster with 256 GPUs and 512 GPUs, respectively. In contrast, the state-of-art 3D Parallelism requires 800 GPUs, and 1600 GPUs, respectively, to fit the same sized models. This represents a 3x reduction in GPUs required to fit models with over a trillion parameters.
716
-
717
- From a system perspective, training models with hundreds of billions and trillions of parameters is extremely challenging. Data parallelism cannot scale the model size much further beyond a billion parameters, model parallelism (with tensor slicing) cannot be used to scale model size efficiently beyond a single node boundary due to massive communication overheads, and pipeline parallelism cannot scale beyond the number of layers available in a model, which limits both the model size and the number of GPUs that it can scale to.
718
-
719
- The only existing parallel technology available that can scale to over a trillion parameters on massively parallel GPU clusters is the 3D parallelism that combines data, model and pipeline parallelism in complex ways. While such a system can be very efficient, it requires major model code refactoring from data scientists to split the model into load balanced pipeline stages. This also makes 3D parallelism inflexible in the type of models that it can support, since models with complex dependency graphs cannot be easily converted into a load balanced pipeline.
720
-
721
- ZeRO-3 Offload address these challenges in two ways:
722
-
723
- i) With ground-breaking memory efficiency, ZeRO-3 and ZeRO-3 Offload are the only DL parallel technology that can efficiently scale to over a trillion parameters by itself, without requiring a hybrid parallelism strategy, greatly simplifying the system stack for DL training.
724
-
725
- ii) ZeRO-3 Offload requires virtually no model refactoring from model scientists, liberating data scientists to scale up complex models to hundreds of billions to trillions of parameters.
726
-
727
- High-performance per-GPU throughput on multiple nodes: ZeRO-3 Offload offers excellent training efficiency for multi-billion and trillion parameter models on multiple nodes. It achieves a sustained throughput of up to 50 Tflops per GPU running on 32 DGX2 nodes comprising 512 NVIDIA V100 GPUs (see Figure 2). In comparison, the standard data parallel training with PyTorch can only achieve 30 TFlops per GPU for a 1.2B parameter model, the largest model that can be trained using data parallelism alone.
728
-
729
- Figure 2. ZeRO-3 Offload: Multi-billion and trillion parameter model throughput on 512 V100 GPUs
730
-
731
- ZeRO-3 Offload obtains high efficiency despite the 50% communication overhead of ZeRO Stage 3 compared to standard data parallel training for a fixed batch size. This is made possible through a communication overlap centric design and implementation, which allows ZeRO-3 Offload to hide nearly all of the communication volume with computation, while taking advantage of a larger batch size for improved efficiency resulting from better GPU memory efficiency.
732
-
733
- Efficient multi-billion parameter model training on a single GPU: ZeRO-3 Offload further democratizes AI by enabling efficient training of multi-billion parameter models on a single GPU. For single GPU training, ZeRO-3 Offload provides benefits over ZeRO-2 Offload along two dimensions. First, ZeRO-3 Offload increases the size of models trainable on a single V100 from 13B to 40B. Second, for ZeRO-3 Offload provides speedups (e.g., 2.3X for 13B) compared to ZeRO-2 Offload for model sizes trainable by both solutions. These results are summarized in Figure 3.
734
-
735
- Figure 3. Multi-billion parameter model training on one V100 GPU
736
-
737
- Super-Linear scalability across GPUs: Additionally, ZeRO-3 Offload also preserves the super-linear scalability characteristics that we have demonstrated with all our previous ZeRO technologies (ZeRO Stage 1, ZeRO Stage 2 and ZeRO Offload). ZeRO-3 Offload can exploit the aggregate PCI-E bandwidth between GPU and CPU across all the GPUs in multi-GPU training configuration, and at the same time, it can also exploit the aggregate CPU compute across all the nodes. As a result, the CPU-GPU-CPU communication time as well as the optimizer update time decreases linearly with number of GPUs and nodes, respectively, allowing ZeRO-3 Offload to exhibit super-linear scaling (see Figure 4).
738
-
739
- Figure 4. ZeRO-3 Offload Superlinear Scalability for a 200B parameter model.
740
-
741
- As with many other existing DeepSpeed features, once the user model has been converted to use DeepSpeed, enabling ZeRO-3 Offload is as easy as turning on a couple of flags in DeepSpeed Config file. Supporting advanced features like weight sharing, or enabling extremely large models that requires to be partitioned across GPUs/nodes to fit in GPU/CPU memory, can be done with just a couple of additional lines of code change using the ZeRO-3 Offload API.
742
-
743
- If you are already a DeepSpeed user, you can find our detailed tutorial on ZeRO-3 Offload below. If you are new to DeepSpeed, we recommend that you start at the getting started page before trying out our ZeRO-3 Offload Tutorial.
744
-
745
- DeepSpeed: Getting Started Page
746
-
747
- ZeRO-3 Offload Documentation, Tutorial
748
-
749
- The DeepSpeed Team is very excited to share ZeRO-3 Offload with the DL community.
750
-
751
- Updated: March 7, 2021
752
-
753
- ---
754
-
755
- ## DeepSpeed: Advancing MoE inference and training to power next-generation AI scale
756
-
757
- **URL:** https://www.deepspeed.ai/2022/01/18/moe-inference.html
758
-
759
- **Contents:**
760
- - DeepSpeed: Advancing MoE inference and training to power next-generation AI scale
761
- - Contents
762
-
763
- Updated: January 18, 2022
764
-
765
- ---
766
-
767
- ## Azure empowers easy-to-use, high-performance, and hyperscale model training using DeepSpeed
768
-
769
- **URL:** https://www.deepspeed.ai/2022/07/25/deepspeed-azure.html
770
-
771
- **Contents:**
772
- - Azure empowers easy-to-use, high-performance, and hyperscale model training using DeepSpeed
773
- - Contents
774
- - Introduction
775
- - Making distributed training faster and easier on Azure using DeepSpeed
776
- - Key Performance Benefits
777
- - Experimental Setup
778
- - Hardware (Azure instances)
779
- - Training setup using AzureML
780
- - Training setup using Azure VMSS
781
- - Performance Evaluation on Various Model Configurations
782
-
783
- Large-scale transformer-based deep learning models trained on large amounts of data have shown great results in recent years in several cognitive tasks and are behind new products and features that augment human capabilities. These models have grown several orders of magnitude in size during the last five years. Starting from a few million parameters of the original transformer model all the way to the latest 530 billion-parameter Megatron-Turing model as shown in Figure 1. There is a growing need for customers to train and fine tune large models at an unprecedented scale.
784
-
785
- Figure 1: Landscape of large models and hardware capabilities
786
-
787
- To train these models, users needed to set up and maintain a complex distributed training infrastructure that usually required several manual and error-prone steps. These lead to a subpar experience both in terms of usability and performance. We recently announced how we are making great strides to simplify this and enable easy-to-use and high-performance training at 1K+ GPU scale on Azure.
788
-
789
- In this extended post, we share the details of how DeepSpeed users can train trillion-parameter models with a new easy-to-use, streamlined, scalable, and high-performance distributed training experience on Azure. We also share details of the experimental setup, model configurations, additional performance trends, and guide our users on how to run these experiments in their own environments.
790
-
791
- We compare the existing manual and error-prone workflow with our proposed easy-to-use workflow for DeepSpeed on Azure in Figure 2. Customers can now use easy-to-use training pipelines to launch training jobs at scale. The new workflow reduces the number of steps from 11 to just 1 if users rely on the recommended AzureML recipes.
792
-
793
- Figure 2: An easy-to-use and streamlined distributed training experience with DeepSpeed on Azure
794
-
795
- For users who have custom environments built using Azure VMs or Azure VMSS, only two steps are needed:
796
-
797
- We already shared a summary of our key performance results in the Azure announcement. We enable the capability to train 2x larger model sizes (2 trillion vs. 1 trillion parameters), scale to 2x more GPUs (1024 vs. 512), and offer up to 1.8x higher compute throughput/GPU (150 TFLOPs vs. 81 TFLOPs) compared to other cloud providers.
798
-
799
- DeepSpeed on Azure offers near-linear scalability both in terms of increase in model size as well as increase in number of GPUs. As shown in Figure 3a, together with the DeepSpeed ZeRO-3, its novel CPU offloading capabilities, and a high-performance Azure stack powered by InfiniBand interconnects and A100 GPUs, we were able to maintain an efficient throughput/GPU (>157 TFLOPs) in a near-linear fashion as the model size increases from 175 billion parameters to 2 trillion parameters. On the other hand, for a given model size, e.g., 175B, we achieve near-linear scaling as we increase the number of GPUs from 128 all the way to 1024 as shown in Figure 3b. The key takeaway is that Azure and DeepSpeed together are breaking the GPU memory wall and enabling our customers to easily and efficiently train trillion-parameter models at scale.
800
-
801
- Figure 3: (a) Near-perfect throughput/GPU as we increase the model size from 175 billion to 2 trillion parameters (BS/GPU=8). (b) Near-perfect performance scaling with the increase in number of GPU devices for the 175B model (BS/GPU=16). The sequence length is 1024 for both cases.
802
-
803
- We share the details of our experimental setup and some of the best practices we followed. The users can either directly use them to reproduce our results or modify them to fit their own setup in terms of model scale as well as the scale of Azure hardware being provisioned.
804
-
805
- We used NDm A100 v4-series instances in our experiments. Each instance includes two socket AMD EPYC 7V12 64-Core CPUs, 1.7TB main memory and eight A100 80GB GPUs. The system has a balanced PCIe topology connecting 4 GPU devices to each CPU socket. Each GPU within the VM is provided with its own dedicated, topology-agnostic 200 Gb/s NVIDIA Mellanox HDR InfiniBand connection providing an accelerated 200 Gbps high speed fabric. The DeepSpeed library exploits offload capabilities where the activation and optimizer states are allocated in the main memory. Hence, 1.7TB memory capacity per node helps us to scale to large model sizes.
806
-
807
- Users can directly use the AzureML studio and use our published recipes to run experiments without any additional setup. This is the easiest and recommended way of running experiments on Azure.
808
-
809
- Existing VMSS customers and others who have custom Azure VM based environments can follow the setup as follows. The scripts to make these steps easy will be released in the coming weeks. A cluster is created using Azure Virtual Machine Scale Sets (VMSS) to provision the desired number of compute nodes running the new Azure HPAI VM image specialized for extreme-scale deep learning applications using the software stack listed in Table 1.
810
-
811
- Table 1: Detailed version information of the software packages in the Azure HPC VM image
812
-
813
- Users can create a VMSS with up to 600 VM instances enabling up to 4,800 A100 GPUs. In addition to the VMSS for the compute nodes, we provision a distinct login node using an inexpensive D4s v4 (or similar) instance with 4-core Intel VCPU, running the same image, for compiling, launching, and monitoring jobs. The login node, compute nodes, and a shared storage filesystem are grouped within an Azure Virtual Network (vnet) allowing VMs to connect to each other over SSH and to shared NFS volume shown in Figure 4.
814
-
815
- Figure 4: Organization of our VMSS-based experimental setup
816
-
817
- We ran our experiments with four different model sizes – 175B, 530B, 1T, and 2T – using the configurations shown in Table 2.
818
-
819
- Table 2: Model configuration
820
-
821
- For each of these configurations, we report peak throughput of the system using TFLOPs/GPU as the main performance metric. To calculate TFLOPs, we use the formula used by the Megatron paper as shown below.
822
-
823
- FLOPs/GPU = 96 * B * s * l * h2 * (1 + s/6h + V/(16*l*h))
824
-
825
- B is batch size, s is sequence length, l is the number of layers, h is hidden size, and V is vocabulary size.
826
-
827
- Figures 5a and 5b show the results of 175B model with sequence length 512 and 1024, respectively. We only scale to 512 GPUs for seq-length 512 as adding more GPUs shows similar performance. On the other hand, with sequence length 1024, we saw linear performance increase to 1024 GPUs. Overall, the peak throughput of 204.49 TFLOPs/GPU was achieved on 256 GPUs with a micro batch size of 32 and sequence length of 512.
828
-
829
- Figure 5: Performance characteristics of 175B model on 512 and 1K GPUs respectively. The colored columns signify different micro batch sizes.
830
-
831
- Next, we report the 530B model scaling. Previous results on the 530B MT-NLG model using DeepSpeed and Megatron-LM on 280 DGX A100 servers on the Selene supercomputer showed the peak throughput of 126 TFLOPS/GPU. However, we were able to surpass that throughput and achieved up to 171.37 TFLOPs/GPU on 128 NDm A100 v4-series A100 systems (i.e., 1024 GPUs) as shown in Figure 6.
832
-
833
- The benefit of this 530B model is its simpler parallelization configuration as there is no tensor/pipeline parallelism. With ZeRO powered data parallelism, there are fewer heuristics required to optimally configure the distributed model. In addition, the consistent steady state performance of more than 140 TFLOPs/GPU for micro batch sizes >1 demonstrates a robust software and hardware platform.
834
-
835
- Figure 6: Throughput achieved with a 530B parameter model on 512 and 1024 GPUs for micro-batch sizes per GPU of 1, 2, 4, and 8, with sequence length 1,024.
836
-
837
- The 1T parameter model contains 128 layers with 160 attention heads. Training such an extreme-scale model is not an easy task. Figure 7 shows the throughput achieved for each of the model configurations we explored on 512 and 1024 GPUs. Peak throughput achieved was 165.36 TFLOPs/GPU for micro batch size of 8 across 1024 GPUs and the model reached steady state performance within the first 3-4 iterations.
838
-
839
- Figure 7: Performance characteristics of 1T parameter model on 512 and 1024 GPUs with 1, 2, 4, and 8 micro batch sizes, with sequence length 1,024.
840
-
841
- The 2T parameter model consists of 160 layers, 32k hidden dimension, and 128 attention heads. Given the large size of the model and the significant time required on 1024 GPUs, we limited our benchmark runs for the 2T model to a batch size of 8 per GPU with a sequence length of 1024. We were able to achieve 157 TFLOPs/GPU on 1,024 GPUs.
842
-
843
- We recognize that DeepSpeed users are diverse and have different environments. In this tutorial, our focus is on making things simpler for users who plan to run large model training experiments on Azure.
844
-
845
- The easiest way to do model training on Azure is via the Azure ML recipes. The job submission and data preparation scripts have been made available here. Users simply need to setup their Azure ML workspace following the guide and submit experiment using the aml_submit.py file.
846
-
847
- Some users have customized environments built on top of Azure VMs and VMSS based clusters. To simplify training on such setups, we are working on an easy-to-use cluster setup script that will be published in the next few weeks. If you already have a cluster setup running, you can use the azure recipes for the 175B and the 1T model. The recipes can easily be modified to train other model configurations.
848
-
849
- This blog post was written by the DeepSpeed team in collaboration with the AzureML and the AzureHPC team. We would like to acknowledge several individuals who made this work possible:
850
-
851
- Updated: July 25, 2022
852
-
853
- ---
854
-
855
- ## DeepSpeed Data Efficiency: A composable library that makes better use of data, increases training efficiency, and improves model quality
856
-
857
- **URL:** https://www.deepspeed.ai/2022/12/11/data-efficiency.html
858
-
859
- **Contents:**
860
- - DeepSpeed Data Efficiency: A composable library that makes better use of data, increases training efficiency, and improves model quality
861
- - Contents
862
- - Efficient Data Sampling via Curriculum Learning
863
- - Motivation
864
- - Design
865
- - Evaluation Results
866
- - Efficient Data Routing via Random Layerwise Token Dropping
867
- - Motivation
868
- - Design
869
- - Evaluation Results
870
-
871
- Recently, large-scale deep learning models are empowering us to achieve more in many ways, such as improving programming efficiency by code generation and providing art inspiration by text-to-image generation. To enable these services and keep improving the quality, deep learning model architecture evolves rapidly, and the model size is also growing at a tremendous speed. For example, from GPT to GPT-3 the model size increased 1500x in 2 years. The increasing model size leads to unprecedented training cost, making it challenging for many AI practitioners to train their own models. On the other hand, a less-emphasized perspective is that data scale is actually increasing at a similar speed as model scale, and the training cost is proportional to both of them. In Figure 1 below we plot the model and data scales of several representative language models in the last 5 years. From the oldest model on the left to the newest models on the right, both the model and data scales increase at similar speed. This demonstrates the importance of improving data efficiency: achieve same model quality with less data and reduced training cost, or achieve better model quality with the same amount of data and similar training cost.
872
-
873
- Figure 1: Model scale (number of parameters) and data scale (number of tokens consumed during training) of representative language models in the last 5 years.
874
-
875
- There are two popular research directions among existing data efficiency techniques: Data sampling techniques aim to improve the convergence speed by sampling the most suitable next data batch from the whole data pool; Data routing techniques aim to reduce the computation by routing each data to only a subset of the model components. These techniques improve data and training efficiency, but existing solutions on them have limitations on extensibility, flexibility, and composability. They are commonly designed for specific training tasks, making them hard to be extended with customized strategies and making them less flexible to be applied on diverse workloads from different users. Furthermore, different techniques are implemented separately, making it challenging to compose multiple solutions to further improve data and training efficiency.
876
-
877
- To address these challenges, we, the DeepSpeed team as part of Microsoft’s AI at Scale initiative, are proud to announce DeepSpeed Data Efficiency Library – a composable framework that makes better use of data, increases training efficiency, and improves model quality. DeepSpeed Data Efficiency takes extensibility, flexibility, and composability into consideration, and it specifically demonstrates the following innovations:
878
-
879
- Efficient data sampling via curriculum learning. Curriculum learning (CL) improves data efficiency by sampling from easier data. We present a general curriculum learning library which enables users to employ curriculum learning to their models at maximum extensibility: users can easily analyze, index, and sample their training data based on various customizable strategies. Using this library, we were able to explore different CL strategies for GPT-3 and BERT pretraining and identify the best solution that provides up to 1.5x data saving while still maintaining similar model quality.
880
-
881
- Efficient data routing via random layerwise token dropping. We present a novel data routing technique called random layerwise token dropping (random-LTD) to skip the computation of a subset of the input tokens at all middle layers. Random-LTD employs a simple yet effective routing strategy and requires minimal model architecture change. It is flexible to apply random-LTD to various tasks (GPT-3/BERT pretraining and GPT/ViT finetuning), and we achieve great data efficiency improvement (up to 1.5x data saving while still maintaining the model quality).
882
-
883
- Seamlessly composing multiple methods. The proposed DeepSpeed Data Efficiency framework seamlessly composes the curriculum learning and random-LTD techniques, and only requires minimal changes on the user code side. Furthermore, by composing both methods we can achieve even better data and training efficiency: for GPT-3 1.3B pretraining, we achieve 2x data and 2x time savings together with better or similar model quality compared to the baseline training. When using the same amount of data, our approach further improves the model quality over the baseline. Users can also extend and contribute to the library by adding additional data efficiency techniques to compose together.
884
-
885
- Each of these advances is explored further in the blog post below. For more about the technical details, please read our papers, “Random-LTD: Random and Layerwise Token Dropping Brings Efficient Training for Large-scale Transformers” which describes the random-LTD technique, and “DeepSpeed Data Efficiency: Improving Deep Learning Model Quality and Training Efficiency via Efficient Data Sampling and Routing” which describes the curriculum learning technique and overall DeepSpeed Data Efficiency framework.
886
-
887
- Curriculum learning aims to improve training convergence speed by presenting relatively easier or simpler examples earlier during training. Building a curriculum learning solution usually requires two components: the difficulty metric (i.e., how to quantify the difficulty of each data sample) and the pacing function (i.e., how to decide the curriculum difficulty range when sampling next training data batch). Curriculum learning has been successfully applied to various training tasks, and last year we also released a specific curriculum learning technique (sequence length warmup) for GPT-style model pretraining (see technical details in our paper “The Stability-Efficiency Dilemma: Investigating Sequence Length Warmup for Training GPT Models” published in NeurIPS 2022). However, one common limitation among existing works is that there does not exist a generalized and extensible curriculum learning library, which allows practitioners to easily apply custom curriculum difficulty metrics, the combination of metrics, and pacing functions.
888
-
889
- To solve the limitation of existing solutions, we design and implement a general curriculum learning library emphasizing the extensibility. It consists of three components as shown in Figure 2 below (top part). First, we use a data analyzer to perform the offline CPU-only data analysis which indexes the whole data pool based on any difficulty metric such as the sequence length, the vocabulary rarity, or anything defined by user. Next, during training, the curriculum scheduler determines the difficulty threshold for the current step based on a pacing function such as linear, rooted, or any strategy provided by users. Then the data sampler will sample the data with desired difficulty from the indexed data pool. Overall, this general implementation would enable users to explore curriculum learning on their workloads with maximum customizability (more technical details in our DeepSpeed Data Efficiency paper).
890
-
891
- Figure 2: Design of the DeepSpeed Data Efficiency framework.
892
-
893
- Using this general and extensible curriculum learning solution for GPT-3 and BERT-Large model pretraining, we are able to easily analyze and index the huge training data based on up to 7 difficulty metrics and enable better data and training efficiency. For GPT-3 pretraining, our solution with the best difficulty metric (combination of truncation-based sequence length and vocabulary rarity) achieves 1.5x data and training cost saving while still maintaining model quality as baseline (Table 1 Case (8) vs. (1)). For BERT-Large pretraining, our solution with the best difficulty metric (vocabulary rarity) achieves 1.5x saving while still maintaining model quality (Table 2 Case (8) vs. (1)). On the other hand, our solutions can further improve model quality when using the same amount of data as baseline (Table 1 Case (2) to (6), Table 2 Case (2) to (6)).
894
-
895
- Table 1: GPT-3 1.3B pretraining data consumption and average evaluation accuracy on 19 tasks.
896
-
897
- Table 2: BERT-Large pretraining data consumption and average GLUE finetuning score on 8 tasks.
898
-
899
- Standard data routing usually feeds the full images/sequences into all layers of a model. However, this process may not be optimal for training efficiency since some parts of an image (or words of a sentence) do not require a frequent feature update. As such, the token dropping method has been proposed, which is illustrated in Figure 3 (b) below, to skip the compute of some tokens/words (i.e., G-2 tokens in Figure 3 (b)) of a sentence in order to save the compute cost.
900
-
901
- Although existing methods show promising results, they also exhibit several caveats: (1) most works solely focus on BERT (encoder-only on text data) pretraining and do not include decoder pretraining and/or other modalities (e.g., images); (2) the ability to skip layers is limited, which bounds the total amount of compute saving. By analyzing existing methods, we found out the potential main issue that limits their skipping and coverage abilities is the loss of attention mechanism for G-2 tokens for all skipped layers, since multi-head attention focuses on different tokens at different layer depths and the attention map aligns with the dependency relation most strongly in the middle of transformer architectures.
902
-
903
- To resolve this main issue, we propose random-LTD, a random and layerwise token dropping mechanism, which processes only a subset of tokens among the entire data batch for all middle layers in order to save compute cost (see more details in our Random-LTD paper). As such, each token rarely bypasses all middle layers and its dependency with other tokens can be captured by the model. The illustration of random-LTD compared to baseline is shown in Figure 3 below, where random-LTD splits the input tokens into two groups and only the first group involves the compute.
904
-
905
- Figure 3: Comparison between baseline, existing token dropping methods, and random-LTD. Note that for random-LTD, only part of the inputs (Group 1) is used for Layer i.
906
-
907
- Random-LTD is simple yet very effective. Particularly, compared to other existing token dropping methods, random-LTD (1) does a purely random selection for each layer for two different groups, as such we do not require any expert design for the selection criterion; (2) is able to apply to all middle layers to achieve better saving ratio; (3) demonstrates great generalizability for both encoder and decoder models; and (4) is easy to use without much modeling change. These advantages enable maximum flexibility when applying random-LTD to various workloads.
908
-
909
- Thanks to its great flexibility, we were able to apply random-LTD method to broader applications, including BERT and GPT pretraining as well as ViT and GPT finetuning tasks. For all cases, random-LTD achieves similar model quality as baseline while using less data, and/or achieve better model quality while using the same amount of data (Table 3 to 6). For GPT-3 and BERT-Large pretraining, random-LTD achieves 1.5-2x data saving while still maintaining the same model quality. For GPT-3 we also tested random-LTD with full data which further improves the model quality compared to baseline.
910
-
911
- Table 3: GPT-3 1.3B pretraining data consumption and average evaluation accuracy on 19 tasks.
912
-
913
- Table 4: BERT-Large pretraining data consumption and average GLUE finetuning score on 8 tasks.
914
-
915
- Table 5: Finetuning result of ViT on ImageNet.
916
-
917
- Table 6: GPT-2 350M finetuning result on the PTB task.
918
-
919
- The curriculum learning and random-LTD techniques are complementary. Inside DeepSpeed Data Efficiency framework, we seamlessly compose the two techniques as shown in Figure 2 above, where curriculum learning helps to sample the next data batch and random-LTD helps to decide how to route each sampled data inside the model. DeepSpeed Data Efficiency solves several complexities when composing the two techniques so that users can easily apply each technique or both to their training pipeline. The composability of DeepSpeed Data Efficiency also applies to data sampling and routing techniques in general, so that it provides a platform to implement and compose additional data efficiency techniques.
920
-
921
- The composed DeepSpeed Data Efficiency solution leverages both data efficiency techniques and achieves even better data and training efficiency. Take the GPT-3 pretraining task as an example, composing CL and random-LTD, with 100% data, leads to the best model quality in our experiments (Table 7 Case (1) to (4)). When pretraining with 50% data, the baseline training results in worse zero-shot and 10-shot evaluation accuracy, and using either CL or random-LTD can only recover part of the 10-shot accuracy loss. On the other hand, the composed data efficiency solution achieves the same or better accuracy results as baseline with 100% data, demonstrating a 2x data and 2x time saving (Case (5) to (8)). Similar benefit such as 2x data saving was also observed when applying our solution to BERT pretraining.
922
-
923
- Table 7: GPT-3 1.3B pretraining data/time consumption and average evaluation accuracy on 19 tasks.
924
-
925
- We are very excited to share DeepSpeed Data Efficiency library with the community and improve it with your feedback. Please find the code, tutorial, and documents at the DeepSpeed GitHub, and website. And for more technical details please read our Random-LTD paper and DeepSpeed Data Efficiency paper. We believe that our composable library and novel data efficiency techniques will help users reduce training cost while maintaining model quality or achieve better quality under similar cost. And we hope DeepSpeed Data Efficiency could become a platform that motivates and accelerates future research on deep learning data efficiency.
926
-
927
- Updated: December 11, 2022
928
-
929
- ---
930
-
931
- ## DeepSpeed Inference: Multi-GPU inference with customized inference kernels and quantization support
932
-
933
- **URL:** https://www.deepspeed.ai/2021/03/15/inference-kernel-optimization.html
934
-
935
- **Contents:**
936
- - DeepSpeed Inference: Multi-GPU inference with customized inference kernels and quantization support
937
- - Contents
938
- - Multi-GPU Inference with Adaptive Parallelism
939
- - Customized Inference Kernels for Boosted Compute Efficiency of Transformer Blocks
940
- - Kernel-Fusion
941
- - Seamless pipeline from training to inference with automatic kernel-injection
942
- - Flexible quantization support
943
- - Performance results
944
-
945
- While DeepSpeed supports training advanced large-scale models, using these trained models in the desired application scenarios is still challenging due to three major limitations in existing inference solutions: 1) lack of support for multi-GPU inference to fit large models and meet latency requirements, 2) limited GPU kernel performance when running inference with small batch sizes, and 3) difficulties in exploiting quantization, which includes both quantizing the model to reduce the model size and latency as well as supporting high-performance inference of quantized models without specialized hardware.
946
-
947
- To handle these challenges, we introduce DeepSpeed Inference, which seamlessly adds high-performance inference support to large models trained in DeepSpeed with three key features: inference-adapted parallelism for multi-GPU inference, inference-optimized kernels tuned for small batch sizes, and flexible support for quantize-aware training and inference kernels for quantized models.
948
-
949
- Parallelism is an effective approach to fit large models and reduce per-device memory consumption for both training and inference. However, simply applying training parallelism choices and degree to inference does not work well. The MP and PP configuration is normally set during the model training, apart from the data parallelism (DP), based on the memory footprint and computation style, and resource budget. On one hand, inference computation intrinsically requires less memory, so it can afford a larger partition per device. It helps reduce the degree of parallelism needed for model deployment. On the other hand, optimizing latency or meeting latency requirements is often a first-class citizen in inference while training optimizes throughput.
950
-
951
- To obtain desired latency, DeepSpeed Inference automatically adapts MP as an effective approach to reduce model latency, and its parallelism degree is often determined first. With MP, we can split the mode and parallelize computational operations across multiple devices (GPUs) to reduce latency, but it reduces computation granularity and increases communication that may hurt throughput. Once the latency target has been met, DeepSpeed can apply pipeline parallelism to maximize the throughput. Overall, DeepSpeed Inference supports flexible adaptation of both parallelism approach and degree choices from training to inference, minimizing latency while saving deployment costs.
952
-
953
- To achieve high compute efficiency, DeepSpeed-inference offers inference kernels tailored for Transformer blocks through operator fusion, taking model-parallelism for multi-GPU into account. The main difference between our kernel-fusion scheme and similar approaches is that we not only fuse element-wise operations (such as bias-add, residual, and activation function), but also merge the General matrix multiply (GeMM) operations with other operations. To do this, we design an efficient implementation for the vector-matrix or skinny matrix-matrix multiplication that allows us to fuse more operations at the reduction boundary of GeMM operations.
954
-
955
- We take two main policies for fusing operations: 1) keeping the access-pattern of inputs and outputs intact throughout the sequence of operations fused together; 2) fusing operations at each all-reduce boundary. The first policy ensures that different thread-blocks won’t encounter transferring data between Streaming-Multiprocessors (SMs). This is due to no straight-forward communication among SMs other than using the main memory which adds the block-synching overhead because of non-deterministic behavior of memory access. The reason behind the second policy is that we cannot continue the execution unless the partial results are reduced among the model-parallel GPUs.
956
-
957
- Figure 1: Transformer Layer with Megatron-style model-parallelism all-reduce components. The figure illustrates the parts of layer fused together with broken lines (width of line shows the fusion depth).
958
-
959
- Figure 1 shows the different components of a Transformer layer, and the groups of operations considered for fusion in our inference optimization. We also consider the NVIDIA Megatron-LM style of parallelism that partitions attention (Attn) and feed-forward (FF) blocks across multiple GPUs. Thus, we include the two all-reduce operations that reduce the results among parallel GPUs after Attn and FF blocks. As Figure 1 shows, we fuse the operations inside a Transformer layer at four main regions:
960
-
961
- To fuse these operations, we exploit shared-memory as an intermediate cache for transferring data between reduction operations used in layer-norm and GeMM, and the element-wise operations. Moreover, we use the warp-level instructions to communicate data between threads when reducing partial computations. In addition, we use a new schedule for GeMM operations, which allows for fusing as many operations as needed for the third kernel-fusion. We also combine the GeMMs for the attention computation in the second kernel-fusion, by using an implicit matrix transformation in order to reduce the memory pressure. Compared to the unfused computation style using cuBLAS GeMM, we improve the performance by 1.5x, 2.9x. 3x, and 1.2x for all these kernel-fusions, respectively.
962
-
963
- To run the model in Inference mode, DeepSpeed simply requires the location of the model checkpoints and the desired parallelism configuration, i.e., MP/PP degree. DeepSpeed Inference kernels can also be enabled for many well-known model architectures such as HuggingFace (Bert and GPT-2) or Megatron GPT-based models using a pre-defined policy map that maps the original parameters to the parameters in the inference kernels. For other transformer-based models, user can specify their own policy map. Note that DS-Inference can run independent of the training pipeline as long as it receives all model checkpoints, and the DeepSpeed Transformer kernels for inference can be injected into any Transformer model if the right mapping policy is defined. For more information on how to enable Transformer inference kernel as well as specifying parallelism, please refer to out inference tutorial.
964
-
965
- To further reduce the inference cost for large-scale models, we created the DeepSpeed Quantization Toolkit, supporting flexible quantize-aware training and high-performance kernels for quantized inference.
966
-
967
- For training, we introduce a novel approach called Mixture of Quantization (MoQ), which is inspired by mixed-precision training while seamlessly applying quantization. With MoQ, we can control the precision of the model by simulating the impact of quantization when updating the parameters at each step of training. Moreover, it supports flexible quantization policies and schedules—we find that by dynamically adjusting the number of quantization bits during training, the final quantized model provides higher accuracy under the same compression ratio. To adapt to different tasks, MoQ can also leverage the second order information of models to detect their sensitivity to precision and adjust the quantization schedule and target accordingly.
968
-
969
- To maximize the performance gains from the quantization model, we provide inference kernels tailored for quantized models that reduce latency through optimizing data movement but do not require specialized hardware. Finally, our toolkit does not require any code changes on the client side, making it easy to use.
970
-
971
- Boosting throughput and reducing inference cost. Figure 3 shows the inference throughput per GPU for the three model sizes corresponding to the three Transformer networks, GPT-2, Turing-NLG, and GPT-3. DeepSpeed Inference increases in per-GPU throughput by 2 to 4 times when using the same precision of FP16 as the baseline. By enabling quantization, we boost throughput further. We reach a throughput improvement of 3x for GPT-2, 5x for Turing-NLG, and 3x for a model that is similar in characteristics and size to GPT-3, which directly translates to 3–5x inference cost reduction on serving these large models. In addition, we achieve these throughput and cost improvements without compromising latency as shown in Figure 5.
972
-
973
- Figure 3: Inference throughput for different model sizes. DeepSpeed Inference achieves 3x to 5x higher throughput than baseline.
974
-
975
- One source of inference cost reduction is through reducing the number of GPUs for hosting large models as shown in Figure 4. The optimized GPU resources comes from 1) using inference-adapted parallelism, allowing users to adjust the model and pipeline parallelism degree from the trained model checkpoints, and 2) shrinking model memory footprint by half with INT8 quantization. As shown in this figure, we use 2x less GPUs to run inference for the 17B model size by adapting the parallelism. Together with INT8 quantization through DeepSpeed MoQ, we use 4x and 2x fewer GPUs for 17B and 175B sizes respectively.
976
-
977
- Figure 4: Number of GPUs used for running inference on the different model sizes shown in Figure 4.
978
-
979
- Reducing inference latency. For the application scenarios where inference latency is critical, we can increase model parallelism degree in DeepSpeed Inference to reduce inference latency further. As Figure 5 depicts, we can reduce the latency by 2.3x compared to PyTorch as we increase the model-parallelism size to 4. Furthermore, we can still have high latency improvement with a fewer number of GPUs by adapting the parallelism at inference and using MoQ to quantize the model. We obtain 1.3x and 1.9x speedups while using 4x and 2x lower resources than baseline, respectively.
980
-
981
- For the application scenarios where inference latency is critical, we can increase model parallelism degree in DeepSpeed Inference to reduce inference latency further. As Figure 5 depicts, we can reduce the latency by 2.3x compared to PyTorch as we increase the model-parallelism size to 4. Furthermore, we can still have high latency improvement with a fewer number of GPUs by adapting the parallelism at inference and using MoQ to quantize the model. We obtain 1.3x and 1.9x speedups while using 4x and 2x lower resources than baseline, respectively.
982
-
983
- Figure 5. Inference latency for the 17B model using different parallelism configuration to optimize latency.
984
-
985
- Updated: March 15, 2021
986
-
987
- ---
988
-
989
- ## Inference Overview and Features
990
-
991
- **URL:** https://www.deepspeed.ai/inference/
992
-
993
- **Contents:**
994
- - Inference Overview and Features
995
- - Contents
996
-
997
- DeepSpeed-Inference v2 is here and it’s called DeepSpeed-FastGen! For the best performance, latest features, and newest model support please see our DeepSpeed-FastGen release blog!
998
-
999
- DeepSpeed-Inference introduces several features to efficiently serve transformer-based PyTorch models. It supports model parallelism (MP) to fit large models that would otherwise not fit in GPU memory. Even for smaller models, MP can be used to reduce latency for inference. To further reduce latency and cost, we introduce inference-customized kernels. Finally, we propose a novel approach to quantize models, called MoQ, to both shrink the model and reduce the inference cost at production. For more details on the inference related optimizations in DeepSpeed, please refer to our blog post.
1000
-
1001
- DeepSpeed provides a seamless inference mode for compatible transformer based models trained using DeepSpeed, Megatron, and HuggingFace, meaning that we don’t require any change on the modeling side such as exporting the model or creating a different checkpoint from your trained checkpoints. To run inference on multi-GPU for compatible models, provide the model parallelism degree and the checkpoint information or the model which is already loaded from a checkpoint, and DeepSpeed will do the rest. It will automatically partition the model as necessary, inject compatible high performance kernels into your model and manage the inter-gpu communication. For list of compatible models please see here.
1002
-
1003
- To get started with DeepSpeed-Inference, please checkout our tutorial.
1004
-
1005
- ---
1006
-
1007
- ## Mixture-of-Quantization: A novel quantization approach for reducing model size with minimal accuracy impact
1008
-
1009
- **URL:** https://www.deepspeed.ai/2021/05/04/MoQ.html
1010
-
1011
- **Contents:**
1012
- - Mixture-of-Quantization: A novel quantization approach for reducing model size with minimal accuracy impact
1013
- - Contents
1014
- - A unified suite for quantization-aware training and inference
1015
- - Quantization methodology
1016
- - Quantized Inference Kernels
1017
- - Ease of use
1018
- - Improving quantization accuracy.
1019
-
1020
- Running large-scale models on multi-GPU might help reduce latency but increases the deployment cost significantly, especially as the model size grows bigger. To mitigate this issue, we resort to model compression techniques and introduce a new methodology that quantizes Transformer networks with a minimal impact on accuracy. Our technique achieves similar or better performance thanFP16 models through customized inference kernels on lower or equal number of GPUs.
1021
-
1022
- Our scheme is flexible in the sense that it provides users the ability to experiment with any quantization configuration, such as the target number of bits used for quantization precision, and the scheduling by which the model gets quantized during training. Furthermore, we combine both the FP16 and quantized precision as a mixed-precision mechanism to smooth the transition from a high to low precision. Finally, we use the second-order gradient (eigenvalue) of the parameters to adjust the quantization schedule during training.
1023
-
1024
- There are two main approaches of applying quantization: offline quantization on the trained model and quantization-aware training (QAT) that reduces the data-precision during training. Unlike the former scheme, QAT gets the model trained by taking the impact of precision loss into account during the training optimization. This will result in significant improvement of the quantized model accuracy. MoQ is designed on top QAT approach, with the difference that we use a mixture of precisions to train the model toward target quantization, as well as defining a scheduling for reducing the precision.
1025
-
1026
- All existing QAT approaches quantize the model with a certain precision (number of bits) from the beginning of training until completion. However, even by using a relatively high quantization precision (8-bit), there will be some drop in model accuracy, which might not be acceptable for some downstream tasks. For instance, the Q8BERT work tries QAT for the BERT network, which results in good accuracy for some tasks while others (like SQuAD) lose 0.8% in the F1 score. Other techniques, such as Q-BERT, use grouped quantization with a large grouping size (128) when quantizing a parameter matrix to gain higher accuracy, but they are still inferior to the baseline.
1027
-
1028
- Here, we present MoQ as a flexible solution for linear quantization that allows users to define a schedule as the model trains. Similar to iterative pruning to inject sparsity, we start quantization from a higher precision (16-bit quantization or FP16) and gradually reduce the quantization bits or the mixed-precision ratio for the FP16 part until reaching a target precision (8-bit). To control the precision transition, we define a hyperparameter, called quantization period, that indicates when the precision reduction should happen. We observe that by using such a schedule, we get the closest accuracy to the baseline. Note that in order to reach a certain precision, we need to define the starting bits and period in a way that within the number of samples to train, the model eventually gets quantized using the target number of bits. Please refer to the quantization tutorial for more information.
1029
-
1030
- In order to dynamically adjust quantization precision, we employ eigenvalue as a metric that shows the sensitivity of training to the precision change. Eigenvalue has been previously used (Q-BERT) for quantization to choose the precision bits on different parts of the network. To combine this with MoQ, we cluster the eigenvalues into several regions based on their absolute values and tune the quantization period for each region accordingly, the higher the magnitude of eigenvalue, the larger the factor and the slower the precision decreases.
1031
-
1032
- Figure 1. Quantization scheduling of one of the GLUE tasks (QNLI), using the eigenvalue of different layers. Different colors show the layers from 0 to 11 for Bert-Base.
1033
-
1034
- Figure 1 shows the result of combining eigenvalue with MoQ for a 12-layer Bert Base model. As we see, the first few layers (0-4) tend to be more sensitive to reduced precision than the last layers, as their quantization period is an order of magnitude larger than the rest. Another observation from this figure is that the neighbor layers reduce the precision in the same way. For instance, layers 9, 10, and 11 on the left chart, and layers 0 and 4 and 1 and 3 on the right chart of Figure 1 get similar schedule. This is due to having similar eigenvalues for these layers throughout the training.
1035
-
1036
- Figure 2: Mixed-precision quantization for the QNLI using target quantization period as 4 bits.
1037
-
1038
- Figure 2 shows another mixed-precision quantization that sets target bits as 4, however the quantization period keeps updated through the eigenvalues of each layer. As we see, the end quantization bits are different for all layers. The first layers still get to 8-bit quantization as the training samples is not enough to decrease the quantization bits. On the other hand, the last layers keep reducing the precision. We finally reduce the average precision to 6 bits for the entire network while maintaining the accuracy of the model (0.3% drop in accuracy).
1039
-
1040
- Figure 3: Mixed-precision quantization with MoQ for Bert SQuAD plus.
1041
-
1042
- As another example, we use eigenvalue-based MoQ to quantize Bert-Large for SQuAD finetuning. Figure 3 shows the number of bits we get to at the end of finetuning on each layer. Here, we see slightly different precision spectrum compared to BertBase on GLUE tasks. As the figure shows, we can reduce the precision on the first few layers more aggressively than the middle ones. Also, the last few layers can tolerate very low precision similar to the beginning layers. This way of quantization finally results in 90.56 F1 Score which is pretty similar to the baseline.
1043
-
1044
- By using other quantization methodologies, after the model is quantized, it can only have performance benefit if there is hardware support for integer-based operations. For this reason, the inputs and output of all GeMM operations need to be quantized. However, since the range of input may vary request by request, finding a range of data for each input at inference time is challenging. On the other hand, using a static range for all inputs can impact the inference accuracy.
1045
-
1046
- To alleviate this problem, we introduce inference custom kernels that neither require the hardware support nor the input quantization. These kernels read quantized parameters and dequantize them on-the-fly and use the floating-point units of GPU cores for the GeMM operations. The main benefit of using these kernels is that they reduce the memory footprint required to load a model so that we can run inference on fewer number of GPUs, while improving the performance by saving the memory bandwidth required to run the inference on GPU.
1047
-
1048
- Regarding the quantization implementation, we use different algorithms to quantize a value based on the range of data and the rounding policy. We support both symmetric and asymmetric quantization as the two mostly used schemes. We applied both techniques for QAT and see very similar results, however since symmetric approach is simpler to implement, we implement our inference kernels based on that. Regarding the rounding, we support stochastic rounding as another option besides the normal rounding. We have seen that for reducing the precision to as low as 4-bit or lower, stochastic rounding is more helpful as it has an unbiased random behavior during training.
1049
-
1050
- For enabling quantization through Deepspeed, we only need to pass the scheduling through a JSON configuration file. To add the impact of quantization, we quantize and dequantize the parameters just before they are updated in the optimizer. Thus, we do not incur any change on the modeling side to quantize a model. Instead, we simulate the quantization impact by lowering the precision of data saved in FP16 format. By using this kind of implementation, we have the full flexibility of changing the precision using the training characteristics such as number of steps, and eigenvalue of the parameters and the original FP16 data format. As shown in this blog post, we can improve the quality of a quantized model by adaptively changing the scheduling of the quantization throughout training. For more information on how to use MoQ scheme, please look at our quantization tutorial.
1051
-
1052
- To show how our quantization scheme preserves accuracy, we have experimented MoQ on several tasks and networks: GLUE tasks on Bert-Base and SQuAD on Bert-Large. Table 1 shows the accuracy results for the baseline without quantization (w/o Quant), basic quantization without using any scheduling during training (Basic Quant), and our MoQ scheme. Without using any scheduling, the accuracy for 8-bit quantization is often inferior to the baseline, and in this workload, it suffers from a drop of 1.02 point in accuracy (ACC). In contrast, MoQ powers 8-bit quantization to obtain comparable accuracy as the FP16 baseline, even with a slightly higher ACC, demonstrating the effectiveness of our quantization approach.
1053
-
1054
- ---
1055
-
1056
- ## DeepSpeed: Accelerating large-scale model inference and training via system optimizations and compression
1057
-
1058
- **URL:** https://www.deepspeed.ai/2021/05/14/inference-release.html
1059
-
1060
- **Contents:**
1061
- - DeepSpeed: Accelerating large-scale model inference and training via system optimizations and compression
1062
- - Contents
1063
-
1064
- Updated: May 14, 2021
1065
-
1066
- ---
1067
-
1068
- ## Autotuning: Automatically discover the optimal DeepSpeed configuration that delivers good training speed
1069
-
1070
- **URL:** https://www.deepspeed.ai/2021/11/16/autotuning.html
1071
-
1072
- **Contents:**
1073
- - Autotuning: Automatically discover the optimal DeepSpeed configuration that delivers good training speed
1074
-
1075
- We introduce a new feature called Autotuning to automatically discover the optimal DeepSpeed configuration that delivers good training speed. One pain point in model training is to figure out good performance-relevant configurations such as micro-batch size to fully utilize the hardware and achieve a high throughput number. This configuration exploring process is commonly done manually but is important since model training is repeated many times and benefits from using a good configuration. Not only is the hand-tuning process time-consuming, but the outcome is hardware-dependent. This means that a good configuration on one hardware might not be the best on another different hardware. The user thus has to hand tune the configuration again. With DeepSpeed, there are more configuration parameters that could potentially affect the training speed, thus making it more tedious to manually tune the configuration.
1076
-
1077
- The DeepSpeed Autotuner mitigates this pain point and automatically discovers the optimal DeepSpeed configuration that delivers good training speed. It not only reduces the time and resources users spend on tuning, but also can discover configurations better than hand-tuned methods. DeepSpeedExamples would demonstrate the effectiveness of autotuning across different models.
1078
-
1079
- Updated: November 16, 2021
1080
-
1081
- ---
1082
-
1083
- ## Contributing
1084
-
1085
- **URL:** https://www.deepspeed.ai/contributing/
1086
-
1087
- **Contents:**
1088
- - Contributing
1089
- - Contents
1090
- - Prerequisites
1091
- - Testing
1092
- - Unit Tests
1093
- - Model Tests
1094
- - Contributor License Agreement
1095
- - Code of Conduct
1096
-
1097
- DeepSpeed welcomes your contributions!
1098
-
1099
- DeepSpeed uses pre-commit to ensure that formatting is consistent across DeepSpeed. First, ensure that pre-commit is installed from either installing DeepSpeed or pip install pre-commit. Next, the pre-commit hooks must be installed once before commits can be made:
1100
-
1101
- Afterwards, our suite of formatting tests run automatically before each git commit. You can also run these manually:
1102
-
1103
- If a formatting test fails, it will fix the modified code in place and abort the git commit. After looking over the changes, you can git add <modified files> and then repeat the previous git commit command.
1104
-
1105
- DeepSpeed tracks two types of tests: unit tests and more costly model convergence tests. The model convergence tests train DeepSpeedExamples and measure end-to-end convergence and related metrics. Unit tests are found in tests/unit/ and the model convergence tests are found in tests/model/.
1106
-
1107
- PyTest is used to execute tests. PyTest can be installed from PyPI via pip install pytest. Simply invoke pytest --forked to run the unit tests:
1108
-
1109
- You can also provide the -v flag to pytest to see additional information about the tests. Note that pytest-forked and the --forked flag are required to test CUDA functionality in distributed tests.
1110
-
1111
- Model tests require four GPUs and training data downloaded for DeepSpeedExamples.
1112
-
1113
- To execute model tests, first install DeepSpeed. The DeepSpeedExamples repository is cloned as part of this process. Next, execute the model test driver:
1114
-
1115
- Note that the --forked flag is not necessary for the model tests.
1116
-
1117
- This project welcomes contributions and suggestions. Most contributions require you to agree to a Contributor License Agreement (CLA) declaring that you have the right to, and actually do, grant us the rights to use your contribution. For details, visit https://cla.opensource.microsoft.com.
1118
-
1119
- When you submit a pull request, a CLA bot will automatically determine whether you need to provide a CLA and decorate the PR appropriately (e.g., status check, comment). Simply follow the instructions provided by the bot. You will only need to do this once across all repos using our CLA.
1120
-
1121
- This project has adopted the Microsoft Open Source Code of Conduct. For more information see the Code of Conduct FAQ or contact opencode@microsoft.com with any additional questions or comments.
1122
-
1123
- **Examples:**
1124
-
1125
- Example 1 (unknown):
1126
- ```unknown
1127
- pre-commit install
1128
- ```
1129
-
1130
- Example 2 (unknown):
1131
- ```unknown
1132
- pre-commit run --all-files
1133
- ```
1134
-
1135
- Example 3 (unknown):
1136
- ```unknown
1137
- pytest --forked tests/unit/
1138
- ```
1139
-
1140
- Example 4 (unknown):
1141
- ```unknown
1142
- cd tests/model/
1143
- pytest run_sanity_check.py
1144
- ```
1145
-
1146
- ---
1147
-
1148
- ## Latest News
1149
-
1150
- **URL:** https://www.deepspeed.ai
1151
-
1152
- **Contents:**
1153
- - Latest News
1154
- - Contents
1155
- - Extreme Speed and Scale for DL Training
1156
- - DeepSpeed Adoption
1157
- - Contributing
1158
- - Contributor License Agreement
1159
- - Code of Conduct
1160
- - Publications
1161
- - Videos
1162
-
1163
- [2025/10] SuperOffload: Unleashing the Power of Large-Scale LLM Training on Superchips
1164
-
1165
- [2025/10] Study of ZenFlow and ZeRO offload performance with DeepSpeed CPU core binding
1166
-
1167
- [2025/08] ZenFlow: Stall-Free Offloading Engine for LLM Training
1168
-
1169
- [2025/06] Arctic Long Sequence Training (ALST) with DeepSpeed: Scalable And Efficient Training For Multi-Million Token Sequences
1170
-
1171
- [2025/06] DeepNVMe: Affordable I/O scaling for Deep Learning Applications
1172
-
1173
- DeepSpeed enabled the world’s most powerful language models (at the time of this writing) such as MT-530B and BLOOM. DeepSpeed offers a confluence of system innovations, that has made large scale DL training effective, and efficient, greatly improved ease of use, and redefined the DL training landscape in terms of scale that is possible. These innovations include ZeRO, 3D-Parallelism, DeepSpeed-MoE, ZeRO-Infinity, etc.
1174
-
1175
- DeepSpeed has been used to train many different large-scale models. Below is a list of several examples that we are aware of (if you’d like to include your model please submit a PR):
1176
-
1177
- DeepSpeed has been integrated with several different popular open-source DL frameworks such as:
1178
-
1179
- DeepSpeed is an integral part of Microsoft’s AI at Scale initiative to enable next-generation AI capabilities at scale.
1180
-
1181
- DeepSpeed welcomes your contributions! Please see our contributing guide for more details on formatting, testing, etc.
1182
-
1183
- This project welcomes contributions and suggestions. Most contributions require you to agree to a Contributor License Agreement (CLA) declaring that you have the right to, and actually do, grant us the rights to use your contribution. For details, visit https://cla.opensource.microsoft.com.
1184
-
1185
- When you submit a pull request, a CLA bot will automatically determine whether you need to provide a CLA and decorate the PR appropriately (e.g., status check, comment). Simply follow the instructions provided by the bot. You will only need to do this once across all repos using our CLA.
1186
-
1187
- This project has adopted the Microsoft Open Source Code of Conduct. For more information see the Code of Conduct FAQ or contact opencode@microsoft.com with any additional questions or comments.
1188
-
1189
- Xinyu Lian, Sam Ade Jacobs, Lev Kurilenko, Masahiro Tanaka, Stas Bekman, Olatunji Ruwase, Minjia Zhang. (2024) Universal Checkpointing: Efficient and Flexible Checkpointing for Large Scale Distributed Training arXiv:2406.18820
1190
-
1191
- ---