@stephen-lord/docs 1.0.0 → 1.0.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (1612) hide show
  1. package/dist-docs/Agent-Skills-for-Context-Engineering/.claude-plugin/marketplace.json +68 -0
  2. package/dist-docs/Agent-Skills-for-Context-Engineering/.cursorindexingignore +3 -0
  3. package/dist-docs/Agent-Skills-for-Context-Engineering/CONTRIBUTING.md +78 -0
  4. package/dist-docs/Agent-Skills-for-Context-Engineering/LICENSE +22 -0
  5. package/dist-docs/Agent-Skills-for-Context-Engineering/README.md +228 -0
  6. package/dist-docs/Agent-Skills-for-Context-Engineering/SKILL.md +104 -0
  7. package/dist-docs/Agent-Skills-for-Context-Engineering/docs/agentskills.md +1264 -0
  8. package/dist-docs/Agent-Skills-for-Context-Engineering/docs/blogs.md +1230 -0
  9. package/dist-docs/Agent-Skills-for-Context-Engineering/docs/claude_research.md +85 -0
  10. package/dist-docs/Agent-Skills-for-Context-Engineering/docs/compression.md +298 -0
  11. package/dist-docs/Agent-Skills-for-Context-Engineering/docs/gemini_research.md +22 -0
  12. package/dist-docs/Agent-Skills-for-Context-Engineering/docs/hncapsule.md +92 -0
  13. package/dist-docs/Agent-Skills-for-Context-Engineering/docs/netflix_context.md +10 -0
  14. package/dist-docs/Agent-Skills-for-Context-Engineering/docs/vercel_tool.md +140 -0
  15. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/book-sft-pipeline/README.md +78 -0
  16. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/book-sft-pipeline/SKILL.md +380 -0
  17. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/book-sft-pipeline/examples/gertrude-stein/README.md +168 -0
  18. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/book-sft-pipeline/examples/gertrude-stein/dataset_sample.jsonl +5 -0
  19. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/book-sft-pipeline/examples/gertrude-stein/pangram/Screenshot 2025-12-27 at 3.05.04/342/200/257AM.png +0 -0
  20. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/book-sft-pipeline/examples/gertrude-stein/pangram/Screenshot 2025-12-27 at 3.05.36/342/200/257AM.png +0 -0
  21. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/book-sft-pipeline/examples/gertrude-stein/pangram/Screenshot 2025-12-27 at 3.07.18/342/200/257AM.png +0 -0
  22. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/book-sft-pipeline/examples/gertrude-stein/sample_outputs.md +63 -0
  23. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/book-sft-pipeline/examples/gertrude-stein/training_config.json +80 -0
  24. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/book-sft-pipeline/references/segmentation-strategies.md +324 -0
  25. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/book-sft-pipeline/references/tinker-format.md +211 -0
  26. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/book-sft-pipeline/references/tinker.txt +3176 -0
  27. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/book-sft-pipeline/scripts/pipeline_example.py +187 -0
  28. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/AGENT.md +35 -0
  29. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/HOW-SKILLS-BUILT-THIS.md +407 -0
  30. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/README.md +209 -0
  31. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/SKILL.md +203 -0
  32. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/SKILLS-MAPPING.md +219 -0
  33. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/agents/AGENTS.md +82 -0
  34. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/agents/scripts/content_ideas.py +132 -0
  35. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/agents/scripts/idea_to_draft.py +181 -0
  36. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/agents/scripts/stale_contacts.py +139 -0
  37. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/agents/scripts/weekly_review.py +121 -0
  38. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/content/CONTENT.md +88 -0
  39. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/content/calendar.md +108 -0
  40. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/content/engagement.jsonl +2 -0
  41. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/content/ideas.jsonl +2 -0
  42. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/content/posts.jsonl +2 -0
  43. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/content/templates/linkedin-post.md +102 -0
  44. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/content/templates/newsletter.md +92 -0
  45. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/content/templates/thread.md +73 -0
  46. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/examples/content-workflow.md +204 -0
  47. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/examples/meeting-prep.md +243 -0
  48. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/identity/IDENTITY.md +46 -0
  49. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/identity/bio-variants.md +101 -0
  50. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/identity/brand.md +165 -0
  51. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/identity/prompts/content-generation.xml +46 -0
  52. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/identity/prompts/reply-generator.xml +40 -0
  53. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/identity/values.yaml +60 -0
  54. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/identity/voice.md +165 -0
  55. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/knowledge/KNOWLEDGE.md +85 -0
  56. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/knowledge/bookmarks.jsonl +2 -0
  57. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/knowledge/competitors.md +117 -0
  58. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/knowledge/learning.yaml +74 -0
  59. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/knowledge/research/_template.md +79 -0
  60. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/network/NETWORK.md +110 -0
  61. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/network/circles.yaml +80 -0
  62. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/network/contacts.jsonl +2 -0
  63. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/network/interactions.jsonl +2 -0
  64. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/network/intros.md +92 -0
  65. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/operations/OPERATIONS.md +75 -0
  66. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/operations/goals.yaml +83 -0
  67. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/operations/meetings.jsonl +2 -0
  68. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/operations/metrics.jsonl +2 -0
  69. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/operations/reviews/_weekly_template.md +114 -0
  70. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/operations/todos.md +76 -0
  71. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/package.json +41 -0
  72. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/references/file-formats.md +386 -0
  73. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/digital-brain-skill/scripts/install.sh +79 -0
  74. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/README.md +620 -0
  75. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/SKILL.md +221 -0
  76. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/docs/agentthinking.md +63 -0
  77. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/docs/interleavedthinking.md +610 -0
  78. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/docs/m2-1.md +224 -0
  79. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/examples/01_basic_capture.py +76 -0
  80. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/examples/02_tool_usage.py +187 -0
  81. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/examples/03_full_optimization.py +1222 -0
  82. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/generated_skills/comprehensive-research-agent/SKILL.md +90 -0
  83. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/generated_skills/comprehensive-research-agent/references/optimization_summary.json +9 -0
  84. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/generated_skills/comprehensive-research-agent/references/optimized_prompt.txt +1 -0
  85. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/generated_skills/comprehensive-research-agent/references/patterns_found.json +205 -0
  86. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/final_prompt.txt +67 -0
  87. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_1/analysis.txt +48 -0
  88. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_1/optimization.txt +15 -0
  89. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_1/optimized_prompt.txt +1 -0
  90. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_1/trace.txt +178 -0
  91. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_10/analysis.txt +47 -0
  92. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_10/trace.txt +162 -0
  93. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_2/analysis.txt +48 -0
  94. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_2/optimization.txt +130 -0
  95. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_2/optimized_prompt.txt +72 -0
  96. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_2/trace.txt +156 -0
  97. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_3/analysis.txt +46 -0
  98. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_3/optimization.txt +147 -0
  99. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_3/optimized_prompt.txt +84 -0
  100. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_3/trace.txt +159 -0
  101. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_4/analysis.txt +46 -0
  102. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_4/optimization.txt +134 -0
  103. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_4/optimized_prompt.txt +67 -0
  104. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_4/trace.txt +165 -0
  105. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_5/analysis.txt +50 -0
  106. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_5/optimization.txt +135 -0
  107. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_5/optimized_prompt.txt +71 -0
  108. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_5/trace.txt +146 -0
  109. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_6/analysis.txt +15 -0
  110. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_6/optimization.txt +15 -0
  111. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_6/optimized_prompt.txt +1 -0
  112. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_6/trace.txt +147 -0
  113. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_7/analysis.txt +46 -0
  114. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_7/optimization.txt +103 -0
  115. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_7/optimized_prompt.txt +45 -0
  116. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_7/trace.txt +134 -0
  117. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_8/analysis.txt +47 -0
  118. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_8/optimization.txt +114 -0
  119. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_8/optimized_prompt.txt +60 -0
  120. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_8/trace.txt +135 -0
  121. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_9/analysis.txt +44 -0
  122. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_9/optimization.txt +106 -0
  123. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_9/optimized_prompt.txt +51 -0
  124. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/iteration_9/trace.txt +170 -0
  125. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/optimization_artifacts/summary.json +11 -0
  126. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/pyproject.toml +70 -0
  127. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/reasoning_trace_optimizer/__init__.py +53 -0
  128. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/reasoning_trace_optimizer/analyzer.py +465 -0
  129. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/reasoning_trace_optimizer/capture.py +417 -0
  130. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/reasoning_trace_optimizer/cli.py +271 -0
  131. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/reasoning_trace_optimizer/loop.py +468 -0
  132. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/reasoning_trace_optimizer/models.py +193 -0
  133. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/reasoning_trace_optimizer/optimizer.py +449 -0
  134. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/reasoning_trace_optimizer/skill_generator.py +502 -0
  135. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/tests/__init__.py +1 -0
  136. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/interleaved_thinking/tests/test_models.py +144 -0
  137. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/.prettierrc +8 -0
  138. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/CONTRIBUTING.md +78 -0
  139. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/LICENSE +21 -0
  140. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/README.md +659 -0
  141. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/agents/evaluator-agent/evaluator-agent.md +177 -0
  142. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/agents/index.md +114 -0
  143. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/agents/orchestrator-agent/orchestrator-agent.md +205 -0
  144. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/agents/research-agent/research-agent.md +183 -0
  145. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/env.example +6 -0
  146. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/eslint.config.js +18 -0
  147. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/examples/basic-evaluation.ts +89 -0
  148. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/examples/full-evaluation-workflow.ts +136 -0
  149. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/examples/generate-rubric.ts +67 -0
  150. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/examples/pairwise-comparison.ts +97 -0
  151. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/package.json +79 -0
  152. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/prompts/agent-system/orchestrator-prompt.md +197 -0
  153. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/prompts/evaluation/direct-scoring-prompt.md +153 -0
  154. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/prompts/evaluation/pairwise-comparison-prompt.md +200 -0
  155. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/prompts/index.md +138 -0
  156. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/prompts/research/research-synthesis-prompt.md +171 -0
  157. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/skills/context-fundamentals/context-fundamentals.md +114 -0
  158. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/skills/index.md +79 -0
  159. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/skills/llm-evaluator/llm-evaluator.md +77 -0
  160. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/skills/tool-design/tool-design.md +198 -0
  161. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/src/agents/evaluator.ts +112 -0
  162. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/src/agents/index.ts +3 -0
  163. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/src/config/index.ts +18 -0
  164. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/src/index.ts +19 -0
  165. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/src/tools/evaluation/direct-score.ts +164 -0
  166. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/src/tools/evaluation/generate-rubric.ts +161 -0
  167. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/src/tools/evaluation/index.ts +9 -0
  168. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/src/tools/evaluation/pairwise-compare.ts +255 -0
  169. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/tests/evaluation.test.ts +233 -0
  170. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/tests/setup.ts +27 -0
  171. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/tests/skills.test.ts +213 -0
  172. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/tools/evaluation/direct-score.md +159 -0
  173. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/tools/evaluation/generate-rubric.md +189 -0
  174. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/tools/evaluation/pairwise-compare.md +182 -0
  175. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/tools/index.md +141 -0
  176. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/tools/orchestration/delegate-to-agent.md +171 -0
  177. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/tools/research/read-url.md +162 -0
  178. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/tools/research/web-search.md +128 -0
  179. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/tsconfig.json +26 -0
  180. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/llm-as-judge-skills/vitest.config.ts +20 -0
  181. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/x-to-book-system/PRD.md +644 -0
  182. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/x-to-book-system/README.md +181 -0
  183. package/dist-docs/Agent-Skills-for-Context-Engineering/examples/x-to-book-system/SKILLS-MAPPING.md +187 -0
  184. package/dist-docs/Agent-Skills-for-Context-Engineering/researcher/example_output.md +75 -0
  185. package/dist-docs/Agent-Skills-for-Context-Engineering/researcher/llm-as-a-judge.md +362 -0
  186. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/advanced-evaluation/SKILL.md +454 -0
  187. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/advanced-evaluation/references/bias-mitigation.md +288 -0
  188. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/advanced-evaluation/references/implementation-patterns.md +315 -0
  189. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/advanced-evaluation/references/metrics-guide.md +331 -0
  190. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/advanced-evaluation/scripts/evaluation_example.py +337 -0
  191. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/bdi-mental-states/SKILL.md +295 -0
  192. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/bdi-mental-states/references/bdi-ontology-core.md +207 -0
  193. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/bdi-mental-states/references/framework-integration.md +582 -0
  194. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/bdi-mental-states/references/rdf-examples.md +315 -0
  195. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/bdi-mental-states/references/sparql-competency.md +420 -0
  196. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/context-compression/SKILL.md +265 -0
  197. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/context-compression/references/evaluation-framework.md +213 -0
  198. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/context-compression/scripts/compression_evaluator.py +658 -0
  199. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/context-degradation/SKILL.md +231 -0
  200. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/context-degradation/references/patterns.md +314 -0
  201. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/context-degradation/scripts/degradation_detector.py +419 -0
  202. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/context-fundamentals/SKILL.md +185 -0
  203. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/context-fundamentals/references/context-components.md +283 -0
  204. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/context-fundamentals/scripts/context_manager.py +370 -0
  205. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/context-optimization/SKILL.md +179 -0
  206. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/context-optimization/references/optimization_techniques.md +272 -0
  207. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/context-optimization/scripts/compaction.py +379 -0
  208. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/evaluation/SKILL.md +231 -0
  209. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/evaluation/references/metrics.md +339 -0
  210. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/evaluation/scripts/evaluator.py +474 -0
  211. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/filesystem-context/SKILL.md +321 -0
  212. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/filesystem-context/references/implementation-patterns.md +549 -0
  213. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/filesystem-context/scripts/filesystem_context.py +353 -0
  214. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/hosted-agents/SKILL.md +279 -0
  215. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/hosted-agents/references/infrastructure-patterns.md +700 -0
  216. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/hosted-agents/scripts/sandbox_manager.py +495 -0
  217. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/memory-systems/SKILL.md +221 -0
  218. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/memory-systems/references/implementation.md +458 -0
  219. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/memory-systems/scripts/memory_store.py +396 -0
  220. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/multi-agent-patterns/SKILL.md +255 -0
  221. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/multi-agent-patterns/references/frameworks.md +433 -0
  222. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/multi-agent-patterns/scripts/coordination.py +439 -0
  223. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/project-development/SKILL.md +342 -0
  224. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/project-development/references/case-studies.md +388 -0
  225. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/project-development/references/pipeline-patterns.md +610 -0
  226. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/project-development/scripts/pipeline_template.py +677 -0
  227. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/tool-design/SKILL.md +311 -0
  228. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/tool-design/references/architectural_reduction.md +210 -0
  229. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/tool-design/references/best_practices.md +176 -0
  230. package/dist-docs/Agent-Skills-for-Context-Engineering/skills/tool-design/scripts/description_generator.py +237 -0
  231. package/dist-docs/Agent-Skills-for-Context-Engineering/template/SKILL.md +98 -0
  232. package/dist-repo/fastapi-backend/.claude/settings.local.json +29 -0
  233. package/dist-repo/fastapi-backend/.claude/skills/tortoise-orm-best-practice/SKILL.md +690 -0
  234. package/dist-repo/fastapi-backend/.editorconfig +12 -0
  235. package/dist-repo/fastapi-backend/.env.example +36 -0
  236. package/dist-repo/fastapi-backend/.python-version +1 -0
  237. package/dist-repo/fastapi-backend/LICENSE +21 -0
  238. package/dist-repo/fastapi-backend/README.md +14 -0
  239. package/dist-repo/fastapi-backend/aerich_config.py +9 -0
  240. package/dist-repo/fastapi-backend/docker-compose.yml +43 -0
  241. package/dist-repo/fastapi-backend/main.py +6 -0
  242. package/dist-repo/fastapi-backend/migrations/models/0_20260118170116_init.py +279 -0
  243. package/dist-repo/fastapi-backend/migrations/models/1_20260118205316_update.py +74 -0
  244. package/dist-repo/fastapi-backend/migrations/models/2_20260118211346_update.py +75 -0
  245. package/dist-repo/fastapi-backend/migrations/models/3_20260118211430_update.py +80 -0
  246. package/dist-repo/fastapi-backend/migrations/models/4_20260118211452_update.py +80 -0
  247. package/dist-repo/fastapi-backend/migrations/models/5_20260118211514_update.py +89 -0
  248. package/dist-repo/fastapi-backend/pyproject.toml +206 -0
  249. package/dist-repo/fastapi-backend/scripts/generate_api_module.py +74 -0
  250. package/dist-repo/fastapi-backend/scripts/lint.py +251 -0
  251. package/dist-repo/fastapi-backend/src/__init__.py +83 -0
  252. package/dist-repo/fastapi-backend/src/api/__init__.py +25 -0
  253. package/dist-repo/fastapi-backend/src/api/api/__init__.py +8 -0
  254. package/dist-repo/fastapi-backend/src/api/api/controller.py +70 -0
  255. package/dist-repo/fastapi-backend/src/api/api/model.py +14 -0
  256. package/dist-repo/fastapi-backend/src/api/api/repository.py +65 -0
  257. package/dist-repo/fastapi-backend/src/api/api/schema.py +39 -0
  258. package/dist-repo/fastapi-backend/src/api/api/service.py +0 -0
  259. package/dist-repo/fastapi-backend/src/api/auditlog/__init__.py +8 -0
  260. package/dist-repo/fastapi-backend/src/api/auditlog/controller.py +54 -0
  261. package/dist-repo/fastapi-backend/src/api/auditlog/model.py +31 -0
  262. package/dist-repo/fastapi-backend/src/api/auditlog/repository.py +0 -0
  263. package/dist-repo/fastapi-backend/src/api/auditlog/schema.py +21 -0
  264. package/dist-repo/fastapi-backend/src/api/auditlog/service.py +0 -0
  265. package/dist-repo/fastapi-backend/src/api/base/__init__.py +8 -0
  266. package/dist-repo/fastapi-backend/src/api/base/controller.py +179 -0
  267. package/dist-repo/fastapi-backend/src/api/base/service.py +259 -0
  268. package/dist-repo/fastapi-backend/src/api/dept/__init__.py +8 -0
  269. package/dist-repo/fastapi-backend/src/api/dept/controller.py +49 -0
  270. package/dist-repo/fastapi-backend/src/api/dept/model.py +24 -0
  271. package/dist-repo/fastapi-backend/src/api/dept/repository.py +109 -0
  272. package/dist-repo/fastapi-backend/src/api/dept/schema.py +37 -0
  273. package/dist-repo/fastapi-backend/src/api/dept/service.py +0 -0
  274. package/dist-repo/fastapi-backend/src/api/file/__init__.py +8 -0
  275. package/dist-repo/fastapi-backend/src/api/file/controller.py +100 -0
  276. package/dist-repo/fastapi-backend/src/api/file/model.py +22 -0
  277. package/dist-repo/fastapi-backend/src/api/file/repository.py +134 -0
  278. package/dist-repo/fastapi-backend/src/api/file/schema.py +36 -0
  279. package/dist-repo/fastapi-backend/src/api/file/service.py +474 -0
  280. package/dist-repo/fastapi-backend/src/api/menu/__init__.py +8 -0
  281. package/dist-repo/fastapi-backend/src/api/menu/controller.py +70 -0
  282. package/dist-repo/fastapi-backend/src/api/menu/model.py +22 -0
  283. package/dist-repo/fastapi-backend/src/api/menu/repository.py +17 -0
  284. package/dist-repo/fastapi-backend/src/api/menu/schema.py +57 -0
  285. package/dist-repo/fastapi-backend/src/api/menu/service.py +0 -0
  286. package/dist-repo/fastapi-backend/src/api/role/__init__.py +8 -0
  287. package/dist-repo/fastapi-backend/src/api/role/controller.py +82 -0
  288. package/dist-repo/fastapi-backend/src/api/role/model.py +19 -0
  289. package/dist-repo/fastapi-backend/src/api/role/repository.py +44 -0
  290. package/dist-repo/fastapi-backend/src/api/role/schema.py +40 -0
  291. package/dist-repo/fastapi-backend/src/api/user/__init__.py +8 -0
  292. package/dist-repo/fastapi-backend/src/api/user/controller.py +59 -0
  293. package/dist-repo/fastapi-backend/src/api/user/model.py +39 -0
  294. package/dist-repo/fastapi-backend/src/api/user/repository.py +89 -0
  295. package/dist-repo/fastapi-backend/src/api/user/schema.py +127 -0
  296. package/dist-repo/fastapi-backend/src/api/user/service.py +180 -0
  297. package/dist-repo/fastapi-backend/src/common/__init__.py +0 -0
  298. package/dist-repo/fastapi-backend/src/common/enums/__init__.py +3 -0
  299. package/dist-repo/fastapi-backend/src/common/enums/request.py +11 -0
  300. package/dist-repo/fastapi-backend/src/common/models/__init__.py +5 -0
  301. package/dist-repo/fastapi-backend/src/common/models/base_model.py +53 -0
  302. package/dist-repo/fastapi-backend/src/common/models/soft_delete_model.py +5 -0
  303. package/dist-repo/fastapi-backend/src/common/models/timestamp_model.py +6 -0
  304. package/dist-repo/fastapi-backend/src/common/models/uuid_model.py +5 -0
  305. package/dist-repo/fastapi-backend/src/common/repositories/__init__.py +4 -0
  306. package/dist-repo/fastapi-backend/src/common/repositories/crud_base_repository.py +288 -0
  307. package/dist-repo/fastapi-backend/src/common/repositories/crud_batch_repository.py +127 -0
  308. package/dist-repo/fastapi-backend/src/common/schemas/__init__.py +19 -0
  309. package/dist-repo/fastapi-backend/src/common/schemas/auth_schema.py +33 -0
  310. package/dist-repo/fastapi-backend/src/common/schemas/response_schema.py +38 -0
  311. package/dist-repo/fastapi-backend/src/common/services/__init__.py +0 -0
  312. package/dist-repo/fastapi-backend/src/config/__init__.py +3 -0
  313. package/dist-repo/fastapi-backend/src/config/settings.py +206 -0
  314. package/dist-repo/fastapi-backend/src/core/__init__.py +0 -0
  315. package/dist-repo/fastapi-backend/src/core/contexts/__init__.py +0 -0
  316. package/dist-repo/fastapi-backend/src/core/contexts/constant.py +14 -0
  317. package/dist-repo/fastapi-backend/src/core/contexts/log_context.py +57 -0
  318. package/dist-repo/fastapi-backend/src/core/contexts/request_log_context.py +28 -0
  319. package/dist-repo/fastapi-backend/src/core/contexts/utils.py +13 -0
  320. package/dist-repo/fastapi-backend/src/core/dependencies/__init__.py +0 -0
  321. package/dist-repo/fastapi-backend/src/core/dependencies/auth_dependency.py +210 -0
  322. package/dist-repo/fastapi-backend/src/core/exceptions/__init__.py +15 -0
  323. package/dist-repo/fastapi-backend/src/core/exceptions/does_not_exist_handle.py +16 -0
  324. package/dist-repo/fastapi-backend/src/core/exceptions/http_exc_handle.py +12 -0
  325. package/dist-repo/fastapi-backend/src/core/exceptions/integrity_handle.py +16 -0
  326. package/dist-repo/fastapi-backend/src/core/exceptions/request_validation_handle.py +18 -0
  327. package/dist-repo/fastapi-backend/src/core/exceptions/response_validation_handle.py +18 -0
  328. package/dist-repo/fastapi-backend/src/core/exceptions/setting_not_found.py +2 -0
  329. package/dist-repo/fastapi-backend/src/core/init_app.py +330 -0
  330. package/dist-repo/fastapi-backend/src/core/middlewares/__init__.py +13 -0
  331. package/dist-repo/fastapi-backend/src/core/middlewares/background_task_middleware.py +11 -0
  332. package/dist-repo/fastapi-backend/src/core/middlewares/http_audit_log_middleware.py +175 -0
  333. package/dist-repo/fastapi-backend/src/core/middlewares/request_logging_middleware.py +67 -0
  334. package/dist-repo/fastapi-backend/src/core/middlewares/security_headers_middleware.py +48 -0
  335. package/dist-repo/fastapi-backend/src/core/middlewares/simple_base_middleware.py +24 -0
  336. package/dist-repo/fastapi-backend/src/core/tasks/__init__.py +0 -0
  337. package/dist-repo/fastapi-backend/src/core/tasks/bgtask.py +31 -0
  338. package/dist-repo/fastapi-backend/src/utils/__init__.py +0 -0
  339. package/dist-repo/fastapi-backend/src/utils/cache.py +224 -0
  340. package/dist-repo/fastapi-backend/src/utils/jwt.py +74 -0
  341. package/dist-repo/fastapi-backend/src/utils/password.py +17 -0
  342. package/dist-repo/fastapi-backend/src/utils/sensitive_word_filter.py +189 -0
  343. package/package.json +11 -4
  344. /package/{dist → dist-docs}/agent-skills/README.md +0 -0
  345. /package/{dist → dist-docs}/agent-skills/README.zh.md +0 -0
  346. /package/{dist → dist-docs}/agent-skills/glm-skills/ASR/LICENSE.txt +0 -0
  347. /package/{dist → dist-docs}/agent-skills/glm-skills/ASR/SKILL.md +0 -0
  348. /package/{dist → dist-docs}/agent-skills/glm-skills/ASR/scripts/asr.ts +0 -0
  349. /package/{dist → dist-docs}/agent-skills/glm-skills/LLM/LICENSE.txt +0 -0
  350. /package/{dist → dist-docs}/agent-skills/glm-skills/LLM/SKILL.md +0 -0
  351. /package/{dist → dist-docs}/agent-skills/glm-skills/LLM/scripts/chat.ts +0 -0
  352. /package/{dist → dist-docs}/agent-skills/glm-skills/TTS/LICENSE.txt +0 -0
  353. /package/{dist → dist-docs}/agent-skills/glm-skills/TTS/SKILL.md +0 -0
  354. /package/{dist → dist-docs}/agent-skills/glm-skills/TTS/tts.ts +0 -0
  355. /package/{dist → dist-docs}/agent-skills/glm-skills/VLM/LICENSE.txt +0 -0
  356. /package/{dist → dist-docs}/agent-skills/glm-skills/VLM/SKILL.md +0 -0
  357. /package/{dist → dist-docs}/agent-skills/glm-skills/VLM/scripts/vlm.ts +0 -0
  358. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/LICENSE.txt +0 -0
  359. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/SKILL.md +0 -0
  360. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/ArsenalSC-OFL.txt +0 -0
  361. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/ArsenalSC-Regular.ttf +0 -0
  362. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/BigShoulders-Bold.ttf +0 -0
  363. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/BigShoulders-OFL.txt +0 -0
  364. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/BigShoulders-Regular.ttf +0 -0
  365. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/Boldonse-OFL.txt +0 -0
  366. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/Boldonse-Regular.ttf +0 -0
  367. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/BricolageGrotesque-Bold.ttf +0 -0
  368. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/BricolageGrotesque-OFL.txt +0 -0
  369. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/BricolageGrotesque-Regular.ttf +0 -0
  370. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/CrimsonPro-Bold.ttf +0 -0
  371. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/CrimsonPro-Italic.ttf +0 -0
  372. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/CrimsonPro-OFL.txt +0 -0
  373. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/CrimsonPro-Regular.ttf +0 -0
  374. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/DMMono-OFL.txt +0 -0
  375. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/DMMono-Regular.ttf +0 -0
  376. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/EricaOne-OFL.txt +0 -0
  377. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/EricaOne-Regular.ttf +0 -0
  378. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/GeistMono-Bold.ttf +0 -0
  379. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/GeistMono-OFL.txt +0 -0
  380. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/GeistMono-Regular.ttf +0 -0
  381. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/Gloock-OFL.txt +0 -0
  382. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/Gloock-Regular.ttf +0 -0
  383. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/IBMPlexMono-Bold.ttf +0 -0
  384. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/IBMPlexMono-OFL.txt +0 -0
  385. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/IBMPlexMono-Regular.ttf +0 -0
  386. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/IBMPlexSerif-Bold.ttf +0 -0
  387. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/IBMPlexSerif-BoldItalic.ttf +0 -0
  388. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/IBMPlexSerif-Italic.ttf +0 -0
  389. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/IBMPlexSerif-Regular.ttf +0 -0
  390. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/InstrumentSans-Bold.ttf +0 -0
  391. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/InstrumentSans-BoldItalic.ttf +0 -0
  392. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/InstrumentSans-Italic.ttf +0 -0
  393. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/InstrumentSans-OFL.txt +0 -0
  394. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/InstrumentSans-Regular.ttf +0 -0
  395. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/InstrumentSerif-Italic.ttf +0 -0
  396. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/InstrumentSerif-Regular.ttf +0 -0
  397. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/Italiana-OFL.txt +0 -0
  398. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/Italiana-Regular.ttf +0 -0
  399. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/JetBrainsMono-Bold.ttf +0 -0
  400. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/JetBrainsMono-OFL.txt +0 -0
  401. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/JetBrainsMono-Regular.ttf +0 -0
  402. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/Jura-Light.ttf +0 -0
  403. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/Jura-Medium.ttf +0 -0
  404. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/Jura-OFL.txt +0 -0
  405. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/LibreBaskerville-OFL.txt +0 -0
  406. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/LibreBaskerville-Regular.ttf +0 -0
  407. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/Lora-Bold.ttf +0 -0
  408. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/Lora-BoldItalic.ttf +0 -0
  409. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/Lora-Italic.ttf +0 -0
  410. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/Lora-OFL.txt +0 -0
  411. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/Lora-Regular.ttf +0 -0
  412. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/NationalPark-Bold.ttf +0 -0
  413. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/NationalPark-OFL.txt +0 -0
  414. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/NationalPark-Regular.ttf +0 -0
  415. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/NothingYouCouldDo-OFL.txt +0 -0
  416. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/NothingYouCouldDo-Regular.ttf +0 -0
  417. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/Outfit-Bold.ttf +0 -0
  418. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/Outfit-OFL.txt +0 -0
  419. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/Outfit-Regular.ttf +0 -0
  420. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/PixelifySans-Medium.ttf +0 -0
  421. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/PixelifySans-OFL.txt +0 -0
  422. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/PoiretOne-OFL.txt +0 -0
  423. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/PoiretOne-Regular.ttf +0 -0
  424. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/RedHatMono-Bold.ttf +0 -0
  425. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/RedHatMono-OFL.txt +0 -0
  426. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/RedHatMono-Regular.ttf +0 -0
  427. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/Silkscreen-OFL.txt +0 -0
  428. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/Silkscreen-Regular.ttf +0 -0
  429. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/SmoochSans-Medium.ttf +0 -0
  430. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/SmoochSans-OFL.txt +0 -0
  431. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/Tektur-Medium.ttf +0 -0
  432. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/Tektur-OFL.txt +0 -0
  433. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/Tektur-Regular.ttf +0 -0
  434. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/WorkSans-Bold.ttf +0 -0
  435. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/WorkSans-BoldItalic.ttf +0 -0
  436. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/WorkSans-Italic.ttf +0 -0
  437. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/WorkSans-OFL.txt +0 -0
  438. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/WorkSans-Regular.ttf +0 -0
  439. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/YoungSerif-OFL.txt +0 -0
  440. /package/{dist → dist-docs}/agent-skills/glm-skills/canvas-design/canvas-fonts/YoungSerif-Regular.ttf +0 -0
  441. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/LICENSE.txt +0 -0
  442. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/SKILL.md +0 -0
  443. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/docx-js.md +0 -0
  444. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/dml-chart.xsd +0 -0
  445. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/dml-chartDrawing.xsd +0 -0
  446. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/dml-diagram.xsd +0 -0
  447. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/dml-lockedCanvas.xsd +0 -0
  448. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/dml-main.xsd +0 -0
  449. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/dml-picture.xsd +0 -0
  450. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/dml-spreadsheetDrawing.xsd +0 -0
  451. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/dml-wordprocessingDrawing.xsd +0 -0
  452. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/pml.xsd +0 -0
  453. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-additionalCharacteristics.xsd +0 -0
  454. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-bibliography.xsd +0 -0
  455. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-commonSimpleTypes.xsd +0 -0
  456. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-customXmlDataProperties.xsd +0 -0
  457. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-customXmlSchemaProperties.xsd +0 -0
  458. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-documentPropertiesCustom.xsd +0 -0
  459. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-documentPropertiesExtended.xsd +0 -0
  460. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-documentPropertiesVariantTypes.xsd +0 -0
  461. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-math.xsd +0 -0
  462. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-relationshipReference.xsd +0 -0
  463. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/sml.xsd +0 -0
  464. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/vml-main.xsd +0 -0
  465. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/vml-officeDrawing.xsd +0 -0
  466. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/vml-presentationDrawing.xsd +0 -0
  467. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/vml-spreadsheetDrawing.xsd +0 -0
  468. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/vml-wordprocessingDrawing.xsd +0 -0
  469. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/wml.xsd +0 -0
  470. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/xml.xsd +0 -0
  471. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/ecma/fouth-edition/opc-contentTypes.xsd +0 -0
  472. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/ecma/fouth-edition/opc-coreProperties.xsd +0 -0
  473. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/ecma/fouth-edition/opc-digSig.xsd +0 -0
  474. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/ecma/fouth-edition/opc-relationships.xsd +0 -0
  475. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/mce/mc.xsd +0 -0
  476. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/microsoft/wml-2010.xsd +0 -0
  477. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/microsoft/wml-2012.xsd +0 -0
  478. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/microsoft/wml-2018.xsd +0 -0
  479. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/microsoft/wml-cex-2018.xsd +0 -0
  480. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/microsoft/wml-cid-2016.xsd +0 -0
  481. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/microsoft/wml-sdtdatahash-2020.xsd +0 -0
  482. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/schemas/microsoft/wml-symex-2015.xsd +0 -0
  483. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/scripts/pack.py +0 -0
  484. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/scripts/unpack.py +0 -0
  485. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/scripts/validate.py +0 -0
  486. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/scripts/validation/__init__.py +0 -0
  487. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/scripts/validation/base.py +0 -0
  488. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/scripts/validation/docx.py +0 -0
  489. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/scripts/validation/pptx.py +0 -0
  490. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml/scripts/validation/redlining.py +0 -0
  491. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/ooxml.md +0 -0
  492. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/scripts/__init__.py +0 -0
  493. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/scripts/document.py +0 -0
  494. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/scripts/templates/comments.xml +0 -0
  495. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/scripts/templates/commentsExtended.xml +0 -0
  496. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/scripts/templates/commentsExtensible.xml +0 -0
  497. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/scripts/templates/commentsIds.xml +0 -0
  498. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/scripts/templates/people.xml +0 -0
  499. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/docx/scripts/utilities.py +0 -0
  500. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pdf/LICENSE.txt +0 -0
  501. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pdf/SKILL.md +0 -0
  502. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pdf/forms.md +0 -0
  503. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pdf/reference.md +0 -0
  504. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pdf/scripts/check_bounding_boxes.py +0 -0
  505. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pdf/scripts/check_bounding_boxes_test.py +0 -0
  506. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pdf/scripts/check_fillable_fields.py +0 -0
  507. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pdf/scripts/convert_pdf_to_images.py +0 -0
  508. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pdf/scripts/create_validation_image.py +0 -0
  509. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pdf/scripts/extract_form_field_info.py +0 -0
  510. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pdf/scripts/fill_fillable_fields.py +0 -0
  511. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pdf/scripts/fill_pdf_form_with_annotations.py +0 -0
  512. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/LICENSE.txt +0 -0
  513. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/SKILL.md +0 -0
  514. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/html2pptx.md +0 -0
  515. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/dml-chart.xsd +0 -0
  516. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/dml-chartDrawing.xsd +0 -0
  517. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/dml-diagram.xsd +0 -0
  518. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/dml-lockedCanvas.xsd +0 -0
  519. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/dml-main.xsd +0 -0
  520. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/dml-picture.xsd +0 -0
  521. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/dml-spreadsheetDrawing.xsd +0 -0
  522. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/dml-wordprocessingDrawing.xsd +0 -0
  523. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/pml.xsd +0 -0
  524. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-additionalCharacteristics.xsd +0 -0
  525. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-bibliography.xsd +0 -0
  526. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-commonSimpleTypes.xsd +0 -0
  527. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-customXmlDataProperties.xsd +0 -0
  528. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-customXmlSchemaProperties.xsd +0 -0
  529. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-documentPropertiesCustom.xsd +0 -0
  530. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-documentPropertiesExtended.xsd +0 -0
  531. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-documentPropertiesVariantTypes.xsd +0 -0
  532. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-math.xsd +0 -0
  533. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-relationshipReference.xsd +0 -0
  534. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/sml.xsd +0 -0
  535. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/vml-main.xsd +0 -0
  536. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/vml-officeDrawing.xsd +0 -0
  537. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/vml-presentationDrawing.xsd +0 -0
  538. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/vml-spreadsheetDrawing.xsd +0 -0
  539. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/vml-wordprocessingDrawing.xsd +0 -0
  540. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/wml.xsd +0 -0
  541. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/xml.xsd +0 -0
  542. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/ecma/fouth-edition/opc-contentTypes.xsd +0 -0
  543. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/ecma/fouth-edition/opc-coreProperties.xsd +0 -0
  544. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/ecma/fouth-edition/opc-digSig.xsd +0 -0
  545. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/ecma/fouth-edition/opc-relationships.xsd +0 -0
  546. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/mce/mc.xsd +0 -0
  547. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/microsoft/wml-2010.xsd +0 -0
  548. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/microsoft/wml-2012.xsd +0 -0
  549. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/microsoft/wml-2018.xsd +0 -0
  550. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/microsoft/wml-cex-2018.xsd +0 -0
  551. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/microsoft/wml-cid-2016.xsd +0 -0
  552. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/microsoft/wml-sdtdatahash-2020.xsd +0 -0
  553. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/schemas/microsoft/wml-symex-2015.xsd +0 -0
  554. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/scripts/pack.py +0 -0
  555. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/scripts/unpack.py +0 -0
  556. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/scripts/validate.py +0 -0
  557. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/scripts/validation/__init__.py +0 -0
  558. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/scripts/validation/base.py +0 -0
  559. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/scripts/validation/docx.py +0 -0
  560. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/scripts/validation/pptx.py +0 -0
  561. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml/scripts/validation/redlining.py +0 -0
  562. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/ooxml.md +0 -0
  563. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/scripts/html2pptx.js +0 -0
  564. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/scripts/inventory.py +0 -0
  565. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/scripts/rearrange.py +0 -0
  566. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/scripts/replace.py +0 -0
  567. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/pptx/scripts/thumbnail.py +0 -0
  568. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/xlsx/LICENSE.txt +0 -0
  569. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/xlsx/SKILL.md +0 -0
  570. /package/{dist → dist-docs}/agent-skills/glm-skills/document-skills/xlsx/recalc.py +0 -0
  571. /package/{dist → dist-docs}/agent-skills/glm-skills/frontend-design/LICENSE +0 -0
  572. /package/{dist → dist-docs}/agent-skills/glm-skills/frontend-design/OPTIMIZATION_SUMMARY.md +0 -0
  573. /package/{dist → dist-docs}/agent-skills/glm-skills/frontend-design/README.md +0 -0
  574. /package/{dist → dist-docs}/agent-skills/glm-skills/frontend-design/SKILL.md +0 -0
  575. /package/{dist → dist-docs}/agent-skills/glm-skills/frontend-design/examples/css/components.css +0 -0
  576. /package/{dist → dist-docs}/agent-skills/glm-skills/frontend-design/examples/css/tokens.css +0 -0
  577. /package/{dist → dist-docs}/agent-skills/glm-skills/frontend-design/examples/typescript/design-tokens.ts +0 -0
  578. /package/{dist → dist-docs}/agent-skills/glm-skills/frontend-design/examples/typescript/sample-components.tsx +0 -0
  579. /package/{dist → dist-docs}/agent-skills/glm-skills/frontend-design/examples/typescript/theme-provider.tsx +0 -0
  580. /package/{dist → dist-docs}/agent-skills/glm-skills/frontend-design/examples/typescript/utils.ts +0 -0
  581. /package/{dist → dist-docs}/agent-skills/glm-skills/frontend-design/package.json +0 -0
  582. /package/{dist → dist-docs}/agent-skills/glm-skills/frontend-design/templates/globals.css +0 -0
  583. /package/{dist → dist-docs}/agent-skills/glm-skills/frontend-design/templates/tailwind.config.js +0 -0
  584. /package/{dist → dist-docs}/agent-skills/glm-skills/image-generation/LICENSE.txt +0 -0
  585. /package/{dist → dist-docs}/agent-skills/glm-skills/image-generation/SKILL.md +0 -0
  586. /package/{dist → dist-docs}/agent-skills/glm-skills/image-generation/scripts/image-generation.ts +0 -0
  587. /package/{dist → dist-docs}/agent-skills/glm-skills/video-generation/LICENSE.txt +0 -0
  588. /package/{dist → dist-docs}/agent-skills/glm-skills/video-generation/SKILL.md +0 -0
  589. /package/{dist → dist-docs}/agent-skills/glm-skills/video-generation/scripts/video.ts +0 -0
  590. /package/{dist → dist-docs}/agent-skills/glm-skills/web-reader/LICENSE.txt +0 -0
  591. /package/{dist → dist-docs}/agent-skills/glm-skills/web-reader/SKILL.md +0 -0
  592. /package/{dist → dist-docs}/agent-skills/glm-skills/web-reader/scripts/web-reader.ts +0 -0
  593. /package/{dist → dist-docs}/agent-skills/glm-skills/web-search/LICENSE.txt +0 -0
  594. /package/{dist → dist-docs}/agent-skills/glm-skills/web-search/SKILL.md +0 -0
  595. /package/{dist → dist-docs}/agent-skills/glm-skills/web-search/scripts/web_search.ts +0 -0
  596. /package/{dist → dist-docs}/agent-skills/openai-skills/docs/render_docx.py +0 -0
  597. /package/{dist → dist-docs}/agent-skills/openai-skills/docs/skill.md +0 -0
  598. /package/{dist → dist-docs}/agent-skills/openai-skills/pdfs/skill.md +0 -0
  599. /package/{dist → dist-docs}/agent-skills/openai-skills/spreadsheets/artifact_tool_spreadsheet_formulas.md +0 -0
  600. /package/{dist → dist-docs}/agent-skills/openai-skills/spreadsheets/artifact_tool_spreadsheets_api.md +0 -0
  601. /package/{dist → dist-docs}/agent-skills/openai-skills/spreadsheets/examples/create_basic_spreadsheet.py +0 -0
  602. /package/{dist → dist-docs}/agent-skills/openai-skills/spreadsheets/examples/create_spreadsheet_with_styling.py +0 -0
  603. /package/{dist → dist-docs}/agent-skills/openai-skills/spreadsheets/examples/features/change_existing_charts.py +0 -0
  604. /package/{dist → dist-docs}/agent-skills/openai-skills/spreadsheets/examples/features/cite_cells.py +0 -0
  605. /package/{dist → dist-docs}/agent-skills/openai-skills/spreadsheets/examples/features/create_area_chart.py +0 -0
  606. /package/{dist → dist-docs}/agent-skills/openai-skills/spreadsheets/examples/features/create_bar_chart.py +0 -0
  607. /package/{dist → dist-docs}/agent-skills/openai-skills/spreadsheets/examples/features/create_doughnut_chart.py +0 -0
  608. /package/{dist → dist-docs}/agent-skills/openai-skills/spreadsheets/examples/features/create_line_chart.py +0 -0
  609. /package/{dist → dist-docs}/agent-skills/openai-skills/spreadsheets/examples/features/create_pie_chart.py +0 -0
  610. /package/{dist → dist-docs}/agent-skills/openai-skills/spreadsheets/examples/features/create_tables.py +0 -0
  611. /package/{dist → dist-docs}/agent-skills/openai-skills/spreadsheets/examples/features/set_cell_borders.py +0 -0
  612. /package/{dist → dist-docs}/agent-skills/openai-skills/spreadsheets/examples/features/set_cell_fills.py +0 -0
  613. /package/{dist → dist-docs}/agent-skills/openai-skills/spreadsheets/examples/features/set_cell_width_height.py +0 -0
  614. /package/{dist → dist-docs}/agent-skills/openai-skills/spreadsheets/examples/features/set_conditional_formatting.py +0 -0
  615. /package/{dist → dist-docs}/agent-skills/openai-skills/spreadsheets/examples/features/set_font_styles.py +0 -0
  616. /package/{dist → dist-docs}/agent-skills/openai-skills/spreadsheets/examples/features/set_merge_cells.py +0 -0
  617. /package/{dist → dist-docs}/agent-skills/openai-skills/spreadsheets/examples/features/set_number_formats.py +0 -0
  618. /package/{dist → dist-docs}/agent-skills/openai-skills/spreadsheets/examples/features/set_text_alignment.py +0 -0
  619. /package/{dist → dist-docs}/agent-skills/openai-skills/spreadsheets/examples/features/set_wrap_text_styles.py +0 -0
  620. /package/{dist → dist-docs}/agent-skills/openai-skills/spreadsheets/examples/read_existing_spreadsheet.py +0 -0
  621. /package/{dist → dist-docs}/agent-skills/openai-skills/spreadsheets/examples/styling_spreadsheet.py +0 -0
  622. /package/{dist → dist-docs}/agent-skills/openai-skills/spreadsheets/skill.md +0 -0
  623. /package/{dist → dist-docs}/agent-skills/openai-skills/spreadsheets/spreadsheet.md +0 -0
  624. /package/{dist → dist-docs}/anthropipc-official-skills/.claude-plugin/marketplace.json +0 -0
  625. /package/{dist → dist-docs}/anthropipc-official-skills/README.md +0 -0
  626. /package/{dist → dist-docs}/anthropipc-official-skills/THIRD_PARTY_NOTICES.md +0 -0
  627. /package/{dist → dist-docs}/anthropipc-official-skills/skills/algorithmic-art/LICENSE.txt +0 -0
  628. /package/{dist → dist-docs}/anthropipc-official-skills/skills/algorithmic-art/SKILL.md +0 -0
  629. /package/{dist → dist-docs}/anthropipc-official-skills/skills/algorithmic-art/templates/generator_template.js +0 -0
  630. /package/{dist → dist-docs}/anthropipc-official-skills/skills/algorithmic-art/templates/viewer.html +0 -0
  631. /package/{dist → dist-docs}/anthropipc-official-skills/skills/brand-guidelines/LICENSE.txt +0 -0
  632. /package/{dist → dist-docs}/anthropipc-official-skills/skills/brand-guidelines/SKILL.md +0 -0
  633. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/LICENSE.txt +0 -0
  634. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/SKILL.md +0 -0
  635. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/ArsenalSC-OFL.txt +0 -0
  636. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/ArsenalSC-Regular.ttf +0 -0
  637. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/BigShoulders-Bold.ttf +0 -0
  638. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/BigShoulders-OFL.txt +0 -0
  639. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/BigShoulders-Regular.ttf +0 -0
  640. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/Boldonse-OFL.txt +0 -0
  641. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/Boldonse-Regular.ttf +0 -0
  642. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/BricolageGrotesque-Bold.ttf +0 -0
  643. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/BricolageGrotesque-OFL.txt +0 -0
  644. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/BricolageGrotesque-Regular.ttf +0 -0
  645. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/CrimsonPro-Bold.ttf +0 -0
  646. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/CrimsonPro-Italic.ttf +0 -0
  647. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/CrimsonPro-OFL.txt +0 -0
  648. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/CrimsonPro-Regular.ttf +0 -0
  649. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/DMMono-OFL.txt +0 -0
  650. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/DMMono-Regular.ttf +0 -0
  651. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/EricaOne-OFL.txt +0 -0
  652. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/EricaOne-Regular.ttf +0 -0
  653. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/GeistMono-Bold.ttf +0 -0
  654. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/GeistMono-OFL.txt +0 -0
  655. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/GeistMono-Regular.ttf +0 -0
  656. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/Gloock-OFL.txt +0 -0
  657. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/Gloock-Regular.ttf +0 -0
  658. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/IBMPlexMono-Bold.ttf +0 -0
  659. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/IBMPlexMono-OFL.txt +0 -0
  660. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/IBMPlexMono-Regular.ttf +0 -0
  661. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/IBMPlexSerif-Bold.ttf +0 -0
  662. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/IBMPlexSerif-BoldItalic.ttf +0 -0
  663. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/IBMPlexSerif-Italic.ttf +0 -0
  664. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/IBMPlexSerif-Regular.ttf +0 -0
  665. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/InstrumentSans-Bold.ttf +0 -0
  666. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/InstrumentSans-BoldItalic.ttf +0 -0
  667. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/InstrumentSans-Italic.ttf +0 -0
  668. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/InstrumentSans-OFL.txt +0 -0
  669. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/InstrumentSans-Regular.ttf +0 -0
  670. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/InstrumentSerif-Italic.ttf +0 -0
  671. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/InstrumentSerif-Regular.ttf +0 -0
  672. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/Italiana-OFL.txt +0 -0
  673. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/Italiana-Regular.ttf +0 -0
  674. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/JetBrainsMono-Bold.ttf +0 -0
  675. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/JetBrainsMono-OFL.txt +0 -0
  676. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/JetBrainsMono-Regular.ttf +0 -0
  677. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/Jura-Light.ttf +0 -0
  678. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/Jura-Medium.ttf +0 -0
  679. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/Jura-OFL.txt +0 -0
  680. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/LibreBaskerville-OFL.txt +0 -0
  681. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/LibreBaskerville-Regular.ttf +0 -0
  682. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/Lora-Bold.ttf +0 -0
  683. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/Lora-BoldItalic.ttf +0 -0
  684. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/Lora-Italic.ttf +0 -0
  685. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/Lora-OFL.txt +0 -0
  686. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/Lora-Regular.ttf +0 -0
  687. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/NationalPark-Bold.ttf +0 -0
  688. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/NationalPark-OFL.txt +0 -0
  689. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/NationalPark-Regular.ttf +0 -0
  690. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/NothingYouCouldDo-OFL.txt +0 -0
  691. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/NothingYouCouldDo-Regular.ttf +0 -0
  692. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/Outfit-Bold.ttf +0 -0
  693. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/Outfit-OFL.txt +0 -0
  694. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/Outfit-Regular.ttf +0 -0
  695. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/PixelifySans-Medium.ttf +0 -0
  696. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/PixelifySans-OFL.txt +0 -0
  697. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/PoiretOne-OFL.txt +0 -0
  698. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/PoiretOne-Regular.ttf +0 -0
  699. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/RedHatMono-Bold.ttf +0 -0
  700. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/RedHatMono-OFL.txt +0 -0
  701. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/RedHatMono-Regular.ttf +0 -0
  702. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/Silkscreen-OFL.txt +0 -0
  703. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/Silkscreen-Regular.ttf +0 -0
  704. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/SmoochSans-Medium.ttf +0 -0
  705. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/SmoochSans-OFL.txt +0 -0
  706. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/Tektur-Medium.ttf +0 -0
  707. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/Tektur-OFL.txt +0 -0
  708. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/Tektur-Regular.ttf +0 -0
  709. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/WorkSans-Bold.ttf +0 -0
  710. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/WorkSans-BoldItalic.ttf +0 -0
  711. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/WorkSans-Italic.ttf +0 -0
  712. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/WorkSans-OFL.txt +0 -0
  713. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/WorkSans-Regular.ttf +0 -0
  714. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/YoungSerif-OFL.txt +0 -0
  715. /package/{dist → dist-docs}/anthropipc-official-skills/skills/canvas-design/canvas-fonts/YoungSerif-Regular.ttf +0 -0
  716. /package/{dist → dist-docs}/anthropipc-official-skills/skills/doc-coauthoring/SKILL.md +0 -0
  717. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/LICENSE.txt +0 -0
  718. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/SKILL.md +0 -0
  719. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/docx-js.md +0 -0
  720. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/dml-chart.xsd +0 -0
  721. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/dml-chartDrawing.xsd +0 -0
  722. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/dml-diagram.xsd +0 -0
  723. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/dml-lockedCanvas.xsd +0 -0
  724. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/dml-main.xsd +0 -0
  725. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/dml-picture.xsd +0 -0
  726. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/dml-spreadsheetDrawing.xsd +0 -0
  727. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/dml-wordprocessingDrawing.xsd +0 -0
  728. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/pml.xsd +0 -0
  729. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-additionalCharacteristics.xsd +0 -0
  730. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-bibliography.xsd +0 -0
  731. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-commonSimpleTypes.xsd +0 -0
  732. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-customXmlDataProperties.xsd +0 -0
  733. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-customXmlSchemaProperties.xsd +0 -0
  734. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-documentPropertiesCustom.xsd +0 -0
  735. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-documentPropertiesExtended.xsd +0 -0
  736. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-documentPropertiesVariantTypes.xsd +0 -0
  737. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-math.xsd +0 -0
  738. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-relationshipReference.xsd +0 -0
  739. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/sml.xsd +0 -0
  740. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/vml-main.xsd +0 -0
  741. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/vml-officeDrawing.xsd +0 -0
  742. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/vml-presentationDrawing.xsd +0 -0
  743. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/vml-spreadsheetDrawing.xsd +0 -0
  744. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/vml-wordprocessingDrawing.xsd +0 -0
  745. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/wml.xsd +0 -0
  746. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/xml.xsd +0 -0
  747. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/ecma/fouth-edition/opc-contentTypes.xsd +0 -0
  748. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/ecma/fouth-edition/opc-coreProperties.xsd +0 -0
  749. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/ecma/fouth-edition/opc-digSig.xsd +0 -0
  750. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/ecma/fouth-edition/opc-relationships.xsd +0 -0
  751. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/mce/mc.xsd +0 -0
  752. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/microsoft/wml-2010.xsd +0 -0
  753. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/microsoft/wml-2012.xsd +0 -0
  754. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/microsoft/wml-2018.xsd +0 -0
  755. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/microsoft/wml-cex-2018.xsd +0 -0
  756. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/microsoft/wml-cid-2016.xsd +0 -0
  757. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/microsoft/wml-sdtdatahash-2020.xsd +0 -0
  758. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/schemas/microsoft/wml-symex-2015.xsd +0 -0
  759. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/scripts/pack.py +0 -0
  760. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/scripts/unpack.py +0 -0
  761. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/scripts/validate.py +0 -0
  762. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/scripts/validation/__init__.py +0 -0
  763. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/scripts/validation/base.py +0 -0
  764. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/scripts/validation/docx.py +0 -0
  765. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/scripts/validation/pptx.py +0 -0
  766. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml/scripts/validation/redlining.py +0 -0
  767. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/ooxml.md +0 -0
  768. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/scripts/__init__.py +0 -0
  769. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/scripts/document.py +0 -0
  770. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/scripts/templates/comments.xml +0 -0
  771. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/scripts/templates/commentsExtended.xml +0 -0
  772. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/scripts/templates/commentsExtensible.xml +0 -0
  773. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/scripts/templates/commentsIds.xml +0 -0
  774. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/scripts/templates/people.xml +0 -0
  775. /package/{dist → dist-docs}/anthropipc-official-skills/skills/docx/scripts/utilities.py +0 -0
  776. /package/{dist → dist-docs}/anthropipc-official-skills/skills/frontend-design/LICENSE.txt +0 -0
  777. /package/{dist → dist-docs}/anthropipc-official-skills/skills/frontend-design/SKILL.md +0 -0
  778. /package/{dist → dist-docs}/anthropipc-official-skills/skills/internal-comms/LICENSE.txt +0 -0
  779. /package/{dist → dist-docs}/anthropipc-official-skills/skills/internal-comms/SKILL.md +0 -0
  780. /package/{dist → dist-docs}/anthropipc-official-skills/skills/internal-comms/examples/3p-updates.md +0 -0
  781. /package/{dist → dist-docs}/anthropipc-official-skills/skills/internal-comms/examples/company-newsletter.md +0 -0
  782. /package/{dist → dist-docs}/anthropipc-official-skills/skills/internal-comms/examples/faq-answers.md +0 -0
  783. /package/{dist → dist-docs}/anthropipc-official-skills/skills/internal-comms/examples/general-comms.md +0 -0
  784. /package/{dist → dist-docs}/anthropipc-official-skills/skills/mcp-builder/LICENSE.txt +0 -0
  785. /package/{dist → dist-docs}/anthropipc-official-skills/skills/mcp-builder/SKILL.md +0 -0
  786. /package/{dist → dist-docs}/anthropipc-official-skills/skills/mcp-builder/reference/evaluation.md +0 -0
  787. /package/{dist → dist-docs}/anthropipc-official-skills/skills/mcp-builder/reference/mcp_best_practices.md +0 -0
  788. /package/{dist → dist-docs}/anthropipc-official-skills/skills/mcp-builder/reference/node_mcp_server.md +0 -0
  789. /package/{dist → dist-docs}/anthropipc-official-skills/skills/mcp-builder/reference/python_mcp_server.md +0 -0
  790. /package/{dist → dist-docs}/anthropipc-official-skills/skills/mcp-builder/scripts/connections.py +0 -0
  791. /package/{dist → dist-docs}/anthropipc-official-skills/skills/mcp-builder/scripts/evaluation.py +0 -0
  792. /package/{dist → dist-docs}/anthropipc-official-skills/skills/mcp-builder/scripts/example_evaluation.xml +0 -0
  793. /package/{dist → dist-docs}/anthropipc-official-skills/skills/mcp-builder/scripts/requirements.txt +0 -0
  794. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pdf/LICENSE.txt +0 -0
  795. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pdf/SKILL.md +0 -0
  796. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pdf/forms.md +0 -0
  797. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pdf/reference.md +0 -0
  798. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pdf/scripts/check_bounding_boxes.py +0 -0
  799. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pdf/scripts/check_bounding_boxes_test.py +0 -0
  800. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pdf/scripts/check_fillable_fields.py +0 -0
  801. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pdf/scripts/convert_pdf_to_images.py +0 -0
  802. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pdf/scripts/create_validation_image.py +0 -0
  803. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pdf/scripts/extract_form_field_info.py +0 -0
  804. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pdf/scripts/fill_fillable_fields.py +0 -0
  805. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pdf/scripts/fill_pdf_form_with_annotations.py +0 -0
  806. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/LICENSE.txt +0 -0
  807. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/SKILL.md +0 -0
  808. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/html2pptx.md +0 -0
  809. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/dml-chart.xsd +0 -0
  810. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/dml-chartDrawing.xsd +0 -0
  811. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/dml-diagram.xsd +0 -0
  812. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/dml-lockedCanvas.xsd +0 -0
  813. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/dml-main.xsd +0 -0
  814. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/dml-picture.xsd +0 -0
  815. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/dml-spreadsheetDrawing.xsd +0 -0
  816. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/dml-wordprocessingDrawing.xsd +0 -0
  817. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/pml.xsd +0 -0
  818. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-additionalCharacteristics.xsd +0 -0
  819. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-bibliography.xsd +0 -0
  820. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-commonSimpleTypes.xsd +0 -0
  821. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-customXmlDataProperties.xsd +0 -0
  822. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-customXmlSchemaProperties.xsd +0 -0
  823. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-documentPropertiesCustom.xsd +0 -0
  824. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-documentPropertiesExtended.xsd +0 -0
  825. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-documentPropertiesVariantTypes.xsd +0 -0
  826. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-math.xsd +0 -0
  827. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-relationshipReference.xsd +0 -0
  828. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/sml.xsd +0 -0
  829. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/vml-main.xsd +0 -0
  830. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/vml-officeDrawing.xsd +0 -0
  831. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/vml-presentationDrawing.xsd +0 -0
  832. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/vml-spreadsheetDrawing.xsd +0 -0
  833. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/vml-wordprocessingDrawing.xsd +0 -0
  834. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/wml.xsd +0 -0
  835. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/xml.xsd +0 -0
  836. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/ecma/fouth-edition/opc-contentTypes.xsd +0 -0
  837. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/ecma/fouth-edition/opc-coreProperties.xsd +0 -0
  838. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/ecma/fouth-edition/opc-digSig.xsd +0 -0
  839. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/ecma/fouth-edition/opc-relationships.xsd +0 -0
  840. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/mce/mc.xsd +0 -0
  841. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/microsoft/wml-2010.xsd +0 -0
  842. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/microsoft/wml-2012.xsd +0 -0
  843. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/microsoft/wml-2018.xsd +0 -0
  844. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/microsoft/wml-cex-2018.xsd +0 -0
  845. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/microsoft/wml-cid-2016.xsd +0 -0
  846. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/microsoft/wml-sdtdatahash-2020.xsd +0 -0
  847. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/schemas/microsoft/wml-symex-2015.xsd +0 -0
  848. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/scripts/pack.py +0 -0
  849. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/scripts/unpack.py +0 -0
  850. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/scripts/validate.py +0 -0
  851. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/scripts/validation/__init__.py +0 -0
  852. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/scripts/validation/base.py +0 -0
  853. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/scripts/validation/docx.py +0 -0
  854. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/scripts/validation/pptx.py +0 -0
  855. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml/scripts/validation/redlining.py +0 -0
  856. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/ooxml.md +0 -0
  857. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/scripts/html2pptx.js +0 -0
  858. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/scripts/inventory.py +0 -0
  859. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/scripts/rearrange.py +0 -0
  860. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/scripts/replace.py +0 -0
  861. /package/{dist → dist-docs}/anthropipc-official-skills/skills/pptx/scripts/thumbnail.py +0 -0
  862. /package/{dist → dist-docs}/anthropipc-official-skills/skills/skill-creator/LICENSE.txt +0 -0
  863. /package/{dist → dist-docs}/anthropipc-official-skills/skills/skill-creator/SKILL.md +0 -0
  864. /package/{dist → dist-docs}/anthropipc-official-skills/skills/skill-creator/references/output-patterns.md +0 -0
  865. /package/{dist → dist-docs}/anthropipc-official-skills/skills/skill-creator/references/workflows.md +0 -0
  866. /package/{dist → dist-docs}/anthropipc-official-skills/skills/skill-creator/scripts/init_skill.py +0 -0
  867. /package/{dist → dist-docs}/anthropipc-official-skills/skills/skill-creator/scripts/package_skill.py +0 -0
  868. /package/{dist → dist-docs}/anthropipc-official-skills/skills/skill-creator/scripts/quick_validate.py +0 -0
  869. /package/{dist → dist-docs}/anthropipc-official-skills/skills/slack-gif-creator/LICENSE.txt +0 -0
  870. /package/{dist → dist-docs}/anthropipc-official-skills/skills/slack-gif-creator/SKILL.md +0 -0
  871. /package/{dist → dist-docs}/anthropipc-official-skills/skills/slack-gif-creator/core/easing.py +0 -0
  872. /package/{dist → dist-docs}/anthropipc-official-skills/skills/slack-gif-creator/core/frame_composer.py +0 -0
  873. /package/{dist → dist-docs}/anthropipc-official-skills/skills/slack-gif-creator/core/gif_builder.py +0 -0
  874. /package/{dist → dist-docs}/anthropipc-official-skills/skills/slack-gif-creator/core/validators.py +0 -0
  875. /package/{dist → dist-docs}/anthropipc-official-skills/skills/slack-gif-creator/requirements.txt +0 -0
  876. /package/{dist → dist-docs}/anthropipc-official-skills/skills/theme-factory/LICENSE.txt +0 -0
  877. /package/{dist → dist-docs}/anthropipc-official-skills/skills/theme-factory/SKILL.md +0 -0
  878. /package/{dist → dist-docs}/anthropipc-official-skills/skills/theme-factory/theme-showcase.pdf +0 -0
  879. /package/{dist → dist-docs}/anthropipc-official-skills/skills/theme-factory/themes/arctic-frost.md +0 -0
  880. /package/{dist → dist-docs}/anthropipc-official-skills/skills/theme-factory/themes/botanical-garden.md +0 -0
  881. /package/{dist → dist-docs}/anthropipc-official-skills/skills/theme-factory/themes/desert-rose.md +0 -0
  882. /package/{dist → dist-docs}/anthropipc-official-skills/skills/theme-factory/themes/forest-canopy.md +0 -0
  883. /package/{dist → dist-docs}/anthropipc-official-skills/skills/theme-factory/themes/golden-hour.md +0 -0
  884. /package/{dist → dist-docs}/anthropipc-official-skills/skills/theme-factory/themes/midnight-galaxy.md +0 -0
  885. /package/{dist → dist-docs}/anthropipc-official-skills/skills/theme-factory/themes/modern-minimalist.md +0 -0
  886. /package/{dist → dist-docs}/anthropipc-official-skills/skills/theme-factory/themes/ocean-depths.md +0 -0
  887. /package/{dist → dist-docs}/anthropipc-official-skills/skills/theme-factory/themes/sunset-boulevard.md +0 -0
  888. /package/{dist → dist-docs}/anthropipc-official-skills/skills/theme-factory/themes/tech-innovation.md +0 -0
  889. /package/{dist → dist-docs}/anthropipc-official-skills/skills/web-artifacts-builder/LICENSE.txt +0 -0
  890. /package/{dist → dist-docs}/anthropipc-official-skills/skills/web-artifacts-builder/SKILL.md +0 -0
  891. /package/{dist → dist-docs}/anthropipc-official-skills/skills/web-artifacts-builder/scripts/bundle-artifact.sh +0 -0
  892. /package/{dist → dist-docs}/anthropipc-official-skills/skills/web-artifacts-builder/scripts/init-artifact.sh +0 -0
  893. /package/{dist → dist-docs}/anthropipc-official-skills/skills/web-artifacts-builder/scripts/shadcn-components.tar.gz +0 -0
  894. /package/{dist → dist-docs}/anthropipc-official-skills/skills/webapp-testing/LICENSE.txt +0 -0
  895. /package/{dist → dist-docs}/anthropipc-official-skills/skills/webapp-testing/SKILL.md +0 -0
  896. /package/{dist → dist-docs}/anthropipc-official-skills/skills/webapp-testing/examples/console_logging.py +0 -0
  897. /package/{dist → dist-docs}/anthropipc-official-skills/skills/webapp-testing/examples/element_discovery.py +0 -0
  898. /package/{dist → dist-docs}/anthropipc-official-skills/skills/webapp-testing/examples/static_html_automation.py +0 -0
  899. /package/{dist → dist-docs}/anthropipc-official-skills/skills/webapp-testing/scripts/with_server.py +0 -0
  900. /package/{dist → dist-docs}/anthropipc-official-skills/skills/xlsx/LICENSE.txt +0 -0
  901. /package/{dist → dist-docs}/anthropipc-official-skills/skills/xlsx/SKILL.md +0 -0
  902. /package/{dist → dist-docs}/anthropipc-official-skills/skills/xlsx/recalc.py +0 -0
  903. /package/{dist → dist-docs}/anthropipc-official-skills/spec/agent-skills-spec.md +0 -0
  904. /package/{dist → dist-docs}/anthropipc-official-skills/template/SKILL.md +0 -0
  905. /package/{dist → dist-docs}/awesome-claude-skills/.claude-plugin/marketplace.json +0 -0
  906. /package/{dist → dist-docs}/awesome-claude-skills/CONTRIBUTING.md +0 -0
  907. /package/{dist → dist-docs}/awesome-claude-skills/README.md +0 -0
  908. /package/{dist → dist-docs}/awesome-claude-skills/artifacts-builder/LICENSE.txt +0 -0
  909. /package/{dist → dist-docs}/awesome-claude-skills/artifacts-builder/SKILL.md +0 -0
  910. /package/{dist → dist-docs}/awesome-claude-skills/artifacts-builder/scripts/bundle-artifact.sh +0 -0
  911. /package/{dist → dist-docs}/awesome-claude-skills/artifacts-builder/scripts/init-artifact.sh +0 -0
  912. /package/{dist → dist-docs}/awesome-claude-skills/artifacts-builder/scripts/shadcn-components.tar.gz +0 -0
  913. /package/{dist → dist-docs}/awesome-claude-skills/brand-guidelines/LICENSE.txt +0 -0
  914. /package/{dist → dist-docs}/awesome-claude-skills/brand-guidelines/SKILL.md +0 -0
  915. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/LICENSE.txt +0 -0
  916. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/SKILL.md +0 -0
  917. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/ArsenalSC-OFL.txt +0 -0
  918. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/ArsenalSC-Regular.ttf +0 -0
  919. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/BigShoulders-Bold.ttf +0 -0
  920. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/BigShoulders-OFL.txt +0 -0
  921. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/BigShoulders-Regular.ttf +0 -0
  922. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/Boldonse-OFL.txt +0 -0
  923. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/Boldonse-Regular.ttf +0 -0
  924. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/BricolageGrotesque-Bold.ttf +0 -0
  925. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/BricolageGrotesque-OFL.txt +0 -0
  926. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/BricolageGrotesque-Regular.ttf +0 -0
  927. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/CrimsonPro-Bold.ttf +0 -0
  928. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/CrimsonPro-Italic.ttf +0 -0
  929. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/CrimsonPro-OFL.txt +0 -0
  930. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/CrimsonPro-Regular.ttf +0 -0
  931. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/DMMono-OFL.txt +0 -0
  932. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/DMMono-Regular.ttf +0 -0
  933. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/EricaOne-OFL.txt +0 -0
  934. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/EricaOne-Regular.ttf +0 -0
  935. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/GeistMono-Bold.ttf +0 -0
  936. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/GeistMono-OFL.txt +0 -0
  937. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/GeistMono-Regular.ttf +0 -0
  938. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/Gloock-OFL.txt +0 -0
  939. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/Gloock-Regular.ttf +0 -0
  940. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/IBMPlexMono-Bold.ttf +0 -0
  941. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/IBMPlexMono-OFL.txt +0 -0
  942. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/IBMPlexMono-Regular.ttf +0 -0
  943. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/IBMPlexSerif-Bold.ttf +0 -0
  944. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/IBMPlexSerif-BoldItalic.ttf +0 -0
  945. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/IBMPlexSerif-Italic.ttf +0 -0
  946. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/IBMPlexSerif-Regular.ttf +0 -0
  947. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/InstrumentSans-Bold.ttf +0 -0
  948. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/InstrumentSans-BoldItalic.ttf +0 -0
  949. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/InstrumentSans-Italic.ttf +0 -0
  950. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/InstrumentSans-OFL.txt +0 -0
  951. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/InstrumentSans-Regular.ttf +0 -0
  952. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/InstrumentSerif-Italic.ttf +0 -0
  953. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/InstrumentSerif-Regular.ttf +0 -0
  954. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/Italiana-OFL.txt +0 -0
  955. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/Italiana-Regular.ttf +0 -0
  956. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/JetBrainsMono-Bold.ttf +0 -0
  957. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/JetBrainsMono-OFL.txt +0 -0
  958. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/JetBrainsMono-Regular.ttf +0 -0
  959. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/Jura-Light.ttf +0 -0
  960. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/Jura-Medium.ttf +0 -0
  961. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/Jura-OFL.txt +0 -0
  962. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/LibreBaskerville-OFL.txt +0 -0
  963. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/LibreBaskerville-Regular.ttf +0 -0
  964. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/Lora-Bold.ttf +0 -0
  965. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/Lora-BoldItalic.ttf +0 -0
  966. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/Lora-Italic.ttf +0 -0
  967. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/Lora-OFL.txt +0 -0
  968. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/Lora-Regular.ttf +0 -0
  969. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/NationalPark-Bold.ttf +0 -0
  970. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/NationalPark-OFL.txt +0 -0
  971. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/NationalPark-Regular.ttf +0 -0
  972. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/NothingYouCouldDo-OFL.txt +0 -0
  973. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/NothingYouCouldDo-Regular.ttf +0 -0
  974. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/Outfit-Bold.ttf +0 -0
  975. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/Outfit-OFL.txt +0 -0
  976. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/Outfit-Regular.ttf +0 -0
  977. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/PixelifySans-Medium.ttf +0 -0
  978. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/PixelifySans-OFL.txt +0 -0
  979. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/PoiretOne-OFL.txt +0 -0
  980. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/PoiretOne-Regular.ttf +0 -0
  981. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/RedHatMono-Bold.ttf +0 -0
  982. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/RedHatMono-OFL.txt +0 -0
  983. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/RedHatMono-Regular.ttf +0 -0
  984. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/Silkscreen-OFL.txt +0 -0
  985. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/Silkscreen-Regular.ttf +0 -0
  986. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/SmoochSans-Medium.ttf +0 -0
  987. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/SmoochSans-OFL.txt +0 -0
  988. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/Tektur-Medium.ttf +0 -0
  989. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/Tektur-OFL.txt +0 -0
  990. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/Tektur-Regular.ttf +0 -0
  991. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/WorkSans-Bold.ttf +0 -0
  992. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/WorkSans-BoldItalic.ttf +0 -0
  993. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/WorkSans-Italic.ttf +0 -0
  994. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/WorkSans-OFL.txt +0 -0
  995. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/WorkSans-Regular.ttf +0 -0
  996. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/YoungSerif-OFL.txt +0 -0
  997. /package/{dist → dist-docs}/awesome-claude-skills/canvas-design/canvas-fonts/YoungSerif-Regular.ttf +0 -0
  998. /package/{dist → dist-docs}/awesome-claude-skills/changelog-generator/SKILL.md +0 -0
  999. /package/{dist → dist-docs}/awesome-claude-skills/competitive-ads-extractor/SKILL.md +0 -0
  1000. /package/{dist → dist-docs}/awesome-claude-skills/connect/SKILL.md +0 -0
  1001. /package/{dist → dist-docs}/awesome-claude-skills/connect-apps/SKILL.md +0 -0
  1002. /package/{dist → dist-docs}/awesome-claude-skills/connect-apps-plugin/.claude-plugin/plugin.json +0 -0
  1003. /package/{dist → dist-docs}/awesome-claude-skills/connect-apps-plugin/README.md +0 -0
  1004. /package/{dist → dist-docs}/awesome-claude-skills/connect-apps-plugin/commands/setup.md +0 -0
  1005. /package/{dist → dist-docs}/awesome-claude-skills/content-research-writer/SKILL.md +0 -0
  1006. /package/{dist → dist-docs}/awesome-claude-skills/developer-growth-analysis/SKILL.md +0 -0
  1007. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/LICENSE.txt +0 -0
  1008. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/SKILL.md +0 -0
  1009. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/docx-js.md +0 -0
  1010. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/dml-chart.xsd +0 -0
  1011. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/dml-chartDrawing.xsd +0 -0
  1012. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/dml-diagram.xsd +0 -0
  1013. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/dml-lockedCanvas.xsd +0 -0
  1014. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/dml-main.xsd +0 -0
  1015. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/dml-picture.xsd +0 -0
  1016. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/dml-spreadsheetDrawing.xsd +0 -0
  1017. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/dml-wordprocessingDrawing.xsd +0 -0
  1018. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/pml.xsd +0 -0
  1019. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-additionalCharacteristics.xsd +0 -0
  1020. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-bibliography.xsd +0 -0
  1021. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-commonSimpleTypes.xsd +0 -0
  1022. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-customXmlDataProperties.xsd +0 -0
  1023. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-customXmlSchemaProperties.xsd +0 -0
  1024. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-documentPropertiesCustom.xsd +0 -0
  1025. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-documentPropertiesExtended.xsd +0 -0
  1026. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-documentPropertiesVariantTypes.xsd +0 -0
  1027. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-math.xsd +0 -0
  1028. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/shared-relationshipReference.xsd +0 -0
  1029. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/sml.xsd +0 -0
  1030. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/vml-main.xsd +0 -0
  1031. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/vml-officeDrawing.xsd +0 -0
  1032. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/vml-presentationDrawing.xsd +0 -0
  1033. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/vml-spreadsheetDrawing.xsd +0 -0
  1034. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/vml-wordprocessingDrawing.xsd +0 -0
  1035. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/wml.xsd +0 -0
  1036. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/ISO-IEC29500-4_2016/xml.xsd +0 -0
  1037. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/ecma/fouth-edition/opc-contentTypes.xsd +0 -0
  1038. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/ecma/fouth-edition/opc-coreProperties.xsd +0 -0
  1039. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/ecma/fouth-edition/opc-digSig.xsd +0 -0
  1040. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/ecma/fouth-edition/opc-relationships.xsd +0 -0
  1041. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/mce/mc.xsd +0 -0
  1042. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/microsoft/wml-2010.xsd +0 -0
  1043. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/microsoft/wml-2012.xsd +0 -0
  1044. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/microsoft/wml-2018.xsd +0 -0
  1045. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/microsoft/wml-cex-2018.xsd +0 -0
  1046. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/microsoft/wml-cid-2016.xsd +0 -0
  1047. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/microsoft/wml-sdtdatahash-2020.xsd +0 -0
  1048. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/schemas/microsoft/wml-symex-2015.xsd +0 -0
  1049. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/scripts/pack.py +0 -0
  1050. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/scripts/unpack.py +0 -0
  1051. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/scripts/validate.py +0 -0
  1052. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/scripts/validation/__init__.py +0 -0
  1053. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/scripts/validation/base.py +0 -0
  1054. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/scripts/validation/docx.py +0 -0
  1055. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/scripts/validation/pptx.py +0 -0
  1056. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml/scripts/validation/redlining.py +0 -0
  1057. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/ooxml.md +0 -0
  1058. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/scripts/__init__.py +0 -0
  1059. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/scripts/document.py +0 -0
  1060. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/scripts/templates/comments.xml +0 -0
  1061. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/scripts/templates/commentsExtended.xml +0 -0
  1062. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/scripts/templates/commentsExtensible.xml +0 -0
  1063. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/scripts/templates/commentsIds.xml +0 -0
  1064. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/scripts/templates/people.xml +0 -0
  1065. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/docx/scripts/utilities.py +0 -0
  1066. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pdf/LICENSE.txt +0 -0
  1067. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pdf/SKILL.md +0 -0
  1068. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pdf/forms.md +0 -0
  1069. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pdf/reference.md +0 -0
  1070. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pdf/scripts/check_bounding_boxes.py +0 -0
  1071. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pdf/scripts/check_bounding_boxes_test.py +0 -0
  1072. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pdf/scripts/check_fillable_fields.py +0 -0
  1073. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pdf/scripts/convert_pdf_to_images.py +0 -0
  1074. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pdf/scripts/create_validation_image.py +0 -0
  1075. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pdf/scripts/extract_form_field_info.py +0 -0
  1076. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pdf/scripts/fill_fillable_fields.py +0 -0
  1077. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pdf/scripts/fill_pdf_form_with_annotations.py +0 -0
  1078. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/LICENSE.txt +0 -0
  1079. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/SKILL.md +0 -0
  1080. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/html2pptx.md +0 -0
  1081. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/dml-chart.xsd +0 -0
  1082. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/dml-chartDrawing.xsd +0 -0
  1083. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/dml-diagram.xsd +0 -0
  1084. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/dml-lockedCanvas.xsd +0 -0
  1085. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/dml-main.xsd +0 -0
  1086. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/dml-picture.xsd +0 -0
  1087. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/dml-spreadsheetDrawing.xsd +0 -0
  1088. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/dml-wordprocessingDrawing.xsd +0 -0
  1089. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/pml.xsd +0 -0
  1090. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-additionalCharacteristics.xsd +0 -0
  1091. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-bibliography.xsd +0 -0
  1092. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-commonSimpleTypes.xsd +0 -0
  1093. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-customXmlDataProperties.xsd +0 -0
  1094. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-customXmlSchemaProperties.xsd +0 -0
  1095. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-documentPropertiesCustom.xsd +0 -0
  1096. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-documentPropertiesExtended.xsd +0 -0
  1097. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-documentPropertiesVariantTypes.xsd +0 -0
  1098. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-math.xsd +0 -0
  1099. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/shared-relationshipReference.xsd +0 -0
  1100. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/sml.xsd +0 -0
  1101. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/vml-main.xsd +0 -0
  1102. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/vml-officeDrawing.xsd +0 -0
  1103. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/vml-presentationDrawing.xsd +0 -0
  1104. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/vml-spreadsheetDrawing.xsd +0 -0
  1105. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/vml-wordprocessingDrawing.xsd +0 -0
  1106. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/wml.xsd +0 -0
  1107. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/ISO-IEC29500-4_2016/xml.xsd +0 -0
  1108. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/ecma/fouth-edition/opc-contentTypes.xsd +0 -0
  1109. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/ecma/fouth-edition/opc-coreProperties.xsd +0 -0
  1110. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/ecma/fouth-edition/opc-digSig.xsd +0 -0
  1111. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/ecma/fouth-edition/opc-relationships.xsd +0 -0
  1112. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/mce/mc.xsd +0 -0
  1113. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/microsoft/wml-2010.xsd +0 -0
  1114. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/microsoft/wml-2012.xsd +0 -0
  1115. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/microsoft/wml-2018.xsd +0 -0
  1116. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/microsoft/wml-cex-2018.xsd +0 -0
  1117. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/microsoft/wml-cid-2016.xsd +0 -0
  1118. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/microsoft/wml-sdtdatahash-2020.xsd +0 -0
  1119. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/schemas/microsoft/wml-symex-2015.xsd +0 -0
  1120. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/scripts/pack.py +0 -0
  1121. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/scripts/unpack.py +0 -0
  1122. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/scripts/validate.py +0 -0
  1123. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/scripts/validation/__init__.py +0 -0
  1124. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/scripts/validation/base.py +0 -0
  1125. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/scripts/validation/docx.py +0 -0
  1126. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/scripts/validation/pptx.py +0 -0
  1127. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml/scripts/validation/redlining.py +0 -0
  1128. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/ooxml.md +0 -0
  1129. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/scripts/html2pptx.js +0 -0
  1130. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/scripts/inventory.py +0 -0
  1131. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/scripts/rearrange.py +0 -0
  1132. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/scripts/replace.py +0 -0
  1133. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/pptx/scripts/thumbnail.py +0 -0
  1134. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/xlsx/LICENSE.txt +0 -0
  1135. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/xlsx/SKILL.md +0 -0
  1136. /package/{dist → dist-docs}/awesome-claude-skills/document-skills/xlsx/recalc.py +0 -0
  1137. /package/{dist → dist-docs}/awesome-claude-skills/domain-name-brainstormer/SKILL.md +0 -0
  1138. /package/{dist → dist-docs}/awesome-claude-skills/file-organizer/SKILL.md +0 -0
  1139. /package/{dist → dist-docs}/awesome-claude-skills/image-enhancer/SKILL.md +0 -0
  1140. /package/{dist → dist-docs}/awesome-claude-skills/internal-comms/LICENSE.txt +0 -0
  1141. /package/{dist → dist-docs}/awesome-claude-skills/internal-comms/SKILL.md +0 -0
  1142. /package/{dist → dist-docs}/awesome-claude-skills/internal-comms/examples/3p-updates.md +0 -0
  1143. /package/{dist → dist-docs}/awesome-claude-skills/internal-comms/examples/company-newsletter.md +0 -0
  1144. /package/{dist → dist-docs}/awesome-claude-skills/internal-comms/examples/faq-answers.md +0 -0
  1145. /package/{dist → dist-docs}/awesome-claude-skills/internal-comms/examples/general-comms.md +0 -0
  1146. /package/{dist → dist-docs}/awesome-claude-skills/invoice-organizer/SKILL.md +0 -0
  1147. /package/{dist → dist-docs}/awesome-claude-skills/langsmith-fetch/SKILL.md +0 -0
  1148. /package/{dist → dist-docs}/awesome-claude-skills/lead-research-assistant/SKILL.md +0 -0
  1149. /package/{dist → dist-docs}/awesome-claude-skills/mcp-builder/LICENSE.txt +0 -0
  1150. /package/{dist → dist-docs}/awesome-claude-skills/mcp-builder/SKILL.md +0 -0
  1151. /package/{dist → dist-docs}/awesome-claude-skills/mcp-builder/reference/evaluation.md +0 -0
  1152. /package/{dist → dist-docs}/awesome-claude-skills/mcp-builder/reference/mcp_best_practices.md +0 -0
  1153. /package/{dist → dist-docs}/awesome-claude-skills/mcp-builder/reference/node_mcp_server.md +0 -0
  1154. /package/{dist → dist-docs}/awesome-claude-skills/mcp-builder/reference/python_mcp_server.md +0 -0
  1155. /package/{dist → dist-docs}/awesome-claude-skills/mcp-builder/scripts/connections.py +0 -0
  1156. /package/{dist → dist-docs}/awesome-claude-skills/mcp-builder/scripts/evaluation.py +0 -0
  1157. /package/{dist → dist-docs}/awesome-claude-skills/mcp-builder/scripts/example_evaluation.xml +0 -0
  1158. /package/{dist → dist-docs}/awesome-claude-skills/mcp-builder/scripts/requirements.txt +0 -0
  1159. /package/{dist → dist-docs}/awesome-claude-skills/meeting-insights-analyzer/SKILL.md +0 -0
  1160. /package/{dist → dist-docs}/awesome-claude-skills/raffle-winner-picker/SKILL.md +0 -0
  1161. /package/{dist → dist-docs}/awesome-claude-skills/skill-creator/LICENSE.txt +0 -0
  1162. /package/{dist → dist-docs}/awesome-claude-skills/skill-creator/SKILL.md +0 -0
  1163. /package/{dist → dist-docs}/awesome-claude-skills/skill-creator/scripts/init_skill.py +0 -0
  1164. /package/{dist → dist-docs}/awesome-claude-skills/skill-creator/scripts/package_skill.py +0 -0
  1165. /package/{dist → dist-docs}/awesome-claude-skills/skill-creator/scripts/quick_validate.py +0 -0
  1166. /package/{dist → dist-docs}/awesome-claude-skills/skill-share/SKILL.md +0 -0
  1167. /package/{dist → dist-docs}/awesome-claude-skills/slack-gif-creator/LICENSE.txt +0 -0
  1168. /package/{dist → dist-docs}/awesome-claude-skills/slack-gif-creator/SKILL.md +0 -0
  1169. /package/{dist → dist-docs}/awesome-claude-skills/slack-gif-creator/core/color_palettes.py +0 -0
  1170. /package/{dist → dist-docs}/awesome-claude-skills/slack-gif-creator/core/easing.py +0 -0
  1171. /package/{dist → dist-docs}/awesome-claude-skills/slack-gif-creator/core/frame_composer.py +0 -0
  1172. /package/{dist → dist-docs}/awesome-claude-skills/slack-gif-creator/core/gif_builder.py +0 -0
  1173. /package/{dist → dist-docs}/awesome-claude-skills/slack-gif-creator/core/typography.py +0 -0
  1174. /package/{dist → dist-docs}/awesome-claude-skills/slack-gif-creator/core/validators.py +0 -0
  1175. /package/{dist → dist-docs}/awesome-claude-skills/slack-gif-creator/core/visual_effects.py +0 -0
  1176. /package/{dist → dist-docs}/awesome-claude-skills/slack-gif-creator/requirements.txt +0 -0
  1177. /package/{dist → dist-docs}/awesome-claude-skills/slack-gif-creator/templates/bounce.py +0 -0
  1178. /package/{dist → dist-docs}/awesome-claude-skills/slack-gif-creator/templates/explode.py +0 -0
  1179. /package/{dist → dist-docs}/awesome-claude-skills/slack-gif-creator/templates/fade.py +0 -0
  1180. /package/{dist → dist-docs}/awesome-claude-skills/slack-gif-creator/templates/flip.py +0 -0
  1181. /package/{dist → dist-docs}/awesome-claude-skills/slack-gif-creator/templates/kaleidoscope.py +0 -0
  1182. /package/{dist → dist-docs}/awesome-claude-skills/slack-gif-creator/templates/morph.py +0 -0
  1183. /package/{dist → dist-docs}/awesome-claude-skills/slack-gif-creator/templates/move.py +0 -0
  1184. /package/{dist → dist-docs}/awesome-claude-skills/slack-gif-creator/templates/pulse.py +0 -0
  1185. /package/{dist → dist-docs}/awesome-claude-skills/slack-gif-creator/templates/shake.py +0 -0
  1186. /package/{dist → dist-docs}/awesome-claude-skills/slack-gif-creator/templates/slide.py +0 -0
  1187. /package/{dist → dist-docs}/awesome-claude-skills/slack-gif-creator/templates/spin.py +0 -0
  1188. /package/{dist → dist-docs}/awesome-claude-skills/slack-gif-creator/templates/wiggle.py +0 -0
  1189. /package/{dist → dist-docs}/awesome-claude-skills/slack-gif-creator/templates/zoom.py +0 -0
  1190. /package/{dist → dist-docs}/awesome-claude-skills/tailored-resume-generator/SKILL.md +0 -0
  1191. /package/{dist → dist-docs}/awesome-claude-skills/template-skill/SKILL.md +0 -0
  1192. /package/{dist → dist-docs}/awesome-claude-skills/theme-factory/LICENSE.txt +0 -0
  1193. /package/{dist → dist-docs}/awesome-claude-skills/theme-factory/SKILL.md +0 -0
  1194. /package/{dist → dist-docs}/awesome-claude-skills/theme-factory/theme-showcase.pdf +0 -0
  1195. /package/{dist → dist-docs}/awesome-claude-skills/theme-factory/themes/arctic-frost.md +0 -0
  1196. /package/{dist → dist-docs}/awesome-claude-skills/theme-factory/themes/botanical-garden.md +0 -0
  1197. /package/{dist → dist-docs}/awesome-claude-skills/theme-factory/themes/desert-rose.md +0 -0
  1198. /package/{dist → dist-docs}/awesome-claude-skills/theme-factory/themes/forest-canopy.md +0 -0
  1199. /package/{dist → dist-docs}/awesome-claude-skills/theme-factory/themes/golden-hour.md +0 -0
  1200. /package/{dist → dist-docs}/awesome-claude-skills/theme-factory/themes/midnight-galaxy.md +0 -0
  1201. /package/{dist → dist-docs}/awesome-claude-skills/theme-factory/themes/modern-minimalist.md +0 -0
  1202. /package/{dist → dist-docs}/awesome-claude-skills/theme-factory/themes/ocean-depths.md +0 -0
  1203. /package/{dist → dist-docs}/awesome-claude-skills/theme-factory/themes/sunset-boulevard.md +0 -0
  1204. /package/{dist → dist-docs}/awesome-claude-skills/theme-factory/themes/tech-innovation.md +0 -0
  1205. /package/{dist → dist-docs}/awesome-claude-skills/video-downloader/SKILL.md +0 -0
  1206. /package/{dist → dist-docs}/awesome-claude-skills/video-downloader/scripts/download_video.py +0 -0
  1207. /package/{dist → dist-docs}/awesome-claude-skills/webapp-testing/LICENSE.txt +0 -0
  1208. /package/{dist → dist-docs}/awesome-claude-skills/webapp-testing/SKILL.md +0 -0
  1209. /package/{dist → dist-docs}/awesome-claude-skills/webapp-testing/examples/console_logging.py +0 -0
  1210. /package/{dist → dist-docs}/awesome-claude-skills/webapp-testing/examples/element_discovery.py +0 -0
  1211. /package/{dist → dist-docs}/awesome-claude-skills/webapp-testing/examples/static_html_automation.py +0 -0
  1212. /package/{dist → dist-docs}/awesome-claude-skills/webapp-testing/scripts/with_server.py +0 -0
  1213. /package/{dist → dist-docs}/claude-code-settings/.claude-plugin/marketplace.json +0 -0
  1214. /package/{dist → dist-docs}/claude-code-settings/.claude-plugin/plugin.json +0 -0
  1215. /package/{dist → dist-docs}/claude-code-settings/.github/copilot-instructions.md +0 -0
  1216. /package/{dist → dist-docs}/claude-code-settings/.mcp.json +0 -0
  1217. /package/{dist → dist-docs}/claude-code-settings/LICENSE +0 -0
  1218. /package/{dist → dist-docs}/claude-code-settings/README.md +0 -0
  1219. /package/{dist → dist-docs}/claude-code-settings/agents/command-creator.md +0 -0
  1220. /package/{dist → dist-docs}/claude-code-settings/agents/deep-reflector.md +0 -0
  1221. /package/{dist → dist-docs}/claude-code-settings/agents/github-issue-fixer.md +0 -0
  1222. /package/{dist → dist-docs}/claude-code-settings/agents/insight-documenter.md +0 -0
  1223. /package/{dist → dist-docs}/claude-code-settings/agents/instruction-reflector.md +0 -0
  1224. /package/{dist → dist-docs}/claude-code-settings/agents/pr-reviewer.md +0 -0
  1225. /package/{dist → dist-docs}/claude-code-settings/agents/ui-engineer.md +0 -0
  1226. /package/{dist → dist-docs}/claude-code-settings/commands/cc/create-command.md +0 -0
  1227. /package/{dist → dist-docs}/claude-code-settings/commands/eureka.md +0 -0
  1228. /package/{dist → dist-docs}/claude-code-settings/commands/gh/fix-issue.md +0 -0
  1229. /package/{dist → dist-docs}/claude-code-settings/commands/gh/review-pr.md +0 -0
  1230. /package/{dist → dist-docs}/claude-code-settings/commands/reflection-harder.md +0 -0
  1231. /package/{dist → dist-docs}/claude-code-settings/commands/reflection.md +0 -0
  1232. /package/{dist → dist-docs}/claude-code-settings/commands/think-harder.md +0 -0
  1233. /package/{dist → dist-docs}/claude-code-settings/commands/think-ultra.md +0 -0
  1234. /package/{dist → dist-docs}/claude-code-settings/commands/translate.md +0 -0
  1235. /package/{dist → dist-docs}/claude-code-settings/config.json +0 -0
  1236. /package/{dist → dist-docs}/claude-code-settings/guidances/github-copilot.md +0 -0
  1237. /package/{dist → dist-docs}/claude-code-settings/guidances/litellm_config.yaml +0 -0
  1238. /package/{dist → dist-docs}/claude-code-settings/guidances/llm-gateway-litellm.md +0 -0
  1239. /package/{dist → dist-docs}/claude-code-settings/plugins/autonomous-skill/.claude-plugin/plugin.json +0 -0
  1240. /package/{dist → dist-docs}/claude-code-settings/plugins/autonomous-skill/skills/autonomous-skill/SKILL.md +0 -0
  1241. /package/{dist → dist-docs}/claude-code-settings/plugins/autonomous-skill/skills/autonomous-skill/scripts/run-session.sh +0 -0
  1242. /package/{dist → dist-docs}/claude-code-settings/plugins/autonomous-skill/skills/autonomous-skill/templates/executor-prompt.md +0 -0
  1243. /package/{dist → dist-docs}/claude-code-settings/plugins/autonomous-skill/skills/autonomous-skill/templates/initializer-prompt.md +0 -0
  1244. /package/{dist → dist-docs}/claude-code-settings/plugins/autonomous-skill/skills/autonomous-skill/templates/task-list-template.md +0 -0
  1245. /package/{dist → dist-docs}/claude-code-settings/plugins/codex-skill/.claude-plugin/plugin.json +0 -0
  1246. /package/{dist → dist-docs}/claude-code-settings/plugins/codex-skill/skills/codex-skill/SKILL.md +0 -0
  1247. /package/{dist → dist-docs}/claude-code-settings/plugins/kiro-skill/.claude-plugin/plugin.json +0 -0
  1248. /package/{dist → dist-docs}/claude-code-settings/plugins/kiro-skill/commands/kiro/design.md +0 -0
  1249. /package/{dist → dist-docs}/claude-code-settings/plugins/kiro-skill/commands/kiro/execute.md +0 -0
  1250. /package/{dist → dist-docs}/claude-code-settings/plugins/kiro-skill/commands/kiro/spec.md +0 -0
  1251. /package/{dist → dist-docs}/claude-code-settings/plugins/kiro-skill/commands/kiro/task.md +0 -0
  1252. /package/{dist → dist-docs}/claude-code-settings/plugins/kiro-skill/commands/kiro/vibe.md +0 -0
  1253. /package/{dist → dist-docs}/claude-code-settings/plugins/kiro-skill/skills/kiro-skill/SKILL.md +0 -0
  1254. /package/{dist → dist-docs}/claude-code-settings/plugins/kiro-skill/skills/kiro-skill/helpers/kiro-identity.md +0 -0
  1255. /package/{dist → dist-docs}/claude-code-settings/plugins/kiro-skill/skills/kiro-skill/helpers/workflow-diagrams.md +0 -0
  1256. /package/{dist → dist-docs}/claude-code-settings/plugins/nanobanana-skill/.claude-plugin/plugin.json +0 -0
  1257. /package/{dist → dist-docs}/claude-code-settings/plugins/nanobanana-skill/skills/nanobanana-skill/SKILL.md +0 -0
  1258. /package/{dist → dist-docs}/claude-code-settings/plugins/nanobanana-skill/skills/nanobanana-skill/nanobanana.py +0 -0
  1259. /package/{dist → dist-docs}/claude-code-settings/plugins/nanobanana-skill/skills/nanobanana-skill/requirements.txt +0 -0
  1260. /package/{dist → dist-docs}/claude-code-settings/plugins/spec-kit-skill/.claude-plugin/plugin.json +0 -0
  1261. /package/{dist → dist-docs}/claude-code-settings/plugins/spec-kit-skill/skills/spec-kit-skill/SKILL.md +0 -0
  1262. /package/{dist → dist-docs}/claude-code-settings/plugins/spec-kit-skill/skills/spec-kit-skill/helpers/detection-logic.md +0 -0
  1263. /package/{dist → dist-docs}/claude-code-settings/plugins/spec-kit-skill/skills/spec-kit-skill/scripts/detect-phase.sh +0 -0
  1264. /package/{dist → dist-docs}/claude-code-settings/plugins/youtube-transcribe-skill/.claude-plugin/plugin.json +0 -0
  1265. /package/{dist → dist-docs}/claude-code-settings/plugins/youtube-transcribe-skill/skills/youtube-transcribe-skill/SKILL.md +0 -0
  1266. /package/{dist → dist-docs}/claude-code-settings/settings/README.md +0 -0
  1267. /package/{dist → dist-docs}/claude-code-settings/settings/azure-foundry-settings.json +0 -0
  1268. /package/{dist → dist-docs}/claude-code-settings/settings/azure-settings.json +0 -0
  1269. /package/{dist → dist-docs}/claude-code-settings/settings/copilot-settings.json +0 -0
  1270. /package/{dist → dist-docs}/claude-code-settings/settings/deepseek-settings.json +0 -0
  1271. /package/{dist → dist-docs}/claude-code-settings/settings/litellm-settings.json +0 -0
  1272. /package/{dist → dist-docs}/claude-code-settings/settings/minimax.json +0 -0
  1273. /package/{dist → dist-docs}/claude-code-settings/settings/openrouter-settings.json +0 -0
  1274. /package/{dist → dist-docs}/claude-code-settings/settings/qwen-settings.json +0 -0
  1275. /package/{dist → dist-docs}/claude-code-settings/settings/siliconflow-settings.json +0 -0
  1276. /package/{dist → dist-docs}/claude-code-settings/settings/vertex-settings.json +0 -0
  1277. /package/{dist → dist-docs}/claude-code-settings/settings.json +0 -0
  1278. /package/{dist → dist-docs}/claude-code-settings/skills/autonomous-skill +0 -0
  1279. /package/{dist → dist-docs}/claude-code-settings/skills/codex-skill +0 -0
  1280. /package/{dist → dist-docs}/claude-code-settings/skills/kiro-skill +0 -0
  1281. /package/{dist → dist-docs}/claude-code-settings/skills/nanobanana-skill +0 -0
  1282. /package/{dist → dist-docs}/claude-code-settings/skills/spec-kit-skill +0 -0
  1283. /package/{dist → dist-docs}/claude-code-settings/skills/youtube-transcribe-skill +0 -0
  1284. /package/{dist → dist-docs}/claude-code-settings/status-line.sh +0 -0
  1285. /package/{dist → dist-docs}/claude-code-skills/About.md +0 -0
  1286. /package/{dist → dist-docs}/claude-code-skills/Detail.md +0 -0
  1287. /package/{dist → dist-docs}/claude-code-skills/agent-skills-architecture.png +0 -0
  1288. /package/{dist → dist-docs}/claude-code-skills/agent-skills-context-window.png +0 -0
  1289. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.agent/workflows/ui-ux-pro-max.md +0 -0
  1290. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.claude/skills/ui-ux-pro-max/SKILL.md +0 -0
  1291. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.claude/skills/ui-ux-pro-max/data/charts.csv +0 -0
  1292. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.claude/skills/ui-ux-pro-max/data/colors.csv +0 -0
  1293. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.claude/skills/ui-ux-pro-max/data/icons.csv +0 -0
  1294. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.claude/skills/ui-ux-pro-max/data/landing.csv +0 -0
  1295. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.claude/skills/ui-ux-pro-max/data/products.csv +0 -0
  1296. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.claude/skills/ui-ux-pro-max/data/prompts.csv +0 -0
  1297. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.claude/skills/ui-ux-pro-max/data/react-performance.csv +0 -0
  1298. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.claude/skills/ui-ux-pro-max/data/stacks/flutter.csv +0 -0
  1299. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.claude/skills/ui-ux-pro-max/data/stacks/html-tailwind.csv +0 -0
  1300. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.claude/skills/ui-ux-pro-max/data/stacks/jetpack-compose.csv +0 -0
  1301. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.claude/skills/ui-ux-pro-max/data/stacks/nextjs.csv +0 -0
  1302. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.claude/skills/ui-ux-pro-max/data/stacks/nuxt-ui.csv +0 -0
  1303. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.claude/skills/ui-ux-pro-max/data/stacks/nuxtjs.csv +0 -0
  1304. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.claude/skills/ui-ux-pro-max/data/stacks/react-native.csv +0 -0
  1305. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.claude/skills/ui-ux-pro-max/data/stacks/react.csv +0 -0
  1306. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.claude/skills/ui-ux-pro-max/data/stacks/shadcn.csv +0 -0
  1307. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.claude/skills/ui-ux-pro-max/data/stacks/svelte.csv +0 -0
  1308. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.claude/skills/ui-ux-pro-max/data/stacks/swiftui.csv +0 -0
  1309. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.claude/skills/ui-ux-pro-max/data/stacks/vue.csv +0 -0
  1310. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.claude/skills/ui-ux-pro-max/data/styles.csv +0 -0
  1311. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.claude/skills/ui-ux-pro-max/data/typography.csv +0 -0
  1312. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.claude/skills/ui-ux-pro-max/data/ui-reasoning.csv +0 -0
  1313. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.claude/skills/ui-ux-pro-max/data/ux-guidelines.csv +0 -0
  1314. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.claude/skills/ui-ux-pro-max/data/web-interface.csv +0 -0
  1315. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.claude/skills/ui-ux-pro-max/scripts/core.py +0 -0
  1316. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.claude/skills/ui-ux-pro-max/scripts/design_system.py +0 -0
  1317. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.claude/skills/ui-ux-pro-max/scripts/search.py +0 -0
  1318. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.claude-plugin/marketplace.json +0 -0
  1319. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.codebuddy/commands/ui-ux-pro-max.md +0 -0
  1320. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.codex/skills/ui-ux-pro-max/SKILL.md +0 -0
  1321. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.codex/skills/ui-ux-pro-max/data/charts.csv +0 -0
  1322. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.codex/skills/ui-ux-pro-max/data/colors.csv +0 -0
  1323. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.codex/skills/ui-ux-pro-max/data/icons.csv +0 -0
  1324. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.codex/skills/ui-ux-pro-max/data/landing.csv +0 -0
  1325. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.codex/skills/ui-ux-pro-max/data/products.csv +0 -0
  1326. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.codex/skills/ui-ux-pro-max/data/prompts.csv +0 -0
  1327. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.codex/skills/ui-ux-pro-max/data/stacks/flutter.csv +0 -0
  1328. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.codex/skills/ui-ux-pro-max/data/stacks/html-tailwind.csv +0 -0
  1329. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.codex/skills/ui-ux-pro-max/data/stacks/nextjs.csv +0 -0
  1330. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.codex/skills/ui-ux-pro-max/data/stacks/nuxt-ui.csv +0 -0
  1331. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.codex/skills/ui-ux-pro-max/data/stacks/nuxtjs.csv +0 -0
  1332. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.codex/skills/ui-ux-pro-max/data/stacks/react-native.csv +0 -0
  1333. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.codex/skills/ui-ux-pro-max/data/stacks/react.csv +0 -0
  1334. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.codex/skills/ui-ux-pro-max/data/stacks/shadcn.csv +0 -0
  1335. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.codex/skills/ui-ux-pro-max/data/stacks/svelte.csv +0 -0
  1336. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.codex/skills/ui-ux-pro-max/data/stacks/swiftui.csv +0 -0
  1337. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.codex/skills/ui-ux-pro-max/data/stacks/vue.csv +0 -0
  1338. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.codex/skills/ui-ux-pro-max/data/styles.csv +0 -0
  1339. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.codex/skills/ui-ux-pro-max/data/typography.csv +0 -0
  1340. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.codex/skills/ui-ux-pro-max/data/ui-reasoning.csv +0 -0
  1341. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.codex/skills/ui-ux-pro-max/data/ux-guidelines.csv +0 -0
  1342. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.codex/skills/ui-ux-pro-max/scripts/core.py +0 -0
  1343. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.codex/skills/ui-ux-pro-max/scripts/design_system.py +0 -0
  1344. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.codex/skills/ui-ux-pro-max/scripts/search.py +0 -0
  1345. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.cursor/commands/ui-ux-pro-max.md +0 -0
  1346. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.gemini/skills/ui-ux-pro-max/SKILL.md +0 -0
  1347. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.gemini/skills/ui-ux-pro-max/data/charts.csv +0 -0
  1348. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.gemini/skills/ui-ux-pro-max/data/colors.csv +0 -0
  1349. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.gemini/skills/ui-ux-pro-max/data/icons.csv +0 -0
  1350. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.gemini/skills/ui-ux-pro-max/data/landing.csv +0 -0
  1351. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.gemini/skills/ui-ux-pro-max/data/products.csv +0 -0
  1352. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.gemini/skills/ui-ux-pro-max/data/prompts.csv +0 -0
  1353. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.gemini/skills/ui-ux-pro-max/data/react-performance.csv +0 -0
  1354. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.gemini/skills/ui-ux-pro-max/data/stacks/flutter.csv +0 -0
  1355. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.gemini/skills/ui-ux-pro-max/data/stacks/html-tailwind.csv +0 -0
  1356. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.gemini/skills/ui-ux-pro-max/data/stacks/nextjs.csv +0 -0
  1357. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.gemini/skills/ui-ux-pro-max/data/stacks/nuxt-ui.csv +0 -0
  1358. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.gemini/skills/ui-ux-pro-max/data/stacks/nuxtjs.csv +0 -0
  1359. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.gemini/skills/ui-ux-pro-max/data/stacks/react-native.csv +0 -0
  1360. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.gemini/skills/ui-ux-pro-max/data/stacks/react.csv +0 -0
  1361. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.gemini/skills/ui-ux-pro-max/data/stacks/shadcn.csv +0 -0
  1362. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.gemini/skills/ui-ux-pro-max/data/stacks/svelte.csv +0 -0
  1363. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.gemini/skills/ui-ux-pro-max/data/stacks/swiftui.csv +0 -0
  1364. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.gemini/skills/ui-ux-pro-max/data/stacks/vue.csv +0 -0
  1365. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.gemini/skills/ui-ux-pro-max/data/styles.csv +0 -0
  1366. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.gemini/skills/ui-ux-pro-max/data/typography.csv +0 -0
  1367. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.gemini/skills/ui-ux-pro-max/data/ui-reasoning.csv +0 -0
  1368. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.gemini/skills/ui-ux-pro-max/data/ux-guidelines.csv +0 -0
  1369. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.gemini/skills/ui-ux-pro-max/data/web-interface.csv +0 -0
  1370. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.gemini/skills/ui-ux-pro-max/scripts/core.py +0 -0
  1371. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.gemini/skills/ui-ux-pro-max/scripts/design_system.py +0 -0
  1372. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.gemini/skills/ui-ux-pro-max/scripts/search.py +0 -0
  1373. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.github/prompts/ui-ux-pro-max.prompt.md +0 -0
  1374. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.kiro/steering/ui-ux-pro-max.md +0 -0
  1375. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.opencode/skills/ui-ux-pro-max/SKILL.md +0 -0
  1376. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.qoder/rules/charts.md +0 -0
  1377. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.qoder/rules/color.md +0 -0
  1378. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.qoder/rules/flutter.md +0 -0
  1379. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.qoder/rules/html-tailwind.md +0 -0
  1380. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.qoder/rules/icons.md +0 -0
  1381. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.qoder/rules/loading.md +0 -0
  1382. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.qoder/rules/nextjs.md +0 -0
  1383. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.qoder/rules/nuxt-ui.md +0 -0
  1384. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.qoder/rules/nuxtjs.md +0 -0
  1385. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.qoder/rules/products.md +0 -0
  1386. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.qoder/rules/prompts.md +0 -0
  1387. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.qoder/rules/react-native.md +0 -0
  1388. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.qoder/rules/react.md +0 -0
  1389. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.qoder/rules/shadcn.md +0 -0
  1390. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.qoder/rules/styles.md +0 -0
  1391. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.qoder/rules/svelte.md +0 -0
  1392. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.qoder/rules/swiftui.md +0 -0
  1393. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.qoder/rules/typography.md +0 -0
  1394. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.qoder/rules/ui-ux-pro-max.md +0 -0
  1395. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.qoder/rules/ux-guideline.md +0 -0
  1396. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.qoder/rules/vue.md +0 -0
  1397. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.qoder/skills/ui-ux-pro-max.md +0 -0
  1398. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.roo/commands/ui-ux-pro-max.md +0 -0
  1399. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.roo/rules/ui-ux-pro-max.md +0 -0
  1400. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.shared/ui-ux-pro-max/data/charts.csv +0 -0
  1401. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.shared/ui-ux-pro-max/data/colors.csv +0 -0
  1402. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.shared/ui-ux-pro-max/data/icons.csv +0 -0
  1403. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.shared/ui-ux-pro-max/data/landing.csv +0 -0
  1404. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.shared/ui-ux-pro-max/data/products.csv +0 -0
  1405. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.shared/ui-ux-pro-max/data/prompts.csv +0 -0
  1406. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.shared/ui-ux-pro-max/data/react-performance.csv +0 -0
  1407. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.shared/ui-ux-pro-max/data/stacks/flutter.csv +0 -0
  1408. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.shared/ui-ux-pro-max/data/stacks/html-tailwind.csv +0 -0
  1409. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.shared/ui-ux-pro-max/data/stacks/jetpack-compose.csv +0 -0
  1410. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.shared/ui-ux-pro-max/data/stacks/nextjs.csv +0 -0
  1411. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.shared/ui-ux-pro-max/data/stacks/nuxt-ui.csv +0 -0
  1412. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.shared/ui-ux-pro-max/data/stacks/nuxtjs.csv +0 -0
  1413. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.shared/ui-ux-pro-max/data/stacks/react-native.csv +0 -0
  1414. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.shared/ui-ux-pro-max/data/stacks/react.csv +0 -0
  1415. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.shared/ui-ux-pro-max/data/stacks/shadcn.csv +0 -0
  1416. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.shared/ui-ux-pro-max/data/stacks/svelte.csv +0 -0
  1417. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.shared/ui-ux-pro-max/data/stacks/swiftui.csv +0 -0
  1418. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.shared/ui-ux-pro-max/data/stacks/vue.csv +0 -0
  1419. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.shared/ui-ux-pro-max/data/styles.csv +0 -0
  1420. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.shared/ui-ux-pro-max/data/typography.csv +0 -0
  1421. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.shared/ui-ux-pro-max/data/ui-reasoning.csv +0 -0
  1422. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.shared/ui-ux-pro-max/data/ux-guidelines.csv +0 -0
  1423. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.shared/ui-ux-pro-max/data/web-interface.csv +0 -0
  1424. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.shared/ui-ux-pro-max/scripts/core.py +0 -0
  1425. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.shared/ui-ux-pro-max/scripts/design_system.py +0 -0
  1426. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.shared/ui-ux-pro-max/scripts/search.py +0 -0
  1427. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.trae/skills/ui-ux-pro-max/SKILL.md +0 -0
  1428. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.trae/skills/ui-ux-pro-max/data/charts.csv +0 -0
  1429. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.trae/skills/ui-ux-pro-max/data/colors.csv +0 -0
  1430. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.trae/skills/ui-ux-pro-max/data/icons.csv +0 -0
  1431. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.trae/skills/ui-ux-pro-max/data/landing.csv +0 -0
  1432. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.trae/skills/ui-ux-pro-max/data/products.csv +0 -0
  1433. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.trae/skills/ui-ux-pro-max/data/prompts.csv +0 -0
  1434. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.trae/skills/ui-ux-pro-max/data/stacks/flutter.csv +0 -0
  1435. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.trae/skills/ui-ux-pro-max/data/stacks/html-tailwind.csv +0 -0
  1436. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.trae/skills/ui-ux-pro-max/data/stacks/nextjs.csv +0 -0
  1437. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.trae/skills/ui-ux-pro-max/data/stacks/nuxt-ui.csv +0 -0
  1438. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.trae/skills/ui-ux-pro-max/data/stacks/nuxtjs.csv +0 -0
  1439. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.trae/skills/ui-ux-pro-max/data/stacks/react-native.csv +0 -0
  1440. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.trae/skills/ui-ux-pro-max/data/stacks/react.csv +0 -0
  1441. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.trae/skills/ui-ux-pro-max/data/stacks/shadcn.csv +0 -0
  1442. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.trae/skills/ui-ux-pro-max/data/stacks/svelte.csv +0 -0
  1443. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.trae/skills/ui-ux-pro-max/data/stacks/swiftui.csv +0 -0
  1444. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.trae/skills/ui-ux-pro-max/data/stacks/vue.csv +0 -0
  1445. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.trae/skills/ui-ux-pro-max/data/styles.csv +0 -0
  1446. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.trae/skills/ui-ux-pro-max/data/typography.csv +0 -0
  1447. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.trae/skills/ui-ux-pro-max/data/ui-reasoning.csv +0 -0
  1448. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.trae/skills/ui-ux-pro-max/data/ux-guidelines.csv +0 -0
  1449. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.trae/skills/ui-ux-pro-max/scripts/core.py +0 -0
  1450. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.trae/skills/ui-ux-pro-max/scripts/design_system.py +0 -0
  1451. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.trae/skills/ui-ux-pro-max/scripts/search.py +0 -0
  1452. /package/{dist → dist-docs}/ui-ux-pro-max-skill/.windsurf/workflows/ui-ux-pro-max.md +0 -0
  1453. /package/{dist → dist-docs}/ui-ux-pro-max-skill/CLAUDE.md +0 -0
  1454. /package/{dist → dist-docs}/ui-ux-pro-max-skill/LICENSE +0 -0
  1455. /package/{dist → dist-docs}/ui-ux-pro-max-skill/README.md +0 -0
  1456. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/README.md +0 -0
  1457. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.agent/workflows/ui-ux-pro-max.md +0 -0
  1458. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.claude/skills/ui-ux-pro-max/SKILL.md +0 -0
  1459. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.claude/skills/ui-ux-pro-max/data/charts.csv +0 -0
  1460. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.claude/skills/ui-ux-pro-max/data/colors.csv +0 -0
  1461. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.claude/skills/ui-ux-pro-max/data/icons.csv +0 -0
  1462. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.claude/skills/ui-ux-pro-max/data/landing.csv +0 -0
  1463. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.claude/skills/ui-ux-pro-max/data/products.csv +0 -0
  1464. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.claude/skills/ui-ux-pro-max/data/prompts.csv +0 -0
  1465. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.claude/skills/ui-ux-pro-max/data/react-performance.csv +0 -0
  1466. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.claude/skills/ui-ux-pro-max/data/stacks/flutter.csv +0 -0
  1467. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.claude/skills/ui-ux-pro-max/data/stacks/html-tailwind.csv +0 -0
  1468. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.claude/skills/ui-ux-pro-max/data/stacks/jetpack-compose.csv +0 -0
  1469. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.claude/skills/ui-ux-pro-max/data/stacks/nextjs.csv +0 -0
  1470. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.claude/skills/ui-ux-pro-max/data/stacks/nuxt-ui.csv +0 -0
  1471. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.claude/skills/ui-ux-pro-max/data/stacks/nuxtjs.csv +0 -0
  1472. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.claude/skills/ui-ux-pro-max/data/stacks/react-native.csv +0 -0
  1473. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.claude/skills/ui-ux-pro-max/data/stacks/react.csv +0 -0
  1474. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.claude/skills/ui-ux-pro-max/data/stacks/shadcn.csv +0 -0
  1475. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.claude/skills/ui-ux-pro-max/data/stacks/svelte.csv +0 -0
  1476. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.claude/skills/ui-ux-pro-max/data/stacks/swiftui.csv +0 -0
  1477. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.claude/skills/ui-ux-pro-max/data/stacks/vue.csv +0 -0
  1478. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.claude/skills/ui-ux-pro-max/data/styles.csv +0 -0
  1479. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.claude/skills/ui-ux-pro-max/data/typography.csv +0 -0
  1480. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.claude/skills/ui-ux-pro-max/data/ui-reasoning.csv +0 -0
  1481. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.claude/skills/ui-ux-pro-max/data/ux-guidelines.csv +0 -0
  1482. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.claude/skills/ui-ux-pro-max/data/web-interface.csv +0 -0
  1483. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.claude/skills/ui-ux-pro-max/scripts/core.py +0 -0
  1484. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.claude/skills/ui-ux-pro-max/scripts/design_system.py +0 -0
  1485. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.claude/skills/ui-ux-pro-max/scripts/search.py +0 -0
  1486. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.codex/skills/ui-ux-pro-max/SKILL.md +0 -0
  1487. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.codex/skills/ui-ux-pro-max/data/charts.csv +0 -0
  1488. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.codex/skills/ui-ux-pro-max/data/colors.csv +0 -0
  1489. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.codex/skills/ui-ux-pro-max/data/icons.csv +0 -0
  1490. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.codex/skills/ui-ux-pro-max/data/landing.csv +0 -0
  1491. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.codex/skills/ui-ux-pro-max/data/products.csv +0 -0
  1492. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.codex/skills/ui-ux-pro-max/data/prompts.csv +0 -0
  1493. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.codex/skills/ui-ux-pro-max/data/stacks/flutter.csv +0 -0
  1494. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.codex/skills/ui-ux-pro-max/data/stacks/html-tailwind.csv +0 -0
  1495. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.codex/skills/ui-ux-pro-max/data/stacks/nextjs.csv +0 -0
  1496. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.codex/skills/ui-ux-pro-max/data/stacks/nuxt-ui.csv +0 -0
  1497. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.codex/skills/ui-ux-pro-max/data/stacks/nuxtjs.csv +0 -0
  1498. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.codex/skills/ui-ux-pro-max/data/stacks/react-native.csv +0 -0
  1499. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.codex/skills/ui-ux-pro-max/data/stacks/react.csv +0 -0
  1500. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.codex/skills/ui-ux-pro-max/data/stacks/shadcn.csv +0 -0
  1501. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.codex/skills/ui-ux-pro-max/data/stacks/svelte.csv +0 -0
  1502. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.codex/skills/ui-ux-pro-max/data/stacks/swiftui.csv +0 -0
  1503. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.codex/skills/ui-ux-pro-max/data/stacks/vue.csv +0 -0
  1504. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.codex/skills/ui-ux-pro-max/data/styles.csv +0 -0
  1505. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.codex/skills/ui-ux-pro-max/data/typography.csv +0 -0
  1506. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.codex/skills/ui-ux-pro-max/data/ui-reasoning.csv +0 -0
  1507. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.codex/skills/ui-ux-pro-max/data/ux-guidelines.csv +0 -0
  1508. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.codex/skills/ui-ux-pro-max/scripts/core.py +0 -0
  1509. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.codex/skills/ui-ux-pro-max/scripts/design_system.py +0 -0
  1510. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.codex/skills/ui-ux-pro-max/scripts/search.py +0 -0
  1511. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.cursor/commands/ui-ux-pro-max.md +0 -0
  1512. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.gemini/skills/ui-ux-pro-max/SKILL.md +0 -0
  1513. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.gemini/skills/ui-ux-pro-max/data/charts.csv +0 -0
  1514. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.gemini/skills/ui-ux-pro-max/data/colors.csv +0 -0
  1515. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.gemini/skills/ui-ux-pro-max/data/icons.csv +0 -0
  1516. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.gemini/skills/ui-ux-pro-max/data/landing.csv +0 -0
  1517. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.gemini/skills/ui-ux-pro-max/data/products.csv +0 -0
  1518. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.gemini/skills/ui-ux-pro-max/data/prompts.csv +0 -0
  1519. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.gemini/skills/ui-ux-pro-max/data/react-performance.csv +0 -0
  1520. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.gemini/skills/ui-ux-pro-max/data/stacks/flutter.csv +0 -0
  1521. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.gemini/skills/ui-ux-pro-max/data/stacks/html-tailwind.csv +0 -0
  1522. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.gemini/skills/ui-ux-pro-max/data/stacks/nextjs.csv +0 -0
  1523. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.gemini/skills/ui-ux-pro-max/data/stacks/nuxt-ui.csv +0 -0
  1524. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.gemini/skills/ui-ux-pro-max/data/stacks/nuxtjs.csv +0 -0
  1525. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.gemini/skills/ui-ux-pro-max/data/stacks/react-native.csv +0 -0
  1526. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.gemini/skills/ui-ux-pro-max/data/stacks/react.csv +0 -0
  1527. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.gemini/skills/ui-ux-pro-max/data/stacks/shadcn.csv +0 -0
  1528. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.gemini/skills/ui-ux-pro-max/data/stacks/svelte.csv +0 -0
  1529. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.gemini/skills/ui-ux-pro-max/data/stacks/swiftui.csv +0 -0
  1530. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.gemini/skills/ui-ux-pro-max/data/stacks/vue.csv +0 -0
  1531. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.gemini/skills/ui-ux-pro-max/data/styles.csv +0 -0
  1532. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.gemini/skills/ui-ux-pro-max/data/typography.csv +0 -0
  1533. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.gemini/skills/ui-ux-pro-max/data/ui-reasoning.csv +0 -0
  1534. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.gemini/skills/ui-ux-pro-max/data/ux-guidelines.csv +0 -0
  1535. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.gemini/skills/ui-ux-pro-max/data/web-interface.csv +0 -0
  1536. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.gemini/skills/ui-ux-pro-max/scripts/core.py +0 -0
  1537. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.gemini/skills/ui-ux-pro-max/scripts/design_system.py +0 -0
  1538. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.gemini/skills/ui-ux-pro-max/scripts/search.py +0 -0
  1539. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.github/prompts/ui-ux-pro-max.prompt.md +0 -0
  1540. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.kiro/steering/ui-ux-pro-max.md +0 -0
  1541. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.opencode/skills/ui-ux-pro-max/SKILL.md +0 -0
  1542. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.qoder/rules/ui-ux-pro-max.md +0 -0
  1543. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.qoder/skills/ui-ux-pro-max.md +0 -0
  1544. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.roo/commands/ui-ux-pro-max.md +0 -0
  1545. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.roo/rules/ui-ux-pro-max.md +0 -0
  1546. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.shared/ui-ux-pro-max/data/charts.csv +0 -0
  1547. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.shared/ui-ux-pro-max/data/colors.csv +0 -0
  1548. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.shared/ui-ux-pro-max/data/icons.csv +0 -0
  1549. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.shared/ui-ux-pro-max/data/landing.csv +0 -0
  1550. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.shared/ui-ux-pro-max/data/products.csv +0 -0
  1551. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.shared/ui-ux-pro-max/data/prompts.csv +0 -0
  1552. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.shared/ui-ux-pro-max/data/react-performance.csv +0 -0
  1553. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.shared/ui-ux-pro-max/data/stacks/flutter.csv +0 -0
  1554. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.shared/ui-ux-pro-max/data/stacks/html-tailwind.csv +0 -0
  1555. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.shared/ui-ux-pro-max/data/stacks/jetpack-compose.csv +0 -0
  1556. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.shared/ui-ux-pro-max/data/stacks/nextjs.csv +0 -0
  1557. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.shared/ui-ux-pro-max/data/stacks/nuxt-ui.csv +0 -0
  1558. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.shared/ui-ux-pro-max/data/stacks/nuxtjs.csv +0 -0
  1559. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.shared/ui-ux-pro-max/data/stacks/react-native.csv +0 -0
  1560. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.shared/ui-ux-pro-max/data/stacks/react.csv +0 -0
  1561. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.shared/ui-ux-pro-max/data/stacks/shadcn.csv +0 -0
  1562. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.shared/ui-ux-pro-max/data/stacks/svelte.csv +0 -0
  1563. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.shared/ui-ux-pro-max/data/stacks/swiftui.csv +0 -0
  1564. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.shared/ui-ux-pro-max/data/stacks/vue.csv +0 -0
  1565. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.shared/ui-ux-pro-max/data/styles.csv +0 -0
  1566. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.shared/ui-ux-pro-max/data/typography.csv +0 -0
  1567. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.shared/ui-ux-pro-max/data/ui-reasoning.csv +0 -0
  1568. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.shared/ui-ux-pro-max/data/ux-guidelines.csv +0 -0
  1569. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.shared/ui-ux-pro-max/data/web-interface.csv +0 -0
  1570. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.shared/ui-ux-pro-max/scripts/core.py +0 -0
  1571. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.shared/ui-ux-pro-max/scripts/design_system.py +0 -0
  1572. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.shared/ui-ux-pro-max/scripts/search.py +0 -0
  1573. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.trae/skills/ui-ux-pro-max/SKILL.md +0 -0
  1574. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.trae/skills/ui-ux-pro-max/data/charts.csv +0 -0
  1575. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.trae/skills/ui-ux-pro-max/data/colors.csv +0 -0
  1576. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.trae/skills/ui-ux-pro-max/data/icons.csv +0 -0
  1577. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.trae/skills/ui-ux-pro-max/data/landing.csv +0 -0
  1578. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.trae/skills/ui-ux-pro-max/data/products.csv +0 -0
  1579. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.trae/skills/ui-ux-pro-max/data/prompts.csv +0 -0
  1580. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.trae/skills/ui-ux-pro-max/data/stacks/flutter.csv +0 -0
  1581. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.trae/skills/ui-ux-pro-max/data/stacks/html-tailwind.csv +0 -0
  1582. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.trae/skills/ui-ux-pro-max/data/stacks/nextjs.csv +0 -0
  1583. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.trae/skills/ui-ux-pro-max/data/stacks/nuxt-ui.csv +0 -0
  1584. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.trae/skills/ui-ux-pro-max/data/stacks/nuxtjs.csv +0 -0
  1585. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.trae/skills/ui-ux-pro-max/data/stacks/react-native.csv +0 -0
  1586. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.trae/skills/ui-ux-pro-max/data/stacks/react.csv +0 -0
  1587. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.trae/skills/ui-ux-pro-max/data/stacks/shadcn.csv +0 -0
  1588. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.trae/skills/ui-ux-pro-max/data/stacks/svelte.csv +0 -0
  1589. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.trae/skills/ui-ux-pro-max/data/stacks/swiftui.csv +0 -0
  1590. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.trae/skills/ui-ux-pro-max/data/stacks/vue.csv +0 -0
  1591. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.trae/skills/ui-ux-pro-max/data/styles.csv +0 -0
  1592. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.trae/skills/ui-ux-pro-max/data/typography.csv +0 -0
  1593. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.trae/skills/ui-ux-pro-max/data/ui-reasoning.csv +0 -0
  1594. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.trae/skills/ui-ux-pro-max/data/ux-guidelines.csv +0 -0
  1595. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.trae/skills/ui-ux-pro-max/scripts/core.py +0 -0
  1596. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.trae/skills/ui-ux-pro-max/scripts/design_system.py +0 -0
  1597. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.trae/skills/ui-ux-pro-max/scripts/search.py +0 -0
  1598. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/assets/.windsurf/workflows/ui-ux-pro-max.md +0 -0
  1599. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/bun.lock +0 -0
  1600. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/package-lock.json +0 -0
  1601. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/package.json +0 -0
  1602. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/src/commands/init.ts +0 -0
  1603. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/src/commands/update.ts +0 -0
  1604. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/src/commands/versions.ts +0 -0
  1605. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/src/index.ts +0 -0
  1606. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/src/types/index.ts +0 -0
  1607. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/src/utils/detect.ts +0 -0
  1608. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/src/utils/extract.ts +0 -0
  1609. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/src/utils/github.ts +0 -0
  1610. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/src/utils/logger.ts +0 -0
  1611. /package/{dist → dist-docs}/ui-ux-pro-max-skill/cli/tsconfig.json +0 -0
  1612. /package/{dist → dist-docs}/ui-ux-pro-max-skill/screenshots/website.png +0 -0
@@ -0,0 +1,3176 @@
1
+ # TINKER DOCUMENTATION
2
+ This file contains the complete Tinker documentation and SDK reference.
3
+
4
+ ## Table of Contents
5
+
6
+ 1. Documentation (MDX files)
7
+ 2. Type Definitions (from tinker.types)
8
+
9
+ ---
10
+
11
+ # PART 1: DOCUMENTATION
12
+
13
+ ## File: index.mdx
14
+
15
+ # Tinker: a training API for researchers and developers
16
+
17
+ Tinker lets you focus on what matters in LLM fine-tuning – your data and algorithms – while we handle the heavy lifting of distributed training.
18
+
19
+ You write a simple loop that runs on your CPU-only machine, including the data or environment and the loss function. We figure out how to make the training work on a bunch of GPUs, doing the exact computation you specified, efficiently. To change the model you're working with, you only need to change a single string in your code.
20
+
21
+ Tinker gives you full control over the training loop and all the algorithmic details. It's not a magic black box that makes fine-tuning "easy". It's a clean abstraction that shields you from the complexity of distributed training while preserving your control.
22
+
23
+ Here's how the division of responsibilities works in practice:
24
+
25
+ | **You focus on** | **You write** | **We handle** |
26
+ |---|---|---|
27
+ | **Datasets and RL environments**<br />Your custom training data | **Simple Python script**<br />Runs on your CPU | **Efficient distributed training of large models**<br />Llama 70B, Qwen 235B |
28
+ | **Training logic**<br />Your loss functions, training loop, and evals | **API calls**<br />`forward_backward()`<br />`optim_step()`<br />`sample()`<br />`save_state()` | **Reliability**<br />Hardware failures handled transparently |
29
+
30
+ ## Features
31
+
32
+ What the Tinker service currently supports:
33
+
34
+ - Tinker lets you fine-tune open-weight models like the Qwen and Llama series, including large mixture-of-experts models like Qwen3-235B-A22B.
35
+ - Tinker supports vision-language models (VLMs) like Qwen3-VL for image understanding tasks. See [Vision Inputs](/rendering#vision-inputs) for details.
36
+ - Tinker implements low-rank adaptation (LoRA) fine-tuning, not full fine-tuning. However, we believe that LoRA gives the same performance as full fine-tuning for many important use cases, especially in RL (see [LoRA Without Regret](https://thinkingmachines.ai/blog/lora/)).
37
+ - You can download the weights of your trained model to use outside of Tinker, for example with your inference provider of choice.
38
+
39
+ ## A quick look at functionality
40
+
41
+ Tinker's main functionality is contained in a few key functions:
42
+
43
+ - `forward_backward`: feed in your data and loss function, and we'll compute and accumulate the gradients for you.
44
+ - `optim_step`: update your model using the accumulated gradients
45
+ - `sample`: Generate outputs from your trained model
46
+ - other functions for saving and loading weights and optimizer state
47
+
48
+ ## What's next?
49
+
50
+ Some features we expect to support in the future:
51
+
52
+ - Full fine-tuning
53
+
54
+
55
+ ---
56
+
57
+ ## File: losses.mdx
58
+
59
+ import { CookbookLink } from '../components/CookbookLink'
60
+
61
+ # Loss functions in Tinker
62
+
63
+ For most use cases, you can use the Tinker API's built-in loss functions by passing in a string identifier to `forward_backward`, which supports cross-entropy and policy gradient objectives. When you need more control, `forward_backward_custom` enables arbitrary differentiable loss functions at the cost of an additional forward pass; we explain both approaches in this doc.
64
+
65
+ When you call `forward_backward`, you specify a loss function using a string that selects from a predetermined set of options, comprising the most common losses used for language model training.
66
+ - **Input:** `forward_backward` expects a certain set of input tensors, passed in via `datum.loss_fn_inputs`, which is a dict mapping `str` to either a numpy or torch tensor
67
+ - **Output:** `forward_backward` returns a `ForwardBackwardOutput`, which has a set of output tensors in `fwd_bwd_result.loss_fn_outputs`
68
+
69
+ For an example of using `forward_backward`, see `rl/train.py` in the Cookbook:
70
+ ```python
71
+ import tinker
72
+ import torch
73
+ from tinker import TensorData
74
+
75
+ # Create training data with required inputs
76
+ datum = tinker.Datum(
77
+ model_input=input_tokens,
78
+ loss_fn_inputs={
79
+ "target_tokens": TensorData.from_torch(torch.tensor(target_tokens)),
80
+ "logprobs": TensorData.from_torch(torch.tensor(sampling_logprobs)), # Reference logprobs
81
+ "advantages": TensorData.from_torch(torch.tensor(advantages)),
82
+ }
83
+ )
84
+
85
+ # Option 1: Use importance sampling REINFORCE
86
+ fwd_bwd_result = await training_client.forward_backward_async(
87
+ [datum], loss_fn="importance_sampling"
88
+ )
89
+
90
+ # Option 2: Use PPO with clipping
91
+ fwd_bwd_result = await training_client.forward_backward_async(
92
+ [datum], loss_fn="ppo"
93
+ )
94
+ ```
95
+
96
+ ## Basic loss functions
97
+
98
+ Currently, the Tinker API supports `cross_entropy` (for supervised learning), `importance_sampling`, `ppo`, `cispo` and `dro` for RL. We denote the training model as $p_{\theta}$, the sampling distribution as $q$, and advantages as $A$. Also, for notation simplicity we omit the query and denote the full model completion sequence of tokens as $x$.
99
+
100
+ All losses are applied at the token level and tensors below have shape `(N,)` where `N` is `model_input.length`. They can be provided as `numpy.ndarray` or `torch.Tensor`, and the return values will use the same tensor type.
101
+
102
+ ### Supervised learning: `cross_entropy`
103
+
104
+ For SL, we implement the standard cross-entropy loss (i.e., negative log-likelihood), which optimizes the policy $p_\theta$ to maximize the log-probability of the tokens $x$:
105
+
106
+ $$
107
+ \mathcal{L(\theta)} = -\mathbb{E}_x[\log p_\theta(x)]
108
+ $$
109
+
110
+ where `weights` is either 0 or 1, typically generated from `renderer.build_supervised_example()` which returns `(model_input, weights)` (i.e., to specify the desired assistant turns to train on).
111
+
112
+ This is implemented as:
113
+
114
+ ```python
115
+ # Apply weights and compute elementwise loss
116
+ elementwise_loss = -target_logprobs * weights
117
+ # Apply sum reduction to get the total loss
118
+ loss = elementwise_loss.sum() # scalar
119
+ ```
120
+
121
+ - **Input tensors:**
122
+ - `target_tokens: array[(N,), int]` - Target token IDs
123
+ - `weights: array[(N,), float]` - Token-level loss weights (typically from the renderer)
124
+ - **Output tensors:**
125
+ - `logprobs: array[(N,), float]` - Log probabilities of predicted tokens
126
+ - **Output diagnostics:**
127
+ - `loss:sum` (scalar) - Sum of weighted cross-entropy losses
128
+
129
+ ### Policy gradient: `importance_sampling`
130
+
131
+ For RL, we implement a common variant of the policy gradient objective, used in practical settings where the *learner policy* $p$ may differ from the *sampling policy* $q$, which is common due to, e.g., [non-determinism](https://thinkingmachines.ai/blog/defeating-nondeterminism-in-llm-inference/). The issue is that if these policies differ, then the objective:
132
+
133
+ $$
134
+ \mathcal{L}(\theta) = \mathbb{E}_{x\sim p_\theta}\bigl[A(x)\bigr]
135
+ $$
136
+
137
+ is not computed in an unbiased why due to $x \sim q$ (sampler) not exactly matching the desired $x \sim p_\theta$ (learner). To correct the bias, we use a modified "importance sampling" objective:
138
+
139
+ $$
140
+ \mathcal{L}_{\text{IS}}(\theta) = \mathbb{E}_{x\sim q}\Bigl[\frac{p_\theta(x)}{q(x)}A(x)\Bigr],
141
+ $$
142
+
143
+ which yields the correct expected reward. In the formula above:
144
+
145
+ - $\log p_\theta(x)$ – `target_logprobs` is from the learner, on the forward part of the `forward_backward` pass.
146
+ - $\log q(x)$ – `sampling_logprobs` is from the sampler, recorded during sampling as a correction term.
147
+
148
+ This is implemented as:
149
+
150
+ ```python
151
+ # Compute probability ratio
152
+ prob_ratio = torch.exp(target_logprobs - sampling_logprobs)
153
+ # Compute importance-weighted loss
154
+ loss = -(prob_ratio * advantages).sum()
155
+ ```
156
+
157
+ - **Input tensors:**
158
+ - `target_tokens: array[(N,), int]` - Target token IDs (from the sampler $q$)
159
+ - `logprobs: array[(N,), float]` - `sampling_logprobs` for the tokens
160
+ - `advantages: array[(N,), float]` - Advantage values for RL (positive to reinforce, negative to discourage)
161
+ - **Output tensors:**
162
+ - `logprobs: array[(N,), float]` - `target_logprobs` for the tokens
163
+ - **Output diagnostics:**
164
+ - `loss:sum` (scalar) - Sum of importance-weighted policy gradient losses $\mathcal L_{\text{IS}}$
165
+
166
+ ### Proximal Policy Optimization: `ppo`
167
+
168
+ PPO ([Schulman et al., 2017](https://arxiv.org/abs/1707.06347)) addresses issues with standard policy gradient methods by introducing a clipping objective that limits policy updates within a close neighborhood of the sampling distribution. This prevents updates that are too large in policy space, especially when taking multiple gradient steps on the same rollout distribution.
169
+
170
+ The objective clips the importance ratio $\frac{p_\theta(x)}{q(x)}$ to prevent large policy updates, where $p_\theta$ is the learner policy and $q$ is the sampling policy. Note that the PPO clipping and loss computation is applied token-wise, computing the loss for each token independently.
171
+
172
+ The PPO clipping objective is:
173
+
174
+ $$
175
+ \mathcal{L}_{\text{CLIP}}(\theta) = -\mathbb{E}_{x \sim q}\left[\text{clip}\left(\frac{p_\theta(x)}{q(x)}, 1-\epsilon_{\text{low}}, 1+\epsilon_{\text{high}}\right) \cdot A(x)\right]
176
+ $$
177
+
178
+ The final PPO loss combines the clipped and unclipped objectives:
179
+
180
+ $$
181
+ \mathcal{L}_{\text{PPO}}(\theta) = -\mathbb{E}_{x \sim q}\left[\min\left(\frac{p_\theta(x)}{q(x)} \cdot A(x), \text{clip}\left(\frac{p_\theta(x)}{q(x)}, 1-\epsilon_{\text{low}}, 1+\epsilon_{\text{high}}\right) \cdot A(x)\right)\right]
182
+ $$
183
+
184
+ where $\epsilon_{\text{low}}$ and $\epsilon_{\text{high}}$ are hyperparameters (currently fixed to 0.2 in Tinker).
185
+
186
+ This is implemented as:
187
+
188
+ ```python
189
+ # Compute probability ratio
190
+ prob_ratio = torch.exp(target_logprobs - sampling_logprobs)
191
+ # Apply clipping
192
+ clipped_ratio = torch.clamp(prob_ratio, clip_low_threshold, clip_high_threshold)
193
+ # Compute both objectives
194
+ unclipped_objective = prob_ratio * advantages
195
+ clipped_objective = clipped_ratio * advantages
196
+ # Take minimum (most conservative)
197
+ ppo_objective = torch.min(unclipped_objective, clipped_objective)
198
+ # PPO loss is negative of objective
199
+ loss = -ppo_objective.sum()
200
+ ```
201
+
202
+
203
+ **Example with custom clipping thresholds:**
204
+ ```python
205
+ fwd_bwd_result = await training_client.forward_backward_async(
206
+ data=data,
207
+ loss_fn="ppo",
208
+ loss_fn_config={"clip_low_threshold": 0.9, "clip_high_threshold": 1.1}
209
+ )
210
+ ```
211
+
212
+ **Additional Notes:**
213
+ - The loss formulation above is quite general, since the user can organize the data generation and advantage estimation in their own code. For example, the main RL training scripts in the Tinker Cookbook use group-based rollouts with per-group advantage centering similar to GRPO ([Shao et al., 2024](https://arxiv.org/abs/2402.03300)).
214
+ - The functional implementations of REINFORCE and PPO do not use an additional KL term like the original GRPO work, which has been noted to be mathematically inconsistent ([Zhang et al., 2025](https://arxiv.org/abs/2505.17508); [Tang et al., 2025](https://arxiv.org/abs/2506.09477)). However, it is possible to include a KL regularization term as part of the reward, which is mathematically correct and we provide this option in our RL training <CookbookLink path="tinker_cookbook/rl/train.py">code and examples</CookbookLink> (consider the incorporate_kl_penalty function).
215
+ - Notice that for all objectives we sum the token-level losses over the sequence length unlike some other loss implementations. If you would like to explore different aggregation schemes, you can include that in the advantage tensor computation.
216
+
217
+ ### Clipped Importance Sampling Policy Optimization: `cispo`
218
+
219
+ CISPO ([Chen et al., 2024](https://arxiv.org/abs/2506.13585); [Khatri et al., 2024](https://arxiv.org/abs/2510.13786)) is a policy gradient method that uses a clipped importance ratio as a coefficient for the policy gradient. Unlike PPO which clips the objective directly, CISPO clips the ratio and uses it to weight the log probability. Mathematically the objective is:
220
+ The CISPO objective is:
221
+
222
+ $$
223
+ \mathcal{L}_{\text{CISPO}}(\theta) = \mathbb{E}_{x \sim q}\left[\textbf{sg}\left( \text{clip}\left(\frac{p_\theta(x)}{q(x)}, 1-\epsilon_{\text{low}}, 1+\epsilon_{\text{high}}\right) \right) \cdot \log p_\theta(x) \cdot A(x)\right]
224
+ $$
225
+
226
+ This is implemented as:
227
+
228
+ ```python
229
+ # Compute probability ratio
230
+ prob_ratio = torch.exp(target_logprobs - sampling_logprobs)
231
+ # Apply clipping
232
+ clipped_ratio = torch.clamp(prob_ratio, clip_low_threshold, clip_high_threshold)
233
+ # Compute CISPO objective (detach the clipped ratio)
234
+ cispo_objective = clipped_ratio.detach() * target_logprobs * advantages
235
+ # CISPO loss is negative of objective
236
+ loss = -cispo_objective.sum()
237
+ ```
238
+
239
+
240
+ Similarly to the PPO objective you can pass loss function parameters in the following way:
241
+
242
+ ```python
243
+ fwd_bwd_result = await training_client.forward_backward_async(
244
+ data=data,
245
+ loss_fn="cispo",
246
+ loss_fn_config={"clip_low_threshold": 0.8, "clip_high_threshold": 1.2}
247
+ )
248
+ ```
249
+
250
+ ### Direct Reward Optimization: `dro`
251
+
252
+ DRO ([Richemond et al., 2024](https://arxiv.org/abs/2405.19107); [Kimi Team et al., 2025](https://arxiv.org/abs/2501.12599)) is a general off-policy (and even offline) reinforcement learning method that uses a quadratic penalty term to constrain the policy update. Notice that this loss uses a dfferent (soft) formulation of the advantage estimation, which needs to be implemented on the client side.
253
+ The DRO objective is:
254
+
255
+ $$
256
+ \mathcal{L}_{\text{DRO}}(\theta) = \mathbb{E}_{x \sim q}\left[\log p_\theta(x) \cdot A(x) - \frac{1}{2}\beta \left(\log \frac{p_\theta(x)}{q(x)}\right)^2\right]
257
+ $$
258
+
259
+
260
+ This is implemented as:
261
+
262
+ ```python
263
+ # Compute quadratic penalty term
264
+ quadratic_term = (target_logprobs - sampling_logprobs) ** 2
265
+ # Compute DRO objective
266
+ dro_objective = target_logprobs * advantages - 0.5 * beta * quadratic_term
267
+ # DRO loss is negative of objective
268
+ loss = -dro_objective.sum()
269
+ ```
270
+
271
+ And similarly to other objectives, can specify the loss hyper-parameter as:
272
+
273
+ ```python
274
+ fwd_bwd_result = await training_client.forward_backward_async(
275
+ data=data,
276
+ loss_fn="dro",
277
+ loss_fn_config={"beta": 0.05}
278
+ )
279
+ ```
280
+
281
+ ## Flexible loss functions: `forward_backward_custom`
282
+
283
+ For use cases outside of the above, we've provided the more flexible (but slower) methods `forward_backward_custom` and `forward_backward_custom_async` to compute a more general class of loss functions.
284
+
285
+ ### Usage
286
+
287
+ Here's a simple example of a custom loss function:
288
+
289
+ ```python
290
+ def logprob_squared_loss(data: list[Datum], logprobs: list[torch.Tensor]) -> tuple[torch.Tensor, dict[str, float]]:
291
+ loss = (logprobs ** 2).sum()
292
+ return loss, {"logprob_squared_loss": loss.item()}
293
+ ```
294
+
295
+ You can call this loss function with `forward_backward_custom` like:
296
+
297
+ ```python
298
+ loss, metrics = training_client.forward_backward_custom(data, logprob_squared_loss)
299
+ ```
300
+
301
+ You can also define loss functions which operate on multiple sequences at a time. For example, a loss function that computes the variance across the sequences (although practically useless) can be implemented as:
302
+
303
+ ```python
304
+ def variance_loss(data: list[Datum], logprobs: list[torch.Tensor]) -> tuple[torch.Tensor, dict[str, float]]:
305
+ flat_logprobs = torch.cat(logprobs)
306
+ variance = torch.var(flat_logprobs)
307
+ return variance, {"variance_loss": variance.item()}
308
+ ```
309
+
310
+ A more practical use case would be to compute a Bradley-Terry loss on pairwise comparison data -- a classic approach in RL from human feedback, as introduced and popularized by [Learning to Summarize](https://arxiv.org/abs/2009.01325). Similarly, we can also implement [Direct Preference Optimization](https://arxiv.org/abs/2305.18290), which also computes a loss involving pairs of sequences; see the [DPO guide](/preferences/dpo-guide) for more details.
311
+
312
+ If you're using a custom loss function that you think is generally useful, please let us know, and we'll add it to the list of built-in loss functions.
313
+
314
+ We detail the `async` version of methods in the [Async and Futures](./async) of these docs.
315
+
316
+ ### How `forward_backward_custom` works
317
+
318
+ ---
319
+
320
+ ## File: publish-weights.mdx
321
+
322
+ # Publishing weights
323
+
324
+ If you've trained a model that you'd like to share with the community, you can
325
+ publish any number of checkpoints you've previously saved.
326
+
327
+ Once published, your checkpoint can be loaded by any tinker user and used to
328
+ further train a new model or be sampled against.
329
+
330
+ ### Publishing
331
+
332
+ ```bash
333
+ tinker checkpoint publish $TINKER_CHECKPOINT_PATH
334
+ ```
335
+
336
+ where `$TINKER_CHECKPOINT_PATH` is a checkpoint path in the form of `tinker://14bdf3a1-0b95-55c7-8659-5edb1bc870af:train:17/weights/checkpoint_id_to_publish`.
337
+
338
+ You may confirm your checkpoint is published by dumping the checkpoint info and checking the `Public` property:
339
+
340
+ ```bash
341
+ tinker checkpoint info tinker://14bdf3a1-0b95-55c7-8659-5edb1bc870af/weights/checkpoint_id_to_publish
342
+ Checkpoint: weights/checkpoint_id_to_publish
343
+ ┏━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┓
344
+ ┃ Property ┃ Value ┃
345
+ ┡━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┩
346
+ │ Checkpoint ID │ weights/checkpoint_id_to_publish │
347
+ │ Type │ training │
348
+ │ Tinker Path │ tinker://14bdf3a1-0b95-55c7-8659-5edb1bc870af/weights/checkpoint_id_to_publish │
349
+ │ Size │ 342.4 MB │
350
+ │ Public │ No │
351
+ │ Created │ 23 minutes ago │
352
+ │ Training Run ID │ 14bdf3a1-0b95-55c7-8659-5edb1bc870af │
353
+ └─────────────────┴────────────────────────────────────────────────────────────────────────────────┘
354
+ ```
355
+
356
+ ### Unpublishing
357
+
358
+ ```bash
359
+ tinker checkpoint unpublish $TINKER_CHECKPOINT_PATH`
360
+ ```
361
+
362
+ ### Loading public weights
363
+
364
+ Loading public weights is exactly the same as loading a non-public one:
365
+
366
+ ```python
367
+ ckpt_path = ...
368
+ training_client = service_client.create_training_client_from_state(ckpt_path)
369
+ ```
370
+
371
+
372
+ ---
373
+
374
+ ## File: supervised-learning.mdx
375
+
376
+ import { CookbookLink } from '../components/CookbookLink'
377
+
378
+ # Cookbook: Supervised learning
379
+
380
+ This section takes you through examples from the Tinker Cookbook that relate to supervised learning.
381
+
382
+ In general, supervised learning (SL) means learning an input-output mapping from labeled data. In the context of language model fine-tuning, this means **minimizing a weighted cross-entropy loss** on token sequences---equivalently, maximizing the log-probability of the specified target tokens.
383
+
384
+ There are a few ways that SL is commonly used in LLM fine-tuning pipelines:
385
+
386
+ - *Instruction tuning*: This is the first step in post-training pipelines, applied to the base (raw, pretrained) model. Typically, we do SL on a high-quality dataset that demonstrates the correct format and style, while boosting the model's reasoning and instruction-following.
387
+ - *Context distillation* / *prompt distillation*: let's say we have a generic model that can do chat / instruction following / reasoning, but we want to adjust how it behaves in a certain scenario. We can add some instructions to the system message of our model. However, the system message might grow impractically long and start ignoring some of its instructions. So it's often better to create a supervised dataset on a narrow prompt distribution, with a shorter set of instructions that that are targeted at these prompts.
388
+
389
+ We'll cover both of these use cases in this documentation and related Cookbook code.
390
+
391
+ The library code implementing supervised learning can be found in the <CookbookLink path="tinker_cookbook/supervised">`supervised`</CookbookLink> directory.
392
+
393
+
394
+ ---
395
+
396
+ ## File: preferences.mdx
397
+
398
+ import { CookbookLink } from '../components/CookbookLink'
399
+
400
+ # Preferences
401
+
402
+ # Learning from Preferences
403
+
404
+ In this section, we focus on learning from **pairwise feedback**, where we have preference data indicating which of two completions is better for a given prompt. This kind of feedback is a natural fit for tasks where there's not a simple correctness criterion that can be computed programmatically. These preferences might be collected from human evaluators or generated bya model.
405
+
406
+ ## Two Approaches to Preference Learning
407
+
408
+ When you have pairwise preference data, there are two main approaches:
409
+
410
+ 1. **Direct Preference Optimization (DPO)**: Directly update the policy to prefer chosen responses over rejected ones, without needing a separate reward model. This is simpler and computationally cheaper. See the [DPO Guide](/preferences/dpo-guide) for details.
411
+
412
+ 2. **Reinforcement Learning from Human Feedback (RLHF)**: Train a reward model on preference data, then use reinforcement learning to optimize the policy against this reward model. This two-stage approach provides more flexibility. See the the [RLHF example](/preferences/rlhf-example) for details.
413
+
414
+
415
+ ---
416
+
417
+ ## File: docs-outline.mdx
418
+
419
+ # Navigating these docs
420
+
421
+ These docs provide guides to both Tinker and the Tinker Cookbook.
422
+
423
+ The first half, "Using the Tinker API", walks you through the fundamentals of Tinker:
424
+
425
+ - [Installation](./install) explains how to install both `tinker` and `tinker-cookbook`, and points you to the Tinker Console for your API key.
426
+ - [Training and Sampling](./training-sampling) takes you through your first training run: setting up your training data, performing the run, and sampling from the model to test the run.
427
+ - [Loss Functions](./losses) starts to get into the detail. Tinker supports a variety of built-in loss function, but also allows you to use arbitrary differentiable loss functions.
428
+ - [Saving and Loading](./save-load) explains the checkpoint types available in Tinker, and how to restart a run from a checkpoint.
429
+ - [Async and Futures](./async) explains Tinker's `sync` and `async` API variants, and how Futures works as Tinker's requests structure.
430
+ - [Model Lineup](./model-lineup) is regularly updated with the models available to fine-tune in Tinker.
431
+
432
+ The second half, "The Tinker Cookbook", provides recipes for how to use the Tinker API for research and applications. You are welcome to adapt these directly for your own use cases.
433
+
434
+ - [Rendering](./rendering) explains how we convert from a conversation data structure to a list of tokens.
435
+ - [Supervised Learning](./supervised-learning) explains basic SL and walks you through your first SL training loop. We make some suggestions for hyperparameter selection and detail how you can run your own hyperparameter sweep. We also show you how to perform prompt distillation.
436
+ - [Reinforcement Learning](./rl) explains the basics of RL and walks you through your first RL run. We explain and provide code for creating your own RL environments and training on them. We provide a simple training loop for you to use and adapt, and explain RL hyperparameters and loss functions in detail.
437
+ - [Preferences](./preferences) is a guide to learning from pairwise feedback, where we have preference data indicating which of two completions is better for a given prompt. We walk you through two approaches to learning from pairwise preference data: direct preference optimization (DPO) and reinforcement learning from human feedback (RLHF).
438
+ - [Evaluations](./evals) explains how you can use Tinker's outputs to run inline and offline evals on your runs.
439
+ - [Completers](./completers) explains how Tinker implements policies, and provides two examples of how to use these in training.
440
+ - [LoRA Primer](./lora-primer) explains the basic background of LoRA, and how to choose hyperparameters.
441
+
442
+
443
+ ---
444
+
445
+ ## File: lora-primer.mdx
446
+
447
+ # LoRA Primer
448
+
449
+ Tinker supports [LoRA fine-tuning](https://arxiv.org/abs/2106.09685), which adjusts a small number of parameters, rather than full fine-tuning, which adjusts all of the parameters of the original model.
450
+
451
+ Our current understanding is that LoRA has equivalent performance to full fine-tuning when doing RL or doing SL on small datasets, while it has worse performance on larger datasets. In more detail:
452
+
453
+ - For supervised fine-tuning on small-to-medium-sized instruction-tuning and reasoning datasets, LoRA performs the same as full fine-tuning.
454
+ - For datasets that exceed LoRA capacity, LoRA underperforms FullFT. Rather than the loss reaching a distinct floor that it can’t go below, LoRA results in worse training efficiency that depends on the relationship between model capacity to dataset size.
455
+ - In some scenarios, LoRA is less tolerant of large batch sizes than full fine-tuning — it pays a larger penalty in loss as batch size increases beyond some point. This penalty is not mitigated by increasing the LoRA rank; it is a property of the product-of-matrices parametrization, which has different training dynamics than optimizing the original weight matrix.
456
+ - Even in small data settings, LoRA performs better when applied to all weight matrices, especially MLP and MoE layers. Attention-only LoRA underperforms even when we match the number of trainable parameters by using higher rank for attention-only LoRA.
457
+ - LoRA performs equivalently to FullFT for reinforcement learning even with small ranks. We find that RL requires very low capacity, a result we anticipated based on information-theoretical arguments.
458
+
459
+ See [LoRA Without Regret](https://thinkingmachines.ai/blog/lora) for more details and experimental results.
460
+
461
+ ## Hyperparameters
462
+
463
+ The learning rate (LR) is usually the most important hyperparameter in your ML experiments.
464
+
465
+
466
+ LoRA requires a much larger LR than full fine-tuning---typically 20-100x larger, depending on model size. People often mistakenly retain their full fine-tuning LR when they port their code to use LoRA, leading them to conclude that LoRA works poorly.
467
+
468
+ **Calculate the correct LoRA learning rate:**
469
+
470
+ We've provided a utility that calculates the factor you should scale the full fine-tuning LR by to get the equivalent LoRA LR:
471
+
472
+ ```python
473
+ from tinker_cookbook.hyperparam_utils import get_lora_lr_over_full_finetune_lr
474
+
475
+ model_name = "meta-llama/Llama-3.1-8B"
476
+ print(get_lora_lr_over_full_finetune_lr(model_name))
477
+ ```
478
+
479
+ Note that for `Llama-3.2-1B`, the factor is 32, while for `Llama-3.1-70B`, the factor is 128.
480
+
481
+ ## What is LoRA exactly?
482
+
483
+ LoRA is short for Low-Rank Adaptation. Given that the original model has a weight matrix $W$, we replace it with a new weight matrix $W'=W + BA$, where $B$ and $A$ are low-rank matrices. If $W$ is an $n \times n$ matrix, then $B$ and $A$ are $n \times r$ and $r \times n$ matrices, respectively, where $r$ is the rank of the low-rank approximation. The default $r$ used by tinker is $32$.
484
+
485
+ The fact that LoRA uses a low-rank approximation of weight matrices is not terribly important. We prefer to think of LoRA as just a random projection of the parameter space that happens to be efficient to implement. When training with RL or small SL datasets, we are only learning a small amount of information, and this reduced set of parameters is more than enough.
486
+
487
+
488
+ ## What rank to use?
489
+
490
+ The default rank used by tinker is $32$. However, if you're doing SL on a large dataset, you should use a larger rank. For supervised learning, as a very rough approximation, LoRA will give good results as long as the number of LoRA parameters is at least as large as the number of completion tokens (i.e., weight=1 tokens). You can calculate the number of LoRA parameters with the following utility:
491
+
492
+ ```python
493
+ from tinker_cookbook.hyperparam_utils import get_lora_param_count
494
+
495
+ model_name = "meta-llama/Llama-3.1-8B"
496
+ print(get_lora_param_count(model_name, lora_rank=32))
497
+ ```
498
+
499
+ For reinforcement learning, we've found that small ranks give equivalent performance to larger ranks and full fine-tuning.
500
+
501
+ Note that conveniently, the optimal learning rate does *not* depend on the LoRA rank. In fact, you can verify that if you train with SL on different ranks (but with the same LR), you'll get exactly the same learning curves for the first few steps of training.
502
+
503
+
504
+ ---
505
+
506
+ ## File: evals.mdx
507
+
508
+ import { Callout } from 'nextra/components'
509
+ import { CookbookLink } from '../components/CookbookLink'
510
+
511
+ # Evaluations
512
+
513
+ Our training scripts will print out training and test loss. Two common workflows for evaluations are to do inline evals during training and to do offline evals on various checkpoints from a run.
514
+
515
+ ## Inline Evals
516
+
517
+ You can add inline evaluations to your training runs by configuring evaluator builders in advance for both supervised fine-tuning and RL training jobs.
518
+
519
+ ### Supervised Fine-Tuning (`supervised.train`)
520
+ Add one or both of the following to your config:
521
+
522
+ - **`evaluator_builders: list[EvaluatorBuilder]`** - Runs evaluations every `eval_every` steps
523
+ - **`infrequent_evaluator_builders: list[EvaluatorBuilder]`** - Runs evaluations every `infrequent_eval_every` steps
524
+
525
+ ### RL Training (`rl.train`)
526
+
527
+ Add the following to your config:
528
+
529
+ - **`evaluator_builders: list[SamplingClientEvaluator]`** - Runs evaluations every `eval_every` steps
530
+
531
+ For implementation guidance and a detailed example, see <CookbookLink path="tinker_cookbook/eval/evaluators.py">here</CookbookLink> and
532
+ <CookbookLink path="tinker_cookbook/eval/inspect_evaluators.py">here</CookbookLink> respectively.
533
+
534
+
535
+ ## Offline evals
536
+
537
+ We support and recommend several ways for creating and running your offline evaluations on your model checkpoints.
538
+
539
+ ### Running Standard Evaluations with Inspect AI.
540
+
541
+ We support running many of the standard cited evaluations using the [Inspect AI library](https://github.com/UKGovernmentBEIS/inspect_ai).
542
+
543
+ We have provided a <CookbookLink path="tinker_cookbook/eval/run_inspect_evals.py">script</CookbookLink> to evaluate models using Tinker's internal sampling functionality as shown below.
544
+
545
+ ```bash
546
+ MODEL_PATH=tinker://FIXME # YOUR MODEL PATH HERE
547
+ python -m tinker_cookbook.eval.run_inspect_evals \
548
+ model_path=$MODEL_PATH \
549
+ model_name=MODEL_NAME \ # YOUR MODEL_NAME HERE
550
+ tasks=inspect_evals/ifeval,inspect_evals/mmlu_0_shot \
551
+ renderer_name=RENDERER_NAME # YOUR RENDERER_NAME HERE
552
+ ```
553
+
554
+ Click [here](https://github.com/UKGovernmentBEIS/inspect_ai/blob/main/docs/evals/listing.yml) to view additional supported evaluations.
555
+
556
+ ### Creating your own Sampling Evaluations
557
+
558
+ We recommend two ways to create your own evaluations:
559
+ - creating your own tasks with Inspect AI and running like above
560
+ - creating your own SamplingClientEvaluator
561
+
562
+ #### Create tasks with Inspect AI
563
+
564
+ In addition to passing in standard evaluations, you can create your own tasks using inspect ai as detailed [here](https://inspect.aisi.org.uk/tasks.html).
565
+
566
+ Here is a toy example of how to create an evaluation with an LLM-as-a-judge where we use a model produced by tinker as a grader.
567
+
568
+ ```python
569
+ import tinker
570
+ from inspect_ai import Task, task
571
+ from inspect_ai.dataset import MemoryDataset, Sample
572
+ from inspect_ai.model import GenerateConfig as InspectAIGenerateConfig
573
+ from inspect_ai.model import Model as InspectAIModel
574
+ from inspect_ai.scorer import model_graded_qa
575
+ from inspect_ai.solver import generate
576
+ from tinker_cookbook.eval.inspect_utils import InspectAPIFromTinkerSampling
577
+
578
+ QA_DATASET = MemoryDataset(
579
+ name="qa_dataset",
580
+ samples=[
581
+ Sample(
582
+ input="What is the capital of France?",
583
+ target="Paris",
584
+ ),
585
+ Sample(
586
+ input="What is the capital of Italy?",
587
+ target="Rome",
588
+ ),
589
+ ],
590
+ )
591
+
592
+ service_client = tinker.ServiceClient()
593
+ sampling_client = service_client.create_sampling_client(
594
+ base_model="meta-llama/Llama-3.1-8B-Instruct"
595
+ )
596
+
597
+ api = InspectAPIFromTinkerSampling(
598
+ renderer_name="llama3",
599
+ model_name="meta-llama/Llama-3.1-8B-Instruct",
600
+ sampling_client=sampling_client,
601
+ verbose=False,
602
+ )
603
+
604
+ GRADER_MODEL = InspectAIModel(api=api, config=InspectAIGenerateConfig())
605
+
606
+
607
+ @task
608
+ def example_lm_as_judge() -> Task:
609
+ """
610
+ Example task using LLM-as-a-judge scoring.
611
+
612
+ Note: The grader model defaults to the model being evaluated.
613
+ To use a different grader model, specify it with --model-grader when using inspect directly.
614
+ """
615
+ return Task(
616
+ name="llm_as_judge",
617
+ dataset=QA_DATASET,
618
+ solver=generate(),
619
+ scorer=model_graded_qa(
620
+ instructions="Grade strictly against the target text as general answer key and rubric. "
621
+ "Respond 'GRADE: C' if correct or 'GRADE: I' otherwise.",
622
+ partial_credit=False,
623
+ # model parameter is optional - if not specified, uses the model being evaluated
624
+ model=GRADER_MODEL,
625
+ ),
626
+ )
627
+ ```
628
+
629
+ Inspect also natively supports replacing our `GRADER_MODEL` with any openai-chat-completion style api (e.g. openrouter).
630
+
631
+ #### Create your own SamplingClientEvaluator
632
+
633
+ Alternatively, you can create your own SamplingClientEvaluator class instead of using Inspect AI. This is a lower
634
+ level abstraction than the above with finer-grain control over running your evaluations.
635
+
636
+ We expose this to interace to allow users more control over their datasets and metrics. To illustrate, see this
637
+ <CookbookLink path="tinker_cookbook/eval/custom_evaluators.py">custom evaluators</CookbookLink> example of how one might create their own complex SamplingClientEvaluator.
638
+
639
+ For a more illustrative toy instructive example see below.
640
+
641
+ ```python
642
+ from typing import Any, Callable
643
+
644
+ import tinker
645
+ from tinker import types
646
+
647
+ from tinker_cookbook import renderers
648
+ from tinker_cookbook.evaluators import SamplingClientEvaluator
649
+ from tinker_cookbook.tokenizer_utils import get_tokenizer
650
+
651
+ class CustomEvaluator(SamplingClientEvaluator):
652
+ """
653
+ A toy SamplingClientEvaluator that runs a custom evaluation and returns its metrics.
654
+ """
655
+
656
+ def __init__(
657
+ self,
658
+ dataset: Any,
659
+ grader_fn: Callable[[str, str], bool],
660
+ model_name: str,
661
+ renderer_name: str,
662
+ ):
663
+ """
664
+ Initialize the CustomEvaluator.
665
+ Args:
666
+ config: Configuration object containing all evaluation parameters
667
+ """
668
+ self.dataset = dataset
669
+ self.grader_fn = grader_fn
670
+
671
+ tokenizer = get_tokenizer(model_name)
672
+ self.renderer = renderers.get_renderer(name=renderer_name, tokenizer=tokenizer)
673
+
674
+ async def __call__(self, sampling_client: tinker.SamplingClient) -> dict[str, float]:
675
+ """
676
+ Run custom evaluation on the given sampling client and return metrics.
677
+ Args:
678
+ sampling_client: The sampling client to evaluate
679
+ Returns:
680
+ Dictionary of metrics from inspect evaluation
681
+ """
682
+
683
+ metrics = {}
684
+
685
+ num_examples = len(self.dataset)
686
+ num_correct = 0
687
+
688
+ sampling_params = types.SamplingParams(
689
+ max_tokens=100,
690
+ temperature=0.7,
691
+ top_p=1.0,
692
+ stop=self.renderer.get_stop_sequences(),
693
+ )
694
+
695
+ for datum in self.dataset:
696
+ model_input: types.ModelInput = self.renderer.build_generation_prompt(
697
+ [renderers.Message(role="user", content=datum["input"])]
698
+ )
699
+ # Generate response
700
+ r: types.SampleResponse = await sampling_client.sample_async(
701
+ prompt=model_input, num_samples=1, sampling_params=sampling_params
702
+ )
703
+ tokens: list[int] = r.sequences[0].tokens
704
+ response: renderers.Message = self.renderer.parse_response(tokens)[0]
705
+ if self.grader_fn(response["content"], datum["output"]):
706
+ num_correct += 1
707
+
708
+ metrics["accuracy"] = num_correct / num_examples
709
+ return metrics
710
+ ```
711
+
712
+ Here is an example of how we can use the above CustomEvaluator on a toy dataset and grader.
713
+
714
+
715
+ ```python
716
+ QA_DATASET = [
717
+ {"input": "What is the capital of France?", "output": "Paris"},
718
+ {"input": "What is the capital of Germany?", "output": "Berlin"},
719
+ {"input": "What is the capital of Italy?", "output": "Rome"},
720
+ ]
721
+
722
+ def grader_fn(response: str, target: str) -> bool:
723
+ return target.lower() in response.lower()
724
+
725
+ evaluator = CustomEvaluator(
726
+ dataset=QA_DATASET,
727
+ grader_fn=grader_fn,
728
+ renderer_name="llama3",
729
+ model_name="meta-llama/Llama-3.1-8B-Instruct",
730
+
731
+ )
732
+
733
+ service_client = tinker.ServiceClient()
734
+ sampling_client = service_client.create_sampling_client(base_model="meta-llama/Llama-3.1-8B-Instruct")
735
+
736
+ async def main():
737
+ result = await evaluator(sampling_client)
738
+ print(result)
739
+
740
+ asyncio.run(main())
741
+ ```
742
+
743
+
744
+ ---
745
+
746
+ ## File: dev-tips.mdx
747
+
748
+ # Developer Tips
749
+
750
+ ## AI-assisted development
751
+
752
+ We've provided a single-file version of the documentation that can be fed to LLMs for development: see [llms.txt](/llms.txt) and [llms-full.txt](/llms-full.txt).
753
+
754
+
755
+ ---
756
+
757
+ ## File: async.mdx
758
+
759
+ # Async and Futures
760
+
761
+ ## Sync and Async APIs
762
+
763
+ Every method in the Tinker Python library has both a synchronous (sync) and an asynchronous (async) version. The async variants end with `_async`:
764
+
765
+ | **Client** | **Sync method** | **Async method** |
766
+ |---|---|---|
767
+ | `ServiceClient` | `create_lora_training_client()` | `create_lora_training_client_async()` |
768
+ | `TrainingClient` | `forward()` | `forward_async()` |
769
+ | `SamplingClient` | `sample()` | `sample_async()` |
770
+ | `RestClient` | `list_training_run_ids()` | `list_training_run_ids_async()` |
771
+
772
+ Tinker's `async` functionality requires an `asyncio` event loop, which you typically run like `asyncio.run(main())`.
773
+
774
+ **When to use each:**
775
+
776
+ - **Async:** Best for high-performance workflows where you need concurrency, especially when waiting on multiple network calls.
777
+ - **Sync:** Simpler for scripts and learning examples. Easier to reason about but blocks on each operation.
778
+
779
+ The Tinker Cookbook generally uses `async` for implementations where performance is critical and sync for pedagogical examples.
780
+
781
+ ## Understanding Futures
782
+
783
+ Most Tinker API methods are **non-blocking**, but may take a little while to run. They return immediately with a `Future` object that acknowledges that your request has been submitted. To get the actual result, you must explicitly wait:
784
+
785
+ **Sync Python:**
786
+ ```python
787
+ future = client.forward_backward(data, loss_fn)
788
+ result = future.result() # Blocks until complete
789
+ ```
790
+
791
+ **Async Python (note the double await):**
792
+ ```python
793
+ future = await client.forward_backward_async(data, loss_fn)
794
+ result = await future
795
+ ```
796
+
797
+ After the first `await`, you're guaranteed that the request has been submitted, which ensures that it'll be ordered correctly relative to other requests. The second `await` waits for the actual computation to finish and returns the numerical outputs. For operations like `forward_backward`, the second `await` also guarantees that operation has been applied to the model---for `forward_backward`, this means that the gradients have been accumulated in the model's optimizer state.
798
+
799
+ ## Performance tips: overlap requests
800
+
801
+ For best performance, you should aim to submit your next request while the current one is running. Doing so is more important with Tinker than with other training systems because Tinker training runs on discrete [clock cycles](./under-the-hood#clock-cycles) (~10 seconds each). If you don't have a request queued when a cycle starts, you'll miss that cycle entirely.
802
+
803
+ **Example pattern for overlapping forward_backward and optim_step:**
804
+ ```python
805
+ # Submit forward_backward
806
+ fwd_bwd_future = await client.forward_backward_async(batch, loss_fn)
807
+
808
+ # Submit optim_step immediately (don't wait for forward_backward to finish)
809
+ optim_future = await client.optim_step_async(adam_params)
810
+
811
+ # Now retrieve results
812
+ fwd_bwd_result = await fwd_bwd_future
813
+ optim_result = await optim_future
814
+ ```
815
+
816
+ This pattern ensures both operations are queued and can be processed in the same [clock cycle](./under-the-hood#clock-cycles). In contrast, if you waited for `forward_backward` to complete before submitting `optim_step`, you might miss the next [clock cycle](./under-the-hood#clock-cycles).
817
+
818
+
819
+ ---
820
+
821
+ ## File: download-weights.mdx
822
+
823
+ # Downloading weights
824
+
825
+ ### CLI
826
+
827
+ ```bash
828
+ tinker checkpoint download $TINKER_CHECKPOINT_PATH
829
+ ```
830
+
831
+ See `tinker checkpoint download --help` for more details.
832
+
833
+ ### SDK
834
+
835
+ You can also download checkpoints using the SDK.
836
+
837
+ Example:
838
+
839
+ ```python
840
+ import tinker
841
+ import urllib.request
842
+
843
+ sc = tinker.ServiceClient()
844
+ rc = sc.create_rest_client()
845
+ future = rc.get_checkpoint_archive_url_from_tinker_path("tinker://<unique_id>/sampler_weights/final")
846
+ checkpoint_archive_url_response = future.result()
847
+
848
+ # `checkpoint_archive_url_response.url` is a signed URL that can be downloaded
849
+ # until checkpoint_archive_url_response.expires
850
+ urllib.request.urlretrieve(checkpoint_archive_url_response.url, "archive.tar")
851
+ ```
852
+
853
+ Replace `<unique_id>` with your Training Run ID. This will save the LoRA adapter weights and config inside the `archive.tar` file.
854
+
855
+
856
+ ---
857
+
858
+ ## File: overview-building.mdx
859
+
860
+ # Overview: Tinker Cookbook
861
+
862
+ The next sections provide a variety of guides for how to use the Tinker API for research and applications.
863
+
864
+ We expect people to use Tinker in a few different ways:
865
+
866
+ 1. You want to define datasets and environments and plug them into existing training code from the Tinker Cookbook.
867
+ 2. You want to write your own training loops from scratch, starting with the basics.
868
+ 3. You want to understand the classes and other concepts in Tinker Cookbook so you can extend them to add new functionality.
869
+
870
+ Different parts of the docs will be tailored to these different approaches.
871
+
872
+ We'll start with a couple of general pages that'll be relevant to almost all of the use cases:
873
+
874
+ - [Rendering to Tokens](./rendering.mdx) -- how we convert from a conversation data structure to a list of tokens (a.k.a. chat templates).
875
+ - [LoRA Primer](./lora-primer.mdx) -- basic background of LoRA, and how to choose hyperparameters. For most fine-tuning applications, LoRA will give results that are roughly the same as full fine-tuning, however, you need to use different learning rates.
876
+
877
+
878
+ ---
879
+
880
+ ## File: save-load.mdx
881
+
882
+ # Saving and loading weights and optimizer state
883
+
884
+ During training, you'll need to save checkpoints for two main purposes: *sampling* (to test your model) and *resuming training* (to continue from where you left off). The `TrainingClient` provides three methods to handle these cases:
885
+
886
+ 1. `save_weights_for_sampler()`: saves a copy of the model weights that can be used for sampling.
887
+ 2. `save_state()`: saves the weights and the optimizer state. You can fully resume training from this checkpoint.
888
+ 3. `load_state()`: load the weights and the optimizer state. You can fully resume training from this checkpoint.
889
+
890
+ Note that (1) is faster and requires less storage space than (2).
891
+
892
+ Both `save_*` functions require a `name` parameter---a string that you can set to identify the checkpoint within the current training run. For example, you can name your checkpoints `"0000"`, `"0001"`, `"step_1000"`, etc.
893
+
894
+ The return value contains a `path` field, which is a fully-qualified path, which will look something like `tinker://<model_id>/<name>`. This path is persistent and can be loaded later by a new `ServiceClient` or `TrainingClient`.
895
+
896
+ ### Example: Saving for sampling
897
+
898
+ ```python
899
+ # Setup
900
+ import tinker
901
+ service_client = tinker.ServiceClient()
902
+ training_client = service_client.create_lora_training_client(
903
+ base_model="meta-llama/Llama-3.2-1B", rank=32
904
+ )
905
+
906
+ # Save a checkpoint that you can use for sampling
907
+ sampling_path = training_client.save_weights_for_sampler(name="0000").result().path
908
+
909
+ # Create a sampling client with that checkpoint
910
+ sampling_client = service_client.create_sampling_client(model_path=sampling_path) #
911
+ ```
912
+
913
+ **Shortcut:** Combine these steps with:
914
+
915
+ ```python
916
+ sampling_client = training_client.save_weights_and_get_sampling_client(name="0000")
917
+ ```
918
+
919
+ ### Example: Saving to resume training
920
+
921
+ Use `save_state()` and `load_state()` when you need to pause and continue training with full optimizer state preserved:
922
+
923
+ ```python
924
+ # Save a checkpoint that you can resume from
925
+ resume_path = training_client.save_state(name="0010").result().path
926
+
927
+ # Load that checkpoint
928
+ training_client.load_state(resume_path)
929
+ ```
930
+
931
+ ### When to use `save_state()` and `load_state()`:
932
+
933
+
934
+ - Multi-step training pipelines (e.g. supervised learning followed by reinforcement learning)
935
+ - Adjusting hyperparameters or data mid-run
936
+ - Recovery from interruptions or failures
937
+ - Any scenario where you need to preserve exact optimizer state (momentum, learning rate schedules, etc.)
938
+
939
+
940
+ ---
941
+
942
+ ## File: training-sampling.mdx
943
+
944
+ import { Callout } from 'nextra/components'
945
+
946
+ # Getting started with training and sampling
947
+
948
+ In this guide, we'll step you through using the Tinker Python library to do the basic operations needed for training and sampling.
949
+ [View the complete Python script →](/quickstart.py.txt)
950
+
951
+ ## Creating the training client
952
+
953
+ The main object we'll be using is the `TrainingClient`, which corresponds to a fine-tuned model that we can train and sample from.
954
+
955
+ First, set your Tinker API key environment variable. In the terminal where you'll run Python, or in your `.bashrc`, put `export TINKER_API_KEY=<your key>`.
956
+
957
+ Then, create a `ServiceInterface`. This lets you find out what base models are available to be fine-tuned.
958
+
959
+ ```python
960
+ import tinker
961
+ service_client = tinker.ServiceClient()
962
+ print("Available models:")
963
+ for item in service_client.get_server_capabilities().supported_models:
964
+ print("- " + item.model_name)
965
+ ```
966
+ You'll see a list of model names:
967
+ ```
968
+ - meta-llama/Llama-3.1-70B
969
+ - meta-llama/Llama-3.1-8B
970
+ ...
971
+ - Qwen/Qwen3-VL-30B-A3B-Instruct
972
+ - Qwen/Qwen3-VL-235B-A22B-Instruct
973
+ ```
974
+ We currently support models from the Qwen3, Qwen3-VL, and Llama3 series. We'll use Qwen3-VL-30B-A3B-Instruct for these examples, as it's a vision-language model that can also handle text-only tasks. See [Available Models in Tinker](/model-lineup) for the full list.
975
+
976
+ Now we can create the `TrainingClient`:
977
+ ```python
978
+ base_model = "Qwen/Qwen3-VL-30B-A3B-Instruct"
979
+ training_client = service_client.create_lora_training_client(
980
+ base_model=base_model
981
+ )
982
+ ```
983
+ As the name suggests, this model was already finetuned for chat/instruction-following. You should check the details of the model you're using in their system cards.
984
+
985
+ ## Preparing the training data
986
+
987
+ Now we can do training updates on the model. This quickstart example won't show best practices for LLM fine-tuning; it's just an API demo. Check out [Rendering](/rendering), [Supervised Fine-tuning](/supervised-learning) and the other Cookbook examples for guidance on how to use Tinker in real applications.
988
+
989
+ For this model, we'll train a model that can translate words into Pig Latin. The rules for Pig Latin are simple:
990
+ - If a word begins with a consonant, move it to the end and add "ay"
991
+ - If a word begins with a vowel, just add "way" to the end
992
+
993
+ Here are some example completions we'd like the model to perform, where the prompt is in green and the model's completion is in red:
994
+
995
+ <div className="example">
996
+ <span className="prompt">English: hello world<br/>
997
+ Pig Latin: </span><span className="completion">ello-hay orld-way</span>
998
+ </div>
999
+
1000
+ Let's create some training examples and convert them to a format expected by Tinker.
1001
+
1002
+ ```python
1003
+ # Create some training examples
1004
+ examples = [
1005
+ {
1006
+ "input": "banana split",
1007
+ "output": "anana-bay plit-say"
1008
+ },
1009
+ {
1010
+ "input": "quantum physics",
1011
+ "output": "uantum-qay ysics-phay"
1012
+ },
1013
+ {
1014
+ "input": "donut shop",
1015
+ "output": "onut-day op-shay"
1016
+ },
1017
+ {
1018
+ "input": "pickle jar",
1019
+ "output": "ickle-pay ar-jay"
1020
+ },
1021
+ {
1022
+ "input": "space exploration",
1023
+ "output": "ace-spay exploration-way"
1024
+ },
1025
+ {
1026
+ "input": "rubber duck",
1027
+ "output": "ubber-ray uck-day"
1028
+ },
1029
+ {
1030
+ "input": "coding wizard",
1031
+ "output": "oding-cay izard-way"
1032
+ },
1033
+ ]
1034
+
1035
+ # Convert examples into the format expected by the training client
1036
+ from tinker import types
1037
+
1038
+ # Get the tokenizer from the training client
1039
+ tokenizer = training_client.get_tokenizer()
1040
+
1041
+ def process_example(example: dict, tokenizer) -> types.Datum:
1042
+ # Format the input with Input/Output template
1043
+ # For most real use cases, you'll want to use a renderer / chat template,
1044
+ # (see later docs) but here, we'll keep it simple.
1045
+ prompt = f"English: {example['input']}\nPig Latin:"
1046
+
1047
+ prompt_tokens = tokenizer.encode(prompt, add_special_tokens=True)
1048
+ prompt_weights = [0] * len(prompt_tokens)
1049
+ # Add a space before the output string, and finish with double newline
1050
+ completion_tokens = tokenizer.encode(f" {example['output']}\n\n", add_special_tokens=False)
1051
+ completion_weights = [1] * len(completion_tokens)
1052
+
1053
+ tokens = prompt_tokens + completion_tokens
1054
+ weights = prompt_weights + completion_weights
1055
+
1056
+ input_tokens = tokens[:-1]
1057
+ target_tokens = tokens[1:] # We're predicting the next token, so targets need to be shifted.
1058
+ weights = weights[1:]
1059
+
1060
+ # A datum is a single training example for the loss function.
1061
+ # It has model_input, which is the input sequence that'll be passed into the LLM,
1062
+ # loss_fn_inputs, which is a dictionary of extra inputs used by the loss function.
1063
+ return types.Datum(
1064
+ model_input=types.ModelInput.from_ints(tokens=input_tokens),
1065
+ loss_fn_inputs=dict(weights=weights, target_tokens=target_tokens)
1066
+ )
1067
+
1068
+ processed_examples = [process_example(ex, tokenizer) for ex in examples]
1069
+
1070
+ # Visualize the first example for debugging purposes
1071
+ datum0 = processed_examples[0]
1072
+ print(f"{'Input':<20} {'Target':<20} {'Weight':<10}")
1073
+ print("-" * 50)
1074
+ for i, (inp, tgt, wgt) in enumerate(zip(datum0.model_input.to_ints(), datum0.loss_fn_inputs['target_tokens'].tolist(), datum0.loss_fn_inputs['weights'].tolist())):
1075
+ print(f"{repr(tokenizer.decode([inp])):<20} {repr(tokenizer.decode([tgt])):<20} {wgt:<10}")
1076
+ ```
1077
+
1078
+ The visualization of the first example is:
1079
+
1080
+ ```
1081
+ Input Target Weight
1082
+ --------------------------------------------------
1083
+ 'English' ':' 0.0
1084
+ ':' ' I' 0.0
1085
+ ' I' ' love' 0.0
1086
+ ' love' ' tink' 0.0
1087
+ ' tink' 'ering' 0.0
1088
+ 'ering' '\n' 0.0
1089
+ '\n' 'P' 0.0
1090
+ 'P' 'ig' 0.0
1091
+ 'ig' ' Latin' 0.0
1092
+ ' Latin' ':' 0.0
1093
+ ':' ' I' 1.0
1094
+ ' I' '-way' 1.0
1095
+ '-way' ' o' 1.0
1096
+ ' o' 've' 1.0
1097
+ 've' '-l' 1.0
1098
+ '-l' 'ay' 1.0
1099
+ 'ay' ' ink' 1.0
1100
+ ' ink' 'ering' 1.0
1101
+ 'ering' '-t' 1.0
1102
+ '-t' 'ay' 1.0
1103
+ 'ay' '<|endoftext|>' 1.0
1104
+ ```
1105
+
1106
+ ## Vision inputs
1107
+
1108
+ The above example is text-only, but adding vision inputs is also straightforward. The `ModelInput` type takes a list of chunks, which can be either `EncodedTextChunk` or `ImageChunk`. For instance:
1109
+
1110
+ ```python
1111
+ image_data = requests.get("https://thinkingmachines.ai/blog/on-policy-distillation/images/chess.png").content
1112
+ model_input = tinker.ModelInput(chunks=[
1113
+ types.EncodedTextChunk(tokens=tokenizer.encode("<|im_start|>user\n<|vision_start|>")),
1114
+ types.ImageChunk(data=image_data, format="png"),
1115
+ types.EncodedTextChunk(tokens=tokenizer.encode("<|vision_end|>What is this?<|im_end|>\n<|im_start|>assistant\n")),
1116
+ ])
1117
+ ```
1118
+
1119
+ Note that Qwen3-VL was trained with special tokens like `<|vision_start|>` and `<|vision_end|>`. The cookbook's `Qwen3VLRenderer` handles these automatically—see [Rendering: Vision Inputs](/rendering#vision-inputs) for details and a complete example.
1120
+
1121
+ ## Performing a training update
1122
+
1123
+ Now we can use this data to perform a training update. We'll do 6 updates on the same batch of data. (Note that this is not typically a good way to train!)
1124
+
1125
+ ```python
1126
+ import numpy as np
1127
+ for _ in range(6):
1128
+ fwdbwd_future = training_client.forward_backward(processed_examples, "cross_entropy")
1129
+ optim_future = training_client.optim_step(types.AdamParams(learning_rate=1e-4))
1130
+
1131
+ # Wait for the results
1132
+ fwdbwd_result = fwdbwd_future.result()
1133
+ optim_result = optim_future.result()
1134
+
1135
+ # fwdbwd_result contains the logprobs of all the tokens we put in. Now we can compute the weighted
1136
+ # average log loss per token.
1137
+ logprobs = np.concatenate([output['logprobs'].tolist() for output in fwdbwd_result.loss_fn_outputs])
1138
+ weights = np.concatenate([example.loss_fn_inputs['weights'].tolist() for example in processed_examples])
1139
+ print(f"Loss per token: {-np.dot(logprobs, weights) / weights.sum():.4f}")
1140
+ ```
1141
+
1142
+ Note that the `forward_backward` and `optim_step` functions immediately return *futures*, which acknowledge that the task has been queued up by the server. For improved speed, we submitted both operations before waiting for the result by calling `result()` on the futures.
1143
+
1144
+
1145
+ ## Sampling from the model
1146
+
1147
+ Now we can test our model by sampling from it. In this case, we'll translate the phrase "coffee break" into Pig Latin.
1148
+
1149
+ ```python
1150
+ # First, create a sampling client. We need to transfer weights
1151
+ sampling_client = training_client.save_weights_and_get_sampling_client(name='pig-latin-model')
1152
+
1153
+ # Now, we can sample from the model.
1154
+ prompt = types.ModelInput.from_ints(tokenizer.encode("English: coffee break\nPig Latin:"))
1155
+ params = types.SamplingParams(max_tokens=20, temperature=0.0, stop=["\n"]) # Greedy sampling
1156
+ future = sampling_client.sample(prompt=prompt, sampling_params=params, num_samples=8)
1157
+ result = future.result()
1158
+ print("Responses:")
1159
+ for i, seq in enumerate(result.sequences):
1160
+ print(f"{i}: {repr(tokenizer.decode(seq.tokens))}")
1161
+ ```
1162
+
1163
+ Since sampling is nondeterministic (sadly, even with temperature=0.0, [due to batching](https://thinkingmachines.ai/blog/defeating-nondeterminism-in-llm-inference/)), the output will be different each time. You should see something like this:
1164
+
1165
+ ```
1166
+ Responses:
1167
+ 0: ' offe-bay eak-bay\n\n'
1168
+ 1: ' offey-coy eak-bray\n\n'
1169
+ 2: ' offecay eakbray\n\n'
1170
+ 3: ' offeec-cay eak-brcay\n\n\n'
1171
+ 4: ' offecay akebay\n\n'
1172
+ 5: ' offee-Cay ake-bay\n\n\n'
1173
+ 6: ' offey-pay eak-bray\n\n'
1174
+ 7: ' offee – cay eak – bray\n\n'
1175
+ ```
1176
+
1177
+ ### Computing logprobs for a sequence
1178
+
1179
+ We can use the sampler to compute logprobs for a given sequence as well. This uses the prefill step and is returned as _prompt logprobs_.
1180
+
1181
+ ```python
1182
+ prompt = types.ModelInput.from_ints(tokenizer.encode("How many r's are in the word strawberry?"))
1183
+ sample_response = sampling_client.sample(
1184
+ prompt=prompt,
1185
+ num_samples=1,
1186
+ sampling_params=tinker.SamplingParams(max_tokens=1), # Must be at least 1 token, represents prefill step
1187
+ include_prompt_logprobs=True,
1188
+ ).result()
1189
+
1190
+ # example: [None, -9.54505, -1.64629, -8.81116, -3.50217, -8.25927, ...]
1191
+ print(sample_response.prompt_logprobs)
1192
+ ```
1193
+
1194
+ The first logprob is `None` (corresponding to the first token), and subsequent entries are logprobs of each token in the prompt.
1195
+
1196
+ The sampling client also has a helper function, which is the same as above:
1197
+
1198
+ ```python
1199
+ sampling_client.compute_logprobs(prompt).result()
1200
+ ```
1201
+
1202
+ ### Top-k logprobs
1203
+
1204
+ For distillation, it may be especially useful to compute _top-k logprobs_ for each token as well, which can get you a sense for what the model "would have said" after each prefix instead of the actual prompt.
1205
+
1206
+ ```python
1207
+ sample_response = sampling_client.sample(
1208
+ prompt=prompt,
1209
+ num_samples=1,
1210
+ sampling_params=tinker.SamplingParams(max_tokens=1),
1211
+ include_prompt_logprobs=True,
1212
+ topk_prompt_logprobs=5,
1213
+ ).result()
1214
+
1215
+ # example:
1216
+ # [None,
1217
+ # [(14924, -1.17005), (755, -2.23255), (2, -2.73255), (791, -3.67005), (16309, -4.29505)],
1218
+ # [(25, -1.64629), (3137, -2.39629), (11630, -2.89629), (21460, -3.83379), (14881, -4.02129)],
1219
+ # [(41, -3.49866), (42, -3.49866), (49, -4.24866), (38, -4.37366), (54, -4.49866)],
1220
+ # [(311, -1.00217), (656, -2.25217), (2057, -2.75217), (649, -3.25217), (10470, -3.37717)],
1221
+ # ...]
1222
+ sample_response.topk_prompt_logprobs
1223
+ ```
1224
+
1225
+ For each position in the response, this returns a list of `(token_id, logprob)` pairs for the top-k most likely tokens at that position.
1226
+
1227
+ ## Putting it together: Sampling from an image
1228
+
1229
+ Here's a complete example that creates a training client, saves weights for sampling, and asks a question about an image. You can copy-paste it into an iPython notebook:
1230
+
1231
+ ```python
1232
+ import requests
1233
+ import tinker
1234
+ from transformers import AutoTokenizer
1235
+
1236
+ model_name = "Qwen/Qwen3-VL-30B-A3B-Instruct"
1237
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
1238
+
1239
+ service_client = tinker.ServiceClient()
1240
+ training_client = await service_client.create_lora_training_client_async(base_model=model_name, rank=32)
1241
+ sampling_client = await training_client.save_weights_and_get_sampling_client_async(name="sampler")
1242
+
1243
+ # Grab an image and ask a question
1244
+ image_data = requests.get("https://thinkingmachines.ai/blog/on-policy-distillation/images/chess.png").content
1245
+ model_input = tinker.ModelInput(chunks=[
1246
+ tinker.types.EncodedTextChunk(tokens=tokenizer.encode("<|im_start|>user\n<|vision_start|>")),
1247
+ tinker.types.ImageChunk(data=image_data, format="png"),
1248
+ tinker.types.EncodedTextChunk(tokens=tokenizer.encode("<|vision_end|>What is this?<|im_end|>\n<|im_start|>assistant\n")),
1249
+ ])
1250
+
1251
+ result = await sampling_client.sample_async(prompt=model_input, num_samples=1, sampling_params=tinker.types.SamplingParams(max_tokens=100))
1252
+ print(tokenizer.decode(result.sequences[0].tokens))
1253
+ ```
1254
+
1255
+ For higher-level abstractions that handle special tokens automatically, see [Rendering: Vision Inputs](/rendering#vision-inputs).
1256
+
1257
+
1258
+ ---
1259
+
1260
+ ## File: rendering.mdx
1261
+
1262
+ import { CookbookLink } from '../components/CookbookLink'
1263
+
1264
+
1265
+ # Rendering to tokens
1266
+
1267
+ Rendering converts list-of-message datatypes into their token representations for model training and inference. While similar to [chat templates](https://huggingface.co/docs/transformers/en/chat_templating), Tinker's rendering system is designed for the full training lifecycle--not just inference--supporting supervised learning, reinforcement learning, and deployment.
1268
+
1269
+
1270
+ ## The Renderer class
1271
+
1272
+ The Renderer class is the main interface used for rendering. It can be found in <CookbookLink path="tinker_cookbook/renderers.py">`renderers.py`</CookbookLink>.
1273
+
1274
+ **Eample conversation:**
1275
+
1276
+ ```python
1277
+ messages =[
1278
+ {'role': 'system', 'content': 'Answer concisely; at most one sentence per response'},
1279
+ {'role': 'user', 'content': 'What is the longest-lived rodent species?'},
1280
+ {'role': 'assistant', 'content': 'The naked mole rat, which can live over 30 years.'},
1281
+ {'role': 'user', 'content': 'How do they live so long?'},
1282
+ {'role': 'assistant', 'content': 'They evolved multiple protective mechanisms including special hyaluronic acid that prevents cancer, extremely stable proteins, and efficient DNA repair systems that work together to prevent aging.'}
1283
+ ]
1284
+ ```
1285
+
1286
+ We'll use this conversation throughout the examples below.
1287
+
1288
+ ## Inference: Generating messages
1289
+
1290
+ Our model maps tokens to tokens, but with the renderer, it can map messages to messages. To sample messages from the model, we need to use three methods from the renderer:
1291
+
1292
+ - `build_generation_prompt`
1293
+ - `get_stop_sequences`
1294
+ - `parse_response`
1295
+
1296
+
1297
+ `build_generation_prompt` converts a conversation into a prompt that we can use to sample from the assistant. This is used during reinforcement learning and at deployment time.
1298
+
1299
+
1300
+ **Example: Generate an alternative assistant response**
1301
+
1302
+ Let's remove the last assistant message and call `build_generation_prompt` to get a prompt that we can use to sample an alternative response from the assistant:
1303
+
1304
+ ```python
1305
+ from tinker_cookbook import renderers, tokenizer_utils
1306
+ tokenizer = tokenizer_utils.get_tokenizer('Qwen/Qwen3-30B-A3B')
1307
+ renderer = renderers.get_renderer('qwen3', tokenizer)
1308
+ prompt = renderer.build_generation_prompt(messages[:-1])
1309
+ print(prompt)
1310
+ print('-'*10)
1311
+ print(tokenizer.decode(prompt.to_ints()))
1312
+ ```
1313
+
1314
+ **Output:**
1315
+ ```
1316
+ ModelInput(chunks=[EncodedTextChunk(tokens=[151644, 8948, 198, 2610, 525, 264, 10950, 17847, 13, 151645, 198, 151644, 8948, 198, 16141, 3529, 285, 974, 26, 518, 1429, 825, 11652, 817, 2033, 151645, 198, 151644, 872, 198, 3838, 374, 279, 22032, 61854, 20589, 306, 9419, 30, 151645, 198, 151644, 77091, 198, 785, 19020, 34651, 11244, 11, 892, 646, 3887, 916, 220, 18, 15, 1635, 13, 151645, 198, 151644, 872, 198, 10234, 30, 151645, 198, 151644, 77091, 198], type='encoded_text')])
1317
+ ----------
1318
+ <|im_start|>system
1319
+ Answer concisely; at most one sentence per response<|im_end|>
1320
+ <|im_start|>user
1321
+ What is the longest-lived rodent species?<|im_end|>
1322
+ <|im_start|>assistant
1323
+ The naked mole rat, which can live over 30 years.<|im_end|>
1324
+ <|im_start|>user
1325
+ How do they live so long?<|im_end|>
1326
+ <|im_start|>assistant
1327
+
1328
+ ```
1329
+
1330
+ You can see that the prompt is a `ModelInput` object, which is a list of `EncodedTextChunk` objects (but contains different objects in multi-modal data).
1331
+
1332
+
1333
+ **Sampling and parsing the response:**
1334
+
1335
+ Given that we're providing messages as input, we probably want a message output, rather than a token output. For that, we can use `parse_response`.
1336
+
1337
+ ```python
1338
+ import tinker
1339
+ from tinker.types import SamplingParams
1340
+ service_client = tinker.ServiceClient()
1341
+ sampling_client = service_client.create_sampling_client(base_model='Qwen/Qwen3-30B-A3B')
1342
+ stop_sequences = renderer.get_stop_sequences()
1343
+ print(f"Stop sequences: {stop_sequences}")
1344
+ sampling_params = SamplingParams(max_tokens=100, temperature=0.5, stop=stop_sequences)
1345
+ output = sampling_client.sample(prompt, sampling_params=sampling_params, num_samples=1).result()
1346
+ print(f"Sampled tokens: {output.sequences[0].tokens}")
1347
+ sampled_message, parse_success = renderer.parse_response(output.sequences[0].tokens)
1348
+ print(f"Sampled message: {sampled_message}")
1349
+ print(f"Parse success: {parse_success}")
1350
+ ```
1351
+
1352
+ **Output:**
1353
+
1354
+ ```
1355
+ Stop sequences: [151645]
1356
+ Sampled tokens: [45, 7741, 34651, 31410, 614, 4911, 76665, 11, 2670, 264, 7548, 11050, 22077, 1849, 323, 264, 1602, 3347, 40761, 4379, 11, 892, 16792, 311, 862, 57119, 13, 151645]
1357
+ Sampled message: {'role': 'assistant', 'content': 'Naked mole rats have unique adaptations, including a highly efficient immune system and a very low metabolic rate, which contribute to their longevity.'}
1358
+ Parse success: True
1359
+ ```
1360
+
1361
+ You can see that the there is one stop sequence, `151645`, which you can verify is the `<|im_end|>` token. The output is parsed successfully into a message.
1362
+
1363
+
1364
+ ## Training: Supervised learning
1365
+
1366
+ For supervised learning (and some other algorithms like [DPO](/preferences/dpo-guide)), we need to distinguish between **prompt tokens** (context) and **completion tokens** (what the model should learn to generate). We want to provide a target assistant message, and the renderer needs to tell us which tokens are part of the prompt and completion.
1367
+
1368
+ We can use `build_supervised_example` to get a `ModelInput` and per-token loss weights:
1369
+
1370
+ ```python
1371
+ model_input, weights = renderer.build_supervised_example(messages)
1372
+
1373
+ from tinker_cookbook.utils.format_colorized import format_colorized
1374
+ print(format_colorized(model_input.to_ints(), weights, tokenizer))
1375
+ ```
1376
+
1377
+ We get the following output:
1378
+
1379
+ <div className="example">
1380
+ <span className="prompt">&lt;|im_start|&gt;system↵<br />Answer concisely; at most one sentence per response&lt;|im_end|&gt;↵<br />&lt;|im_start|&gt;user↵<br />What is the longest-lived rodent species?&lt;|im_end|&gt;↵<br />&lt;|im_start|&gt;assistant↵<br />The naked mole rat, which can live over 30 years.&lt;|im_end|&gt;↵<br />&lt;|im_start|&gt;user↵<br />How do they live so long?&lt;|im_end|&gt;↵<br />&lt;|im_start|&gt;assistant↵<br /></span>
1381
+ <span className="completion">They evolved multiple protective mechanisms including special hyaluronic acid that prevents cancer, extremely stable proteins, and efficient DNA repair systems that work together to prevent aging.&lt;|im_end|&gt;<br /></span>
1382
+ </div>
1383
+ The green text is part of the prompt (i.e. with `weight=0`, so no loss is computed on these) and red is part of the completion (i.e. with `weight=1`, so the model is trained to predict these). Note that the ↵ have been inserted for clarity to show newlines; these are not actually part of the token sequence.
1384
+
1385
+ The key insight here is that only the final assistant message is treated as the completion. All previous context, including the first assistant response, is part of the prompt, so the model learns to continue conversations rather than just answer single questions.
1386
+
1387
+ ## Vision Inputs
1388
+
1389
+ Tinker supports vision-language models (VLMs) like `Qwen/Qwen3-VL-30B-A3B-Instruct` and `Qwen/Qwen3-VL-235B-A22B-Instruct`. For low-level `ImageChunk` usage, see [Vision inputs](/training-sampling#vision-inputs) in the Training and Sampling guide. This section covers the higher-level message abstractions.
1390
+
1391
+ ### Multimodal messages
1392
+
1393
+ For VLMs, message content can be either a string or a list of content parts:
1394
+
1395
+ ```python
1396
+ from tinker_cookbook.renderers import Message, TextPart, ImagePart
1397
+
1398
+ # Text-only message (standard)
1399
+ text_message = Message(role='user', content='What is this?')
1400
+
1401
+ # Multimodal message with image
1402
+ multimodal_message = Message(
1403
+ role='user',
1404
+ content=[
1405
+ ImagePart(type='image', image='https://thinkingmachines.ai/blog/on-policy-distillation/images/chess.png'),
1406
+ TextPart(type='text', text='What is in this image?'),
1407
+ ]
1408
+ )
1409
+ ```
1410
+
1411
+ For lower-level control using `ImageChunk` directly, see [Vision inputs](/training-sampling#vision-inputs) in the Training and Sampling guide.
1412
+
1413
+ ### Using Qwen3VLRenderer
1414
+
1415
+ The `Qwen3VLRenderer` handles Qwen's vision special tokens (`<|vision_start|>`, `<|vision_end|>`) automatically:
1416
+
1417
+ ```python
1418
+ from tinker_cookbook import renderers, tokenizer_utils
1419
+ from tinker_cookbook.image_processing_utils import get_image_processor
1420
+
1421
+ model_name = "Qwen/Qwen3-VL-235B-A22B-Instruct"
1422
+ tokenizer = tokenizer_utils.get_tokenizer(model_name)
1423
+ image_processor = get_image_processor(model_name)
1424
+
1425
+ renderer = renderers.Qwen3VLRenderer(tokenizer, image_processor)
1426
+
1427
+ messages = [
1428
+ {
1429
+ 'role': 'user',
1430
+ 'content': [
1431
+ {'type': 'image', 'image': 'https://thinkingmachines.ai/blog/on-policy-distillation/images/chess.png'},
1432
+ {'type': 'text', 'text': 'What is in this image?'},
1433
+ ]
1434
+ }
1435
+ ]
1436
+
1437
+ prompt = renderer.build_generation_prompt(messages)
1438
+ ```
1439
+
1440
+ For a complete example of training a VLM image classifier, see the <CookbookLink path="tinker_cookbook/recipes/vlm_classifier">VLM Classifier recipe</CookbookLink> in the cookbook.
1441
+
1442
+ ## Multi-turn RL and the Extension Property
1443
+
1444
+ When using renderers in multi-turn RL, an important consideration is whether consecutive timesteps satisfy the **extension property**—where each observation is a prefix extension of the previous observation plus action. This affects compute efficiency (O(T) vs O(T^2)) and KV-cache reuse.
1445
+
1446
+ Some renderers, like `Qwen3Renderer`, have options that affect this property. For example, `strip_thinking_from_history` controls whether `<think>` blocks are preserved in conversation history.
1447
+
1448
+ See the [Sequence Extension](/rl/sequence-extension) documentation for details on how this works and the tradeoffs involved.
1449
+
1450
+ ## Appendix: Why not Jinja templates?
1451
+
1452
+ In our experience, the Jinja2 templates are harder to write than Python code, especially when we need to get the whitespace exactly right. They are also unwieldy for supervised learning, where you need to put different labels on different tokens.
1453
+
1454
+
1455
+ ---
1456
+
1457
+ ## File: completers.mdx
1458
+
1459
+ import { CookbookLink } from '../components/CookbookLink'
1460
+
1461
+ # Completers
1462
+
1463
+ The concept of policies is crucial to the RL training process. In the Tinker Cookbook, policies are implemented as `Completers`. Completers are abstractions that represent models or policies that can be sampled from, providing different levels of structure depending on your use case.
1464
+
1465
+ ## Overview of Completer Types
1466
+
1467
+ The Tinker Cookbook provides two main types of completers, each designed for different use cases:
1468
+
1469
+ 1. **TokenCompleter**: Operates on tokens and is used by RL algorithms
1470
+ 2. **MessageCompleter**: Operates on messages and needs to be used with a renderer
1471
+
1472
+ The choice between these depends on whether you're working at the token level for RL training or at the message level for interacting with and evaluating the model.
1473
+
1474
+ ### TokenCompleter
1475
+
1476
+ The `TokenCompleter` is the foundational interface used by RL algorithms because they work directly with tokens.
1477
+
1478
+ ```python
1479
+ class TokenCompleter:
1480
+ async def __call__(
1481
+ self, model_input: types.ModelInput, stop: StopCondition
1482
+ ) -> TokensWithLogprobs:
1483
+ ```
1484
+
1485
+ This interface takes:
1486
+ - `model_input`: The input to the model (of type `types.ModelInput`)
1487
+ - `stop`: Stop conditions, either a list of strings or token IDs (combined into a `StopCondition` class). When training with reinforcement learning, this should be defined by the `initial_observation` function of the environment.
1488
+
1489
+ It returns a `TokensWithLogprobs` object containing:
1490
+ - `tokens`: The generated token sequence
1491
+ - `maybe_logprobs`: Optional log probabilities for each token
1492
+
1493
+ ### MessageCompleter
1494
+
1495
+ The `MessageCompleter` operates at a higher level with structured messages, similarly to standard chat APIs. It takes a list of messages and returns a single assistant message response.
1496
+
1497
+ ```python
1498
+ class MessageCompleter:
1499
+ async def __call__(self, messages: list[renderers.Message]) -> renderers.Message:
1500
+ ```
1501
+
1502
+ For training purposes the `TokenCompleter` is the class we will use for RL training as we need to optimize the same same set of tokens during the update step that the model output during rollout. The `MessageCompleter` is useful for sampling where we need to use the model output for semantic purposes such as Judge models or multi-agent environments.
1503
+
1504
+ The Tinker Cookbook uses two concrete implementations of these interfaces - <CookbookLink path="tinker_cookbook/completers.py">`TinkerTokenCompleter`</CookbookLink> and <CookbookLink path="tinker_cookbook/completers.py">`TinkerMessageCompleter`</CookbookLink> which are both wrappers around a `tinker.SamplingClient`. While the TinkerTokenCompleter operates directly on tokens, the TinkerMessageCompleter needs to be instantiated with a renderer to make it compatible with the inputs expected by the samping client.
1505
+
1506
+
1507
+ ---
1508
+
1509
+ ## File: install.mdx
1510
+
1511
+ # Installing Tinker
1512
+
1513
+ Install the Tinker SDK with:
1514
+
1515
+ ```bash
1516
+ pip install tinker
1517
+ ```
1518
+
1519
+ Installation makes two components available: the python SDK and the tinker CLI.
1520
+
1521
+ #### Python SDK
1522
+
1523
+ The python SDK provides low-level operations like `forward_backward`, `sample`, `optim_step`, and `save_state`.
1524
+
1525
+ #### Tinker CLI
1526
+
1527
+ The tinker CLI is available as `tinker` or through `python -m tinker`. The CLI provides management functionality similar to that of the web console.
1528
+
1529
+ Run `tinker --help` to see which functionality is available.
1530
+
1531
+ ## Tinker Cookbook
1532
+
1533
+ We also release [tinker-cookbook](https://github.com/thinking-machines-lab/tinker-cookbook), which is a collection of training code and experiment tools built on top of Tinker.
1534
+ For the Cookbook, we'd recommend doing a local editable install, as you'll probably want to browse and edit the code:
1535
+
1536
+ ```bash
1537
+ git clone https://github.com/thinking-machines-lab/tinker-cookbook.git
1538
+ cd tinker-cookbook
1539
+ # Switch to your virtual environment
1540
+ pip install -e .
1541
+ ```
1542
+
1543
+ ## Getting an API key
1544
+
1545
+ Create an API key from the [console](https://tinker-console.thinkingmachines.ai). You'll then want to set the `TINKER_API_KEY` environment variable to your newly generated API key.
1546
+
1547
+
1548
+ ---
1549
+
1550
+ ## File: rl.mdx
1551
+
1552
+ import { CookbookLink } from '../components/CookbookLink'
1553
+
1554
+ # Reinforcement learning
1555
+
1556
+ Reinforcement learning (RL) means learning from trial and error. Whereas in supervised learning, we're given input-output pairs, in RL, we're given inputs (prompts) and reward functions (i.e., a function for scoring candidate outputs). RL algorithms need to discover what good outputs look like.
1557
+
1558
+ Here are a few different types of RL training that we support in the Tinker Cookbook:
1559
+
1560
+ - *RL with Verifiable Rewards*: this is when we do RL on a reward function that checks model outputs using a program. Typically, the reward function checks the candidate answer against a reference answer, or, in coding cases, it may check if the candidate solution passes some unit tests. RLVR is especially suitable for teaching models to do reasoning (with chain-of-thought) and multi-step tool use (e.g., debugging and iterative modification pf programs).
1561
+ - *RL on Human Feedback*: here, we assume we have an objective that can't be calculated by a simple program, and it requires some human judgement. For example, we typically want to optimize our models for helpfulness, which includes being clear, informative, and interesting. For RLHF, we train a *preference model* using supervised learning to match human judgement, scoring or ranking candidate outputs. Then we do RL on the preference model's scores. See the [Preferences](/preferences) section for more details.
1562
+
1563
+ We'll first show how to do small RL runs in the RLVR setting, then we'll show you how to define your own RL environments and train on them, then we'll provide examples for larger-scale or more complicated training setups.
1564
+
1565
+
1566
+ We anticipate that people will want to use Tinker for RL in a few different ways:
1567
+
1568
+ - Creating a specialist model that's SoTA at a specific skill, which existing models haven't been trained on. In this case, you'll want to start with a post-trained model that's already strong, and then do RL on an environment you've defined. See [RL Environments](/rl/rl-envs).
1569
+ - Doing research on post-training pipelines. In this case, you'll probably want to chain together SL and RL and runs with different data mixes, environments, and reward functions. See our [RLHF example](/preferences/rlhf-example).
1570
+ - Doing research on RL algorithms. Here, you'll probably want to find some existing environments to use as benchmarks, and either modify our provided training code (<CookbookLink path="tinker_cookbook/rl/train.py">rl/train.py</CookbookLink>) or write your own minimal training loop. We've provided a [minimal training loop](/rl/rl-loops) that you can use as a starting point.
1571
+
1572
+
1573
+ ---
1574
+
1575
+ ## File: under-the-hood.mdx
1576
+
1577
+ # Under the Hood
1578
+
1579
+ This page explains some implementation details of Tinker, which are important for understanding how to speed up your code.
1580
+
1581
+ ## Clock Cycles
1582
+
1583
+ In Tinker, after you call `ServiceClient.create_lora_training_client`, your training job gets assigned to a pool of machines that working together -- a *worker pool* -- which are doing forward-backward operations repeatedly in lock-step.
1584
+ Each of these steps of the worker pool is called a *clock cycle*.
1585
+ In each clock cycle, we do forward-backward and an optimizer step operation, each of which may involve multiple LoRA models that are being trained by this pool.
1586
+ You can think of this pool as a single large training run that is time-shared between multiple different LoRA models, often from different users.
1587
+
1588
+ With multi-tenancy -- sharing the same worker pool between multiple models -- we can run the training system efficiently even if users are training with small batch sizes, or if they have other delays in their training loops that would otherwise leave the worker pool idle. Small batch sizes can often give better *sample efficiency*, so this setup lets us achieve both high compute efficiency and high sample efficiency.
1589
+
1590
+ The downside is that it can sometimes lead to worse *latency*: even if training with a small batch, you'll still see the same step time as a large batch. (Still, note that we'll only charge you for the compute you use.) Also, if your training loop is implemented naively, you might have to wait multiple clock cycles per batch, because you might miss a clock cycle between operations.
1591
+
1592
+ ### Overlapping `forward_backward` and `optim_step` Requests
1593
+
1594
+ As mentioned in the [Async and Futures](/async) section, you should submit your `forward_backward` and `optim_step` requests together before waiting for either of them. This way, they'll end up on the same clock cycle. If you write the code naively, you'll end up using *three* clock cycles per training step. Here's a recap of the example from the [Async and Futures](/async) section:
1595
+
1596
+ **❌ Naive implementation (uses 3 clock cycles):**
1597
+ ```python
1598
+ # Submit forward_backward, gets queued for clock cycle N
1599
+ fwd_bwd_future = await client.forward_backward_async(batch, loss_fn)
1600
+
1601
+ # Wait for it to complete, and for client to receive the result
1602
+ # Due to communication latency, this happens a little after cycle N+1 started
1603
+ fwd_bwd_result = await fwd_bwd_future
1604
+
1605
+ # Submit optim_step, gets queued for clock cycle N+2
1606
+ optim_future = await client.optim_step_async(adam_params)
1607
+
1608
+ # Wait for it to complete, and for client to receive the result
1609
+ # This happens a little after cycle N+2 finishes
1610
+ optim_result = await optim_future
1611
+
1612
+ # Total: forward_backward on cycle N, optim_step on cycle N+2
1613
+ # This takes 3 clock cycles (plus the time we waited before cycle N started)
1614
+ ```
1615
+
1616
+ **✓ Better implementation (uses 1 clock cycle):**
1617
+ ```python
1618
+ # Submit both requests immediately. They'll both be slotted into the same clock cycle N
1619
+ fwd_bwd_future = await client.forward_backward_async(batch, loss_fn)
1620
+ optim_future = await client.optim_step_async(adam_params)
1621
+
1622
+ # Now wait for results - both operations happen on cycle N
1623
+ fwd_bwd_result = await fwd_bwd_future
1624
+ optim_result = await optim_future
1625
+
1626
+ # Total: both operations on cycle N
1627
+ # This takes 1 clock cycle
1628
+ ```
1629
+
1630
+ ### Pipelining to Maximize Clock Cycle Efficiency
1631
+
1632
+ To maximize efficiency and avoid missing clock cycles, you should **pipeline your training loop**: submit the next batch before waiting for the current batch to complete. This ensures there's always a request queued when a new clock cycle starts.
1633
+
1634
+ We've created a demonstration script that shows the difference between pipelined and non-pipelined training:
1635
+
1636
+ [View the clock cycles demonstration script →](/clock_cycles.py.txt)
1637
+
1638
+ The script includes two versions:
1639
+
1640
+ - **Non-pipelined**: Submits a batch, waits for it to complete, then submits the next. This approach typically wastes clock cycles because there's a gap between when one batch finishes and the next is submitted, often using 2 clock cycles per training step.
1641
+
1642
+ - **Pipelined**: Submits the next batch *before* waiting for the previous batch to complete. This approach often uses exactly 1 clock cycle per step, achieving maximum efficiency. Though it might sometimes take more than 1 clock cycle per step if the server is heavily loaded, or due to subtleties of our current implementation. (For example, if there are no other users, we might start the clock cycle after receiving the first `forward_backward` but before receiving the `optim_step`. Then we'll do `optim_step` on the next cycle. This causes an extra clock cycle but doesn't cause a slowdown.)
1643
+
1644
+ Running the script will show you the performance comparison, including total time and clock cycles used. The pipelined version typically saves both time and clock cycles.
1645
+
1646
+
1647
+ ---
1648
+
1649
+ ## File: model-lineup.mdx
1650
+
1651
+ # Available Models in Tinker
1652
+
1653
+ The table below shows the models that are currently available in Tinker. We plan to update this list as new models are released.
1654
+
1655
+ ## What model should I use?
1656
+
1657
+ - In general, use MoE models, which are more cost effective than the dense models.
1658
+ - Use Base models only if you're doing research or are running the full post-training pipeline yourself
1659
+ - If you want to create a model that is good at a specific task or domain, use an existing post-trained model model, and fine-tune it on your own data or environment.
1660
+ - If you care about latency, use one of the Instruction models, which will start outputting tokens without a chain-of-thought.
1661
+ - If you care about intelligence and robustness, use one of the Hybrid or Reasoning models, which can use long chain-of-thought.
1662
+
1663
+ ## Full Listing
1664
+
1665
+ | Model Name | Training Type | Architecture | Size |
1666
+ | ----------------------------------------------------------------------------------------------- | ------------- | ------------ | --------- |
1667
+ | [Qwen/Qwen3-VL-235B-A22B-Instruct](https://huggingface.co/Qwen/Qwen3-VL-235B-A22B-Instruct) | Vision | MoE | Large |
1668
+ | [Qwen/Qwen3-VL-30B-A3B-Instruct](https://huggingface.co/Qwen/Qwen3-VL-30B-A3B-Instruct) | Vision | MoE | Medium |
1669
+ | [Qwen/Qwen3-235B-A22B-Instruct-2507](https://huggingface.co/Qwen/Qwen3-235B-A22B-Instruct-2507) | Instruction | MoE | Large |
1670
+ | [Qwen/Qwen3-30B-A3B-Instruct-2507](https://huggingface.co/Qwen/Qwen3-30B-A3B-Instruct-2507) | Instruction | MoE | Medium |
1671
+ | [Qwen/Qwen3-30B-A3B](https://huggingface.co/Qwen/Qwen3-30B-A3B) | Hybrid | MoE | Medium |
1672
+ | [Qwen/Qwen3-30B-A3B-Base](https://huggingface.co/Qwen/Qwen3-30B-A3B-Base) | Base | MoE | Medium |
1673
+ | [Qwen/Qwen3-32B](https://huggingface.co/Qwen/Qwen3-32B) | Hybrid | Dense | Medium |
1674
+ | [Qwen/Qwen3-8B](https://huggingface.co/Qwen/Qwen3-8B) | Hybrid | Dense | Small |
1675
+ | [Qwen/Qwen3-8B-Base](https://huggingface.co/Qwen/Qwen3-8B-Base) | Base | Dense | Small |
1676
+ | [Qwen/Qwen3-4B-Instruct-2507](https://huggingface.co/Qwen/Qwen3-4B-Instruct-2507) | Instruction | Dense | Compact |
1677
+ | [openai/gpt-oss-120b](https://huggingface.co/openai/gpt-oss-120b) | Reasoning | MoE | Medium |
1678
+ | [openai/gpt-oss-20b](https://huggingface.co/openai/gpt-oss-20b) | Reasoning | MoE | Small |
1679
+ | [deepseek-ai/DeepSeek-V3.1](https://huggingface.co/deepseek-ai/DeepSeek-V3.1) | Hybrid | MoE | Large |
1680
+ | [deepseek-ai/DeepSeek-V3.1-Base](https://huggingface.co/deepseek-ai/DeepSeek-V3.1-Base) | Base | MoE | Large |
1681
+ | [meta-llama/Llama-3.1-70B](https://huggingface.co/meta-llama/Llama-3.1-70B) | Base | Dense | Large |
1682
+ | [meta-llama/Llama-3.3-70B-Instruct](https://huggingface.co/meta-llama/Llama-3.3-70B-Instruct) | Instruction | Dense | Large |
1683
+ | [meta-llama/Llama-3.1-8B](https://huggingface.co/meta-llama/Llama-3.1-8B) | Base | Dense | Small |
1684
+ | [meta-llama/Llama-3.1-8B-Instruct](https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct) | Instruction | Dense | Small |
1685
+ | [meta-llama/Llama-3.2-3B](https://huggingface.co/meta-llama/Llama-3.2-3B) | Base | Dense | Compact |
1686
+ | [meta-llama/Llama-3.2-1B](https://huggingface.co/meta-llama/Llama-3.2-1B) | Base | Dense | Compact |
1687
+ | [moonshotai/Kimi-K2-Thinking](https://huggingface.co/moonshotai/Kimi-K2-Thinking) | Reasoning | MoE | Large |
1688
+
1689
+ ## Legend
1690
+
1691
+ ### Training Types
1692
+ - **Base**: Foundation models trained on raw text data, suitable for post-training research and custom fine-tuning.
1693
+ - **Instruction**: Models fine-tuned for following instructions and chat, optimized for fast inference.
1694
+ - **Reasoning**: Models that always use chain-of-thought reasoning before their "visible" output that responds to the prompt.
1695
+ - **Hybrid**: Models that can operate in both thinking and non-thinking modes, where the non-thinking mode requires using a special renderer or argument that disables chain-of-thought.
1696
+ - **Vision**: Vision-language models (VLMs) that can process images alongside text. See [Vision Inputs](/rendering#vision-inputs) for usage.
1697
+
1698
+ ### Architecture
1699
+ - **Dense**: Standard transformer architecture with all parameters active
1700
+ - **MoE**: Mixture of Experts architecture with sparse activation
1701
+
1702
+ ### Model Sizes
1703
+
1704
+ - **Compact**: 1B-4B parameters
1705
+ - **Small**: 8B parameters
1706
+ - **Medium**: 30B-32B parameters
1707
+ - **Large**: 70B+ parameters
1708
+
1709
+ Note that the MoE models are much more cost effective than the dense models as their cost is proportional to the number of active parameters and not the total number of parameters.
1710
+
1711
+
1712
+ ---
1713
+
1714
+ ## File: preferences/dpo-guide.mdx
1715
+
1716
+ import { Callout } from 'nextra/components'
1717
+ import { CookbookLink } from '../../components/CookbookLink'
1718
+
1719
+ # Direct Preference Optimization (DPO)
1720
+
1721
+ Direct Preference Optimization (DPO) is a method for training language models to align with human preferences without requiring a separate reward model. Instead of using reinforcement learning with human feedback (RLHF), DPO directly optimizes the model to prefer chosen responses over rejected ones using a simple classification loss.
1722
+
1723
+ ## DPO Algorithm Details
1724
+
1725
+ The core DPO loss is computed as:
1726
+
1727
+ $$
1728
+ \mathcal{L}_{\theta} = -\mathbb{E}_{x, y_\text{chosen}, y_\text{rejected} \sim \mathcal{D}}\left[\log\sigma\left(\beta\log \frac{\pi_{\theta}(y_\text{chosen}|x)}{\pi_{\text{ref}}(y_\text{chosen}|x)} - \beta\log \frac{\pi_{\theta}(y_\text{rejected}|x)}{\pi_{\text{ref}}(y_\text{rejected}|x)}\right)\right]
1729
+ $$
1730
+
1731
+ Where:
1732
+ - $\pi_{\theta}$ is the current policy
1733
+ - $\pi_{\text{ref}}$ is the reference model (typically the initial model before DPO training)
1734
+ - $\beta$ is the DPO beta parameter
1735
+ - Where $\mathcal{D}$ is a dataset of prompts $x$, a chosen response $y_{\text{chosen}}$ and a rejected response $y_{\text{rejected}}$
1736
+
1737
+ This optimizes the classical constrianed RLHF objective, where the reference model constrains deviation from the initial distribution.
1738
+
1739
+ <Callout type="info">
1740
+ **DPO vs RLHF**: DPO eliminates the need for a separate reward model by directly optimizing the policy to prefer chosen responses. This makes training simpler and computationally cheaper than classical RLHF.
1741
+ </Callout>
1742
+
1743
+
1744
+ ## Running DPO Training
1745
+
1746
+ The implementation is in <CookbookLink path="tinker_cookbook/preference/train_dpo.py">train_dpo.py</CookbookLink> with a CLI interface in <CookbookLink path="tinker_cookbook/recipes/preference/dpo/train.py">train.py</CookbookLink>. You can run it from the command line:
1747
+
1748
+ ```bash
1749
+ python -m tinker_cookbook.recipes.preference.train \
1750
+ log_path=/tmp/dpo-hhh-experiment \
1751
+ model_name=meta-llama/Llama-3.2-1B \
1752
+ dataset=hhh \
1753
+ renderer_name=role_colon \
1754
+ learning_rate=1e-5 \
1755
+ dpo_beta=0.1
1756
+ ```
1757
+
1758
+ ### Key Parameters
1759
+
1760
+ - `log_relpath`: Directory where results and checkpoints are saved
1761
+ - `model_name`: Base model used as initialization and for the reference policy
1762
+ - `dataset`: Dataset name (`hhh`, `helpsteer3`, `ultrafeedback`)
1763
+ - `renderer_name`: How conversations are formatted (see [Rendering](../rendering.mdx))
1764
+ - `learning_rate`: Learning rate for optimization
1765
+ - `dpo_beta`: DPO beta parameter (controls the strength of preference learning)
1766
+
1767
+ ### Available Datasets
1768
+
1769
+ There are several pre-defined datasets:
1770
+
1771
+ - **`hhh`**: Anthropic's Helpful-Harmless-Honest dataset
1772
+ - **`helpsteer3`**: NVIDIA's HelpSteer3 preference dataset
1773
+ - **`ultrafeedback`**: UltraFeedback binarized preferences dataset
1774
+
1775
+ These are implemented as `DPODatasetBuilder` classes and you can implement a custom dataset builder following the `tinker_cookbook.preference.preference_datasets` interface.
1776
+
1777
+ ## Training Process
1778
+
1779
+ During training, you'll see output like this showing the DPO metrics:
1780
+
1781
+ ```
1782
+ Step 50
1783
+ ┏━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━┓
1784
+ ┃ Metric ┃ Value ┃
1785
+ ┡━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━┩
1786
+ │ accuracy │ 0.568627 │
1787
+ │ batch_time │ 27.953704 │
1788
+ │ chosen_reward │ 0.053621 │
1789
+ │ dpo_loss │ 0.683825 │
1790
+ │ learning_rate │ 0.000009 │
1791
+ │ margin │ 0.002147 │
1792
+ │ num_pairs │ 255 │
1793
+ │ num_tokens │ 112638 │
1794
+ │ progress │ 0.081210 │
1795
+ │ rejected_reward │ 0.032152 │
1796
+ │ test/nll │ 1.871778 │
1797
+ └────────────────────────────────┴───────────┘
1798
+ ```
1799
+
1800
+ The key metrics are:
1801
+ - **`dpo_loss`**: The DPO classification loss
1802
+ - **`accuracy`**: Accuracy of the implicit reward model evaluated on the preference dataset
1803
+ - **`margin`**: Average difference between chosen and rejected rewards
1804
+ - **`chosen_reward`/`rejected_reward`**: Average rewards for chosen/rejected responses
1805
+
1806
+ ## Evaluating DPO Models
1807
+
1808
+ After training, you can evaluate your DPO model using the inspect evaluation framework:
1809
+
1810
+ ```bash
1811
+ MODEL_PATH=tinker://YOUR_MODEL_PATH_HERE
1812
+ python -m tinker_cookbook.eval.run_inspect_evals \
1813
+ model_path=$MODEL_PATH \
1814
+ model_name=meta-llama/Llama-3.2-1B \
1815
+ tasks=inspect_evals/ifeval \
1816
+ renderer_name=role_colon
1817
+ ```
1818
+
1819
+ This will evaluate the model on various benchmarks to measure the impact of preference optimization.
1820
+
1821
+ ## Tips for DPO Training
1822
+
1823
+ 1. **Beta Parameter**: Start with `dpo_beta=0.1` and adjust based on your dataset.
1824
+
1825
+ 2. **Learning Rate**: Use a lower learning rate than supervised fine-tuning (typically 1e-5 to 1e-6).
1826
+
1827
+ 3. **Base Model**: The base model should already be in-distribution with the preference data. Either start with a ligh SFT phase or collect on-policy preferences. While training would still work. sharp distribution mis-match will create strange model behaviors.
1828
+
1829
+
1830
+ ---
1831
+
1832
+ ## File: preferences/rlhf-example.mdx
1833
+
1834
+ import { CookbookLink } from '../../components/CookbookLink'
1835
+
1836
+ # Reinforcement Learning from Human Feedback
1837
+
1838
+ We've provided a script that shows how to run a standard pipeline for reinforcement learning from human feedback (RLHF) in <CookbookLink path="tinker_cookbook/recipes/preference/rlhf/rlhf_pipeline.py">rlhf_pipeline.py</CookbookLink>.
1839
+
1840
+ ```bash
1841
+ python -m recipes.preference.rlhf.rlhf_pipeline
1842
+ ```
1843
+
1844
+ ## Training the initial policy via supervised learning
1845
+
1846
+ First, we train the policy on the [no_robots dataset](https://huggingface.co/datasets/HuggingFaceH4/no_robots) from Huggingface, which is a basic instruction following dataset with human-written answers, which was designed to match the methodology from [InstructGPT](https://arxiv.org/abs/2203.02155).
1847
+
1848
+
1849
+ ## Training the preference model via supervised learning
1850
+
1851
+ We train the preference model on the [HHH dataset](https://huggingface.co/datasets/Anthropic/hh-rlhf) from Anthropic, which is a dataset of pairwise comparisons of completions. We train a model that sees a pair of completions, A and B, and outputs which one is preferred.
1852
+
1853
+ ## Training the policy via reinforcement learning
1854
+
1855
+ Taking the initial policy, and the preference model we just trained, we can now train the policy via reinforcement learning. This RL is a form of self-play, where we use the preference model to grade match-ups between the policy and itself. In particular, for each prompt, we sample multiple completions, and use the preference model to grade all pairs of completions. We then give the policy a reward based on the win fraction.
1856
+
1857
+
1858
+ ---
1859
+
1860
+ ## File: rl/rl-basic.mdx
1861
+
1862
+ import { CookbookLink } from '../../components/CookbookLink'
1863
+
1864
+ # Your First RL Run
1865
+
1866
+ We've provided a minimal script that runs RL on the [GSM8K dataset](https://huggingface.co/datasets/openai/gsm8k): <CookbookLink path="tinker_cookbook/recipes/rl_basic.py">rl_basic.py</CookbookLink>. You can run the minimal RL script from the command line as follows:
1867
+
1868
+ ```bash
1869
+ python -m tinker_cookbook.recipes.rl_basic
1870
+ ```
1871
+
1872
+ This script will fine-tune the Llama-3.1-8B base (pretrained) model on this dataset with the following reward function:
1873
+
1874
+ $$
1875
+ 1[\text{answer is correct}] + 0.1 \times (1[\text{answer is formatted correctly}] - 1)
1876
+ $$
1877
+
1878
+ The training should take about 1 minute per iteration and climb to about 63% accuracy after 15 iterations (`env/all/correct`). You can look at the printouts for some other metrics of interest:
1879
+
1880
+ - `ac_tokens_per_turn`: the number of each tokens in each generated completion
1881
+ - `env/all/format`: the fraction of completions that are formatted correctly
1882
+ - `env/all/reward/total`: mean total reward (combining format and correctness as defined above)
1883
+ - `entropy`: per-token entropy (mean negative log-probability of sampled tokens)
1884
+ - `kl_sample_train_{v1,v2}`: two different approximations/estimators of KL divergence between the sampler's and learner's probability distribution (contributed to by numerical differences and rounding noise)
1885
+ - `progress/done_frac`: what fraction of the total number of iterations we've completed so far
1886
+ - `time/...`: time for different parts of the training loop
1887
+
1888
+ You can also look at the `log_path` directory for more detailed metrics. There are several files of interest, which are mostly the same as in the [Supervised Learning](/supervised-learning/sl-basic) case.
1889
+
1890
+
1891
+ ---
1892
+
1893
+ ## File: rl/sequence-extension.mdx
1894
+
1895
+ import { CookbookLink } from '../../components/CookbookLink'
1896
+
1897
+ # Sequence Extension Property in Multi-Turn RL
1898
+
1899
+ When running reinforcement learning with multi-turn conversations, the way you render observations at each timestep has important implications for compute efficiency. This document explains the **extension property** and how it affects training and sampling.
1900
+
1901
+ ## What is the Extension Property?
1902
+
1903
+ A sequence of observations has the **extension property** if each successive observation contains all previous observations and actions as a prefix. In other words, the context grows monotonically by appending new tokens to the end.
1904
+
1905
+ When this property holds, multiple timesteps can be merged into a single training datum, the KV-cache can be reused during sampling, and compute scales as O(T) rather than O(T^2) for a trajectory of length T.
1906
+
1907
+ ## Example 1: Qwen3 with Thinking Visible (Extension Holds)
1908
+
1909
+ When using `Qwen3Renderer` with `strip_thinking_from_history=False`, the full conversation history (including `<think>` blocks) is preserved at each timestep. Consider a two-turn math conversation:
1910
+
1911
+ **Timestep 1:**
1912
+ <div className="example">
1913
+ <span className="prompt">User: What is 2+2?<br/><br/>Assistant: </span><span className="completion">&lt;think&gt;Let me calculate...&lt;/think&gt; 4<br/><br/>User:</span>
1914
+ </div>
1915
+
1916
+ **Timestep 2:**
1917
+ <div className="example">
1918
+ <span className="prompt">User: What is 2+2?<br/><br/>Assistant: &lt;think&gt;Let me calculate...&lt;/think&gt; 4<br/><br/>User: What is 3+3?<br/><br/>Assistant: </span><span className="completion">&lt;think&gt;Let me calculate...&lt;/think&gt; 6<br/><br/>User:</span>
1919
+ </div>
1920
+
1921
+ Notice that the observation (green) at timestep 2 contains the entire timestep 1 sequence as a prefix. The new observation just appends `What is 3+3?\n\nAssistant: ` to the end. This is the **extension property**.
1922
+
1923
+ Because extension holds, the RL code can merge both timesteps into a **single Datum**:
1924
+
1925
+ <div className="example">
1926
+ <span className="prompt">User: What is 2+2?<br/><br/>Assistant: </span><span className="completion">&lt;think&gt;Let me calculate...&lt;/think&gt; 4<br/><br/>User:</span><span className="prompt"> What is 3+3?<br/><br/>Assistant: </span><span className="completion">&lt;think&gt;Let me calculate...&lt;/think&gt; 6<br/><br/>User:</span>
1927
+ </div>
1928
+
1929
+ Green = observation tokens (loss weight = 0). Red = action tokens (loss weight > 0).
1930
+
1931
+ ## Example 2: Qwen3 with Thinking Hidden (Extension Breaks)
1932
+
1933
+ When using `Qwen3Renderer` with the default `strip_thinking_from_history=True`, the `<think>...</think>` blocks are stripped from previous assistant messages. This matches how Qwen3 models were post-trained by the Qwen team.
1934
+
1935
+ **Timestep 1:**
1936
+ <div className="example">
1937
+ <span className="prompt">User: What is 2+2?<br/><br/>Assistant: </span><span className="completion">&lt;think&gt;Let me calculate...&lt;/think&gt; 4<br/><br/>User:</span>
1938
+ </div>
1939
+
1940
+ **Timestep 2:**
1941
+ <div className="example">
1942
+ <span className="prompt">User: What is 2+2?<br/><br/>Assistant: 4<br/><br/>User: What is 3+3?<br/><br/>Assistant: </span><span className="completion">&lt;think&gt;Let me calculate...&lt;/think&gt; 6<br/><br/>User:</span>
1943
+ </div>
1944
+
1945
+ The observation at timestep 2 is **not** an extension of timestep 1's full sequence. The `<think>Let me calculate...</think>` portion was stripped, so the prefix doesn't match. The RL code must create **two separate Datums**:
1946
+
1947
+ **Datum 1:**
1948
+ <div className="example">
1949
+ <span className="prompt">User: What is 2+2?<br/><br/>Assistant: </span><span className="completion">&lt;think&gt;Let me calculate...&lt;/think&gt; 4<br/><br/>User:</span>
1950
+ </div>
1951
+
1952
+ **Datum 2:**
1953
+ <div className="example">
1954
+ <span className="prompt">User: What is 2+2?<br/><br/>Assistant: 4<br/><br/>User: What is 3+3?<br/><br/>Assistant: </span><span className="completion">&lt;think&gt;Let me calculate...&lt;/think&gt; 6<br/><br/>User:</span>
1955
+ </div>
1956
+
1957
+ This results in more compute during training (two forward/backward passes instead of one) and prevents KV-cache reuse during sampling. For a trajectory of T timesteps, compute scales as O(T²) instead of O(T).
1958
+
1959
+ ## The Tradeoff
1960
+
1961
+ **Keeping thinking visible** (`strip_thinking_from_history=False`) gives you O(T) compute scaling, allows packing sequences together in training batches, and enables KV-cache reuse during sampling. The downside is that context grows faster since all thinking tokens are retained, so you may hit context length limits sooner.
1962
+
1963
+ **Stripping thinking** (`strip_thinking_from_history=True`, the default) keeps context smaller but breaks the extension property, leading to O(T²) compute scaling.
1964
+
1965
+ Note that while stripping thinking matches Qwen3's original post-training distribution, with RL fine-tuning the model should quickly adapt to the new situation where thinking is preserved. So "distribution match" might not be a major concern in practice.
1966
+
1967
+ ## How the RL Code Handles This
1968
+
1969
+ The RL training code in <CookbookLink path="tinker_cookbook/rl/data_processing.py">`data_processing.py`</CookbookLink> automatically detects whether consecutive timesteps satisfy the extension property. The key function is `trajectory_to_data`:
1970
+
1971
+ ```python
1972
+ def trajectory_to_data(traj: Trajectory, traj_advantage: float) -> list[tinker.Datum]:
1973
+ """
1974
+ Return one or more Datum objects corresponding to the trajectory.
1975
+ If the sequence grows by appending, i.e., each successive observation contains
1976
+ the previous observation+action as a prefix, then we can return a single Datum.
1977
+ However, if we get a sequence that's not an extension of the previous sequence,
1978
+ then that results in a new Datum.
1979
+ """
1980
+ ```
1981
+
1982
+ When rendering your conversations, be aware of whether your renderer has the extension property. For `Qwen3Renderer`:
1983
+ - `strip_thinking_from_history=False` → Extension holds
1984
+ - `strip_thinking_from_history=True` (default) → Extension breaks
1985
+
1986
+ **Note on sampling:** The training code automatically merges timesteps when possible. Sampling infrastructure doesn't yet adjust billing based on KV-cache hits, but this is planned for a future release.
1987
+
1988
+ ## Advanced: Periodic Compaction
1989
+
1990
+ A hybrid approach is to use **periodic compaction**: keep thinking visible most of the time (preserving extension), but periodically clear old thinking blocks from the context.
1991
+
1992
+ **How it works:**
1993
+ - For turns 1-10, keep all thinking visible (extension holds, single datum)
1994
+ - At turn 11, strip thinking from turns 1-10 (extension breaks once, new datum starts)
1995
+ - For turns 11-20, keep thinking visible again (extension holds)
1996
+ - Repeat every N turns
1997
+
1998
+ Here's what the datums look like with compaction every 3 turns:
1999
+
2000
+ **Datum 1 (turns 1-3):**
2001
+ <div className="example">
2002
+ <span className="prompt">User: Q1<br/>Assistant: </span><span className="completion">&lt;think&gt;...&lt;/think&gt; A1<br/>User:</span><span className="prompt"> Q2<br/>Assistant: </span><span className="completion">&lt;think&gt;...&lt;/think&gt; A2<br/>User:</span><span className="prompt"> Q3<br/>Assistant: </span><span className="completion">&lt;think&gt;...&lt;/think&gt; A3<br/>User:</span>
2003
+ </div>
2004
+
2005
+ **Datum 2 (turns 4-6, thinking from turns 1-3 stripped):**
2006
+ <div className="example">
2007
+ <span className="prompt">User: Q1<br/>Assistant: A1<br/>User: Q2<br/>Assistant: A2<br/>User: Q3<br/>Assistant: A3<br/>User: Q4<br/>Assistant: </span><span className="completion">&lt;think&gt;...&lt;/think&gt; A4<br/>User:</span><span className="prompt"> Q5<br/>Assistant: </span><span className="completion">&lt;think&gt;...&lt;/think&gt; A5<br/>User:</span><span className="prompt"> Q6<br/>Assistant: </span><span className="completion">&lt;think&gt;...&lt;/think&gt; A6<br/>User:</span>
2008
+ </div>
2009
+
2010
+ This approach breaks extension only every N timesteps instead of every timestep, keeps context size bounded (old thinking doesn't accumulate forever), and amortizes the recomputation cost over N turns.
2011
+
2012
+ To implement this, you would modify your environment or renderer to periodically transform the conversation history, stripping `<think>` blocks from messages older than N turns.
2013
+
2014
+ ## Summary
2015
+
2016
+ For `Qwen3Renderer`:
2017
+ - `strip_thinking_from_history=False` → Extension holds → Use for long trajectories where compute efficiency matters
2018
+ - `strip_thinking_from_history=True` (default) → Extension breaks → Use for short trajectories, or when you want minimal changes from base model behavior
2019
+ - Periodic compaction → Best of both worlds when you need efficiency with bounded context
2020
+
2021
+ When designing your RL environment, consider how many turns you expect and whether the O(T) vs O(T²) difference will be significant for your use case.
2022
+
2023
+
2024
+ ---
2025
+
2026
+ ## File: rl/rl-hyperparams.mdx
2027
+
2028
+ # RL Hyperparameters
2029
+
2030
+ This guide covers the key hyperparameters for reinforcement learning training, from core settings to advanced configurations.
2031
+
2032
+ ## Core Hyperparameters
2033
+
2034
+ ### Learning Rate
2035
+
2036
+ Similar to the [supervised learning setting](../supervised-learning/sl-hyperparams), the learning rate is the most critical hyperparameter choice. We recommend using the guidance presented there as a starting point for RL experiments as well.
2037
+
2038
+
2039
+ ### Batch and Group Sizes
2040
+
2041
+ As described in our [RL environments](../rl/rl-envs.mdx) documentation, we use two key parameters:
2042
+
2043
+ - **`batch_size`**: The number of unique environments or problems used for training
2044
+ - **`group_size`**: The number of rollouts performed per unique environment
2045
+
2046
+ If you have limited environments or problems available for training, increase the `group_size` to generate more training data. While the total number of rollouts depends on both parameters, we recommend scaling learning rates proportionally to $\text{LR} \propto \sqrt{\text{batch\_size}}$.
2047
+
2048
+ ## Multiple Updates per Sampling Iteration
2049
+
2050
+ The `num_substeps` parameter controls how many policy weight updates are performed on data sampled from the last policy iteration, similar to PPO and GRPO.
2051
+
2052
+ ### How it works:
2053
+
2054
+ - **`num_substeps = 1` (default)**: Each batch of collected trajectories is used for exactly one optimizer update
2055
+ - **`num_substeps > 1`**: The batch of unique environments is split into `num_substeps` mini-batches, where each environment/problem has `group_size` rollouts (we pack all rollouts for a particular environment/problem in the same minibatch). We do a single update step on each mini-batch. Note that our implementation still takes only a single epoch through the data.
2056
+
2057
+ ### Usage Guidelines:
2058
+
2059
+ - The batch size must be divisible by `num_substeps`
2060
+ - Our experiments show that `num_substeps = 1` already gives decent performance, but if you would like to experiment with this parameter, we recommend starting with a low value of 2-4 and using the PPO objective.
2061
+ - Higher values can lead to update steps that are too out-of-distribution for the policy. Consider limiting the number of updates or decreasing the learning rate when using multiple update steps.
2062
+
2063
+ ## Advanced Training Configurations
2064
+
2065
+ ⚠️ **Note**: These features are experimental and may be subject to instabilities. They are currently disabled by default.
2066
+
2067
+ ### Streaming Minibatch Training
2068
+
2069
+ Enable streaming minibatch training by specifying the `StreamMinibatchConfig`. This approach overlaps trajectory sampling and model training, improving overall throughput by submitting training requests as soon as enough rollouts complete, without waiting for all sampling jobs to finish.
2070
+
2071
+ **Configuration Parameters:**
2072
+
2073
+ - **`groups_per_batch`**: Same as batch size
2074
+ - **`num_minibatches`**: Number of minibatches per substep—controls how many individual forward-backward requests we submit. This controls how the work is split.
2075
+
2076
+
2077
+ **Important**: This remains on-policy training and is strictly a pipeline efficiency improvement.
2078
+
2079
+ ### Async Off-Policy Training
2080
+
2081
+ Async training allows the model to train on trajectories generated with slightly older model versions, enabling higher throughput at the cost of some off-policy bias. While Tinker doesn't currently support in-flight weight changes, it supports the "off-by-K" async RL approach where multiple model iterations generate data simultaneously. Configure this by setting the `AsyncConfig` object.
2082
+
2083
+ **Configuration Parameters:**
2084
+
2085
+ - **`max_steps_off_policy`**: Maximum age (in training steps) of trajectories before they're discarded. Essentially, trajectories from policy iterations older than `max_steps_off_policy` steps will not be used.
2086
+ - **`groups_per_batch`**: Number of new trajectory groups to accumulate (with a `group_size` number of rollouts each) before updating the current iteration of the model. Note: This is separate from the batch size used for dataset construction.
2087
+
2088
+ **Usage Guidelines:**
2089
+
2090
+ - Async RL is appropriate for applications with long and heterogeneous rollouts, such as very long CoT models, multi-hop tool use, or agentic workflows
2091
+ - Start with a small value for `max_steps_off_policy` (less than 5)
2092
+
2093
+
2094
+
2095
+ ## Monitoring and Run Health
2096
+
2097
+ Using policy-gradient algorithms with off-policy data can significantly degrade performance or even crash the policy, making monitoring essential during training.
2098
+
2099
+ ### KL Divergence Monitoring
2100
+
2101
+ The current implementation logs the KL divergence between the data generation policy and the current learner: $\mathbb{D}_{KL}[\pi_{\text{sampler}}(\cdot|x)||\pi_{\theta}(\cdot|x)]$ using two separate estimators ([Schulman 2020](http://joschu.net/blog/kl-approx.html)):
2102
+
2103
+ - `kl_sample_train_v1`
2104
+ - `kl_sample_train_v2`
2105
+
2106
+
2107
+ A few important notes to keep in mind:
2108
+ - Even with full on-policy training, the divergence between sampling and learning policies will not be exactly zero ([He 2025](https://thinkingmachines.ai/blog/defeating-nondeterminism-in-llm-inference/)) due to implementation details
2109
+ - In our experience training is stable with KL divergence below 0.01
2110
+ - If KL divergence crosses a recommended threshold, this indicates a numerical instability or potential issue with the training run
2111
+
2112
+
2113
+ ---
2114
+
2115
+ ## File: rl/rl-loops.mdx
2116
+
2117
+ import { CookbookLink } from '../../components/CookbookLink'
2118
+
2119
+ # Reinforcement Learning Training Loop
2120
+
2121
+ We've provided a simple RL training loop in <CookbookLink path="tinker_cookbook/recipes/rl_loop.py">rl_loop.py</CookbookLink>, which avoids using our environment classes and instead defines the data loading and rollouts in a more self-contained way. This is for people who like to write their own training loops or learn about how things work under the hood. Our more performant implementation in <CookbookLink path="tinker_cookbook/rl/train.py">rl/train.py</CookbookLink> does basically the same thing, but with some performance optimizations, and with some additional features like periodic evals.
2122
+
2123
+ You can run the RL training loop using:
2124
+ ```
2125
+ python -m tinker_cookbook.recipes.rl_loop
2126
+ ```
2127
+
2128
+ The default config should write the results to `/tmp/tinker-examples/rl-loop`. The experiment should be completed after 57 steps of training. You can plot the reward curve as follows:
2129
+ ```python
2130
+ import pandas
2131
+ import matplotlib.pyplot as plt
2132
+
2133
+ metrics_path = "/tmp/tinker-examples/rl-loop/metrics.jsonl"
2134
+ df = pandas.read_json(metrics_path, lines=True)
2135
+ plt.plot(df["reward/total"], label="reward/total")
2136
+ plt.legend()
2137
+ plt.show()
2138
+ ```
2139
+
2140
+ You should see a plot like this:
2141
+ ![Reward as a function of steps](./images/rl_loop_reward.png)
2142
+
2143
+
2144
+ ---
2145
+
2146
+ ## File: rl/rl-envs.mdx
2147
+
2148
+ import { CookbookLink } from '../../components/CookbookLink'
2149
+
2150
+ # RL Environments
2151
+
2152
+ Here, we'll explain how to create your own RL environments and train on them. First, lets look at the basic classes, which can be found in <CookbookLink path="tinker_cookbook/rl/types.py">`tinker_cookbook.rl.types`</CookbookLink>. As you can see, there's an `Env` interface, corresponding to an RL environment. To write an environment, you need to implement two methods: `initial_observation` and `step`.
2153
+
2154
+ ```python
2155
+ class Env:
2156
+ """
2157
+ Stateful environment that a single agent interacts with.
2158
+ Discard after running for one episode.
2159
+ """
2160
+
2161
+ async def initial_observation(self) -> tuple[Observation, StopCondition]:
2162
+ raise NotImplementedError
2163
+
2164
+ async def step(self, action: Action) -> StepResult:
2165
+ raise NotImplementedError
2166
+ ```
2167
+
2168
+ Note that this `Env` operates on *tokens*, rather than strings or messages. Why define it this way, when it's usually more natural to define the logic in terms of strings or messages? We've defined `Env` this way because this interface is what's needed by the *training* code, which needs to know the exact tokens that were sampled, and their logprobs.
2169
+
2170
+ We need to write two more small classes to use this environment in the RL training code. First, since the environment is discarded after a single episode, we need to be able to instantiate new environments in the training loop. We actually build a *group* of environments at a time, which enables multi-agent training or objectives that compare multiple samples (for example, a reward model that acts on a pair of samples).
2171
+
2172
+ ```python
2173
+ class EnvGroupBuilder:
2174
+ """
2175
+ Builds a group of environments.
2176
+ """
2177
+
2178
+ async def make_envs(self) -> Sequence[Env]:
2179
+ raise NotImplementedError
2180
+ ```
2181
+
2182
+ This object creates a group of environments. Often it does the trivial thing of returning a list of copies of the same environment.
2183
+
2184
+ Finally, we need a dataset of these EnvGroupBuilders.
2185
+
2186
+ ```python
2187
+ class RLDataset:
2188
+ """
2189
+ Dataset of EnvGroupBuilders.
2190
+ """
2191
+
2192
+ def get_batch(self, index: int) -> list[EnvGroupBuilder]:
2193
+ raise NotImplementedError
2194
+ ```
2195
+
2196
+
2197
+ That's a lot of classes! But their combination gives us a lot of flexibility. In previous implementations (like OpenAI Gym), the dataset is implicitly part of the environment; this structure is more modular and gives us more control over the data loading.
2198
+
2199
+ ## Building a simple example
2200
+
2201
+ You can find an example of writing a new RL environment in the <CookbookLink path="tinker_cookbook/recipes/multiplayer_rl/twenty_questions">Twenty Questions</CookbookLink> directory.
2202
+ Here, we define a multi-step environment, where we're training a question-asking agent, which asks questions to another agent to guess a hidden word.
2203
+ In this case, the answerer model is fixed and is Llama-3.1-8B-Instruct.
2204
+ The player model (which we fine-tune) is also based on that same model.
2205
+
2206
+ You can run the training script as follows:
2207
+
2208
+ ```bash
2209
+ python -m tinker_cookbook.recipes.twenty_questions.train
2210
+ ```
2211
+
2212
+
2213
+ ---
2214
+
2215
+ ## File: supervised-learning/sl-hyperparams.mdx
2216
+
2217
+ # Supervised Learning Hyperparameters
2218
+
2219
+ Successful LLM fine-tuning requires careful hyperparameter tuning. While the most accurate approach is to sweep over ranges and selecting values that minimize loss or maximize eval performance for each hyperparameter, this is often time-consuming and expensive. This guide provides some starting recommendations for the most important hyperparameters.
2220
+
2221
+
2222
+ ## Learning rate
2223
+
2224
+ The most important hyperparameter is generally the learning rate (LR). Our current best estimate of optimal LR for a model $m$ is the following:
2225
+
2226
+ $$ LR(m) = lr_{base} · M_{LoRA} · \Big(\frac{2000}{H_m}\Big)^{P_m} $$
2227
+
2228
+ where $lr_{base}$ is a constant base LR, $M_{LoRA}$ is a multiplier applied when using LoRA (1 if using full-finetuning), $H_m$ is the hidden size of the model $m$, and $P_m$ is a model-specific exponent adjustment. Importantly, this function is independent of the LoRA rank.
2229
+
2230
+ Our current best estimates are the following: $lr_{base} = 5e-5$,
2231
+ $M_{LoRA} = 10$, $P_m = 0.0775$ for Qwen models and $P_m = 0.781$ for Llama models.
2232
+
2233
+ ### Getting the recommended learning rate
2234
+ You can use the following function to get the recommended LR for any Llama or Qwen model:
2235
+ ```
2236
+ from tinker_cookbook.hyperparam_utils import get_lr
2237
+ model_name = "meta-llama/Llama-3.2-1B"
2238
+ recommended_lr = get_lr(model_name)
2239
+ print(f"Recommended LR: {recommended_lr}")
2240
+ ```
2241
+ ### Validation
2242
+ We validated this formula across diverse supervised fine-tuning experiments, varying datasets, dataset sizes, batch_sizes and lora_ranks.
2243
+
2244
+ Using our LR estimates resulted in \<0.5% regret compared to exhaustive hyperparameter sweeps, where regret is defined as:
2245
+
2246
+ We can define the regret of using any lr as the following:
2247
+ $$regret(lr') = \frac{loss(lr') - min_{lr} loss(lr)}{min_{lr} loss(lr)}$$
2248
+
2249
+
2250
+ ## Batch size
2251
+
2252
+ Batch size is the second-most important hyperparameter; it significantly affects both training efficiency and final performance.
2253
+
2254
+ For small batch sizes, there's a phenomenon of *perfect scaling*, where the LR and batchsize should be varied together as $LR \propto \sqrt{B}$, and the learning curve only depends on $\frac{LR}{\sqrt{B}}$. See [Shallue et al. (2018)](https://arxiv.org/abs/1811.03600) for an example in the training-from-scratch setting.
2255
+
2256
+ When fine-tuning LLMs, we're often in a regime where smaller batch sizes give better performance, at the cost of longer training time; moreover, the $LR \propto \sqrt{B}$ scaling doesn't always hold. When doing SL fine-tuning, we recommend using smaller batch sizes like 128, depending on your tolerance for longer training time.
2257
+
2258
+ For best results, you should aim for at least 100 steps of training (but usually get best results with 1000 or more).
2259
+
2260
+ ⚠️ Note: Our batch size recommendations are based on preliminary findings and ongoing research. We're not confident about them!
2261
+
2262
+
2263
+ ---
2264
+
2265
+ ## File: supervised-learning/sl-basic.mdx
2266
+
2267
+ import { CookbookLink } from '../../components/CookbookLink'
2268
+
2269
+ # Basic Supervised Learning
2270
+
2271
+ This guide walks you through running your first supervised learning experiment using Tinker's built-in training loop.
2272
+
2273
+ ## Quick start
2274
+
2275
+ We've provided an implementation of supervised learning in <CookbookLink path="tinker_cookbook/supervised/train.py">train_cli.py</CookbookLink>. To use this training loop, you'll need to create a `Config` object with the data and parameters.
2276
+
2277
+ We've provided a ready-to-run example that fine-tunes Llama-3.1-8B on a small instruction-following dataset in <CookbookLink path="tinker_cookbook/recipes/sl_basic.py">sl_basic.py</CookbookLink>. You can run it from the command line as follows:
2278
+
2279
+ ```bash
2280
+ python -m tinker_cookbook.recipes.sl_basic
2281
+ ```
2282
+
2283
+ This script fine-tunes the base (pretrained) model on a small dataset called [NoRobots](https://huggingface.co/datasets/HuggingFaceH4/no_robots), created by Hugging Face.
2284
+
2285
+ ### What you'll see during training
2286
+
2287
+ - Each step you should see a printout of the train and test loss, along with other stats like timing.
2288
+ - The training script will also print out what the data looks like, with predicted tokens (weight=1) in green and context tokens (weight=0) in yellow.
2289
+ - The training script will write various logs and checkpoint info to the `log_path` directory, which is set to `/tmp/tinker-examples/sl_basic` in the example script.
2290
+
2291
+ ### Understanding the output files
2292
+ Looking at the `log_path` directory, you will find several files of interest:
2293
+ - `metrics.jsonl`: the training metrics that also were printed to the console. You can load and plot them like this:
2294
+
2295
+ ```python
2296
+ import pandas
2297
+ import matplotlib.pyplot as plt
2298
+ df = pandas.read_json("/tmp/tinker-examples/sl_basic/metrics.jsonl", lines=True)
2299
+ plt.plot(df['train_mean_nll'], label='train_loss')
2300
+ plt.plot(df['test/nll'].dropna(), label='test_loss')
2301
+ plt.legend()
2302
+ plt.show()
2303
+ ```
2304
+ You should see a plot like this:
2305
+ ![Train and test loss as a function of steps](./images/train_test_loss.png)
2306
+
2307
+
2308
+ - `checkpoints.jsonl`: the checkpoints that were saved during training. Recall from [Saving and Loading](/save-load) that there are (currently) two kinds of checkpoints: one that has "/sampler_weights/" in the path (used for sampling), and the other that has "/weights/" in the path (includes full optimizer state, used for resuming training). If you interrupt the training script, then run it again, it will ask you if you want to resume training. If you choose to do so, it'll load the last (full state) checkpoint from this file.
2309
+ - `config.json`: the configuration that you used for training.
2310
+
2311
+ In the `sl_basic` script, you'll see that there's also some disabled code (under `if 0:`) that shows how to use your own dataset, specified as a JSONL file, provided in the format of <CookbookLink path="example-data/conversations.jsonl">conversations.jsonl</CookbookLink>.
2312
+
2313
+
2314
+ ---
2315
+
2316
+ ## File: supervised-learning/prompt-distillation.mdx
2317
+
2318
+ import { CookbookLink } from '../../components/CookbookLink'
2319
+
2320
+ # Prompt Distillation
2321
+
2322
+ Prompt distillation is a training technique in which a model is optimized to behave as though it had been provided with a long and complex prompt, without requiring access to that prompt during inference.
2323
+
2324
+ At a high level, this procedure involves two main steps:
2325
+ - **Creation of distillation data**: A teacher prompt, which is typically lengthy and highly detailed, provides explicit, step-by-step instructions. A teacher model uses this prompt to generate responses for a set of queries.
2326
+ - **Training the student model**: A student model is then trained (or fine-tuned) on the distilled dataset, thereby learning to reproduce the essential behaviors and reasoning encoded in the teacher’s instructions.
2327
+
2328
+ ---
2329
+
2330
+ ## Overview
2331
+
2332
+ Let $f_T$ and $f_S$ denote the teacher and student models, respectively. Given an instruction prompt $P$ and a query $q_i$, the teacher model generates a response $r_i$:
2333
+
2334
+ $$
2335
+ r_i = f_T([P, q_i])
2336
+ $$
2337
+
2338
+ Here, the prompt $P$ and the query $q_i$ are concatenated to form the input to the teacher model $f_T$. For a dataset of queries $Q = \{q_i \mid 1 \leq i \leq D\}$, we obtain a corresponding set of teacher responses $R = \{r_i \mid 1 \leq i \leq D\}$.
2339
+
2340
+ The distillation training dataset is defined as the set of query–response pairs (excluding the original prompt):
2341
+
2342
+ $$
2343
+ T = \{(q_i, r_i) \mid 1 \leq i \leq D\}.
2344
+ $$
2345
+
2346
+ The student model $f_S$ is then trained to minimize the cross-entropy loss:
2347
+
2348
+ $$
2349
+ \ell(f_S(q_i), r_i) = \ell(f_S(q_i), f_T([P, q_i])).
2350
+ $$
2351
+
2352
+ ---
2353
+
2354
+ ## Example
2355
+
2356
+ The Tinker Cookbook provides a prompt distillation recipe tailored for a language classification task. The objective is straightforward: given a text query, the model should predict a two-character code corresponding to the language of the input. The set of possible labels is:
2357
+ ```
2358
+ ar (Arabic), de (German), el (Greek), en (English), es (Spanish), fr (French), hi (Hindi), ru (Russian), tr (Turkish), ur (Urdu), vi (Vietnamese), zh (Chinese - Simplified), ot (Other/Unknown).
2359
+ ```
2360
+
2361
+ The recipe in <CookbookLink path="tinker_cookbook/recipes/prompt_distillation/create_data.py">recipes/prompt_distillation/create_data.py</CookbookLink> also includes handling strategies for inputs containing code, numerical content, or multiple languages.
2362
+
2363
+ In the example below, the same model (`Qwen/Qwen3-30B-A3B`) is used as both teacher and student, though in general they need not be identical.
2364
+
2365
+ ---
2366
+
2367
+ ### Step 1: Generate Training Data
2368
+
2369
+ Create prompt distillation data using the teacher model using <CookbookLink path="tinker_cookbook/recipes/prompt_distillation/create_data.py">recipes/prompt_distillation/create_data.py</CookbookLink>:
2370
+
2371
+ ```bash
2372
+ python -m tinker_cookbook.recipes.prompt_distillation.create_data \
2373
+ output_file=/tmp/tinker-datasets/prompt_distillation_lang.jsonl
2374
+ ```
2375
+
2376
+ This command will:
2377
+ - Use the configured teacher model to generate language classification examples
2378
+ - Save the distilled dataset to the specified output file
2379
+ - Create diverse training examples suitable for student model fine-tuning
2380
+
2381
+ ### Step 2: Train the Student Model
2382
+
2383
+ Fine-tune a student model on the distillation data using <CookbookLink path="tinker_cookbook/recipes/prompt_distillation/train.py">recipes/prompt_distillation/train.py</CookbookLink>:
2384
+
2385
+ ```bash
2386
+ python -m tinker_cookbook.recipes.prompt_distillation.train
2387
+ ```
2388
+
2389
+ The training script will:
2390
+ - Load the generated distillation dataset
2391
+ - Apply optimized training configurations
2392
+ - Fine-tune the student model for language classification
2393
+
2394
+ ### Step 3: Test Your Model
2395
+
2396
+ Once training is complete, you can test your distilled model by sampling from the trained model to verify its performance on language classification tasks.
2397
+
2398
+ ## Advanced Configuration
2399
+
2400
+ The prompt distillation recipe can be customized for different scenarios:
2401
+
2402
+ - **Teacher model selection**: Choose different base models based on your requirements
2403
+ - **Sampling strategies**: Adjust temperature and other generation parameters
2404
+ - **Data volume**: Scale the number of generated examples based on your needs
2405
+ - **Training hyperparameters**: Fine-tune learning rates and other training settings
2406
+
2407
+
2408
+ ---
2409
+
2410
+ ## File: supervised-learning/sweep-case-study.mdx
2411
+
2412
+ import { CookbookLink } from '../../components/CookbookLink'
2413
+
2414
+ # Sweep case study
2415
+
2416
+ In [Supervised Learning Hyperparameters](./sl-hyperparams), we introduced default hyperparameters as a starting point. While defaults are useful, optimal values are often task-specific. A hyperparameter sweep---systematically testing values across a range---is a more reliable way to identify the best settings for your use case.
2417
+
2418
+ This guide demonstrates how to sweep over the **learning rate (LR)** to find an optimal value.
2419
+
2420
+ ## Why sweep the learning rate?
2421
+
2422
+ The learning rate is typically the most impactful hyperparameter. While our default recommendations perform well (usually \<0.5% regret), you can often achieve even better results by sweeping to find the task-specific optimum.
2423
+
2424
+
2425
+ ## Setup
2426
+
2427
+ We use the simple supervised learning training loop in
2428
+ <CookbookLink path="tinker_cookbook/recipes/sl_loop.py">sl_loop.py</CookbookLink>, which trains a Llama-3.1-8B model.
2429
+
2430
+ To retrieve the model’s default learning rate recommendation:
2431
+ ```
2432
+ from tinker_cookbook.hyperparam_utils import get_lr
2433
+ print(get_lr("meta-llama/Llama-3.1-8B"))
2434
+ ```
2435
+ This should output
2436
+ ```
2437
+ 0.0002856415043086949 # ≈ 2.8e-4
2438
+ ```
2439
+ This default value provides a baseline. A common best practice is to sweep one order of magnitude above and below the default. For this case, we sweep over: $LR \in [1e-5, 3e-5, 1e-4, 3e-4, 1e-3, 3e-3]$
2440
+
2441
+
2442
+
2443
+ ## Running the sweep
2444
+ Launch experiments in parallel, using separate terminal windows for each LR value. For example:
2445
+ ```bash
2446
+ python -m tinker_cookbook.recipes.sl_loop learning_rate=0.003 log_path=/tmp/sft-lr-sweep/lr-0.003
2447
+ python -m tinker_cookbook.recipes.sl_loop learning_rate=0.001 log_path=/tmp/sft-lr-sweep/lr-0.001
2448
+ python -m tinker_cookbook.recipes.sl_loop learning_rate=0.0003 log_path=/tmp/sft-lr-sweep/lr-0.0003
2449
+ python -m tinker_cookbook.recipes.sl_loop learning_rate=0.0001 log_path=/tmp/sft-lr-sweep/lr-0.0001
2450
+ python -m tinker_cookbook.recipes.sl_loop learning_rate=0.00003 log_path=/tmp/sft-lr-sweep/lr-0.00003
2451
+ python -m tinker_cookbook.recipes.sl_loop learning_rate=0.00001 log_path=/tmp/sft-lr-sweep/lr-0.00001
2452
+ ```
2453
+ You can also automate this process by writing a script that spawns multiple tmux windows and launches experiments programmatically. This is especially useful for larger sweeps.
2454
+
2455
+
2456
+ ## Collecting Results
2457
+ After the experiments are complete, you can read the `metrics.jsonl` files:
2458
+ ```python
2459
+ from glob import glob
2460
+ import pandas
2461
+ import os
2462
+ import json
2463
+
2464
+ data = []
2465
+ for fname in sorted(glob(os.path.expanduser("/tmp/sft-lr-sweep/*/metrics.jsonl"))):
2466
+ df = pandas.read_json(fname, lines=True)
2467
+ # make sure the experiment is completed
2468
+ if len(df) == 0 or df["progress"].iloc[-1] < 0.98:
2469
+ continue
2470
+ config_fname = fname.replace("metrics.jsonl", "config.json")
2471
+ with open(config_fname, "rb") as f:
2472
+ metadata = json.load(f)
2473
+ data.append({
2474
+ "fname": fname,
2475
+ "learning_rate": metadata["learning_rate"],
2476
+ "final_loss": df["train_mean_nll"].iloc[-1].item()
2477
+ })
2478
+
2479
+ print(f"Read metrics for {len(data)} experiments")
2480
+ ```
2481
+ If all the experiments are completed, the above code should print:
2482
+ ```
2483
+ Read metrics for 6 experiments
2484
+ ```
2485
+
2486
+ ## Visualizing the Sweep
2487
+ Plot the `final_loss` as a function of `learning_rate`:
2488
+ ```python
2489
+ import matplotlib.pyplot as plt
2490
+ df = pandas.DataFrame(data)
2491
+ plt.plot(df["learning_rate"], df["final_loss"], marker='o')
2492
+ plt.axhline(y=df["final_loss"].min(), color="green", linestyle="--")
2493
+ plt.ylim(1.65, 1.8)
2494
+ plt.xscale("log")
2495
+ plt.xlabel("Learning Rate (log scale)")
2496
+ plt.ylabel("Final Loss")
2497
+ plt.title("Final Loss vs Learning Rate")
2498
+ plt.show()
2499
+ ```
2500
+ You should see a U-shaped curve, similar to this:
2501
+ ![final_loss_vs_lr](./images/lr_sweep.png)
2502
+
2503
+ If the full U-curve is not visible in your setting, expand the sweep range by adding more LR values.
2504
+
2505
+
2506
+ ## Determining the Optimal LR
2507
+ The optimal learning rate is the one that minimizes the loss. The plot above shows that the optimal LR is `3e-4` which you can also calculate by finding the minima:
2508
+ ```
2509
+ optimal_lr = df["learning_rate"][df["final_loss"].idxmin()]
2510
+ print(f"The optimal LR is {optimal_lr:.2e}")
2511
+ ```
2512
+ Expected output:
2513
+ ```
2514
+ The optimal LR is 3.00e-04
2515
+ ```
2516
+
2517
+ Note that the optimal LR in our sweep (`3e-4`) is very close to the default LR (`2.8e-4`). However, task-specific sweeps can still provide marginal improvements and greater confidence in your hyperparameter choices.
2518
+
2519
+ ## Next steps
2520
+ Now that you've identified the optimal learning rate:
2521
+ 1. Retrain with the optimal LR for your production run
2522
+ 2. Consider sweeping other hyperparameters like batch size, warmup steps, or weight decay
2523
+ 3. Use the optimal LR as a baseline for future experiments on similar tasks
2524
+
2525
+
2526
+ ---
2527
+
2528
+ ## File: supervised-learning/sl-loop.mdx
2529
+
2530
+ import { CookbookLink } from '../../components/CookbookLink'
2531
+
2532
+ # Supervised Learning Training Loop
2533
+
2534
+ We've provided a simple SL training loop in <CookbookLink path="tinker_cookbook/recipes/sl_loop.py">sl_loop.py</CookbookLink>, which avoids using our dataset classes and instead defines the data loading in a more self-contained way. This is for people who like to write their own training loops or learn about how things work under the hood. Our more performant implementation in <CookbookLink path="tinker_cookbook/supervised/train.py">supervised/train.py</CookbookLink> does basically the same thing, but with some performance optimizations, and with some additional features like periodic evals.
2535
+
2536
+
2537
+ ---
2538
+
2539
+ ## File: compatible-apis/openai.mdx
2540
+
2541
+ # OpenAI API Compatible Inference (in beta)
2542
+
2543
+ OpenAI-compatible inference lets you interact with any model checkpoint in Tinker, using an endpoint compatible with the [OpenAI Completions API](https://platform.openai.com/docs/api-reference/chat). It’s designed to let you easily “poke at” your model while you're training it.
2544
+
2545
+ For inference within your training runs (e.g. RL), we recommend using Tinker’s standard [sampling client](/training-sampling).
2546
+
2547
+ Currently, OpenAI-compatible inference is meant for testing and internal use with low internal traffic, rather than large, high-throughput, user-facing deployments. Latency and throughput may vary by model and may change without notice during the beta. If you need higher or more stable throughput, contact the Tinker team in [our Discord](https://discord.gg/KqqEZNX88c) for guidance on larger-scale setups.
2548
+
2549
+ ## Use Cases
2550
+
2551
+ OpenAI-compatible inference is designed for
2552
+ - **Fast feedback while training**: Start sampling very quickly from any sampler checkpoint obtained during training.
2553
+ - **Sampling while training continues**: Sample even while the training job is still running on that experiment.
2554
+ - **Developer &amp; internal workflows**: Intended for testing, evaluation, and internal tools.
2555
+
2556
+ We will release production-grade inference soon and will update our users then.
2557
+
2558
+ ## Using OpenAI compatible inference from an OpenAI client
2559
+
2560
+ The new interface exposes an OpenAI-compatible HTTP API. You can use any OpenAI SDK or HTTP client that lets you override the base URL.
2561
+
2562
+ 1\. Set the base URL of your OpenAI-compatible client to:
2563
+
2564
+ ```
2565
+ https://tinker.thinkingmachines.dev/services/tinker-prod/oai/api/v1
2566
+ ```
2567
+
2568
+ 2\. Use a Tinker sampler weight path as the model name. For example:
2569
+
2570
+ ```
2571
+ tinker://0034d8c9-0a88-52a9-b2b7-bce7cb1e6fef:train:0/sampler_weights/000080
2572
+ ```
2573
+
2574
+ Any valid Tinker sampler checkpoint path works here. You can keep training and sample from the same checkpoint simultaneously.
2575
+
2576
+ 3\. Authenticate with your Tinker API key, by passing the same key used for Tinker as the API key to the OpenAI client.
2577
+
2578
+ **Note:** We support both `/completions` and `/chat/completions` endpoints. Chat requests are rendered with the model’s default Hugging Face chat template; if your checkpoint expects a different renderer, render the prompt yourself (see [Rendering](/rendering)) and use `/completions`.
2579
+
2580
+ ## Code Example
2581
+
2582
+ ```py
2583
+ from os import getenv
2584
+ from openai import OpenAI
2585
+
2586
+ BASE_URL = "https://tinker.thinkingmachines.dev/services/tinker-prod/oai/api/v1"
2587
+ MODEL_PATH = "tinker://0034d8c9-0a88-52a9-b2b7-bce7cb1e6fef:train:0/sampler_weights/000080"
2588
+
2589
+ api_key = getenv("TINKER_API_KEY")
2590
+
2591
+ client = OpenAI(
2592
+ base_url=BASE_URL,
2593
+ api_key=api_key,
2594
+ )
2595
+
2596
+ response = client.completions.create(
2597
+ model=MODEL_PATH,
2598
+ prompt="The capital of France is",
2599
+ max_tokens=50,
2600
+ temperature=0.7,
2601
+ top_p=0.9,
2602
+ )
2603
+
2604
+ print(f"{response.choices[0].text}")
2605
+ ```
2606
+
2607
+ Notes:
2608
+
2609
+ * `BASE_URL` points to the OpenAI compatible inference endpoint.
2610
+ * `MODEL_PATH` is a sampler checkpoint path from Tinker (`tinker://0034d8c9-0a88-52a9-b2b7-bce7cb1e6fef:train:0/sampler_weights/000080`).
2611
+ * The rest of the arguments (`prompt`, `max_tokens`, `temperature`, `top_p`) behave like they do in the OpenAI Completions API.
2612
+ * You can swap `MODEL_PATH` to any other sampler checkpoint to compare runs quickly in your evals or notebooks.
2613
+
2614
+ ## Related docs
2615
+
2616
+ * [Getting a `TINKER_API_KEY`](/install)
2617
+
2618
+ * [Security and Privacy](https://thinkingmachines.ai/legal/terms/)
2619
+
2620
+ * [Training and Sampling](/training-sampling)
2621
+
2622
+
2623
+ ---
2624
+
2625
+ # PART 2: TYPE DEFINITIONS
2626
+
2627
+ Total types collected: 30
2628
+
2629
+ ## Type: AdamParams
2630
+
2631
+ ```python
2632
+ class AdamParams(StrictBase):
2633
+ learning_rate: float = 0.0001
2634
+ """Learning rate for the optimizer"""
2635
+
2636
+ beta1: float = 0.9
2637
+ """Coefficient used for computing running averages of gradient"""
2638
+
2639
+ beta2: float = 0.95
2640
+ """Coefficient used for computing running averages of gradient square"""
2641
+
2642
+ eps: float = 1e-12
2643
+ """Term added to the denominator to improve numerical stability"""
2644
+ ```
2645
+
2646
+ ## Type: CreateModelResponse
2647
+
2648
+ ```python
2649
+ class CreateModelResponse(BaseModel):
2650
+ model_id: ModelID
2651
+
2652
+ type: Literal["create_model"] = "create_model"
2653
+ ```
2654
+
2655
+ ## Type: Datum
2656
+
2657
+ ```python
2658
+ class Datum(StrictBase):
2659
+ loss_fn_inputs: LossFnInputs
2660
+ """Dictionary mapping field names to tensor data"""
2661
+
2662
+ model_input: ModelInput
2663
+
2664
+ @model_validator(mode="before")
2665
+ @classmethod
2666
+ def convert_tensors(cls, data: Any) -> Any:
2667
+ """Convert torch.Tensor and numpy arrays to TensorData in loss_fn_inputs during construction."""
2668
+ if isinstance(data, dict) and "loss_fn_inputs" in data:
2669
+ loss_fn_inputs = data["loss_fn_inputs"]
2670
+ if isinstance(loss_fn_inputs, dict):
2671
+ converted_inputs = {}
2672
+ for key, value in loss_fn_inputs.items():
2673
+ converted_inputs[key] = cls._maybe_convert_array(key, value)
2674
+ data = dict(data) # Make a copy
2675
+ data["loss_fn_inputs"] = converted_inputs
2676
+ return data
2677
+
2678
+ @classmethod
2679
+ def _maybe_convert_array(cls, key: str, value: Any) -> Any:
2680
+ """Convert torch.Tensor, numpy array, or 1-D list to TensorData if needed."""
2681
+ if _HAVE_TORCH and isinstance(value, torch.Tensor):
2682
+ return TensorData.from_torch(value)
2683
+ elif isinstance(value, np.ndarray):
2684
+ return TensorData.from_numpy(value)
2685
+ elif isinstance(value, list):
2686
+ # assume it's 1d and infer the dtype from the key
2687
+ return TensorData(data=value, dtype=_key_to_type[key], shape=[len(value)])
2688
+ else:
2689
+ return value
2690
+
2691
+
2692
+ _key_to_type = {
2693
+ "target_tokens": "int64",
2694
+ "weights": "float32",
2695
+ "advantages": "float32",
2696
+ "logprobs": "float32",
2697
+ "clip_low_threshold": "float32",
2698
+ "clip_high_threshold": "float32",
2699
+ }
2700
+ ```
2701
+
2702
+ ## Type: EncodedTextChunk
2703
+
2704
+ ```python
2705
+ class EncodedTextChunk(StrictBase):
2706
+ tokens: Sequence[int]
2707
+ """Array of token IDs"""
2708
+
2709
+ type: Literal["encoded_text"] = "encoded_text"
2710
+
2711
+ @property
2712
+ def length(self) -> int:
2713
+ return len(self.tokens)
2714
+ ```
2715
+
2716
+ ## Type: ForwardBackwardInput
2717
+
2718
+ ```python
2719
+ class ForwardBackwardInput(StrictBase):
2720
+ data: List[Datum]
2721
+ """Array of input data for the forward/backward pass"""
2722
+
2723
+ loss_fn: LossFnType
2724
+ """Fully qualified function path for the loss function"""
2725
+
2726
+ loss_fn_config: Optional[Dict[str, float]] = None
2727
+ """Optional configuration parameters for the loss function (e.g., PPO clip thresholds, DPO beta)"""
2728
+ ```
2729
+
2730
+ ## Type: ForwardBackwardOutput
2731
+
2732
+ ```python
2733
+ class ForwardBackwardOutput(BaseModel):
2734
+ loss_fn_output_type: str
2735
+ """The type of the ForwardBackward output. Can be one of [...] TODO"""
2736
+
2737
+ loss_fn_outputs: List[LossFnOutput]
2738
+ """Dictionary mapping field names to tensor data"""
2739
+
2740
+ metrics: Dict[str, float]
2741
+ """Training metrics as key-value pairs"""
2742
+ ```
2743
+
2744
+ ## Type: GetInfoResponse
2745
+
2746
+ ```python
2747
+ class GetInfoResponse(BaseModel):
2748
+ type: Optional[Literal["get_info"]] = None
2749
+
2750
+ model_data: ModelData
2751
+
2752
+ model_id: ModelID
2753
+
2754
+ is_lora: Optional[bool] = None
2755
+
2756
+ lora_rank: Optional[int] = None
2757
+
2758
+ model_name: Optional[str] = None
2759
+
2760
+ if PYDANTIC_V2:
2761
+ # allow fields with a `model_` prefix
2762
+ model_config = ConfigDict(protected_namespaces=tuple())
2763
+ ```
2764
+
2765
+ ## Type: GetServerCapabilitiesResponse
2766
+
2767
+ ```python
2768
+ class GetServerCapabilitiesResponse(BaseModel):
2769
+ supported_models: List[SupportedModel]
2770
+ ```
2771
+
2772
+ ## Type: ImageAssetPointerChunk
2773
+
2774
+ ```python
2775
+ class ImageAssetPointerChunk(StrictBase):
2776
+ format: Literal["png", "jpeg"]
2777
+ """Image format"""
2778
+
2779
+ location: str
2780
+ """Path or URL to the image asset"""
2781
+
2782
+ expected_tokens: int | None = None
2783
+ """Expected number of tokens this image represents.
2784
+ This is only advisory: the tinker backend will compute the number of tokens
2785
+ from the image, and we can fail requests quickly if the tokens does not
2786
+ match expected_tokens."""
2787
+
2788
+ type: Literal["image_asset_pointer"] = "image_asset_pointer"
2789
+
2790
+ @property
2791
+ def length(self) -> int:
2792
+ if self.expected_tokens is None:
2793
+ raise ValueError("ImageAssetPointerChunk expected_tokens needs to be set in order to compute the length")
2794
+ return self.expected_tokens
2795
+ ```
2796
+
2797
+ ## Type: ImageChunk
2798
+
2799
+ ```python
2800
+ class ImageChunk(StrictBase):
2801
+ data: bytes
2802
+ """Image data as bytes"""
2803
+
2804
+ format: Literal["png", "jpeg"]
2805
+ """Image format"""
2806
+
2807
+ expected_tokens: int | None = None
2808
+ """Expected number of tokens this image represents.
2809
+ This is only advisory: the tinker backend will compute the number of tokens
2810
+ from the image, and we can fail requests quickly if the tokens does not
2811
+ match expected_tokens."""
2812
+
2813
+ type: Literal["image"] = "image"
2814
+
2815
+ @field_validator("data", mode="before")
2816
+ @classmethod
2817
+ def validate_data(cls, value: Union[bytes, str]) -> bytes:
2818
+ """Deserialize base64 string to bytes if needed."""
2819
+ if isinstance(value, str):
2820
+ return base64.b64decode(value)
2821
+ return value
2822
+
2823
+ @field_serializer("data")
2824
+ def serialize_data(self, value: bytes) -> str:
2825
+ """Serialize bytes to base64 string for JSON."""
2826
+ return base64.b64encode(value).decode("utf-8")
2827
+
2828
+ @property
2829
+ def length(self) -> int:
2830
+ if self.expected_tokens is None:
2831
+ raise ValueError("ImageChunk expected_tokens needs to be set in order to compute the length")
2832
+ return self.expected_tokens
2833
+ ```
2834
+
2835
+ ## Type: LoadWeightsResponse
2836
+
2837
+ ```python
2838
+ class LoadWeightsResponse(BaseModel):
2839
+ path: Optional[str] = None
2840
+ """A tinker URI for model weights at a specific step"""
2841
+
2842
+ type: Optional[Literal["load_weights"]] = None
2843
+ ```
2844
+
2845
+ ## Type: LoraConfig
2846
+
2847
+ ```python
2848
+ class LoraConfig(StrictBase):
2849
+ rank: int
2850
+ """LoRA rank (dimension of low-rank matrices)"""
2851
+
2852
+ seed: Optional[int] = None
2853
+ """Seed used for initialization of LoRA weights.
2854
+
2855
+ Useful if you need deterministic or reproducible initialization of weights.
2856
+ """
2857
+
2858
+ train_unembed: bool = True
2859
+ """Whether to add lora to the unembedding layer"""
2860
+
2861
+ train_mlp: bool = True
2862
+ """Whether to add loras to the MLP layers (including MoE layers)"""
2863
+
2864
+ train_attn: bool = True
2865
+ """Whether to add loras to the attention layers"""
2866
+ ```
2867
+
2868
+ ## Type: LossFnInputs
2869
+
2870
+ ```python
2871
+ LossFnInputs: TypeAlias = Dict[str, TensorData]
2872
+ ```
2873
+
2874
+ ## Type: LossFnOutput
2875
+
2876
+ ```python
2877
+ LossFnOutput: TypeAlias = Dict[str, TensorData]
2878
+ ```
2879
+
2880
+ ## Type: LossFnType
2881
+
2882
+ ```python
2883
+ LossFnType: TypeAlias = Literal["cross_entropy", "importance_sampling", "ppo", "cispo", "dro"]
2884
+ ```
2885
+
2886
+ ## Type: ModelData
2887
+
2888
+ ```python
2889
+ class ModelData(BaseModel):
2890
+ arch: Optional[str] = None
2891
+
2892
+ model_name: Optional[str] = None
2893
+
2894
+ tokenizer_id: Optional[str] = None
2895
+ ```
2896
+
2897
+ ## Type: ModelID
2898
+
2899
+ ```python
2900
+ ModelID: TypeAlias = str
2901
+ ```
2902
+
2903
+ ## Type: ModelInput
2904
+
2905
+ ```python
2906
+ class ModelInput(StrictBase):
2907
+ chunks: List[ModelInputChunk]
2908
+ """Sequence of input chunks (formerly TokenSequence)"""
2909
+
2910
+
2911
+ @classmethod
2912
+ def from_ints(cls, tokens: List[int]) -> "ModelInput":
2913
+ """
2914
+ Create a ModelInput from a list of ints (tokens).
2915
+ """
2916
+ return cls(chunks=[EncodedTextChunk(tokens=tokens)])
2917
+
2918
+ def to_ints(self) -> List[int]:
2919
+ """
2920
+ Convert the ModelInput to a list of ints (tokens)
2921
+ Throws exception if there are any non-token chunks
2922
+ """
2923
+ if not all(isinstance(chunk, EncodedTextChunk) for chunk in self.chunks):
2924
+ raise ValueError(f"to_ints only supported for ModelInput with EncodedTextChunks, got {[type(chunk) for chunk in self.chunks]}")
2925
+ return [token for chunk in self.chunks for token in chunk.tokens]
2926
+
2927
+ @property
2928
+ def length(self) -> int:
2929
+ """
2930
+ Return the total context length used by this ModelInput.
2931
+ """
2932
+ return sum(chunk.length for chunk in self.chunks)
2933
+
2934
+ @classmethod
2935
+ def empty(cls) -> "ModelInput":
2936
+ """
2937
+ Create an empty ModelInput.
2938
+ """
2939
+ return cls(chunks=[])
2940
+
2941
+ def append(self, chunk: ModelInputChunk) -> "ModelInput":
2942
+ """
2943
+ Add a new chunk, return a new ModelInput.
2944
+ """
2945
+ return ModelInput(chunks=self.chunks + [chunk])
2946
+
2947
+ def append_int(self, token: int) -> "ModelInput":
2948
+ """
2949
+ Add a new token, return a new ModelInput.
2950
+ """
2951
+ return self.append(EncodedTextChunk(tokens=[token]))
2952
+ ```
2953
+
2954
+ ## Type: ModelInputChunk
2955
+
2956
+ ```python
2957
+ ModelInputChunk: TypeAlias = Annotated[
2958
+ Union[EncodedTextChunk, ImageAssetPointerChunk, ImageChunk], PropertyInfo(discriminator="type")
2959
+ ]
2960
+ ```
2961
+
2962
+ ## Type: OptimStepResponse
2963
+
2964
+ ```python
2965
+ class OptimStepResponse(BaseModel):
2966
+ metrics: Optional[Dict[str, float]] = None
2967
+ """Optimization step metrics as key-value pairs"""
2968
+ ```
2969
+
2970
+ ## Type: SampleResponse
2971
+
2972
+ ```python
2973
+ class SampleResponse(BaseModel):
2974
+ sequences: Sequence[SampledSequence]
2975
+
2976
+ type: Literal["sample"] = "sample"
2977
+
2978
+ prompt_logprobs: Optional[List[Optional[float]]] = None
2979
+ """
2980
+ If prompt_logprobs was set to true in the request, logprobs are computed for
2981
+ every token in the prompt. The `prompt_logprobs` response contains a float32
2982
+ value for every token in the prompt.
2983
+ """
2984
+
2985
+ topk_prompt_logprobs: Optional[list[Optional[list[tuple[int, float]]]]] = None
2986
+ """
2987
+ If topk_prompt_logprobs was set to a positive integer k in the request,
2988
+ the top-k logprobs are computed for every token in the prompt. The
2989
+ `topk_prompt_logprobs` response contains, for every token in the prompt,
2990
+ a list of up to k (token_id, logprob) tuples.
2991
+ """
2992
+ ```
2993
+
2994
+ ## Type: SampledSequence
2995
+
2996
+ ```python
2997
+ class SampledSequence(BaseModel):
2998
+ stop_reason: StopReason
2999
+ """Reason why sampling stopped"""
3000
+
3001
+ tokens: List[int]
3002
+ """List of generated token IDs"""
3003
+
3004
+ logprobs: Optional[List[float]] = None
3005
+ """Log probabilities for each token (optional)"""
3006
+ ```
3007
+
3008
+ ## Type: SamplingParams
3009
+
3010
+ ```python
3011
+ class SamplingParams(BaseModel):
3012
+ max_tokens: Optional[int] = None
3013
+ """Maximum number of tokens to generate"""
3014
+
3015
+ seed: Optional[int] = None
3016
+ """Random seed for reproducible generation"""
3017
+
3018
+ stop: Union[str, Sequence[str], Sequence[int], None] = None
3019
+ """Stop sequences for generation"""
3020
+
3021
+ temperature: float = 1
3022
+ """Sampling temperature"""
3023
+
3024
+ top_k: int = -1
3025
+ """Top-k sampling parameter (-1 for no limit)"""
3026
+
3027
+ top_p: float = 1
3028
+ """Nucleus sampling probability"""
3029
+ ```
3030
+
3031
+ ## Type: SaveWeightsForSamplerResponse
3032
+
3033
+ ```python
3034
+ class SaveWeightsForSamplerResponse(BaseModel):
3035
+ path: str
3036
+ """A tinker URI for model weights for sampling at a specific step"""
3037
+
3038
+ type: Optional[Literal["save_weights_for_sampler"]] = None
3039
+ ```
3040
+
3041
+ ## Type: SaveWeightsResponse
3042
+
3043
+ ```python
3044
+ class SaveWeightsResponse(BaseModel):
3045
+ path: str
3046
+ """A tinker URI for model weights at a specific step"""
3047
+
3048
+ type: Optional[Literal["save_weights"]] = None
3049
+ ```
3050
+
3051
+ ## Type: StopReason
3052
+
3053
+ ```python
3054
+ StopReason: TypeAlias = Literal["length", "stop"]
3055
+ ```
3056
+
3057
+ ## Type: SupportedModel
3058
+
3059
+ ```python
3060
+ class SupportedModel(BaseModel):
3061
+ model_name: Optional[str] = None
3062
+ ```
3063
+
3064
+ ## Type: TensorData
3065
+
3066
+ ```python
3067
+ class TensorData(StrictBase):
3068
+ data: Union[List[int], List[float]]
3069
+ """Flattened tensor data as array of numbers."""
3070
+
3071
+ dtype: TensorDtype
3072
+
3073
+ shape: Optional[List[int]] = None
3074
+ """Optional.
3075
+
3076
+ The shape of the tensor (see PyTorch tensor.shape). The shape of a
3077
+ one-dimensional list of length N is `(N,)`. Can usually be inferred if not
3078
+ provided, and is generally inferred as a 1D tensor.
3079
+ """
3080
+
3081
+ @classmethod
3082
+ def from_numpy(cls, array: npt.NDArray[Any]) -> "TensorData":
3083
+ return cls(
3084
+ data=array.flatten().tolist(),
3085
+ dtype=_convert_numpy_dtype_to_tensor(array.dtype),
3086
+ shape=list(array.shape),
3087
+ )
3088
+
3089
+ @classmethod
3090
+ def from_torch(cls, tensor: "torch.Tensor") -> "TensorData":
3091
+ return cls(
3092
+ data=tensor.flatten().tolist(),
3093
+ dtype=_convert_torch_dtype_to_tensor(tensor.dtype),
3094
+ shape=list(tensor.shape),
3095
+ )
3096
+
3097
+ def to_numpy(self) -> npt.NDArray[Any]:
3098
+ """Convert TensorData to numpy array."""
3099
+ numpy_dtype = _convert_tensor_dtype_to_numpy(self.dtype)
3100
+ arr = np.array(self.data, dtype=numpy_dtype)
3101
+ if self.shape is not None:
3102
+ arr = arr.reshape(self.shape)
3103
+ return arr
3104
+
3105
+ def to_torch(self) -> "torch.Tensor":
3106
+ """Convert TensorData to torch tensor."""
3107
+ if not _HAVE_TORCH:
3108
+ raise ImportError("PyTorch is not installed. Cannot convert to torch tensor.")
3109
+
3110
+ torch_dtype = _convert_tensor_dtype_to_torch(self.dtype)
3111
+ tensor = torch.tensor(self.data, dtype=torch_dtype)
3112
+ if self.shape is not None:
3113
+ tensor = tensor.reshape(self.shape)
3114
+ return tensor
3115
+
3116
+ def tolist(self) -> List[Any]:
3117
+ return self.to_numpy().tolist()
3118
+
3119
+
3120
+ def _convert_tensor_dtype_to_numpy(dtype: TensorDtype) -> npt.DTypeLike:
3121
+ """Convert TensorDtype to numpy dtype-like."""
3122
+ if dtype == "float32":
3123
+ return np.float32
3124
+ elif dtype == "int64":
3125
+ return np.int64
3126
+ else:
3127
+ raise ValueError(f"Unsupported TensorDtype: {dtype}")
3128
+
3129
+
3130
+ def _convert_tensor_dtype_to_torch(dtype: TensorDtype) -> "torch.dtype":
3131
+ """Convert TensorDtype to torch dtype."""
3132
+ if not _HAVE_TORCH:
3133
+ raise ImportError("PyTorch is not installed. Cannot convert to torch dtype.")
3134
+ import torch
3135
+
3136
+ if dtype == "float32":
3137
+ return torch.float32
3138
+ elif dtype == "int64":
3139
+ return torch.int64
3140
+ else:
3141
+ raise ValueError(f"Unsupported TensorDtype: {dtype}")
3142
+
3143
+
3144
+ def _convert_numpy_dtype_to_tensor(dtype: np.dtype[Any]) -> TensorDtype:
3145
+ """Convert numpy dtype to TensorDtype."""
3146
+ if dtype.kind == "f":
3147
+ return "float32"
3148
+ elif dtype.kind == "i":
3149
+ return "int64"
3150
+ else:
3151
+ raise ValueError(f"Unsupported numpy dtype: {dtype}")
3152
+
3153
+
3154
+ def _convert_torch_dtype_to_tensor(dtype: "torch.dtype") -> TensorDtype:
3155
+ """Convert torch dtype to TensorDtype."""
3156
+ # torch.dtype objects have .is_floating_point
3157
+ if getattr(dtype, "is_floating_point", False):
3158
+ return "float32"
3159
+ else:
3160
+ return "int64"
3161
+ ```
3162
+
3163
+ ## Type: TensorDtype
3164
+
3165
+ ```python
3166
+ TensorDtype: TypeAlias = Literal["int64", "float32"]
3167
+ ```
3168
+
3169
+ ## Type: UnloadModelResponse
3170
+
3171
+ ```python
3172
+ class UnloadModelResponse(BaseModel):
3173
+ model_id: ModelID
3174
+
3175
+ type: Optional[Literal["unload_model"]] = None
3176
+ ```