@synsci/cli-darwin-x64 1.1.58 → 1.1.60
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/bin/skills/grpo-rl-training/README.md +1 -1
- package/bin/skills/hugging-face-evaluation/examples/.env.example +7 -0
- package/bin/skills/hugging-face-evaluation/scripts/test_extraction.py +0 -0
- package/bin/skills/hugging-face-model-trainer/scripts/convert_to_gguf.py +0 -0
- package/bin/skills/hugging-face-model-trainer/scripts/estimate_cost.py +0 -0
- package/bin/skills/hugging-face-paper-publisher/scripts/paper_manager.py +0 -0
- package/bin/skills/hugging-face-trackio/.claude-plugin/plugin.json +19 -0
- package/bin/skills/modal/SKILL.md +316 -275
- package/bin/skills/modal/references/advanced-patterns.md +598 -0
- package/bin/skills/modal/references/examples-catalog.md +423 -0
- package/bin/skills/prime-intellect-lab/README.md +69 -0
- package/bin/skills/prime-intellect-lab/SKILL.md +598 -0
- package/bin/skills/prime-intellect-lab/templates/basic_rl_training.toml +82 -0
- package/bin/skills/tensorpool/SKILL.md +519 -0
- package/bin/synsc +0 -0
- package/package.json +1 -1
- package/bin/skills/modal/references/advanced-usage.md +0 -503
|
@@ -0,0 +1,423 @@
|
|
|
1
|
+
# Modal Examples Catalog
|
|
2
|
+
|
|
3
|
+
Complete catalog of Modal's official example library with implementation notes. Each example is production-ready code you can adapt for your workloads.
|
|
4
|
+
|
|
5
|
+
**Source**: https://modal.com/docs/examples
|
|
6
|
+
|
|
7
|
+
## LLM Inference & Serving
|
|
8
|
+
|
|
9
|
+
### `llm_inference` — Deploy an OpenAI-compatible LLM service
|
|
10
|
+
- **GPU**: H100 or A100-80GB
|
|
11
|
+
- **Key pattern**: Class-based app with `@modal.enter()` for model loading, `@modal.web_server()` for OpenAI-compatible API
|
|
12
|
+
- **Features**: vLLM backend, streaming responses, OpenAI-compatible `/v1/chat/completions` endpoint, Volume for model weights, auto-scaling
|
|
13
|
+
- **When to use**: Default choice for serving any HF model as an API
|
|
14
|
+
|
|
15
|
+
### `very_large_models` — Deploy really big language models
|
|
16
|
+
- **GPU**: H200:4-8 or B200:4-8
|
|
17
|
+
- **Key pattern**: SGLang + multi-GPU tensor parallelism for models that don't fit on a single GPU
|
|
18
|
+
- **Features**: DeepSeek V3, Kimi-K2, GLM 4.7/5 support, multi-GPU serving, auto-scaling
|
|
19
|
+
- **When to use**: Serving 100B+ parameter models (DeepSeek V3, Kimi-K2, etc.)
|
|
20
|
+
|
|
21
|
+
### `ministral3_inference` — Cut Ministral 3 cold start times by 10x with snapshots
|
|
22
|
+
- **GPU**: Any
|
|
23
|
+
- **Key pattern**: Memory snapshots to pre-load model weights into container memory
|
|
24
|
+
- **Features**: `modal.Sandbox` snapshots, near-zero cold starts for large models
|
|
25
|
+
- **When to use**: When cold start latency is critical (user-facing APIs)
|
|
26
|
+
|
|
27
|
+
### `vllm_throughput` — Optimize tokens per second with vLLM
|
|
28
|
+
- **GPU**: H100 or A100
|
|
29
|
+
- **Key pattern**: Batch processing with vLLM for maximum throughput
|
|
30
|
+
- **Features**: Offline batch inference, throughput optimization (~30K input tok/s per H100), parallel processing
|
|
31
|
+
- **When to use**: Processing large datasets (evaluation, synthetic data generation)
|
|
32
|
+
|
|
33
|
+
### `sglang_low_latency` — Low-latency inference with SGLang
|
|
34
|
+
- **GPU**: H100:2
|
|
35
|
+
- **Key pattern**: SGLang runtime with speculative decoding (EAGLE-3) for optimized inference
|
|
36
|
+
- **Features**: Speculative decoding, low latency, dual-GPU, streaming
|
|
37
|
+
- **When to use**: When you need lower latency than vLLM, latency-critical applications
|
|
38
|
+
|
|
39
|
+
### `llama_cpp` — Run GGUF models with llama.cpp
|
|
40
|
+
- **GPU**: Optional (CPU inference supported)
|
|
41
|
+
- **Key pattern**: llama.cpp server on Modal
|
|
42
|
+
- **Features**: GGUF quantized models, CPU/GPU hybrid, low resource usage
|
|
43
|
+
- **When to use**: Serving quantized models, budget-friendly inference
|
|
44
|
+
|
|
45
|
+
### `trtllm_latency` — Low-latency inference with TensorRT-LLM
|
|
46
|
+
- **GPU**: H100
|
|
47
|
+
- **Key pattern**: TensorRT-LLM compiled model serving
|
|
48
|
+
- **Features**: Optimized kernels, lowest possible latency
|
|
49
|
+
- **When to use**: Production inference where every millisecond counts
|
|
50
|
+
|
|
51
|
+
### `trtllm_throughput` — High-throughput batch inference with TensorRT-LLM
|
|
52
|
+
- **GPU**: H100
|
|
53
|
+
- **Key pattern**: TensorRT-LLM for batch processing
|
|
54
|
+
- **Features**: Maximum tokens/sec throughput
|
|
55
|
+
- **When to use**: Batch processing with TensorRT optimization
|
|
56
|
+
|
|
57
|
+
## Training & Fine-Tuning
|
|
58
|
+
|
|
59
|
+
### `grpo_verl` — Train a model to solve math problems using GRPO and verl
|
|
60
|
+
- **GPU**: H100:8 (multi-GPU)
|
|
61
|
+
- **Key pattern**: verl framework for GRPO RL training
|
|
62
|
+
- **Features**: Math reasoning reward, multi-GPU training, Volume checkpointing
|
|
63
|
+
- **When to use**: RL-based training for reasoning tasks (math, logic)
|
|
64
|
+
|
|
65
|
+
### `grpo_trl` — Train a model to solve coding problems using GRPO and TRL
|
|
66
|
+
- **GPU**: A100 or H100
|
|
67
|
+
- **Key pattern**: TRL's GRPOTrainer on Modal
|
|
68
|
+
- **Features**: Code generation reward, sandbox-based evaluation, RL training
|
|
69
|
+
- **When to use**: RL-based training for code generation
|
|
70
|
+
|
|
71
|
+
### `unsloth_finetune` — Efficient LLM fine-tuning with Unsloth
|
|
72
|
+
- **GPU**: A100 or L40S
|
|
73
|
+
- **Key pattern**: Unsloth for 2x faster LoRA training
|
|
74
|
+
- **Features**: Memory efficient, 2x speed, QLoRA support
|
|
75
|
+
- **When to use**: LoRA fine-tuning when Tinker doesn't support your model
|
|
76
|
+
|
|
77
|
+
### `hp_sweep_gpt` — Train SLM with early-stopping grid search
|
|
78
|
+
- **GPU**: A100
|
|
79
|
+
- **Key pattern**: Grid search over hyperparameters with `.map()`
|
|
80
|
+
- **Features**: Parallel HP sweep, early stopping, Volume for checkpoints
|
|
81
|
+
- **When to use**: Hyperparameter optimization, training from scratch
|
|
82
|
+
|
|
83
|
+
### `long-training` — Run long, resumable training jobs
|
|
84
|
+
- **GPU**: A100 or H100
|
|
85
|
+
- **Key pattern**: Checkpointing to Volume with resume-on-preemption
|
|
86
|
+
- **Features**: Volume checkpointing, preemption handling, long-running jobs
|
|
87
|
+
- **When to use**: Multi-hour/day training runs that need fault tolerance
|
|
88
|
+
|
|
89
|
+
### `llm-finetuning` — Full LLM fine-tuning pipeline
|
|
90
|
+
- **GPU**: A100-80GB or H100
|
|
91
|
+
- **Key pattern**: End-to-end fine-tuning with evaluation
|
|
92
|
+
- **Features**: Data loading, training loop, evaluation, model upload
|
|
93
|
+
- **When to use**: Complete fine-tuning workflow on Modal
|
|
94
|
+
|
|
95
|
+
### `flan_t5_finetune` — Fine-tune Flan-T5
|
|
96
|
+
- **GPU**: A10G or L40S
|
|
97
|
+
- **Key pattern**: Seq2seq model fine-tuning
|
|
98
|
+
- **Features**: Flan-T5 training, evaluation
|
|
99
|
+
- **When to use**: Seq2seq tasks (summarization, translation, Q&A)
|
|
100
|
+
|
|
101
|
+
### `diffusers_lora_finetune` — Custom pet art from Flux with Hugging Face and Gradio
|
|
102
|
+
- **GPU**: A100-80GB
|
|
103
|
+
- **Key pattern**: Diffusers LoRA training for image generation
|
|
104
|
+
- **Features**: Image LoRA, custom datasets, Flux models, Gradio UI
|
|
105
|
+
- **When to use**: Fine-tuning image generation models
|
|
106
|
+
|
|
107
|
+
## Multimodal & Vision
|
|
108
|
+
|
|
109
|
+
### `flux` — Serve diffusion models fast with torch.compile
|
|
110
|
+
- **GPU**: A100 or H100
|
|
111
|
+
- **Key pattern**: torch.compile for optimized image generation
|
|
112
|
+
- **Features**: Compilation cache, fast generation, various Flux variants
|
|
113
|
+
- **When to use**: Fast image generation with Flux models
|
|
114
|
+
|
|
115
|
+
### `image_to_image` — Edit images with Flux Kontext
|
|
116
|
+
- **GPU**: A100 or H100
|
|
117
|
+
- **Key pattern**: Image-to-image with Flux Kontext
|
|
118
|
+
- **Features**: Image editing, style transfer
|
|
119
|
+
- **When to use**: Image editing tasks
|
|
120
|
+
|
|
121
|
+
### `image_to_video` — Animate images with LTX-Video
|
|
122
|
+
- **GPU**: A100 or H100
|
|
123
|
+
- **Key pattern**: Image animation pipeline
|
|
124
|
+
- **Features**: Image-to-video generation
|
|
125
|
+
- **When to use**: Creating video from static images
|
|
126
|
+
|
|
127
|
+
### `ltx` — Generate video clips with LTX-Video
|
|
128
|
+
- **GPU**: A100 or H100
|
|
129
|
+
- **Key pattern**: Text-to-video generation
|
|
130
|
+
- **Features**: LTX-Video model, video clips
|
|
131
|
+
- **When to use**: Text-to-video generation
|
|
132
|
+
|
|
133
|
+
### `text_to_image` — Stable Diffusion CLI/API/UI
|
|
134
|
+
- **GPU**: A10G or L40S
|
|
135
|
+
- **Key pattern**: Stable Diffusion serving with multiple interfaces
|
|
136
|
+
- **Features**: CLI, REST API, Gradio UI
|
|
137
|
+
- **When to use**: Standard text-to-image generation
|
|
138
|
+
|
|
139
|
+
### `finetune_yolo` — Fine-tune and serve YOLO models
|
|
140
|
+
- **GPU**: T4 or A10G
|
|
141
|
+
- **Key pattern**: YOLO training + serving pipeline
|
|
142
|
+
- **Features**: Object detection, model serving
|
|
143
|
+
- **When to use**: Computer vision / object detection
|
|
144
|
+
|
|
145
|
+
### `segment_anything` — Segment Anything Model
|
|
146
|
+
- **GPU**: A10G
|
|
147
|
+
- **Key pattern**: SAM inference
|
|
148
|
+
- **Features**: Zero-shot segmentation
|
|
149
|
+
- **When to use**: Image segmentation tasks
|
|
150
|
+
|
|
151
|
+
### `comfyapp` — Run Flux on ComfyUI as an API
|
|
152
|
+
- **GPU**: A100 or H100
|
|
153
|
+
- **Key pattern**: ComfyUI workflow as API
|
|
154
|
+
- **Features**: ComfyUI, workflow automation, Flux
|
|
155
|
+
- **When to use**: Complex image generation pipelines via ComfyUI
|
|
156
|
+
|
|
157
|
+
### `blender_video` — Build a 3D render farm with Blender
|
|
158
|
+
- **GPU**: Optional (CPU rendering supported)
|
|
159
|
+
- **Key pattern**: Parallel rendering with Blender on Modal
|
|
160
|
+
- **Features**: 3D rendering, frame parallelism, render farm
|
|
161
|
+
- **When to use**: Distributed 3D rendering, animation pipelines
|
|
162
|
+
|
|
163
|
+
## Audio & Speech
|
|
164
|
+
|
|
165
|
+
### `llm-voice-chat` — Voice chat with LLMs
|
|
166
|
+
- **GPU**: A10G or L40S
|
|
167
|
+
- **Key pattern**: Real-time voice interaction with WebSocket
|
|
168
|
+
- **Features**: Moshi model, WebSocket streaming, real-time audio
|
|
169
|
+
- **When to use**: Voice chatbot / real-time audio interaction
|
|
170
|
+
|
|
171
|
+
### `streaming_kyutai_stt` — Transcribe speech with Kyutai STT
|
|
172
|
+
- **GPU**: L4 or A10G
|
|
173
|
+
- **Key pattern**: Streaming speech-to-text
|
|
174
|
+
- **Features**: Real-time transcription, Kyutai STT, low latency
|
|
175
|
+
- **When to use**: Real-time transcription (live captions, dictation)
|
|
176
|
+
|
|
177
|
+
### `music-video-gen` — Star in custom music videos
|
|
178
|
+
- **GPU**: A100 or H100
|
|
179
|
+
- **Key pattern**: Multi-model pipeline (image gen + video gen + audio)
|
|
180
|
+
- **Features**: Music video generation, multi-model orchestration
|
|
181
|
+
- **When to use**: Creative AI pipelines combining audio, image, and video generation
|
|
182
|
+
|
|
183
|
+
### `generate_music` — Make music with ACE-Step
|
|
184
|
+
- **GPU**: A10G or L40S
|
|
185
|
+
- **Key pattern**: Music generation pipeline
|
|
186
|
+
- **Features**: ACE-Step model, music generation
|
|
187
|
+
- **When to use**: AI music generation
|
|
188
|
+
|
|
189
|
+
### `chatterbox_tts` — Generate speech with Chatterbox
|
|
190
|
+
- **GPU**: A10G
|
|
191
|
+
- **Key pattern**: Text-to-speech pipeline
|
|
192
|
+
- **Features**: Chatterbox TTS, voice synthesis
|
|
193
|
+
- **When to use**: Text-to-speech generation
|
|
194
|
+
|
|
195
|
+
### `batched_whisper` — High throughput batched transcription
|
|
196
|
+
- **GPU**: A10G or L40S
|
|
197
|
+
- **Key pattern**: Batch Whisper with `@modal.batched()`
|
|
198
|
+
- **Features**: Dynamic batching, high throughput, parallel processing
|
|
199
|
+
- **When to use**: Transcribing large audio datasets
|
|
200
|
+
|
|
201
|
+
### `fine_tune_asr` — Fine-tune Whisper to recognize new words
|
|
202
|
+
- **GPU**: A10G or A100
|
|
203
|
+
- **Key pattern**: Whisper fine-tuning pipeline
|
|
204
|
+
- **Features**: Custom vocabulary, domain-specific ASR
|
|
205
|
+
- **When to use**: Adapting Whisper to domain-specific terminology
|
|
206
|
+
|
|
207
|
+
## Sandboxes & Code Execution
|
|
208
|
+
|
|
209
|
+
### `agent` — Sandbox a LangGraph agent's code
|
|
210
|
+
- **GPU**: T4 or A10G
|
|
211
|
+
- **Key pattern**: LangGraph agent with Modal Sandbox for code execution
|
|
212
|
+
- **Features**: GPU sandbox, LangGraph, tool use, file access
|
|
213
|
+
- **When to use**: Building coding agents that need to execute code
|
|
214
|
+
|
|
215
|
+
### `opencode_server` — Run a background coding agent with OpenCode
|
|
216
|
+
- **GPU**: Optional
|
|
217
|
+
- **Key pattern**: OpenCode coding agent running in a Modal Sandbox
|
|
218
|
+
- **Features**: Background coding agent, sandbox isolation, persistent sessions
|
|
219
|
+
- **When to use**: Deploying autonomous coding agents in sandboxes
|
|
220
|
+
|
|
221
|
+
### `modal-vibe` — Deploy vibe coding at scale
|
|
222
|
+
- **GPU**: Optional
|
|
223
|
+
- **Key pattern**: React frontend + LLM + Modal Sandboxes for scalable AI coding platform
|
|
224
|
+
- **Features**: Code execution at scale, React UI, LLM integration, sandbox isolation
|
|
225
|
+
- **When to use**: Building scalable AI-powered coding platforms
|
|
226
|
+
|
|
227
|
+
### `safe_code_execution` — Run Node.js, Ruby, and more in a Sandbox
|
|
228
|
+
- **GPU**: Optional
|
|
229
|
+
- **Key pattern**: Multi-language sandboxed code execution at scale
|
|
230
|
+
- **Features**: Secure execution, Node.js/Ruby/Python, scaling, multi-language
|
|
231
|
+
- **When to use**: Building coding platforms, code evaluation systems, multi-language sandboxes
|
|
232
|
+
|
|
233
|
+
### `simple_code_interpreter` — Build a stateful, sandboxed code interpreter
|
|
234
|
+
- **GPU**: Optional
|
|
235
|
+
- **Key pattern**: Stateful sandbox with session persistence
|
|
236
|
+
- **Features**: Jupyter-like sessions, state persistence, file I/O
|
|
237
|
+
- **When to use**: Interactive code interpreter (like ChatGPT Code Interpreter)
|
|
238
|
+
|
|
239
|
+
### `jupyter_sandbox` — Run a sandboxed Jupyter notebook
|
|
240
|
+
- **GPU**: Optional
|
|
241
|
+
- **Key pattern**: Jupyter notebook in Modal Sandbox
|
|
242
|
+
- **Features**: Jupyter server, file access, isolated execution
|
|
243
|
+
- **When to use**: Providing sandboxed Jupyter environments
|
|
244
|
+
|
|
245
|
+
### `anthropic_computer_use` — Control a sandboxed computer with an LLM
|
|
246
|
+
- **GPU**: Optional
|
|
247
|
+
- **Key pattern**: Virtual desktop in sandbox controlled by LLM
|
|
248
|
+
- **Features**: Computer use, screenshot, mouse/keyboard control
|
|
249
|
+
- **When to use**: Computer use / GUI automation agents
|
|
250
|
+
|
|
251
|
+
## RAG & Embeddings
|
|
252
|
+
|
|
253
|
+
### `chat_with_pdf_vision` — RAG Chat with PDFs
|
|
254
|
+
- **GPU**: A10G or L40S
|
|
255
|
+
- **Key pattern**: Vision-based PDF parsing + RAG
|
|
256
|
+
- **Features**: PDF parsing, multimodal embeddings, chat interface
|
|
257
|
+
- **When to use**: Document Q&A, PDF chatbots
|
|
258
|
+
|
|
259
|
+
### `amazon_embeddings` — Embed millions of documents with TEI
|
|
260
|
+
- **GPU**: A10G or L40S
|
|
261
|
+
- **Key pattern**: HuggingFace Text Embeddings Inference (TEI) for bulk embedding
|
|
262
|
+
- **Features**: High throughput, batch processing, `.map()` parallelism
|
|
263
|
+
- **When to use**: Large-scale embedding generation
|
|
264
|
+
|
|
265
|
+
### `mongodb-search` — Turn satellite images into vectors and store them in MongoDB
|
|
266
|
+
- **GPU**: T4 or A10G
|
|
267
|
+
- **Key pattern**: Image embedding + MongoDB Atlas vector search + GeoJSON
|
|
268
|
+
- **Features**: Satellite image embeddings, MongoDB Atlas, vector + geospatial search
|
|
269
|
+
- **When to use**: Image search with geospatial queries, visual similarity + location
|
|
270
|
+
|
|
271
|
+
### `potus_speech_qanda` — Retrieval-Augmented Generation (RAG) for Q&A
|
|
272
|
+
- **GPU**: None (CPU)
|
|
273
|
+
- **Key pattern**: Basic RAG pipeline with OpenAI
|
|
274
|
+
- **Features**: Document indexing, question answering
|
|
275
|
+
- **When to use**: Simple RAG setup with OpenAI
|
|
276
|
+
|
|
277
|
+
## Web Apps & Endpoints
|
|
278
|
+
|
|
279
|
+
### `basic_web` — Serving web endpoints
|
|
280
|
+
- **GPU**: None (CPU)
|
|
281
|
+
- **Key pattern**: FastAPI + `@modal.asgi_app()` / `@modal.fastapi_endpoint()`
|
|
282
|
+
- **Features**: REST API, ASGI/WSGI support, auto-scaling
|
|
283
|
+
- **When to use**: Learning Modal web endpoints, simple APIs
|
|
284
|
+
|
|
285
|
+
### `serve_streamlit` — Deploy Streamlit apps
|
|
286
|
+
- **GPU**: Optional
|
|
287
|
+
- **Key pattern**: `@modal.web_server()` for Streamlit
|
|
288
|
+
- **Features**: Streamlit, interactive dashboards
|
|
289
|
+
- **When to use**: Data apps, ML demos with Streamlit
|
|
290
|
+
|
|
291
|
+
### `fasthtml_app` — FastHTML applications
|
|
292
|
+
- **GPU**: None (CPU)
|
|
293
|
+
- **Key pattern**: FastHTML on Modal
|
|
294
|
+
- **Features**: FastHTML, lightweight web apps
|
|
295
|
+
- **When to use**: Simple web applications
|
|
296
|
+
|
|
297
|
+
### `mcp_server_stateless` — Deploy a stateless MCP with FastMCP
|
|
298
|
+
- **GPU**: Optional
|
|
299
|
+
- **Key pattern**: FastMCP server on Modal
|
|
300
|
+
- **Features**: MCP protocol, tool serving, stateless
|
|
301
|
+
- **When to use**: Deploying MCP tool servers
|
|
302
|
+
|
|
303
|
+
### `webrtc_yolo` — Serverless WebRTC with YOLO detection
|
|
304
|
+
- **GPU**: T4 or A10G
|
|
305
|
+
- **Key pattern**: WebRTC + YOLO on Modal
|
|
306
|
+
- **Features**: Real-time video, object detection, WebRTC
|
|
307
|
+
- **When to use**: Real-time video processing apps
|
|
308
|
+
|
|
309
|
+
### `fastrtc_flip_webcam` — WebRTC quickstart with FastRTC
|
|
310
|
+
- **GPU**: Optional
|
|
311
|
+
- **Key pattern**: FastRTC framework on Modal
|
|
312
|
+
- **Features**: WebRTC, fast setup, real-time communication
|
|
313
|
+
- **When to use**: Getting started with WebRTC on Modal
|
|
314
|
+
|
|
315
|
+
### `webscraper` — Simple web scraper
|
|
316
|
+
- **GPU**: None (CPU)
|
|
317
|
+
- **Key pattern**: Web scraping with parallel processing
|
|
318
|
+
- **Features**: Scraping, data collection, parallelism
|
|
319
|
+
- **When to use**: Web scraping and data collection tasks
|
|
320
|
+
|
|
321
|
+
## Data & Infrastructure
|
|
322
|
+
|
|
323
|
+
### `s3_bucket_mount` — Parallel processing of Parquet files on S3
|
|
324
|
+
- **GPU**: None (CPU)
|
|
325
|
+
- **Key pattern**: `modal.CloudBucketMount` for S3 access
|
|
326
|
+
- **Features**: S3 as filesystem, Parquet processing, parallel reads
|
|
327
|
+
- **When to use**: Processing data from S3, large datasets
|
|
328
|
+
|
|
329
|
+
### `cloud_bucket_mount_loras` — Create a LoRA Playground with Modal, Gradio, and S3
|
|
330
|
+
- **GPU**: A10G or L40S
|
|
331
|
+
- **Key pattern**: S3 bucket mount for LoRA weight management
|
|
332
|
+
- **Features**: LoRA weights on S3, Gradio UI, dynamic model loading
|
|
333
|
+
- **When to use**: Managing and serving multiple LoRA adapters
|
|
334
|
+
|
|
335
|
+
### `dbt_duckdb` — Build your own data warehouse with DuckDB, DBT, and Modal
|
|
336
|
+
- **GPU**: None (CPU)
|
|
337
|
+
- **Key pattern**: Data pipeline with DuckDB + DBT
|
|
338
|
+
- **Features**: Data warehouse, ETL, analytics
|
|
339
|
+
- **When to use**: Building data pipelines and warehouses
|
|
340
|
+
|
|
341
|
+
### `doc_ocr_jobs` — Document OCR job queue
|
|
342
|
+
- **GPU**: Optional
|
|
343
|
+
- **Key pattern**: `modal.Queue` for job processing
|
|
344
|
+
- **Features**: Job queue, OCR, async processing
|
|
345
|
+
- **When to use**: Document processing pipelines
|
|
346
|
+
|
|
347
|
+
### `doc_ocr_webapp` — Serve a Document OCR web app
|
|
348
|
+
- **GPU**: Optional
|
|
349
|
+
- **Key pattern**: OCR web app with file upload
|
|
350
|
+
- **Features**: Web UI, OCR, file processing
|
|
351
|
+
- **When to use**: OCR web applications with user-facing interface
|
|
352
|
+
|
|
353
|
+
### `hackernews_alerts` — Deploy a Hacker News Slackbot
|
|
354
|
+
- **GPU**: None (CPU)
|
|
355
|
+
- **Key pattern**: `modal.Cron` + Slack integration
|
|
356
|
+
- **Features**: Scheduled jobs, Slack webhooks, web scraping
|
|
357
|
+
- **When to use**: Scheduled data collection + notifications
|
|
358
|
+
|
|
359
|
+
### `discord_bot` — Deploy and run a Discord bot
|
|
360
|
+
- **GPU**: None (CPU)
|
|
361
|
+
- **Key pattern**: Discord.py on Modal
|
|
362
|
+
- **Features**: Discord bot, persistent service
|
|
363
|
+
- **When to use**: Discord bot deployment
|
|
364
|
+
|
|
365
|
+
### `db_to_sheet` — Sync databases and APIs to a Google Sheet
|
|
366
|
+
- **GPU**: None (CPU)
|
|
367
|
+
- **Key pattern**: Scheduled ETL to Google Sheets
|
|
368
|
+
- **Features**: Database sync, Google Sheets API, scheduled jobs
|
|
369
|
+
- **When to use**: Automated reporting, database-to-spreadsheet sync
|
|
370
|
+
|
|
371
|
+
### `cron_datasette` — Publish explorable data with SQLite and Datasette
|
|
372
|
+
- **GPU**: None (CPU)
|
|
373
|
+
- **Key pattern**: SQLite + Datasette on Modal
|
|
374
|
+
- **Features**: Data exploration, SQLite, Datasette UI
|
|
375
|
+
- **When to use**: Publishing explorable datasets
|
|
376
|
+
|
|
377
|
+
### `algolia_indexer` — Build docsearch with an Algolia crawler
|
|
378
|
+
- **GPU**: None (CPU)
|
|
379
|
+
- **Key pattern**: Web crawling + Algolia indexing
|
|
380
|
+
- **Features**: Documentation search, web crawling, Algolia
|
|
381
|
+
- **When to use**: Building search indexes for documentation sites
|
|
382
|
+
|
|
383
|
+
## Computational Biology
|
|
384
|
+
|
|
385
|
+
### `chai1` — Fold proteins with Chai-1
|
|
386
|
+
- **GPU**: A100 or H100
|
|
387
|
+
- **Key pattern**: Protein folding pipeline
|
|
388
|
+
- **Features**: Chai-1 model, protein structure prediction
|
|
389
|
+
- **When to use**: Protein folding / structural biology
|
|
390
|
+
|
|
391
|
+
### `boltz_predict` — Fold proteins with Boltz-2
|
|
392
|
+
- **GPU**: A100 or H100
|
|
393
|
+
- **Key pattern**: Boltz-2 protein structure prediction
|
|
394
|
+
- **Features**: Latest protein folding model
|
|
395
|
+
- **When to use**: Protein structure prediction
|
|
396
|
+
|
|
397
|
+
### `esm3` — Build a protein folding dashboard
|
|
398
|
+
- **GPU**: A100
|
|
399
|
+
- **Key pattern**: ESM3 protein language model with visualization
|
|
400
|
+
- **Features**: Protein embeddings, sequence analysis, dashboard UI
|
|
401
|
+
- **When to use**: Protein sequence analysis, structural biology dashboards
|
|
402
|
+
|
|
403
|
+
## Networking & Connectivity
|
|
404
|
+
|
|
405
|
+
### `modal_tailscale` — Add Modal Apps to your VPN with Tailscale
|
|
406
|
+
- **GPU**: None
|
|
407
|
+
- **Key pattern**: Tailscale VPN integration with Modal containers
|
|
408
|
+
- **Features**: VPN access, private networking, Tailscale
|
|
409
|
+
- **When to use**: Connecting Modal containers to private networks
|
|
410
|
+
|
|
411
|
+
### `pushgateway` — Publish custom metrics with Prometheus Pushgateway
|
|
412
|
+
- **GPU**: None
|
|
413
|
+
- **Key pattern**: Prometheus metrics from Modal functions
|
|
414
|
+
- **Features**: Custom metrics, Prometheus, monitoring
|
|
415
|
+
- **When to use**: Observability and metrics collection from Modal workloads
|
|
416
|
+
|
|
417
|
+
## Reinforcement Learning
|
|
418
|
+
|
|
419
|
+
### `grpo_verl` — GRPO math training with verl
|
|
420
|
+
- See Training & Fine-Tuning section above
|
|
421
|
+
|
|
422
|
+
### `grpo_trl` — GRPO coding training with TRL
|
|
423
|
+
- See Training & Fine-Tuning section above
|
|
@@ -0,0 +1,69 @@
|
|
|
1
|
+
# Prime Intellect Lab Skill
|
|
2
|
+
|
|
3
|
+
**Hosted RL post-training with environments, verifiers, GEPA prompt optimization, and agentic training**
|
|
4
|
+
|
|
5
|
+
## Skill Structure
|
|
6
|
+
|
|
7
|
+
```
|
|
8
|
+
prime-intellect-lab/
|
|
9
|
+
├── SKILL.md # Main skill documentation (READ THIS FIRST)
|
|
10
|
+
├── README.md # This file
|
|
11
|
+
└── templates/
|
|
12
|
+
└── basic_rl_training.toml # Production-ready training config template
|
|
13
|
+
```
|
|
14
|
+
|
|
15
|
+
## Quick Start
|
|
16
|
+
|
|
17
|
+
1. **Read SKILL.md** — Comprehensive guide with all concepts and workflows
|
|
18
|
+
2. **Set up workspace** — `prime lab setup` creates configs, environments, and agent skills
|
|
19
|
+
3. **Browse environments** — `prime env list` to find training environments
|
|
20
|
+
4. **Copy template** — `cp templates/basic_rl_training.toml configs/rl/my-run.toml`
|
|
21
|
+
5. **Launch training** — `prime rl run configs/rl/my-run.toml`
|
|
22
|
+
|
|
23
|
+
## What's Inside
|
|
24
|
+
|
|
25
|
+
### SKILL.md (Main Documentation)
|
|
26
|
+
- When to use Prime Intellect vs Tinker vs local training
|
|
27
|
+
- Core concepts: environments, hosted training, verifiers, GEPA, Lab agent skills
|
|
28
|
+
- Setup and authentication (`prime login` or `PRIME_API_KEY`)
|
|
29
|
+
- Complete training workflow (env install, baseline eval, config, train, monitor)
|
|
30
|
+
- Full configuration reference (`.toml` fields with `[[env]]` syntax)
|
|
31
|
+
- Available models table (Qwen3, INTELLECT-3 with exact IDs)
|
|
32
|
+
- GEPA prompt optimization (gradient-free system prompt refinement)
|
|
33
|
+
- Custom environment development with `verifiers` library
|
|
34
|
+
- Multi-environment training with weighted `[[env]]` blocks
|
|
35
|
+
- Compute API for direct GPU provisioning
|
|
36
|
+
- Troubleshooting guide (9 common issues)
|
|
37
|
+
- Agent usage instructions including brainstorm skill workflow
|
|
38
|
+
|
|
39
|
+
### Templates
|
|
40
|
+
- **basic_rl_training.toml**: Production-ready training config
|
|
41
|
+
- Small / medium / large run presets
|
|
42
|
+
- Correct `[[env]]` syntax with `id` and `args` fields
|
|
43
|
+
- All key fields documented with comments
|
|
44
|
+
|
|
45
|
+
## When to Use
|
|
46
|
+
|
|
47
|
+
| Task | Use This? |
|
|
48
|
+
|------|-----------|
|
|
49
|
+
| Hosted RL training with environments | Yes |
|
|
50
|
+
| Agentic multi-turn RL | Yes |
|
|
51
|
+
| LoRA on open-weight models via RL | Yes |
|
|
52
|
+
| GEPA prompt optimization | Yes |
|
|
53
|
+
| SFT / supervised fine-tuning | No — use Tinker |
|
|
54
|
+
| Local GPU training | No — use Axolotl/Unsloth |
|
|
55
|
+
| Custom architectures | No — use Modal/Lambda |
|
|
56
|
+
| On-demand GPU clusters | No — use TensorPool |
|
|
57
|
+
|
|
58
|
+
## Version
|
|
59
|
+
|
|
60
|
+
**v1.0.0** — Initial release (February 2026)
|
|
61
|
+
|
|
62
|
+
## Maintained By
|
|
63
|
+
|
|
64
|
+
Synthetic Sciences
|
|
65
|
+
https://syntheticsciences.ai
|
|
66
|
+
|
|
67
|
+
---
|
|
68
|
+
|
|
69
|
+
**License:** MIT
|