@synsci/cli-darwin-x64 1.1.58 → 1.1.60

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -0,0 +1,423 @@
1
+ # Modal Examples Catalog
2
+
3
+ Complete catalog of Modal's official example library with implementation notes. Each example is production-ready code you can adapt for your workloads.
4
+
5
+ **Source**: https://modal.com/docs/examples
6
+
7
+ ## LLM Inference & Serving
8
+
9
+ ### `llm_inference` — Deploy an OpenAI-compatible LLM service
10
+ - **GPU**: H100 or A100-80GB
11
+ - **Key pattern**: Class-based app with `@modal.enter()` for model loading, `@modal.web_server()` for OpenAI-compatible API
12
+ - **Features**: vLLM backend, streaming responses, OpenAI-compatible `/v1/chat/completions` endpoint, Volume for model weights, auto-scaling
13
+ - **When to use**: Default choice for serving any HF model as an API
14
+
15
+ ### `very_large_models` — Deploy really big language models
16
+ - **GPU**: H200:4-8 or B200:4-8
17
+ - **Key pattern**: SGLang + multi-GPU tensor parallelism for models that don't fit on a single GPU
18
+ - **Features**: DeepSeek V3, Kimi-K2, GLM 4.7/5 support, multi-GPU serving, auto-scaling
19
+ - **When to use**: Serving 100B+ parameter models (DeepSeek V3, Kimi-K2, etc.)
20
+
21
+ ### `ministral3_inference` — Cut Ministral 3 cold start times by 10x with snapshots
22
+ - **GPU**: Any
23
+ - **Key pattern**: Memory snapshots to pre-load model weights into container memory
24
+ - **Features**: `modal.Sandbox` snapshots, near-zero cold starts for large models
25
+ - **When to use**: When cold start latency is critical (user-facing APIs)
26
+
27
+ ### `vllm_throughput` — Optimize tokens per second with vLLM
28
+ - **GPU**: H100 or A100
29
+ - **Key pattern**: Batch processing with vLLM for maximum throughput
30
+ - **Features**: Offline batch inference, throughput optimization (~30K input tok/s per H100), parallel processing
31
+ - **When to use**: Processing large datasets (evaluation, synthetic data generation)
32
+
33
+ ### `sglang_low_latency` — Low-latency inference with SGLang
34
+ - **GPU**: H100:2
35
+ - **Key pattern**: SGLang runtime with speculative decoding (EAGLE-3) for optimized inference
36
+ - **Features**: Speculative decoding, low latency, dual-GPU, streaming
37
+ - **When to use**: When you need lower latency than vLLM, latency-critical applications
38
+
39
+ ### `llama_cpp` — Run GGUF models with llama.cpp
40
+ - **GPU**: Optional (CPU inference supported)
41
+ - **Key pattern**: llama.cpp server on Modal
42
+ - **Features**: GGUF quantized models, CPU/GPU hybrid, low resource usage
43
+ - **When to use**: Serving quantized models, budget-friendly inference
44
+
45
+ ### `trtllm_latency` — Low-latency inference with TensorRT-LLM
46
+ - **GPU**: H100
47
+ - **Key pattern**: TensorRT-LLM compiled model serving
48
+ - **Features**: Optimized kernels, lowest possible latency
49
+ - **When to use**: Production inference where every millisecond counts
50
+
51
+ ### `trtllm_throughput` — High-throughput batch inference with TensorRT-LLM
52
+ - **GPU**: H100
53
+ - **Key pattern**: TensorRT-LLM for batch processing
54
+ - **Features**: Maximum tokens/sec throughput
55
+ - **When to use**: Batch processing with TensorRT optimization
56
+
57
+ ## Training & Fine-Tuning
58
+
59
+ ### `grpo_verl` — Train a model to solve math problems using GRPO and verl
60
+ - **GPU**: H100:8 (multi-GPU)
61
+ - **Key pattern**: verl framework for GRPO RL training
62
+ - **Features**: Math reasoning reward, multi-GPU training, Volume checkpointing
63
+ - **When to use**: RL-based training for reasoning tasks (math, logic)
64
+
65
+ ### `grpo_trl` — Train a model to solve coding problems using GRPO and TRL
66
+ - **GPU**: A100 or H100
67
+ - **Key pattern**: TRL's GRPOTrainer on Modal
68
+ - **Features**: Code generation reward, sandbox-based evaluation, RL training
69
+ - **When to use**: RL-based training for code generation
70
+
71
+ ### `unsloth_finetune` — Efficient LLM fine-tuning with Unsloth
72
+ - **GPU**: A100 or L40S
73
+ - **Key pattern**: Unsloth for 2x faster LoRA training
74
+ - **Features**: Memory efficient, 2x speed, QLoRA support
75
+ - **When to use**: LoRA fine-tuning when Tinker doesn't support your model
76
+
77
+ ### `hp_sweep_gpt` — Train SLM with early-stopping grid search
78
+ - **GPU**: A100
79
+ - **Key pattern**: Grid search over hyperparameters with `.map()`
80
+ - **Features**: Parallel HP sweep, early stopping, Volume for checkpoints
81
+ - **When to use**: Hyperparameter optimization, training from scratch
82
+
83
+ ### `long-training` — Run long, resumable training jobs
84
+ - **GPU**: A100 or H100
85
+ - **Key pattern**: Checkpointing to Volume with resume-on-preemption
86
+ - **Features**: Volume checkpointing, preemption handling, long-running jobs
87
+ - **When to use**: Multi-hour/day training runs that need fault tolerance
88
+
89
+ ### `llm-finetuning` — Full LLM fine-tuning pipeline
90
+ - **GPU**: A100-80GB or H100
91
+ - **Key pattern**: End-to-end fine-tuning with evaluation
92
+ - **Features**: Data loading, training loop, evaluation, model upload
93
+ - **When to use**: Complete fine-tuning workflow on Modal
94
+
95
+ ### `flan_t5_finetune` — Fine-tune Flan-T5
96
+ - **GPU**: A10G or L40S
97
+ - **Key pattern**: Seq2seq model fine-tuning
98
+ - **Features**: Flan-T5 training, evaluation
99
+ - **When to use**: Seq2seq tasks (summarization, translation, Q&A)
100
+
101
+ ### `diffusers_lora_finetune` — Custom pet art from Flux with Hugging Face and Gradio
102
+ - **GPU**: A100-80GB
103
+ - **Key pattern**: Diffusers LoRA training for image generation
104
+ - **Features**: Image LoRA, custom datasets, Flux models, Gradio UI
105
+ - **When to use**: Fine-tuning image generation models
106
+
107
+ ## Multimodal & Vision
108
+
109
+ ### `flux` — Serve diffusion models fast with torch.compile
110
+ - **GPU**: A100 or H100
111
+ - **Key pattern**: torch.compile for optimized image generation
112
+ - **Features**: Compilation cache, fast generation, various Flux variants
113
+ - **When to use**: Fast image generation with Flux models
114
+
115
+ ### `image_to_image` — Edit images with Flux Kontext
116
+ - **GPU**: A100 or H100
117
+ - **Key pattern**: Image-to-image with Flux Kontext
118
+ - **Features**: Image editing, style transfer
119
+ - **When to use**: Image editing tasks
120
+
121
+ ### `image_to_video` — Animate images with LTX-Video
122
+ - **GPU**: A100 or H100
123
+ - **Key pattern**: Image animation pipeline
124
+ - **Features**: Image-to-video generation
125
+ - **When to use**: Creating video from static images
126
+
127
+ ### `ltx` — Generate video clips with LTX-Video
128
+ - **GPU**: A100 or H100
129
+ - **Key pattern**: Text-to-video generation
130
+ - **Features**: LTX-Video model, video clips
131
+ - **When to use**: Text-to-video generation
132
+
133
+ ### `text_to_image` — Stable Diffusion CLI/API/UI
134
+ - **GPU**: A10G or L40S
135
+ - **Key pattern**: Stable Diffusion serving with multiple interfaces
136
+ - **Features**: CLI, REST API, Gradio UI
137
+ - **When to use**: Standard text-to-image generation
138
+
139
+ ### `finetune_yolo` — Fine-tune and serve YOLO models
140
+ - **GPU**: T4 or A10G
141
+ - **Key pattern**: YOLO training + serving pipeline
142
+ - **Features**: Object detection, model serving
143
+ - **When to use**: Computer vision / object detection
144
+
145
+ ### `segment_anything` — Segment Anything Model
146
+ - **GPU**: A10G
147
+ - **Key pattern**: SAM inference
148
+ - **Features**: Zero-shot segmentation
149
+ - **When to use**: Image segmentation tasks
150
+
151
+ ### `comfyapp` — Run Flux on ComfyUI as an API
152
+ - **GPU**: A100 or H100
153
+ - **Key pattern**: ComfyUI workflow as API
154
+ - **Features**: ComfyUI, workflow automation, Flux
155
+ - **When to use**: Complex image generation pipelines via ComfyUI
156
+
157
+ ### `blender_video` — Build a 3D render farm with Blender
158
+ - **GPU**: Optional (CPU rendering supported)
159
+ - **Key pattern**: Parallel rendering with Blender on Modal
160
+ - **Features**: 3D rendering, frame parallelism, render farm
161
+ - **When to use**: Distributed 3D rendering, animation pipelines
162
+
163
+ ## Audio & Speech
164
+
165
+ ### `llm-voice-chat` — Voice chat with LLMs
166
+ - **GPU**: A10G or L40S
167
+ - **Key pattern**: Real-time voice interaction with WebSocket
168
+ - **Features**: Moshi model, WebSocket streaming, real-time audio
169
+ - **When to use**: Voice chatbot / real-time audio interaction
170
+
171
+ ### `streaming_kyutai_stt` — Transcribe speech with Kyutai STT
172
+ - **GPU**: L4 or A10G
173
+ - **Key pattern**: Streaming speech-to-text
174
+ - **Features**: Real-time transcription, Kyutai STT, low latency
175
+ - **When to use**: Real-time transcription (live captions, dictation)
176
+
177
+ ### `music-video-gen` — Star in custom music videos
178
+ - **GPU**: A100 or H100
179
+ - **Key pattern**: Multi-model pipeline (image gen + video gen + audio)
180
+ - **Features**: Music video generation, multi-model orchestration
181
+ - **When to use**: Creative AI pipelines combining audio, image, and video generation
182
+
183
+ ### `generate_music` — Make music with ACE-Step
184
+ - **GPU**: A10G or L40S
185
+ - **Key pattern**: Music generation pipeline
186
+ - **Features**: ACE-Step model, music generation
187
+ - **When to use**: AI music generation
188
+
189
+ ### `chatterbox_tts` — Generate speech with Chatterbox
190
+ - **GPU**: A10G
191
+ - **Key pattern**: Text-to-speech pipeline
192
+ - **Features**: Chatterbox TTS, voice synthesis
193
+ - **When to use**: Text-to-speech generation
194
+
195
+ ### `batched_whisper` — High throughput batched transcription
196
+ - **GPU**: A10G or L40S
197
+ - **Key pattern**: Batch Whisper with `@modal.batched()`
198
+ - **Features**: Dynamic batching, high throughput, parallel processing
199
+ - **When to use**: Transcribing large audio datasets
200
+
201
+ ### `fine_tune_asr` — Fine-tune Whisper to recognize new words
202
+ - **GPU**: A10G or A100
203
+ - **Key pattern**: Whisper fine-tuning pipeline
204
+ - **Features**: Custom vocabulary, domain-specific ASR
205
+ - **When to use**: Adapting Whisper to domain-specific terminology
206
+
207
+ ## Sandboxes & Code Execution
208
+
209
+ ### `agent` — Sandbox a LangGraph agent's code
210
+ - **GPU**: T4 or A10G
211
+ - **Key pattern**: LangGraph agent with Modal Sandbox for code execution
212
+ - **Features**: GPU sandbox, LangGraph, tool use, file access
213
+ - **When to use**: Building coding agents that need to execute code
214
+
215
+ ### `opencode_server` — Run a background coding agent with OpenCode
216
+ - **GPU**: Optional
217
+ - **Key pattern**: OpenCode coding agent running in a Modal Sandbox
218
+ - **Features**: Background coding agent, sandbox isolation, persistent sessions
219
+ - **When to use**: Deploying autonomous coding agents in sandboxes
220
+
221
+ ### `modal-vibe` — Deploy vibe coding at scale
222
+ - **GPU**: Optional
223
+ - **Key pattern**: React frontend + LLM + Modal Sandboxes for scalable AI coding platform
224
+ - **Features**: Code execution at scale, React UI, LLM integration, sandbox isolation
225
+ - **When to use**: Building scalable AI-powered coding platforms
226
+
227
+ ### `safe_code_execution` — Run Node.js, Ruby, and more in a Sandbox
228
+ - **GPU**: Optional
229
+ - **Key pattern**: Multi-language sandboxed code execution at scale
230
+ - **Features**: Secure execution, Node.js/Ruby/Python, scaling, multi-language
231
+ - **When to use**: Building coding platforms, code evaluation systems, multi-language sandboxes
232
+
233
+ ### `simple_code_interpreter` — Build a stateful, sandboxed code interpreter
234
+ - **GPU**: Optional
235
+ - **Key pattern**: Stateful sandbox with session persistence
236
+ - **Features**: Jupyter-like sessions, state persistence, file I/O
237
+ - **When to use**: Interactive code interpreter (like ChatGPT Code Interpreter)
238
+
239
+ ### `jupyter_sandbox` — Run a sandboxed Jupyter notebook
240
+ - **GPU**: Optional
241
+ - **Key pattern**: Jupyter notebook in Modal Sandbox
242
+ - **Features**: Jupyter server, file access, isolated execution
243
+ - **When to use**: Providing sandboxed Jupyter environments
244
+
245
+ ### `anthropic_computer_use` — Control a sandboxed computer with an LLM
246
+ - **GPU**: Optional
247
+ - **Key pattern**: Virtual desktop in sandbox controlled by LLM
248
+ - **Features**: Computer use, screenshot, mouse/keyboard control
249
+ - **When to use**: Computer use / GUI automation agents
250
+
251
+ ## RAG & Embeddings
252
+
253
+ ### `chat_with_pdf_vision` — RAG Chat with PDFs
254
+ - **GPU**: A10G or L40S
255
+ - **Key pattern**: Vision-based PDF parsing + RAG
256
+ - **Features**: PDF parsing, multimodal embeddings, chat interface
257
+ - **When to use**: Document Q&A, PDF chatbots
258
+
259
+ ### `amazon_embeddings` — Embed millions of documents with TEI
260
+ - **GPU**: A10G or L40S
261
+ - **Key pattern**: HuggingFace Text Embeddings Inference (TEI) for bulk embedding
262
+ - **Features**: High throughput, batch processing, `.map()` parallelism
263
+ - **When to use**: Large-scale embedding generation
264
+
265
+ ### `mongodb-search` — Turn satellite images into vectors and store them in MongoDB
266
+ - **GPU**: T4 or A10G
267
+ - **Key pattern**: Image embedding + MongoDB Atlas vector search + GeoJSON
268
+ - **Features**: Satellite image embeddings, MongoDB Atlas, vector + geospatial search
269
+ - **When to use**: Image search with geospatial queries, visual similarity + location
270
+
271
+ ### `potus_speech_qanda` — Retrieval-Augmented Generation (RAG) for Q&A
272
+ - **GPU**: None (CPU)
273
+ - **Key pattern**: Basic RAG pipeline with OpenAI
274
+ - **Features**: Document indexing, question answering
275
+ - **When to use**: Simple RAG setup with OpenAI
276
+
277
+ ## Web Apps & Endpoints
278
+
279
+ ### `basic_web` — Serving web endpoints
280
+ - **GPU**: None (CPU)
281
+ - **Key pattern**: FastAPI + `@modal.asgi_app()` / `@modal.fastapi_endpoint()`
282
+ - **Features**: REST API, ASGI/WSGI support, auto-scaling
283
+ - **When to use**: Learning Modal web endpoints, simple APIs
284
+
285
+ ### `serve_streamlit` — Deploy Streamlit apps
286
+ - **GPU**: Optional
287
+ - **Key pattern**: `@modal.web_server()` for Streamlit
288
+ - **Features**: Streamlit, interactive dashboards
289
+ - **When to use**: Data apps, ML demos with Streamlit
290
+
291
+ ### `fasthtml_app` — FastHTML applications
292
+ - **GPU**: None (CPU)
293
+ - **Key pattern**: FastHTML on Modal
294
+ - **Features**: FastHTML, lightweight web apps
295
+ - **When to use**: Simple web applications
296
+
297
+ ### `mcp_server_stateless` — Deploy a stateless MCP with FastMCP
298
+ - **GPU**: Optional
299
+ - **Key pattern**: FastMCP server on Modal
300
+ - **Features**: MCP protocol, tool serving, stateless
301
+ - **When to use**: Deploying MCP tool servers
302
+
303
+ ### `webrtc_yolo` — Serverless WebRTC with YOLO detection
304
+ - **GPU**: T4 or A10G
305
+ - **Key pattern**: WebRTC + YOLO on Modal
306
+ - **Features**: Real-time video, object detection, WebRTC
307
+ - **When to use**: Real-time video processing apps
308
+
309
+ ### `fastrtc_flip_webcam` — WebRTC quickstart with FastRTC
310
+ - **GPU**: Optional
311
+ - **Key pattern**: FastRTC framework on Modal
312
+ - **Features**: WebRTC, fast setup, real-time communication
313
+ - **When to use**: Getting started with WebRTC on Modal
314
+
315
+ ### `webscraper` — Simple web scraper
316
+ - **GPU**: None (CPU)
317
+ - **Key pattern**: Web scraping with parallel processing
318
+ - **Features**: Scraping, data collection, parallelism
319
+ - **When to use**: Web scraping and data collection tasks
320
+
321
+ ## Data & Infrastructure
322
+
323
+ ### `s3_bucket_mount` — Parallel processing of Parquet files on S3
324
+ - **GPU**: None (CPU)
325
+ - **Key pattern**: `modal.CloudBucketMount` for S3 access
326
+ - **Features**: S3 as filesystem, Parquet processing, parallel reads
327
+ - **When to use**: Processing data from S3, large datasets
328
+
329
+ ### `cloud_bucket_mount_loras` — Create a LoRA Playground with Modal, Gradio, and S3
330
+ - **GPU**: A10G or L40S
331
+ - **Key pattern**: S3 bucket mount for LoRA weight management
332
+ - **Features**: LoRA weights on S3, Gradio UI, dynamic model loading
333
+ - **When to use**: Managing and serving multiple LoRA adapters
334
+
335
+ ### `dbt_duckdb` — Build your own data warehouse with DuckDB, DBT, and Modal
336
+ - **GPU**: None (CPU)
337
+ - **Key pattern**: Data pipeline with DuckDB + DBT
338
+ - **Features**: Data warehouse, ETL, analytics
339
+ - **When to use**: Building data pipelines and warehouses
340
+
341
+ ### `doc_ocr_jobs` — Document OCR job queue
342
+ - **GPU**: Optional
343
+ - **Key pattern**: `modal.Queue` for job processing
344
+ - **Features**: Job queue, OCR, async processing
345
+ - **When to use**: Document processing pipelines
346
+
347
+ ### `doc_ocr_webapp` — Serve a Document OCR web app
348
+ - **GPU**: Optional
349
+ - **Key pattern**: OCR web app with file upload
350
+ - **Features**: Web UI, OCR, file processing
351
+ - **When to use**: OCR web applications with user-facing interface
352
+
353
+ ### `hackernews_alerts` — Deploy a Hacker News Slackbot
354
+ - **GPU**: None (CPU)
355
+ - **Key pattern**: `modal.Cron` + Slack integration
356
+ - **Features**: Scheduled jobs, Slack webhooks, web scraping
357
+ - **When to use**: Scheduled data collection + notifications
358
+
359
+ ### `discord_bot` — Deploy and run a Discord bot
360
+ - **GPU**: None (CPU)
361
+ - **Key pattern**: Discord.py on Modal
362
+ - **Features**: Discord bot, persistent service
363
+ - **When to use**: Discord bot deployment
364
+
365
+ ### `db_to_sheet` — Sync databases and APIs to a Google Sheet
366
+ - **GPU**: None (CPU)
367
+ - **Key pattern**: Scheduled ETL to Google Sheets
368
+ - **Features**: Database sync, Google Sheets API, scheduled jobs
369
+ - **When to use**: Automated reporting, database-to-spreadsheet sync
370
+
371
+ ### `cron_datasette` — Publish explorable data with SQLite and Datasette
372
+ - **GPU**: None (CPU)
373
+ - **Key pattern**: SQLite + Datasette on Modal
374
+ - **Features**: Data exploration, SQLite, Datasette UI
375
+ - **When to use**: Publishing explorable datasets
376
+
377
+ ### `algolia_indexer` — Build docsearch with an Algolia crawler
378
+ - **GPU**: None (CPU)
379
+ - **Key pattern**: Web crawling + Algolia indexing
380
+ - **Features**: Documentation search, web crawling, Algolia
381
+ - **When to use**: Building search indexes for documentation sites
382
+
383
+ ## Computational Biology
384
+
385
+ ### `chai1` — Fold proteins with Chai-1
386
+ - **GPU**: A100 or H100
387
+ - **Key pattern**: Protein folding pipeline
388
+ - **Features**: Chai-1 model, protein structure prediction
389
+ - **When to use**: Protein folding / structural biology
390
+
391
+ ### `boltz_predict` — Fold proteins with Boltz-2
392
+ - **GPU**: A100 or H100
393
+ - **Key pattern**: Boltz-2 protein structure prediction
394
+ - **Features**: Latest protein folding model
395
+ - **When to use**: Protein structure prediction
396
+
397
+ ### `esm3` — Build a protein folding dashboard
398
+ - **GPU**: A100
399
+ - **Key pattern**: ESM3 protein language model with visualization
400
+ - **Features**: Protein embeddings, sequence analysis, dashboard UI
401
+ - **When to use**: Protein sequence analysis, structural biology dashboards
402
+
403
+ ## Networking & Connectivity
404
+
405
+ ### `modal_tailscale` — Add Modal Apps to your VPN with Tailscale
406
+ - **GPU**: None
407
+ - **Key pattern**: Tailscale VPN integration with Modal containers
408
+ - **Features**: VPN access, private networking, Tailscale
409
+ - **When to use**: Connecting Modal containers to private networks
410
+
411
+ ### `pushgateway` — Publish custom metrics with Prometheus Pushgateway
412
+ - **GPU**: None
413
+ - **Key pattern**: Prometheus metrics from Modal functions
414
+ - **Features**: Custom metrics, Prometheus, monitoring
415
+ - **When to use**: Observability and metrics collection from Modal workloads
416
+
417
+ ## Reinforcement Learning
418
+
419
+ ### `grpo_verl` — GRPO math training with verl
420
+ - See Training & Fine-Tuning section above
421
+
422
+ ### `grpo_trl` — GRPO coding training with TRL
423
+ - See Training & Fine-Tuning section above
@@ -0,0 +1,69 @@
1
+ # Prime Intellect Lab Skill
2
+
3
+ **Hosted RL post-training with environments, verifiers, GEPA prompt optimization, and agentic training**
4
+
5
+ ## Skill Structure
6
+
7
+ ```
8
+ prime-intellect-lab/
9
+ ├── SKILL.md # Main skill documentation (READ THIS FIRST)
10
+ ├── README.md # This file
11
+ └── templates/
12
+ └── basic_rl_training.toml # Production-ready training config template
13
+ ```
14
+
15
+ ## Quick Start
16
+
17
+ 1. **Read SKILL.md** — Comprehensive guide with all concepts and workflows
18
+ 2. **Set up workspace** — `prime lab setup` creates configs, environments, and agent skills
19
+ 3. **Browse environments** — `prime env list` to find training environments
20
+ 4. **Copy template** — `cp templates/basic_rl_training.toml configs/rl/my-run.toml`
21
+ 5. **Launch training** — `prime rl run configs/rl/my-run.toml`
22
+
23
+ ## What's Inside
24
+
25
+ ### SKILL.md (Main Documentation)
26
+ - When to use Prime Intellect vs Tinker vs local training
27
+ - Core concepts: environments, hosted training, verifiers, GEPA, Lab agent skills
28
+ - Setup and authentication (`prime login` or `PRIME_API_KEY`)
29
+ - Complete training workflow (env install, baseline eval, config, train, monitor)
30
+ - Full configuration reference (`.toml` fields with `[[env]]` syntax)
31
+ - Available models table (Qwen3, INTELLECT-3 with exact IDs)
32
+ - GEPA prompt optimization (gradient-free system prompt refinement)
33
+ - Custom environment development with `verifiers` library
34
+ - Multi-environment training with weighted `[[env]]` blocks
35
+ - Compute API for direct GPU provisioning
36
+ - Troubleshooting guide (9 common issues)
37
+ - Agent usage instructions including brainstorm skill workflow
38
+
39
+ ### Templates
40
+ - **basic_rl_training.toml**: Production-ready training config
41
+ - Small / medium / large run presets
42
+ - Correct `[[env]]` syntax with `id` and `args` fields
43
+ - All key fields documented with comments
44
+
45
+ ## When to Use
46
+
47
+ | Task | Use This? |
48
+ |------|-----------|
49
+ | Hosted RL training with environments | Yes |
50
+ | Agentic multi-turn RL | Yes |
51
+ | LoRA on open-weight models via RL | Yes |
52
+ | GEPA prompt optimization | Yes |
53
+ | SFT / supervised fine-tuning | No — use Tinker |
54
+ | Local GPU training | No — use Axolotl/Unsloth |
55
+ | Custom architectures | No — use Modal/Lambda |
56
+ | On-demand GPU clusters | No — use TensorPool |
57
+
58
+ ## Version
59
+
60
+ **v1.0.0** — Initial release (February 2026)
61
+
62
+ ## Maintained By
63
+
64
+ Synthetic Sciences
65
+ https://syntheticsciences.ai
66
+
67
+ ---
68
+
69
+ **License:** MIT