traicebox 0.1.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/LICENSE +21 -0
- package/dist/00-app-databases-mx6j59m0.sh +28 -0
- package/dist/Caddyfile-at2nzhxs. +18 -0
- package/dist/Dockerfile-bzexf8bh. +7 -0
- package/dist/Dockerfile-tvb2c6ma. +7 -0
- package/dist/bootstrap-client-key-v79zzaxg.sh +39 -0
- package/dist/compose-1b8sxndd.yml +266 -0
- package/dist/compose-vz8yebk4.yml +266 -0
- package/dist/config-tt3vhpk0.yaml +101 -0
- package/dist/config-ymzwdk89.yaml +7 -0
- package/dist/index.js +16487 -0
- package/dist/logging-jyk2svr4.json +35 -0
- package/dist/request_session_metadata_callback-wfkph4zx.py +109 -0
- package/dist/server-bm43enwc.ts +257 -0
- package/dist/server-d2v9t7c1.ts +257 -0
- package/dist/server-dchc55xz.ts +259 -0
- package/dist/server-k3xr84w3.ts +259 -0
- package/dist/start-litellm-1rtmgp2c.sh +16 -0
- package/dist/traicebox-7pysd22b.yaml +2 -0
- package/dist/traicebox-d8k281x5.yaml +2 -0
- package/package.json +50 -0
|
@@ -0,0 +1,101 @@
|
|
|
1
|
+
model_list:
|
|
2
|
+
- model_name: bytedance/seed-oss-36b
|
|
3
|
+
litellm_params:
|
|
4
|
+
model: openai/bytedance/seed-oss-36b
|
|
5
|
+
api_base: http://host.docker.internal:1234/v1
|
|
6
|
+
api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
|
|
7
|
+
- model_name: google/gemma-4-26b-a4b
|
|
8
|
+
litellm_params:
|
|
9
|
+
model: openai/google/gemma-4-26b-a4b
|
|
10
|
+
api_base: http://host.docker.internal:1234/v1
|
|
11
|
+
api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
|
|
12
|
+
- model_name: google/gemma-4-31b
|
|
13
|
+
litellm_params:
|
|
14
|
+
model: openai/google/gemma-4-31b
|
|
15
|
+
api_base: http://host.docker.internal:1234/v1
|
|
16
|
+
api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
|
|
17
|
+
- model_name: google/gemma-4-e2b
|
|
18
|
+
litellm_params:
|
|
19
|
+
model: openai/google/gemma-4-e2b
|
|
20
|
+
api_base: http://host.docker.internal:1234/v1
|
|
21
|
+
api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
|
|
22
|
+
- model_name: google/gemma-4-e4b
|
|
23
|
+
litellm_params:
|
|
24
|
+
model: openai/google/gemma-4-e4b
|
|
25
|
+
api_base: http://host.docker.internal:1234/v1
|
|
26
|
+
api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
|
|
27
|
+
- model_name: ibm/granite-4-h-tiny
|
|
28
|
+
litellm_params:
|
|
29
|
+
model: openai/ibm/granite-4-h-tiny
|
|
30
|
+
api_base: http://host.docker.internal:1234/v1
|
|
31
|
+
api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
|
|
32
|
+
- model_name: liquid/lfm2-24b-a2b
|
|
33
|
+
litellm_params:
|
|
34
|
+
model: openai/liquid/lfm2-24b-a2b
|
|
35
|
+
api_base: http://host.docker.internal:1234/v1
|
|
36
|
+
api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
|
|
37
|
+
- model_name: microsoft/phi-4-reasoning-plus
|
|
38
|
+
litellm_params:
|
|
39
|
+
model: openai/microsoft/phi-4-reasoning-plus
|
|
40
|
+
api_base: http://host.docker.internal:1234/v1
|
|
41
|
+
api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
|
|
42
|
+
- model_name: mistralai/devstral-small-2-2512
|
|
43
|
+
litellm_params:
|
|
44
|
+
model: openai/mistralai/devstral-small-2-2512
|
|
45
|
+
api_base: http://host.docker.internal:1234/v1
|
|
46
|
+
api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
|
|
47
|
+
- model_name: mistralai/devstral-small-2507
|
|
48
|
+
litellm_params:
|
|
49
|
+
model: openai/mistralai/devstral-small-2507
|
|
50
|
+
api_base: http://host.docker.internal:1234/v1
|
|
51
|
+
api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
|
|
52
|
+
- model_name: mistralai/magistral-small-2509
|
|
53
|
+
litellm_params:
|
|
54
|
+
model: openai/mistralai/magistral-small-2509
|
|
55
|
+
api_base: http://host.docker.internal:1234/v1
|
|
56
|
+
api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
|
|
57
|
+
- model_name: mistralai/ministral-3-14b-reasoning
|
|
58
|
+
litellm_params:
|
|
59
|
+
model: openai/mistralai/ministral-3-14b-reasoning
|
|
60
|
+
api_base: http://host.docker.internal:1234/v1
|
|
61
|
+
api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
|
|
62
|
+
- model_name: nvidia/nemotron-3-nano
|
|
63
|
+
litellm_params:
|
|
64
|
+
model: openai/nvidia/nemotron-3-nano
|
|
65
|
+
api_base: http://host.docker.internal:1234/v1
|
|
66
|
+
api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
|
|
67
|
+
- model_name: openai/gpt-oss-20b
|
|
68
|
+
litellm_params:
|
|
69
|
+
model: openai/openai/gpt-oss-20b
|
|
70
|
+
api_base: http://host.docker.internal:1234/v1
|
|
71
|
+
api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
|
|
72
|
+
- model_name: qwen/qwen3-30b-a3b-2507
|
|
73
|
+
litellm_params:
|
|
74
|
+
model: openai/qwen/qwen3-30b-a3b-2507
|
|
75
|
+
api_base: http://host.docker.internal:1234/v1
|
|
76
|
+
api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
|
|
77
|
+
- model_name: qwen/qwen3-4b-2507
|
|
78
|
+
litellm_params:
|
|
79
|
+
model: openai/qwen/qwen3-4b-2507
|
|
80
|
+
api_base: http://host.docker.internal:1234/v1
|
|
81
|
+
api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
|
|
82
|
+
- model_name: qwen/qwen3.5-35b-a3b
|
|
83
|
+
litellm_params:
|
|
84
|
+
model: openai/qwen/qwen3.5-35b-a3b
|
|
85
|
+
api_base: http://host.docker.internal:1234/v1
|
|
86
|
+
api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
|
|
87
|
+
- model_name: text-embedding-nomic-embed-text-v1.5
|
|
88
|
+
litellm_params:
|
|
89
|
+
model: openai/text-embedding-nomic-embed-text-v1.5
|
|
90
|
+
api_base: http://host.docker.internal:1234/v1
|
|
91
|
+
api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
|
|
92
|
+
- model_name: zai-org/glm-4.7-flash
|
|
93
|
+
litellm_params:
|
|
94
|
+
model: openai/zai-org/glm-4.7-flash
|
|
95
|
+
api_base: http://host.docker.internal:1234/v1
|
|
96
|
+
api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
|
|
97
|
+
litellm_settings:
|
|
98
|
+
database_url: os.environ/DATABASE_URL
|
|
99
|
+
callbacks:
|
|
100
|
+
- request_session_metadata_callback.proxy_handler_instance
|
|
101
|
+
- langfuse_otel
|