traicebox 0.1.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -0,0 +1,101 @@
1
+ model_list:
2
+ - model_name: bytedance/seed-oss-36b
3
+ litellm_params:
4
+ model: openai/bytedance/seed-oss-36b
5
+ api_base: http://host.docker.internal:1234/v1
6
+ api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
7
+ - model_name: google/gemma-4-26b-a4b
8
+ litellm_params:
9
+ model: openai/google/gemma-4-26b-a4b
10
+ api_base: http://host.docker.internal:1234/v1
11
+ api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
12
+ - model_name: google/gemma-4-31b
13
+ litellm_params:
14
+ model: openai/google/gemma-4-31b
15
+ api_base: http://host.docker.internal:1234/v1
16
+ api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
17
+ - model_name: google/gemma-4-e2b
18
+ litellm_params:
19
+ model: openai/google/gemma-4-e2b
20
+ api_base: http://host.docker.internal:1234/v1
21
+ api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
22
+ - model_name: google/gemma-4-e4b
23
+ litellm_params:
24
+ model: openai/google/gemma-4-e4b
25
+ api_base: http://host.docker.internal:1234/v1
26
+ api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
27
+ - model_name: ibm/granite-4-h-tiny
28
+ litellm_params:
29
+ model: openai/ibm/granite-4-h-tiny
30
+ api_base: http://host.docker.internal:1234/v1
31
+ api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
32
+ - model_name: liquid/lfm2-24b-a2b
33
+ litellm_params:
34
+ model: openai/liquid/lfm2-24b-a2b
35
+ api_base: http://host.docker.internal:1234/v1
36
+ api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
37
+ - model_name: microsoft/phi-4-reasoning-plus
38
+ litellm_params:
39
+ model: openai/microsoft/phi-4-reasoning-plus
40
+ api_base: http://host.docker.internal:1234/v1
41
+ api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
42
+ - model_name: mistralai/devstral-small-2-2512
43
+ litellm_params:
44
+ model: openai/mistralai/devstral-small-2-2512
45
+ api_base: http://host.docker.internal:1234/v1
46
+ api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
47
+ - model_name: mistralai/devstral-small-2507
48
+ litellm_params:
49
+ model: openai/mistralai/devstral-small-2507
50
+ api_base: http://host.docker.internal:1234/v1
51
+ api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
52
+ - model_name: mistralai/magistral-small-2509
53
+ litellm_params:
54
+ model: openai/mistralai/magistral-small-2509
55
+ api_base: http://host.docker.internal:1234/v1
56
+ api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
57
+ - model_name: mistralai/ministral-3-14b-reasoning
58
+ litellm_params:
59
+ model: openai/mistralai/ministral-3-14b-reasoning
60
+ api_base: http://host.docker.internal:1234/v1
61
+ api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
62
+ - model_name: nvidia/nemotron-3-nano
63
+ litellm_params:
64
+ model: openai/nvidia/nemotron-3-nano
65
+ api_base: http://host.docker.internal:1234/v1
66
+ api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
67
+ - model_name: openai/gpt-oss-20b
68
+ litellm_params:
69
+ model: openai/openai/gpt-oss-20b
70
+ api_base: http://host.docker.internal:1234/v1
71
+ api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
72
+ - model_name: qwen/qwen3-30b-a3b-2507
73
+ litellm_params:
74
+ model: openai/qwen/qwen3-30b-a3b-2507
75
+ api_base: http://host.docker.internal:1234/v1
76
+ api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
77
+ - model_name: qwen/qwen3-4b-2507
78
+ litellm_params:
79
+ model: openai/qwen/qwen3-4b-2507
80
+ api_base: http://host.docker.internal:1234/v1
81
+ api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
82
+ - model_name: qwen/qwen3.5-35b-a3b
83
+ litellm_params:
84
+ model: openai/qwen/qwen3.5-35b-a3b
85
+ api_base: http://host.docker.internal:1234/v1
86
+ api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
87
+ - model_name: text-embedding-nomic-embed-text-v1.5
88
+ litellm_params:
89
+ model: openai/text-embedding-nomic-embed-text-v1.5
90
+ api_base: http://host.docker.internal:1234/v1
91
+ api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
92
+ - model_name: zai-org/glm-4.7-flash
93
+ litellm_params:
94
+ model: openai/zai-org/glm-4.7-flash
95
+ api_base: http://host.docker.internal:1234/v1
96
+ api_key: os.environ/OPENAI_COMPATIBLE_API_KEY
97
+ litellm_settings:
98
+ database_url: os.environ/DATABASE_URL
99
+ callbacks:
100
+ - request_session_metadata_callback.proxy_handler_instance
101
+ - langfuse_otel
@@ -0,0 +1,7 @@
1
+ model_list: []
2
+
3
+ litellm_settings:
4
+ database_url: os.environ/DATABASE_URL
5
+ callbacks:
6
+ - request_session_metadata_callback.proxy_handler_instance
7
+ - langfuse_otel