vue2server7 7.0.65 → 7.0.66
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/package.json +1 -1
- package/test/2.txt +17 -35
package/package.json
CHANGED
package/test/2.txt
CHANGED
|
@@ -1,37 +1,19 @@
|
|
|
1
|
-
curl -X POST "http://maasapp.aip.bj.bob.test:8080/apis/ais-v2/chat/completions" \
|
|
2
|
-
-H "Authorization: Bearer sk-f154d42f-6976-478f-b444-88b90f9dc67b" \
|
|
3
|
-
-H "X-LLM-Application-Tag: proxyai" \
|
|
4
|
-
-H "Content-Type: application/json" \
|
|
5
|
-
-d '{
|
|
6
|
-
"model": "qwen15-32b",
|
|
7
|
-
"messages": [{"role": "user", "content": "测试:请用一句话介绍自己"}],
|
|
8
|
-
"temperature": 0.1,
|
|
9
|
-
"max_tokens": 50,
|
|
10
|
-
"stream": false
|
|
11
|
-
}'
|
|
12
|
-
|
|
13
|
-
|
|
14
|
-
|
|
15
|
-
name: Internal Qwen15-32B (proxyAI Verified)
|
|
16
1
|
models:
|
|
17
|
-
-
|
|
18
|
-
|
|
19
|
-
|
|
20
|
-
apiBase:
|
|
21
|
-
apiKey:
|
|
22
|
-
|
|
23
|
-
|
|
24
|
-
|
|
25
|
-
|
|
26
|
-
|
|
27
|
-
|
|
28
|
-
|
|
29
|
-
|
|
30
|
-
|
|
31
|
-
method: "POST"
|
|
2
|
+
- name: 'qwen15-32b' # 模型名称,对应你截图 Body 里的 model
|
|
3
|
+
provider: 'openai' # 使用 openai 兼容模式
|
|
4
|
+
model: 'qwen15-32b' # 实际调用的模型ID
|
|
5
|
+
apiBase: 'http://maasapp.aip.bj.bob.test:8080/apis/ais-v2' # 对应截图 URL,注意去掉了末尾的 /chat/completions
|
|
6
|
+
apiKey: 'YOUR_CUSTOM_SERVICE_API_KEY' # 对应截图中的 API key
|
|
7
|
+
requestOptions:
|
|
8
|
+
headers:
|
|
9
|
+
# 对应截图 Headers 中的 Authorization
|
|
10
|
+
Authorization: 'Bearer YOUR_CUSTOM_SERVICE_API_KEY'
|
|
11
|
+
# 对应截图 Headers 中的 X-LLM-Application-Tag
|
|
12
|
+
X-LLM-Application-Tag: 'proxyai'
|
|
13
|
+
# 对应截图 Headers 中的 Content-Type
|
|
14
|
+
Content-Type: 'application/json'
|
|
15
|
+
# 对应截图 Body 中的其他参数
|
|
32
16
|
body:
|
|
33
|
-
|
|
34
|
-
|
|
35
|
-
|
|
36
|
-
max_tokens: "{{maxTokens}}" # 用下划线(更安全)
|
|
37
|
-
stream: false # ❗先关掉流式!proxyAI 可能实际发的是 false
|
|
17
|
+
stream: true
|
|
18
|
+
temperature: 0.1
|
|
19
|
+
max_tokens: 20000
|