vue2server7 7.0.64 → 7.0.66

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (2) hide show
  1. package/package.json +1 -1
  2. package/test/2.txt +16 -20
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "vue2server7",
3
- "version": "7.0.64",
3
+ "version": "7.0.66",
4
4
  "description": "",
5
5
  "scripts": {
6
6
  "dev": "nodemon --watch src --ext ts --exec \"ts-node src/app.ts\"",
package/test/2.txt CHANGED
@@ -1,23 +1,19 @@
1
- name: Internal Qwen15-32B
2
1
  models:
3
- - title: "Qwen15-32B (Internal)"
4
- model: "qwen15-32b"
5
- provider: "custom" # ✅ 关键:改为 custom
6
- apiBase: "http://maasapp.aip.bj.bob.test:8080" # 只保留 host:port
7
- apiKey: "sk-f154d42f-6976-478f-b444-88b90f9dc67b" # 先保留测试
8
- headers:
9
- Authorization: "Bearer sk-f154d42f-6976-478f-b444-88b90f9dc67b" # ✅ 显式写完整 Header
10
- X-LLM-Application-Tag: "proxyai"
11
- Content-Type: "application/json"
12
- enableCaching: false
13
- maxTokens: 8192
14
- temperature: 0.1
15
- customRequest:
16
- url: "/apis/ais-v2/chat/completions" # 完整路径
17
- method: "POST"
2
+ - name: 'qwen15-32b' # 模型名称,对应你截图 Body 里的 model
3
+ provider: 'openai' # 使用 openai 兼容模式
4
+ model: 'qwen15-32b' # 实际调用的模型ID
5
+ apiBase: 'http://maasapp.aip.bj.bob.test:8080/apis/ais-v2' # 对应截图 URL,注意去掉了末尾的 /chat/completions
6
+ apiKey: 'YOUR_CUSTOM_SERVICE_API_KEY' # 对应截图中的 API key
7
+ requestOptions:
8
+ headers:
9
+ # 对应截图 Headers 中的 Authorization
10
+ Authorization: 'Bearer YOUR_CUSTOM_SERVICE_API_KEY'
11
+ # 对应截图 Headers 中的 X-LLM-Application-Tag
12
+ X-LLM-Application-Tag: 'proxyai'
13
+ # 对应截图 Headers 中的 Content-Type
14
+ Content-Type: 'application/json'
15
+ # 对应截图 Body 中的其他参数
18
16
  body:
19
- model: "{{model}}"
20
- messages: "{{messages}}"
21
17
  stream: true
22
- temperature: "{{temperature}}"
23
- max_tokens: "{{maxTokens}}"
18
+ temperature: 0.1
19
+ max_tokens: 20000