pro-craft 0.1.15__tar.gz → 0.1.17__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of pro-craft might be problematic. Click here for more details.
- {pro_craft-0.1.15 → pro_craft-0.1.17}/PKG-INFO +1 -1
- {pro_craft-0.1.15 → pro_craft-0.1.17}/pyproject.toml +1 -1
- {pro_craft-0.1.15 → pro_craft-0.1.17}/src/pro_craft/code_helper/coder.py +1 -1
- {pro_craft-0.1.15 → pro_craft-0.1.17}/src/pro_craft/file_manager.py +2 -3
- {pro_craft-0.1.15 → pro_craft-0.1.17}/src/pro_craft/prompt_craft/async_.py +10 -4
- {pro_craft-0.1.15 → pro_craft-0.1.17}/src/pro_craft/prompt_craft/new.py +2 -2
- {pro_craft-0.1.15 → pro_craft-0.1.17}/src/pro_craft/prompt_craft/sync.py +4 -8
- pro_craft-0.1.15/src/pro_craft/server/mcp/weather.py → pro_craft-0.1.17/src/pro_craft/server/mcp/prompt.py +1 -2
- {pro_craft-0.1.15 → pro_craft-0.1.17}/src/pro_craft/server/router/prompt.py +24 -3
- {pro_craft-0.1.15 → pro_craft-0.1.17}/src/pro_craft.egg-info/PKG-INFO +1 -1
- {pro_craft-0.1.15 → pro_craft-0.1.17}/src/pro_craft.egg-info/SOURCES.txt +1 -1
- {pro_craft-0.1.15 → pro_craft-0.1.17}/README.md +0 -0
- {pro_craft-0.1.15 → pro_craft-0.1.17}/setup.cfg +0 -0
- {pro_craft-0.1.15 → pro_craft-0.1.17}/src/pro_craft/__init__.py +0 -0
- {pro_craft-0.1.15 → pro_craft-0.1.17}/src/pro_craft/code_helper/designer.py +0 -0
- {pro_craft-0.1.15 → pro_craft-0.1.17}/src/pro_craft/database.py +0 -0
- {pro_craft-0.1.15 → pro_craft-0.1.17}/src/pro_craft/log.py +0 -0
- {pro_craft-0.1.15 → pro_craft-0.1.17}/src/pro_craft/prompt_craft/__init__.py +0 -0
- {pro_craft-0.1.15 → pro_craft-0.1.17}/src/pro_craft/prompt_craft/evals.py +0 -0
- {pro_craft-0.1.15 → pro_craft-0.1.17}/src/pro_craft/server/mcp/__init__.py +0 -0
- {pro_craft-0.1.15 → pro_craft-0.1.17}/src/pro_craft/server/router/__init__.py +0 -0
- {pro_craft-0.1.15 → pro_craft-0.1.17}/src/pro_craft/utils.py +0 -0
- {pro_craft-0.1.15 → pro_craft-0.1.17}/src/pro_craft.egg-info/dependency_links.txt +0 -0
- {pro_craft-0.1.15 → pro_craft-0.1.17}/src/pro_craft.egg-info/requires.txt +0 -0
- {pro_craft-0.1.15 → pro_craft-0.1.17}/src/pro_craft.egg-info/top_level.txt +0 -0
- {pro_craft-0.1.15 → pro_craft-0.1.17}/tests/test22.py +0 -0
- {pro_craft-0.1.15 → pro_craft-0.1.17}/tests/test_coder.py +0 -0
- {pro_craft-0.1.15 → pro_craft-0.1.17}/tests/test_designer.py +0 -0
|
@@ -2,9 +2,8 @@
|
|
|
2
2
|
from db_help.qdrant import QdrantManager
|
|
3
3
|
from pydantic import BaseModel
|
|
4
4
|
from datetime import datetime
|
|
5
|
-
from pro_craft.utils import get_adler32_hash, embedding_inputs
|
|
6
|
-
from pro_craft
|
|
7
|
-
from pro_craft.utils import extract_
|
|
5
|
+
from pro_craft.utils import get_adler32_hash, embedding_inputs,extract_
|
|
6
|
+
from pro_craft import Intel
|
|
8
7
|
from enum import Enum
|
|
9
8
|
import json
|
|
10
9
|
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# 测试1
|
|
2
2
|
from pro_craft.utils import extract_
|
|
3
|
-
from pro_craft
|
|
3
|
+
from pro_craft import logger as pro_craft_logger
|
|
4
4
|
from llmada.core import BianXieAdapter, ArkAdapter
|
|
5
5
|
from datetime import datetime
|
|
6
6
|
from enum import Enum
|
|
@@ -68,9 +68,10 @@ class AsyncIntel():
|
|
|
68
68
|
def __init__(self,
|
|
69
69
|
database_url = "",
|
|
70
70
|
model_name = "",
|
|
71
|
+
logger = None,
|
|
71
72
|
):
|
|
72
73
|
database_url = database_url or os.getenv("database_url")
|
|
73
|
-
self.logger = logger or
|
|
74
|
+
self.logger = logger or pro_craft_logger
|
|
74
75
|
try:
|
|
75
76
|
assert database_url
|
|
76
77
|
assert 'aio' in database_url
|
|
@@ -240,7 +241,8 @@ class AsyncIntel():
|
|
|
240
241
|
chat_history = s_prompt or system_result
|
|
241
242
|
await self.save_prompt_increment_version(prompt_id,
|
|
242
243
|
new_prompt = chat_history,
|
|
243
|
-
|
|
244
|
+
use_case = " summary ",
|
|
245
|
+
score = 60,
|
|
244
246
|
session = session)
|
|
245
247
|
|
|
246
248
|
async def prompt_finetune_to_sql(
|
|
@@ -289,7 +291,8 @@ class AsyncIntel():
|
|
|
289
291
|
new_prompt = prompt
|
|
290
292
|
await self.save_prompt_increment_version(prompt_id = prompt_id,
|
|
291
293
|
new_prompt = new_prompt,
|
|
292
|
-
|
|
294
|
+
use_case = " finetune ",
|
|
295
|
+
score = 60,
|
|
293
296
|
session = session)
|
|
294
297
|
|
|
295
298
|
|
|
@@ -343,6 +346,7 @@ class AsyncIntel():
|
|
|
343
346
|
prompt_id = prompt_id,
|
|
344
347
|
new_prompt = "做一些处理",
|
|
345
348
|
use_case = input_,
|
|
349
|
+
score = 60,
|
|
346
350
|
session = session
|
|
347
351
|
)
|
|
348
352
|
ai_result = await self.intellect_remove(input_data = input_data,
|
|
@@ -389,6 +393,7 @@ class AsyncIntel():
|
|
|
389
393
|
chat_history = input_prompt + "\nassistant:\n" + ai_result # 用聊天记录作为完整提示词
|
|
390
394
|
await self.save_prompt_increment_version(prompt_id, chat_history,
|
|
391
395
|
use_case = input_,
|
|
396
|
+
score = 60,
|
|
392
397
|
session = session)
|
|
393
398
|
|
|
394
399
|
elif result_obj.action_type == "summary":
|
|
@@ -507,6 +512,7 @@ class AsyncIntel():
|
|
|
507
512
|
chat_history = input_prompt + "\nassistant:\n" + ai_result # 用聊天记录作为完整提示词
|
|
508
513
|
await self.save_prompt_increment_version(prompt_id, chat_history,
|
|
509
514
|
use_case = input_,
|
|
515
|
+
score = 60,
|
|
510
516
|
session = session)
|
|
511
517
|
|
|
512
518
|
elif result_obj.action_type == "summary":
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
# 测试1
|
|
2
2
|
|
|
3
3
|
from pro_craft.utils import extract_
|
|
4
|
-
from pro_craft
|
|
4
|
+
from pro_craft import logger as pro_craft_logger
|
|
5
5
|
from pro_craft.database import Prompt, UseCase, PromptBase
|
|
6
6
|
from pro_craft.utils import create_session, create_async_session
|
|
7
7
|
from llmada.core import BianXieAdapter, ArkAdapter
|
|
@@ -88,7 +88,7 @@ class IntelNew():
|
|
|
88
88
|
else:
|
|
89
89
|
print('Use BianXieAdapter')
|
|
90
90
|
self.llm = BianXieAdapter()
|
|
91
|
-
self.logger = logger or
|
|
91
|
+
self.logger = logger or pro_craft_logger
|
|
92
92
|
|
|
93
93
|
def _get_latest_prompt_version(self,target_prompt_id,session):
|
|
94
94
|
"""
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
# 测试1
|
|
2
2
|
|
|
3
3
|
from pro_craft.utils import extract_
|
|
4
|
-
from pro_craft
|
|
4
|
+
from pro_craft import logger as pro_craft_logger
|
|
5
5
|
from pro_craft.database import Prompt, UseCase, PromptBase
|
|
6
6
|
from pro_craft.utils import create_session, create_async_session
|
|
7
7
|
from llmada.core import BianXieAdapter, ArkAdapter
|
|
@@ -88,7 +88,7 @@ class Intel():
|
|
|
88
88
|
else:
|
|
89
89
|
print('Use BianXieAdapter')
|
|
90
90
|
self.llm = BianXieAdapter()
|
|
91
|
-
self.logger = logger or
|
|
91
|
+
self.logger = logger or pro_craft_logger
|
|
92
92
|
|
|
93
93
|
def _get_latest_prompt_version(self,target_prompt_id,session):
|
|
94
94
|
"""
|
|
@@ -397,14 +397,10 @@ class Intel():
|
|
|
397
397
|
prompt_id = prompt_id,
|
|
398
398
|
new_prompt = "做一些处理",
|
|
399
399
|
use_case = input_,
|
|
400
|
+
score = 60,
|
|
400
401
|
session = session
|
|
401
402
|
)
|
|
402
|
-
ai_result =
|
|
403
|
-
output_format = output_format,
|
|
404
|
-
prompt_id = prompt_id,
|
|
405
|
-
version = version,
|
|
406
|
-
inference_save_case = inference_save_case
|
|
407
|
-
)
|
|
403
|
+
ai_result = "初始化完成"
|
|
408
404
|
return ai_result
|
|
409
405
|
prompt = result_obj.prompt
|
|
410
406
|
if result_obj.action_type == "inference":
|
|
@@ -1,9 +1,8 @@
|
|
|
1
1
|
|
|
2
2
|
|
|
3
3
|
from fastapi import APIRouter
|
|
4
|
-
from pro_craft
|
|
4
|
+
from pro_craft import Intel
|
|
5
5
|
from pro_craft.utils import create_session
|
|
6
|
-
import os
|
|
7
6
|
|
|
8
7
|
def create_router(database_url: str,
|
|
9
8
|
slave_database_url: str,
|
|
@@ -30,7 +29,8 @@ def create_router(database_url: str,
|
|
|
30
29
|
tags=["prompt"] # 这里使用 Depends 确保每次请求都验证
|
|
31
30
|
)
|
|
32
31
|
|
|
33
|
-
@router.get("/push_order"
|
|
32
|
+
@router.get("/push_order",
|
|
33
|
+
description="可选 train,inference,summary,finetune,patch",)
|
|
34
34
|
async def push_order(demand: str, prompt_id: str, action_type: str = "train"):
|
|
35
35
|
result = intels.push_action_order(
|
|
36
36
|
demand=demand,
|
|
@@ -53,4 +53,25 @@ def create_router(database_url: str,
|
|
|
53
53
|
result = intels.sync_prompt_data_to_database(slave_database_url)
|
|
54
54
|
return {"message": "success","result":result}
|
|
55
55
|
|
|
56
|
+
|
|
57
|
+
@router.get("/roll_back")
|
|
58
|
+
async def roll_back(prompt_id:str,version:str):
|
|
59
|
+
with create_session(intels.engine) as session:
|
|
60
|
+
result = intels.get_prompts_from_sql(
|
|
61
|
+
prompt_id=prompt_id,
|
|
62
|
+
version = version,
|
|
63
|
+
session=session
|
|
64
|
+
)
|
|
65
|
+
assert result.version == version
|
|
66
|
+
intels.save_prompt_increment_version(
|
|
67
|
+
prompt_id = prompt_id,
|
|
68
|
+
new_prompt = result.prompt,
|
|
69
|
+
use_case = "roll_back",
|
|
70
|
+
action_type = "inference",
|
|
71
|
+
demand = "",
|
|
72
|
+
score = 61,
|
|
73
|
+
session = session)
|
|
74
|
+
return {"message": "success","result":result.version}
|
|
75
|
+
|
|
56
76
|
return router
|
|
77
|
+
|
|
@@ -18,7 +18,7 @@ src/pro_craft/prompt_craft/evals.py
|
|
|
18
18
|
src/pro_craft/prompt_craft/new.py
|
|
19
19
|
src/pro_craft/prompt_craft/sync.py
|
|
20
20
|
src/pro_craft/server/mcp/__init__.py
|
|
21
|
-
src/pro_craft/server/mcp/
|
|
21
|
+
src/pro_craft/server/mcp/prompt.py
|
|
22
22
|
src/pro_craft/server/router/__init__.py
|
|
23
23
|
src/pro_craft/server/router/prompt.py
|
|
24
24
|
tests/test22.py
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|