pro-craft 0.1.21__tar.gz → 0.1.23__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of pro-craft might be problematic. Click here for more details.
- {pro_craft-0.1.21 → pro_craft-0.1.23}/PKG-INFO +1 -1
- {pro_craft-0.1.21 → pro_craft-0.1.23}/pyproject.toml +1 -1
- {pro_craft-0.1.21 → pro_craft-0.1.23}/src/pro_craft/prompt_craft/async_.py +25 -27
- {pro_craft-0.1.21 → pro_craft-0.1.23}/src/pro_craft.egg-info/PKG-INFO +1 -1
- {pro_craft-0.1.21 → pro_craft-0.1.23}/README.md +0 -0
- {pro_craft-0.1.21 → pro_craft-0.1.23}/setup.cfg +0 -0
- {pro_craft-0.1.21 → pro_craft-0.1.23}/src/pro_craft/__init__.py +0 -0
- {pro_craft-0.1.21 → pro_craft-0.1.23}/src/pro_craft/code_helper/coder.py +0 -0
- {pro_craft-0.1.21 → pro_craft-0.1.23}/src/pro_craft/code_helper/designer.py +0 -0
- {pro_craft-0.1.21 → pro_craft-0.1.23}/src/pro_craft/database.py +0 -0
- {pro_craft-0.1.21 → pro_craft-0.1.23}/src/pro_craft/file_manager.py +0 -0
- {pro_craft-0.1.21 → pro_craft-0.1.23}/src/pro_craft/log.py +0 -0
- {pro_craft-0.1.21 → pro_craft-0.1.23}/src/pro_craft/prompt_craft/__init__.py +0 -0
- {pro_craft-0.1.21 → pro_craft-0.1.23}/src/pro_craft/prompt_craft/evals.py +0 -0
- {pro_craft-0.1.21 → pro_craft-0.1.23}/src/pro_craft/prompt_craft/new.py +0 -0
- {pro_craft-0.1.21 → pro_craft-0.1.23}/src/pro_craft/prompt_craft/sync.py +0 -0
- {pro_craft-0.1.21 → pro_craft-0.1.23}/src/pro_craft/server/mcp/__init__.py +0 -0
- {pro_craft-0.1.21 → pro_craft-0.1.23}/src/pro_craft/server/mcp/prompt.py +0 -0
- {pro_craft-0.1.21 → pro_craft-0.1.23}/src/pro_craft/server/router/__init__.py +0 -0
- {pro_craft-0.1.21 → pro_craft-0.1.23}/src/pro_craft/server/router/prompt.py +0 -0
- {pro_craft-0.1.21 → pro_craft-0.1.23}/src/pro_craft/utils.py +0 -0
- {pro_craft-0.1.21 → pro_craft-0.1.23}/src/pro_craft.egg-info/SOURCES.txt +0 -0
- {pro_craft-0.1.21 → pro_craft-0.1.23}/src/pro_craft.egg-info/dependency_links.txt +0 -0
- {pro_craft-0.1.21 → pro_craft-0.1.23}/src/pro_craft.egg-info/requires.txt +0 -0
- {pro_craft-0.1.21 → pro_craft-0.1.23}/src/pro_craft.egg-info/top_level.txt +0 -0
- {pro_craft-0.1.21 → pro_craft-0.1.23}/tests/test22.py +0 -0
- {pro_craft-0.1.21 → pro_craft-0.1.23}/tests/test_coder.py +0 -0
- {pro_craft-0.1.21 → pro_craft-0.1.23}/tests/test_designer.py +0 -0
|
@@ -23,6 +23,9 @@ from json.decoder import JSONDecodeError
|
|
|
23
23
|
class IntellectRemoveFormatError(Exception):
|
|
24
24
|
pass
|
|
25
25
|
|
|
26
|
+
class IntellectRemoveError(Exception):
|
|
27
|
+
pass
|
|
28
|
+
|
|
26
29
|
|
|
27
30
|
def slog(s, target: str = "target",logger = None):
|
|
28
31
|
COLOR_GREEN = "\033[92m"
|
|
@@ -310,8 +313,8 @@ class AsyncIntel():
|
|
|
310
313
|
async with create_async_session(self.engine) as session:
|
|
311
314
|
|
|
312
315
|
latest_prompt = await self.get_prompts_from_sql(prompt_id=prompt_id,session=session)
|
|
313
|
-
|
|
314
|
-
|
|
316
|
+
if latest_prompt:
|
|
317
|
+
await self.save_prompt_increment_version(prompt_id=latest_prompt.prompt_id,
|
|
315
318
|
new_prompt = latest_prompt.prompt,
|
|
316
319
|
use_case = latest_prompt.use_case,
|
|
317
320
|
action_type=action_type,
|
|
@@ -319,8 +322,21 @@ class AsyncIntel():
|
|
|
319
322
|
score=latest_prompt.score,
|
|
320
323
|
session=session
|
|
321
324
|
)
|
|
325
|
+
return "success"
|
|
326
|
+
else:
|
|
327
|
+
await self.save_prompt_increment_version(prompt_id=prompt_id,
|
|
328
|
+
new_prompt = demand,
|
|
329
|
+
use_case = "init",
|
|
330
|
+
action_type="inference",
|
|
331
|
+
demand=demand,
|
|
332
|
+
score=60,
|
|
333
|
+
session=session
|
|
334
|
+
)
|
|
335
|
+
return "init"
|
|
336
|
+
|
|
337
|
+
|
|
338
|
+
|
|
322
339
|
|
|
323
|
-
return "success"
|
|
324
340
|
|
|
325
341
|
|
|
326
342
|
|
|
@@ -330,7 +346,6 @@ class AsyncIntel():
|
|
|
330
346
|
prompt_id: str,
|
|
331
347
|
version: str = None,
|
|
332
348
|
inference_save_case = True,
|
|
333
|
-
push_patch = False,
|
|
334
349
|
):
|
|
335
350
|
if isinstance(input_data,dict):
|
|
336
351
|
input_ = json.dumps(input_data,ensure_ascii=False)
|
|
@@ -340,18 +355,8 @@ class AsyncIntel():
|
|
|
340
355
|
# 查数据库, 获取最新提示词对象
|
|
341
356
|
async with create_async_session(self.engine) as session:
|
|
342
357
|
result_obj = await self.get_prompts_from_sql(prompt_id=prompt_id,session=session)
|
|
343
|
-
|
|
344
|
-
|
|
345
358
|
if result_obj is None:
|
|
346
|
-
|
|
347
|
-
prompt_id = prompt_id,
|
|
348
|
-
new_prompt = "做一些处理",
|
|
349
|
-
use_case = input_,
|
|
350
|
-
score = 60,
|
|
351
|
-
session = session
|
|
352
|
-
)
|
|
353
|
-
ai_result = "初始化完成"
|
|
354
|
-
return ai_result
|
|
359
|
+
raise IntellectRemoveError("不存在的prompt_id")
|
|
355
360
|
|
|
356
361
|
prompt = result_obj.prompt
|
|
357
362
|
if result_obj.action_type == "inference":
|
|
@@ -427,18 +432,15 @@ class AsyncIntel():
|
|
|
427
432
|
)
|
|
428
433
|
ai_result = await self.llm.aproduct(prompt + output_format + "\n-----input----\n" + input_)
|
|
429
434
|
elif result_obj.action_type == "patch":
|
|
430
|
-
|
|
431
435
|
demand = result_obj.demand
|
|
432
436
|
assert demand
|
|
433
|
-
|
|
434
437
|
chat_history = prompt + demand
|
|
435
438
|
ai_result = await self.llm.aproduct(chat_history + output_format + "\n-----input----\n" + input_)
|
|
436
|
-
|
|
437
|
-
|
|
438
|
-
|
|
439
|
-
|
|
440
|
-
|
|
441
|
-
session = session)
|
|
439
|
+
self.save_prompt_increment_version(prompt_id,
|
|
440
|
+
chat_history,
|
|
441
|
+
use_case = input_,
|
|
442
|
+
score = 60,
|
|
443
|
+
session = session)
|
|
442
444
|
|
|
443
445
|
else:
|
|
444
446
|
raise
|
|
@@ -583,7 +585,6 @@ class AsyncIntel():
|
|
|
583
585
|
ExtraFormats: list[object] = [],
|
|
584
586
|
version: str = None,
|
|
585
587
|
inference_save_case = True,
|
|
586
|
-
push_patch = False,
|
|
587
588
|
):
|
|
588
589
|
|
|
589
590
|
base_format_prompt = """
|
|
@@ -601,7 +602,6 @@ class AsyncIntel():
|
|
|
601
602
|
prompt_id=prompt_id,
|
|
602
603
|
version=version,
|
|
603
604
|
inference_save_case=inference_save_case,
|
|
604
|
-
push_patch = push_patch,
|
|
605
605
|
)
|
|
606
606
|
|
|
607
607
|
try:
|
|
@@ -631,7 +631,6 @@ class AsyncIntel():
|
|
|
631
631
|
ExtraFormats: list[object] = [],
|
|
632
632
|
version: str = None,
|
|
633
633
|
inference_save_case = True,
|
|
634
|
-
push_patch = False,
|
|
635
634
|
):
|
|
636
635
|
|
|
637
636
|
async with create_async_session(self.engine) as session:
|
|
@@ -649,7 +648,6 @@ class AsyncIntel():
|
|
|
649
648
|
ExtraFormats = ExtraFormats,
|
|
650
649
|
version = version,
|
|
651
650
|
inference_save_case = inference_save_case,
|
|
652
|
-
push_patch = push_patch
|
|
653
651
|
)
|
|
654
652
|
)
|
|
655
653
|
results = await asyncio.gather(*tasks, return_exceptions=False)
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|