pro-craft 0.1.17__tar.gz → 0.1.19__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of pro-craft might be problematic. Click here for more details.
- {pro_craft-0.1.17 → pro_craft-0.1.19}/PKG-INFO +1 -1
- {pro_craft-0.1.17 → pro_craft-0.1.19}/pyproject.toml +1 -1
- {pro_craft-0.1.17 → pro_craft-0.1.19}/src/pro_craft/prompt_craft/async_.py +34 -21
- {pro_craft-0.1.17 → pro_craft-0.1.19}/src/pro_craft/prompt_craft/sync.py +24 -13
- {pro_craft-0.1.17 → pro_craft-0.1.19}/src/pro_craft/server/router/prompt.py +1 -1
- {pro_craft-0.1.17 → pro_craft-0.1.19}/src/pro_craft.egg-info/PKG-INFO +1 -1
- {pro_craft-0.1.17 → pro_craft-0.1.19}/README.md +0 -0
- {pro_craft-0.1.17 → pro_craft-0.1.19}/setup.cfg +0 -0
- {pro_craft-0.1.17 → pro_craft-0.1.19}/src/pro_craft/__init__.py +0 -0
- {pro_craft-0.1.17 → pro_craft-0.1.19}/src/pro_craft/code_helper/coder.py +0 -0
- {pro_craft-0.1.17 → pro_craft-0.1.19}/src/pro_craft/code_helper/designer.py +0 -0
- {pro_craft-0.1.17 → pro_craft-0.1.19}/src/pro_craft/database.py +0 -0
- {pro_craft-0.1.17 → pro_craft-0.1.19}/src/pro_craft/file_manager.py +0 -0
- {pro_craft-0.1.17 → pro_craft-0.1.19}/src/pro_craft/log.py +0 -0
- {pro_craft-0.1.17 → pro_craft-0.1.19}/src/pro_craft/prompt_craft/__init__.py +0 -0
- {pro_craft-0.1.17 → pro_craft-0.1.19}/src/pro_craft/prompt_craft/evals.py +0 -0
- {pro_craft-0.1.17 → pro_craft-0.1.19}/src/pro_craft/prompt_craft/new.py +0 -0
- {pro_craft-0.1.17 → pro_craft-0.1.19}/src/pro_craft/server/mcp/__init__.py +0 -0
- {pro_craft-0.1.17 → pro_craft-0.1.19}/src/pro_craft/server/mcp/prompt.py +0 -0
- {pro_craft-0.1.17 → pro_craft-0.1.19}/src/pro_craft/server/router/__init__.py +0 -0
- {pro_craft-0.1.17 → pro_craft-0.1.19}/src/pro_craft/utils.py +0 -0
- {pro_craft-0.1.17 → pro_craft-0.1.19}/src/pro_craft.egg-info/SOURCES.txt +0 -0
- {pro_craft-0.1.17 → pro_craft-0.1.19}/src/pro_craft.egg-info/dependency_links.txt +0 -0
- {pro_craft-0.1.17 → pro_craft-0.1.19}/src/pro_craft.egg-info/requires.txt +0 -0
- {pro_craft-0.1.17 → pro_craft-0.1.19}/src/pro_craft.egg-info/top_level.txt +0 -0
- {pro_craft-0.1.17 → pro_craft-0.1.19}/tests/test22.py +0 -0
- {pro_craft-0.1.17 → pro_craft-0.1.19}/tests/test_coder.py +0 -0
- {pro_craft-0.1.17 → pro_craft-0.1.19}/tests/test_designer.py +0 -0
|
@@ -349,12 +349,7 @@ class AsyncIntel():
|
|
|
349
349
|
score = 60,
|
|
350
350
|
session = session
|
|
351
351
|
)
|
|
352
|
-
ai_result =
|
|
353
|
-
output_format = output_format,
|
|
354
|
-
prompt_id = prompt_id,
|
|
355
|
-
version = version,
|
|
356
|
-
inference_save_case = inference_save_case
|
|
357
|
-
)
|
|
352
|
+
ai_result = "初始化完成"
|
|
358
353
|
return ai_result
|
|
359
354
|
|
|
360
355
|
prompt = result_obj.prompt
|
|
@@ -379,22 +374,39 @@ class AsyncIntel():
|
|
|
379
374
|
demand = result_obj.demand
|
|
380
375
|
|
|
381
376
|
|
|
382
|
-
assert demand
|
|
383
|
-
# 注意, 这里的调整要求使用最初的那个输入, 最好一口气调整好
|
|
384
|
-
chat_history = prompt
|
|
385
|
-
if input_ == before_input: # 输入没变, 说明还是针对同一个输入进行讨论
|
|
386
|
-
|
|
387
|
-
|
|
388
|
-
else:
|
|
389
|
-
|
|
390
|
-
|
|
377
|
+
# assert demand
|
|
378
|
+
# # 注意, 这里的调整要求使用最初的那个输入, 最好一口气调整好
|
|
379
|
+
# chat_history = prompt
|
|
380
|
+
# if input_ == before_input: # 输入没变, 说明还是针对同一个输入进行讨论
|
|
381
|
+
# # input_prompt = chat_history + "\nuser:" + demand
|
|
382
|
+
# input_prompt = chat_history + "\nuser:" + demand + output_format
|
|
383
|
+
# else:
|
|
384
|
+
# # input_prompt = chat_history + "\nuser:" + demand + "\n-----input----\n" + input_
|
|
385
|
+
# input_prompt = chat_history + "\nuser:" + demand + output_format + "\n-----input----\n" + input_
|
|
391
386
|
|
|
392
|
-
ai_result = await self.llm.aproduct(input_prompt)
|
|
393
|
-
chat_history = input_prompt + "\nassistant:\n" + ai_result # 用聊天记录作为完整提示词
|
|
394
|
-
await self.save_prompt_increment_version(prompt_id, chat_history,
|
|
395
|
-
|
|
396
|
-
|
|
397
|
-
|
|
387
|
+
# ai_result = await self.llm.aproduct(input_prompt)
|
|
388
|
+
# chat_history = input_prompt + "\nassistant:\n" + ai_result # 用聊天记录作为完整提示词
|
|
389
|
+
# await self.save_prompt_increment_version(prompt_id, chat_history,
|
|
390
|
+
# use_case = input_,
|
|
391
|
+
# score = 60,
|
|
392
|
+
# session = session)
|
|
393
|
+
|
|
394
|
+
if input_ == before_input:
|
|
395
|
+
new_prompt = prompt + "\nuser:" + demand
|
|
396
|
+
else:
|
|
397
|
+
new_prompt = prompt + "\nuser:" + input_
|
|
398
|
+
|
|
399
|
+
ai_result = await self.llm.aproduct(new_prompt + output_format)
|
|
400
|
+
|
|
401
|
+
save_new_prompt = new_prompt + "\nassistant:\n" + ai_result
|
|
402
|
+
|
|
403
|
+
|
|
404
|
+
await self.save_prompt_increment_version(
|
|
405
|
+
prompt_id,
|
|
406
|
+
new_prompt=save_new_prompt,
|
|
407
|
+
use_case = input_,
|
|
408
|
+
score = 60,
|
|
409
|
+
session = session)
|
|
398
410
|
|
|
399
411
|
elif result_obj.action_type == "summary":
|
|
400
412
|
|
|
@@ -514,6 +526,7 @@ class AsyncIntel():
|
|
|
514
526
|
use_case = input_,
|
|
515
527
|
score = 60,
|
|
516
528
|
session = session)
|
|
529
|
+
|
|
517
530
|
|
|
518
531
|
elif result_obj.action_type == "summary":
|
|
519
532
|
|
|
@@ -419,24 +419,35 @@ class Intel():
|
|
|
419
419
|
# 则训练推广
|
|
420
420
|
|
|
421
421
|
# 新版本 默人修改会 inference 状态
|
|
422
|
-
|
|
422
|
+
prompt = result_obj.prompt
|
|
423
423
|
before_input = result_obj.use_case
|
|
424
424
|
demand = result_obj.demand
|
|
425
|
+
|
|
426
|
+
# assert demand
|
|
427
|
+
# # 注意, 这里的调整要求使用最初的那个输入, 最好一口气调整好
|
|
428
|
+
|
|
429
|
+
# if input_ == before_input: # 输入没变, 说明还是针对同一个输入进行讨论
|
|
430
|
+
# # input_prompt = chat_history + "\nuser:" + demand
|
|
431
|
+
# input_prompt = chat_history + "\nuser:" + demand + output_format
|
|
432
|
+
# else:
|
|
433
|
+
# # input_prompt = chat_history + "\nuser:" + demand + "\n-----input----\n" + input_
|
|
434
|
+
# input_prompt = chat_history + "\nuser:" + demand + output_format + "\n-----input----\n" + input_
|
|
425
435
|
|
|
436
|
+
# ai_result = self.llm.product(input_prompt)
|
|
437
|
+
# chat_history = input_prompt + "\nassistant:\n" + ai_result # 用聊天记录作为完整提示词
|
|
426
438
|
|
|
427
|
-
|
|
428
|
-
|
|
429
|
-
chat_history = prompt
|
|
430
|
-
if input_ == before_input: # 输入没变, 说明还是针对同一个输入进行讨论
|
|
431
|
-
# input_prompt = chat_history + "\nuser:" + demand
|
|
432
|
-
input_prompt = chat_history + "\nuser:" + demand + output_format
|
|
439
|
+
if input_ == before_input:
|
|
440
|
+
new_prompt = prompt + "\nuser:" + demand
|
|
433
441
|
else:
|
|
434
|
-
|
|
435
|
-
|
|
436
|
-
|
|
437
|
-
|
|
438
|
-
|
|
439
|
-
|
|
442
|
+
new_prompt = prompt + "\nuser:" + input_
|
|
443
|
+
|
|
444
|
+
ai_result = self.llm.product(new_prompt + output_format)
|
|
445
|
+
|
|
446
|
+
save_new_prompt = new_prompt + "\nassistant:\n" + ai_result
|
|
447
|
+
|
|
448
|
+
|
|
449
|
+
self.save_prompt_increment_version(prompt_id,
|
|
450
|
+
new_prompt=save_new_prompt,
|
|
440
451
|
use_case = input_,
|
|
441
452
|
score = 60,
|
|
442
453
|
session = session)
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|