llm-dialog-manager 0.4.5__py3-none-any.whl → 0.4.6__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,4 +1,4 @@
1
1
  from .chat_history import ChatHistory
2
2
  from .agent import Agent
3
3
 
4
- __version__ = "0.4.5"
4
+ __version__ = "0.4.6"
@@ -394,16 +394,21 @@ def completion(model: str, messages: List[Dict[str, Union[str, List[Union[str, I
394
394
  if model.endswith("-openai"):
395
395
  model = model[:-7] # Remove last 7 characters ("-openai")
396
396
  client = openai.OpenAI(api_key=api_key, base_url=base_url)
397
- # Set response_format based on json_format
398
- response_format = {"type": "json_object"} if json_format else {"type": "plain_text"}
399
-
400
- response = client.chat.completions.create(
401
- model=model,
402
- messages=formatted_messages,
403
- max_tokens=max_tokens,
404
- temperature=temperature,
405
- response_format=response_format # Added response_format
406
- )
397
+
398
+ # Create base parameters
399
+ params = {
400
+ "model": model,
401
+ "messages": formatted_messages,
402
+ }
403
+
404
+ # Add optional parameters
405
+ if json_format:
406
+ params["response_format"] = {"type": "json_object"}
407
+ if not ("o1" in model or "o3" in model):
408
+ params["max_tokens"] = max_tokens
409
+ params["temperature"] = temperature
410
+
411
+ response = client.chat.completions.create(**params)
407
412
  return response.choices[0].message.content
408
413
 
409
414
  # Release the API key after successful use
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.2
2
2
  Name: llm_dialog_manager
3
- Version: 0.4.5
3
+ Version: 0.4.6
4
4
  Summary: A Python package for managing LLM chat conversation history
5
5
  Author-email: xihajun <work@2333.fun>
6
6
  License: MIT
@@ -0,0 +1,9 @@
1
+ llm_dialog_manager/__init__.py,sha256=klLFvHayR7ew1Oh9xyhAXdXnfs82YnFUEFzw0YvxKJI,86
2
+ llm_dialog_manager/agent.py,sha256=NVQKIMebl4cYkqMaBceZ3qs1vYhq1bum9okAn8VcfCg,27680
3
+ llm_dialog_manager/chat_history.py,sha256=DKKRnj_M6h-4JncnH6KekMTghX7vMgdN3J9uOwXKzMU,10347
4
+ llm_dialog_manager/key_manager.py,sha256=shvxmn4zUtQx_p-x1EFyOmnk-WlhigbpKtxTKve-zXk,4421
5
+ llm_dialog_manager-0.4.6.dist-info/LICENSE,sha256=vWGbYgGuWpWrXL8-xi6pNcX5UzD6pWoIAZmcetyfbus,1064
6
+ llm_dialog_manager-0.4.6.dist-info/METADATA,sha256=-qTRYkfAJMJCQTkRqNrtHjUuN-xGLhLR4CJvSJURgeg,4194
7
+ llm_dialog_manager-0.4.6.dist-info/WHEEL,sha256=In9FTNxeP60KnTkGw7wk6mJPYd_dQSjEZmXdBdMCI-8,91
8
+ llm_dialog_manager-0.4.6.dist-info/top_level.txt,sha256=u2EQEXW0NGAt0AAHT7jx1odXZ4rZfjcgbmJhvKFuMkI,19
9
+ llm_dialog_manager-0.4.6.dist-info/RECORD,,
@@ -1,9 +0,0 @@
1
- llm_dialog_manager/__init__.py,sha256=Vai6V5cLO8X7dkHgnALzKyhtDYdPhkd9oD0HPjzdqZ8,86
2
- llm_dialog_manager/agent.py,sha256=cYK3SMa1zPeSP4QoSxfceQ3YLQYqBTNwDE_ZJdBi7pM,27595
3
- llm_dialog_manager/chat_history.py,sha256=DKKRnj_M6h-4JncnH6KekMTghX7vMgdN3J9uOwXKzMU,10347
4
- llm_dialog_manager/key_manager.py,sha256=shvxmn4zUtQx_p-x1EFyOmnk-WlhigbpKtxTKve-zXk,4421
5
- llm_dialog_manager-0.4.5.dist-info/LICENSE,sha256=vWGbYgGuWpWrXL8-xi6pNcX5UzD6pWoIAZmcetyfbus,1064
6
- llm_dialog_manager-0.4.5.dist-info/METADATA,sha256=TbPdPjwrncfdYIJjFTzOoaWp5p90zYzpzsazVSK3F_8,4194
7
- llm_dialog_manager-0.4.5.dist-info/WHEEL,sha256=In9FTNxeP60KnTkGw7wk6mJPYd_dQSjEZmXdBdMCI-8,91
8
- llm_dialog_manager-0.4.5.dist-info/top_level.txt,sha256=u2EQEXW0NGAt0AAHT7jx1odXZ4rZfjcgbmJhvKFuMkI,19
9
- llm_dialog_manager-0.4.5.dist-info/RECORD,,