hjxdl 0.0.3__py3-none-any.whl → 0.0.5__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
hdl/__init__.py CHANGED
@@ -0,0 +1 @@
1
+ from ._version import *
hdl/_version.py CHANGED
@@ -12,5 +12,5 @@ __version__: str
12
12
  __version_tuple__: VERSION_TUPLE
13
13
  version_tuple: VERSION_TUPLE
14
14
 
15
- __version__ = version = '0.0.3'
16
- __version_tuple__ = version_tuple = (0, 0, 3)
15
+ __version__ = version = '0.0.5'
16
+ __version_tuple__ = version_tuple = (0, 0, 5)
hdl/utils/llm/chat.py CHANGED
@@ -16,14 +16,15 @@ class GGUF_M():
16
16
  server_ip: str = "127.0.0.1",
17
17
  server_port: int = 8000,
18
18
  ):
19
- """Initialize the model from a local path.
20
- Here it does not manage chat histories.
19
+ """Initialize the model with the provided model path and optional parameters.
21
20
 
22
21
  Args:
23
22
  model_path (str): The path to the model.
24
23
  device (str, optional): The device to use for model initialization. Defaults to 'gpu'.
25
24
  generation_kwargs (dict, optional): Additional keyword arguments for model generation. Defaults to {}.
26
- """
25
+ server_ip (str, optional): The IP address of the server. Defaults to "127.0.0.1".
26
+ server_port (int, optional): The port of the server. Defaults to 8000.
27
+ """
27
28
  # 从本地初始化模型
28
29
  super().__init__()
29
30
  self.generation_kwargs = generation_kwargs
@@ -130,6 +131,15 @@ class OpenAI_M():
130
131
  server_ip: str = "172.28.1.2",
131
132
  server_port: int = 8000,
132
133
  ):
134
+ """Initialize the class with the specified parameters.
135
+
136
+ Args:
137
+ model_path (str, optional): Path to the model file. Defaults to None.
138
+ device (str, optional): Device to run the model on. Defaults to 'gpu'.
139
+ generation_kwargs (dict, optional): Additional keyword arguments for model generation. Defaults to {}.
140
+ server_ip (str, optional): IP address of the server. Defaults to "172.28.1.2".
141
+ server_port (int, optional): Port number of the server. Defaults to 8000.
142
+ """
133
143
  self.model_path = model_path
134
144
  self.server_ip = server_ip
135
145
  self.server_port = server_port
@@ -142,6 +152,16 @@ class OpenAI_M():
142
152
  # history: list = [],
143
153
  **kwargs: t.Any,
144
154
  ) -> str:
155
+ """Invoke the chatbot with the given prompt and return the response.
156
+
157
+ Args:
158
+ prompt (str): The prompt to provide to the chatbot.
159
+ stop (list[str], optional): List of strings that indicate the end of the conversation. Defaults to ["USER:", "ASSISTANT:"].
160
+ **kwargs: Additional keyword arguments to pass to the chatbot.
161
+
162
+ Returns:
163
+ str: The response generated by the chatbot.
164
+ """
145
165
  resp = chat_oai_invoke(
146
166
  base_url=self.base_url,
147
167
  model=self.model_path,
@@ -156,6 +176,16 @@ class OpenAI_M():
156
176
  # history: list = [],
157
177
  **kwargs: t.Any,
158
178
  ) -> str:
179
+ """Generate a response by streaming conversation with the OpenAI chat model.
180
+
181
+ Args:
182
+ prompt (str): The prompt to start the conversation.
183
+ stop (list[str], optional): List of strings that indicate when the conversation should stop. Defaults to ["USER:", "ASSISTANT:"].
184
+ **kwargs: Additional keyword arguments to pass to the chat model.
185
+
186
+ Returns:
187
+ str: The response generated by the chat model.
188
+ """
159
189
  resp = chat_oai_stream(
160
190
  base_url=self.base_url,
161
191
  model=self.model_path,
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: hjxdl
3
- Version: 0.0.3
3
+ Version: 0.0.5
4
4
  Summary: A collection of functions for Jupyter notebooks
5
5
  Home-page: https://github.com/huluxiaohuowa/hdl
6
6
  Author: Jianxing Hu
@@ -1,5 +1,5 @@
1
- hdl/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
2
- hdl/_version.py,sha256=hB095avW4HuDZxn8qPHRG1UMzSSonb8ZDAsLxt9hmk8,411
1
+ hdl/__init__.py,sha256=3bp7HUNfj6gqhNMj_EsyToZqjpqggsv9ohT9Sghk2mA,23
2
+ hdl/_version.py,sha256=EJB7__SNK9kQS_SWZB_U4DHJ3P8ftF6etZEihTYnuXE,411
3
3
  hdl/args/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
4
4
  hdl/args/loss_args.py,sha256=s7YzSdd7IjD24rZvvOrxLLFqMZQb9YylxKeyelSdrTk,70
5
5
  hdl/controllers/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
@@ -84,10 +84,10 @@ hdl/utils/database_tools/connect.py,sha256=KUnVG-8raifEJ_N0b3c8LkTTIfn9NIyw8LX6q
84
84
  hdl/utils/general/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
85
85
  hdl/utils/general/glob.py,sha256=8-RCnt6L297wMIfn34ZAMCsGCZUjHG3MGglGZI1cX0g,491
86
86
  hdl/utils/llm/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
87
- hdl/utils/llm/chat.py,sha256=SoODp3Sf6FNYdajoI39swvPJm64fSd6bFYFdta4xD28,5344
87
+ hdl/utils/llm/chat.py,sha256=UFMv_0KQoGLCim0YGQqneA_HUKzQA9R6K69_rTeBmKM,7022
88
88
  hdl/utils/schedulers/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
89
89
  hdl/utils/schedulers/norm_lr.py,sha256=bDwCmdEK-WkgxQMFBiMuchv8Mm7C0-GZJ6usm-PQk14,4461
90
- hjxdl-0.0.3.dist-info/METADATA,sha256=IMDR9eTW1PeEGxs6eZESV2ys8LsI0j7aZp6qO9eXZWg,525
91
- hjxdl-0.0.3.dist-info/WHEEL,sha256=y4mX-SOX4fYIkonsAGA5N0Oy-8_gI4FXw5HNI1xqvWg,91
92
- hjxdl-0.0.3.dist-info/top_level.txt,sha256=-kxwTM5JPhylp06z3zAVO3w6_h7wtBfBo2zgM6YZoTk,4
93
- hjxdl-0.0.3.dist-info/RECORD,,
90
+ hjxdl-0.0.5.dist-info/METADATA,sha256=HsgmiTeotSLmNOMiDXsjrI7puQ3o-aoB0p4aItOhf38,525
91
+ hjxdl-0.0.5.dist-info/WHEEL,sha256=y4mX-SOX4fYIkonsAGA5N0Oy-8_gI4FXw5HNI1xqvWg,91
92
+ hjxdl-0.0.5.dist-info/top_level.txt,sha256=-kxwTM5JPhylp06z3zAVO3w6_h7wtBfBo2zgM6YZoTk,4
93
+ hjxdl-0.0.5.dist-info/RECORD,,
File without changes