llm-ie 0.4.2__tar.gz → 0.4.4__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- llm_ie-0.4.2/README.md → llm_ie-0.4.4/PKG-INFO +49 -1
- llm_ie-0.4.2/PKG-INFO → llm_ie-0.4.4/README.md +32 -17
- {llm_ie-0.4.2 → llm_ie-0.4.4}/pyproject.toml +2 -1
- {llm_ie-0.4.2 → llm_ie-0.4.4}/src/llm_ie/__init__.py +2 -2
- {llm_ie-0.4.2 → llm_ie-0.4.4}/src/llm_ie/engines.py +83 -3
- {llm_ie-0.4.2 → llm_ie-0.4.4}/src/llm_ie/prompt_editor.py +11 -11
- {llm_ie-0.4.2 → llm_ie-0.4.4}/src/llm_ie/asset/PromptEditor_prompts/chat.txt +0 -0
- {llm_ie-0.4.2 → llm_ie-0.4.4}/src/llm_ie/asset/PromptEditor_prompts/comment.txt +0 -0
- {llm_ie-0.4.2 → llm_ie-0.4.4}/src/llm_ie/asset/PromptEditor_prompts/rewrite.txt +0 -0
- {llm_ie-0.4.2 → llm_ie-0.4.4}/src/llm_ie/asset/PromptEditor_prompts/system.txt +0 -0
- {llm_ie-0.4.2 → llm_ie-0.4.4}/src/llm_ie/asset/default_prompts/ReviewFrameExtractor_addition_review_prompt.txt +0 -0
- {llm_ie-0.4.2 → llm_ie-0.4.4}/src/llm_ie/asset/default_prompts/ReviewFrameExtractor_revision_review_prompt.txt +0 -0
- {llm_ie-0.4.2 → llm_ie-0.4.4}/src/llm_ie/asset/default_prompts/SentenceReviewFrameExtractor_addition_review_prompt.txt +0 -0
- {llm_ie-0.4.2 → llm_ie-0.4.4}/src/llm_ie/asset/default_prompts/SentenceReviewFrameExtractor_revision_review_prompt.txt +0 -0
- {llm_ie-0.4.2 → llm_ie-0.4.4}/src/llm_ie/asset/prompt_guide/BasicFrameExtractor_prompt_guide.txt +0 -0
- {llm_ie-0.4.2 → llm_ie-0.4.4}/src/llm_ie/asset/prompt_guide/BinaryRelationExtractor_prompt_guide.txt +0 -0
- {llm_ie-0.4.2 → llm_ie-0.4.4}/src/llm_ie/asset/prompt_guide/MultiClassRelationExtractor_prompt_guide.txt +0 -0
- {llm_ie-0.4.2 → llm_ie-0.4.4}/src/llm_ie/asset/prompt_guide/ReviewFrameExtractor_prompt_guide.txt +0 -0
- {llm_ie-0.4.2 → llm_ie-0.4.4}/src/llm_ie/asset/prompt_guide/SentenceCoTFrameExtractor_prompt_guide.txt +0 -0
- {llm_ie-0.4.2 → llm_ie-0.4.4}/src/llm_ie/asset/prompt_guide/SentenceFrameExtractor_prompt_guide.txt +0 -0
- {llm_ie-0.4.2 → llm_ie-0.4.4}/src/llm_ie/asset/prompt_guide/SentenceReviewFrameExtractor_prompt_guide.txt +0 -0
- {llm_ie-0.4.2 → llm_ie-0.4.4}/src/llm_ie/data_types.py +0 -0
- {llm_ie-0.4.2 → llm_ie-0.4.4}/src/llm_ie/extractors.py +0 -0
|
@@ -1,3 +1,20 @@
|
|
|
1
|
+
Metadata-Version: 2.1
|
|
2
|
+
Name: llm-ie
|
|
3
|
+
Version: 0.4.4
|
|
4
|
+
Summary: An LLM-powered tool that transforms everyday language into robust information extraction pipelines.
|
|
5
|
+
License: MIT
|
|
6
|
+
Author: Enshuo (David) Hsu
|
|
7
|
+
Requires-Python: >=3.11,<4.0
|
|
8
|
+
Classifier: License :: OSI Approved :: MIT License
|
|
9
|
+
Classifier: Programming Language :: Python :: 3
|
|
10
|
+
Classifier: Programming Language :: Python :: 3.11
|
|
11
|
+
Classifier: Programming Language :: Python :: 3.12
|
|
12
|
+
Requires-Dist: colorama (>=0.4.6,<0.5.0)
|
|
13
|
+
Requires-Dist: json_repair (>=0.30,<0.31)
|
|
14
|
+
Requires-Dist: nest_asyncio (>=1.6.0,<2.0.0)
|
|
15
|
+
Requires-Dist: nltk (>=3.8,<4.0)
|
|
16
|
+
Description-Content-Type: text/markdown
|
|
17
|
+
|
|
1
18
|
<div align="center"><img src=doc_asset/readme_img/LLM-IE.png width=500 ></div>
|
|
2
19
|
|
|
3
20
|

|
|
@@ -23,6 +40,7 @@ An LLM-powered tool that transforms everyday language into robust information ex
|
|
|
23
40
|
- Concurrent LLM inferencing to speed up frame and relation extraction.
|
|
24
41
|
- Support for LiteLLM.
|
|
25
42
|
- [v0.4.1](https://github.com/daviden1013/llm-ie/releases/tag/v0.4.1) (Jan 25, 2025): Added filters, table view, and some new features to visualization tool (make sure to update [ie-viz](https://github.com/daviden1013/ie-viz)).
|
|
43
|
+
- [v0.4.3](https://github.com/daviden1013/llm-ie/releases/tag/v0.4.3) (Feb 7, 2025): Added Azure OpenAI support.
|
|
26
44
|
|
|
27
45
|
## Table of Contents
|
|
28
46
|
- [Overview](#overview)
|
|
@@ -83,6 +101,20 @@ inference_engine = OpenAIInferenceEngine(model="gpt-4o-mini")
|
|
|
83
101
|
```
|
|
84
102
|
</details>
|
|
85
103
|
|
|
104
|
+
<details>
|
|
105
|
+
<summary><img src=doc_asset/readme_img/Azure_icon.png width=32 /> Azure OpenAI API</summary>
|
|
106
|
+
|
|
107
|
+
Follow the [Azure AI Services Quickstart](https://learn.microsoft.com/en-us/azure/ai-services/openai/quickstart?tabs=command-line%2Ckeyless%2Ctypescript-keyless%2Cpython-new&pivots=programming-language-python) to set up Endpoint and API key.
|
|
108
|
+
|
|
109
|
+
```python
|
|
110
|
+
from llm_ie.engines import AzureOpenAIInferenceEngine
|
|
111
|
+
|
|
112
|
+
inference_engine = AzureOpenAIInferenceEngine(model="gpt-4o-mini",
|
|
113
|
+
api_version="<your api version>")
|
|
114
|
+
```
|
|
115
|
+
|
|
116
|
+
</details>
|
|
117
|
+
|
|
86
118
|
<details>
|
|
87
119
|
<summary>🤗 Huggingface_hub</summary>
|
|
88
120
|
|
|
@@ -308,6 +340,22 @@ from llm_ie.engines import OpenAIInferenceEngine
|
|
|
308
340
|
inference_engine = OpenAIInferenceEngine(model="gpt-4o-mini")
|
|
309
341
|
```
|
|
310
342
|
|
|
343
|
+
#### <img src=doc_asset/readme_img/Azure_icon.png width=32 /> Azure OpenAI API
|
|
344
|
+
In bash, save the endpoint name and API key to environmental variables `AZURE_OPENAI_ENDPOINT` and `AZURE_OPENAI_API_KEY`.
|
|
345
|
+
```
|
|
346
|
+
export AZURE_OPENAI_API_KEY="<your_API_key>"
|
|
347
|
+
export AZURE_OPENAI_ENDPOINT="<your_endpoint>"
|
|
348
|
+
```
|
|
349
|
+
|
|
350
|
+
In Python, create inference engine and specify model name. For the available models, refer to [OpenAI webpage](https://platform.openai.com/docs/models).
|
|
351
|
+
For more parameters, see [Azure OpenAI reference](https://learn.microsoft.com/en-us/azure/ai-services/openai/quickstart).
|
|
352
|
+
|
|
353
|
+
```python
|
|
354
|
+
from llm_ie.engines import AzureOpenAIInferenceEngine
|
|
355
|
+
|
|
356
|
+
inference_engine = AzureOpenAIInferenceEngine(model="gpt-4o-mini")
|
|
357
|
+
```
|
|
358
|
+
|
|
311
359
|
#### 🤗 huggingface_hub
|
|
312
360
|
The ```model``` can be a model id hosted on the Hugging Face Hub or a URL to a deployed Inference Endpoint. Refer to the [Inference Client](https://huggingface.co/docs/huggingface_hub/en/package_reference/inference_client) documentation for more details.
|
|
313
361
|
|
|
@@ -1130,4 +1178,4 @@ For more information and benchmarks, please check our paper:
|
|
|
1130
1178
|
journal={arXiv preprint arXiv:2411.11779},
|
|
1131
1179
|
year={2024}
|
|
1132
1180
|
}
|
|
1133
|
-
```
|
|
1181
|
+
```
|
|
@@ -1,19 +1,3 @@
|
|
|
1
|
-
Metadata-Version: 2.1
|
|
2
|
-
Name: llm-ie
|
|
3
|
-
Version: 0.4.2
|
|
4
|
-
Summary: An LLM-powered tool that transforms everyday language into robust information extraction pipelines.
|
|
5
|
-
License: MIT
|
|
6
|
-
Author: Enshuo (David) Hsu
|
|
7
|
-
Requires-Python: >=3.11,<4.0
|
|
8
|
-
Classifier: License :: OSI Approved :: MIT License
|
|
9
|
-
Classifier: Programming Language :: Python :: 3
|
|
10
|
-
Classifier: Programming Language :: Python :: 3.11
|
|
11
|
-
Classifier: Programming Language :: Python :: 3.12
|
|
12
|
-
Requires-Dist: colorama (>=0.4.6,<0.5.0)
|
|
13
|
-
Requires-Dist: json_repair (>=0.30,<0.31)
|
|
14
|
-
Requires-Dist: nltk (>=3.8,<4.0)
|
|
15
|
-
Description-Content-Type: text/markdown
|
|
16
|
-
|
|
17
1
|
<div align="center"><img src=doc_asset/readme_img/LLM-IE.png width=500 ></div>
|
|
18
2
|
|
|
19
3
|

|
|
@@ -39,6 +23,7 @@ An LLM-powered tool that transforms everyday language into robust information ex
|
|
|
39
23
|
- Concurrent LLM inferencing to speed up frame and relation extraction.
|
|
40
24
|
- Support for LiteLLM.
|
|
41
25
|
- [v0.4.1](https://github.com/daviden1013/llm-ie/releases/tag/v0.4.1) (Jan 25, 2025): Added filters, table view, and some new features to visualization tool (make sure to update [ie-viz](https://github.com/daviden1013/ie-viz)).
|
|
26
|
+
- [v0.4.3](https://github.com/daviden1013/llm-ie/releases/tag/v0.4.3) (Feb 7, 2025): Added Azure OpenAI support.
|
|
42
27
|
|
|
43
28
|
## Table of Contents
|
|
44
29
|
- [Overview](#overview)
|
|
@@ -99,6 +84,20 @@ inference_engine = OpenAIInferenceEngine(model="gpt-4o-mini")
|
|
|
99
84
|
```
|
|
100
85
|
</details>
|
|
101
86
|
|
|
87
|
+
<details>
|
|
88
|
+
<summary><img src=doc_asset/readme_img/Azure_icon.png width=32 /> Azure OpenAI API</summary>
|
|
89
|
+
|
|
90
|
+
Follow the [Azure AI Services Quickstart](https://learn.microsoft.com/en-us/azure/ai-services/openai/quickstart?tabs=command-line%2Ckeyless%2Ctypescript-keyless%2Cpython-new&pivots=programming-language-python) to set up Endpoint and API key.
|
|
91
|
+
|
|
92
|
+
```python
|
|
93
|
+
from llm_ie.engines import AzureOpenAIInferenceEngine
|
|
94
|
+
|
|
95
|
+
inference_engine = AzureOpenAIInferenceEngine(model="gpt-4o-mini",
|
|
96
|
+
api_version="<your api version>")
|
|
97
|
+
```
|
|
98
|
+
|
|
99
|
+
</details>
|
|
100
|
+
|
|
102
101
|
<details>
|
|
103
102
|
<summary>🤗 Huggingface_hub</summary>
|
|
104
103
|
|
|
@@ -324,6 +323,22 @@ from llm_ie.engines import OpenAIInferenceEngine
|
|
|
324
323
|
inference_engine = OpenAIInferenceEngine(model="gpt-4o-mini")
|
|
325
324
|
```
|
|
326
325
|
|
|
326
|
+
#### <img src=doc_asset/readme_img/Azure_icon.png width=32 /> Azure OpenAI API
|
|
327
|
+
In bash, save the endpoint name and API key to environmental variables `AZURE_OPENAI_ENDPOINT` and `AZURE_OPENAI_API_KEY`.
|
|
328
|
+
```
|
|
329
|
+
export AZURE_OPENAI_API_KEY="<your_API_key>"
|
|
330
|
+
export AZURE_OPENAI_ENDPOINT="<your_endpoint>"
|
|
331
|
+
```
|
|
332
|
+
|
|
333
|
+
In Python, create inference engine and specify model name. For the available models, refer to [OpenAI webpage](https://platform.openai.com/docs/models).
|
|
334
|
+
For more parameters, see [Azure OpenAI reference](https://learn.microsoft.com/en-us/azure/ai-services/openai/quickstart).
|
|
335
|
+
|
|
336
|
+
```python
|
|
337
|
+
from llm_ie.engines import AzureOpenAIInferenceEngine
|
|
338
|
+
|
|
339
|
+
inference_engine = AzureOpenAIInferenceEngine(model="gpt-4o-mini")
|
|
340
|
+
```
|
|
341
|
+
|
|
327
342
|
#### 🤗 huggingface_hub
|
|
328
343
|
The ```model``` can be a model id hosted on the Hugging Face Hub or a URL to a deployed Inference Endpoint. Refer to the [Inference Client](https://huggingface.co/docs/huggingface_hub/en/package_reference/inference_client) documentation for more details.
|
|
329
344
|
|
|
@@ -1146,4 +1161,4 @@ For more information and benchmarks, please check our paper:
|
|
|
1146
1161
|
journal={arXiv preprint arXiv:2411.11779},
|
|
1147
1162
|
year={2024}
|
|
1148
1163
|
}
|
|
1149
|
-
```
|
|
1164
|
+
```
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
[tool.poetry]
|
|
2
2
|
name = "llm-ie"
|
|
3
|
-
version = "0.4.
|
|
3
|
+
version = "0.4.4"
|
|
4
4
|
description = "An LLM-powered tool that transforms everyday language into robust information extraction pipelines."
|
|
5
5
|
authors = ["Enshuo (David) Hsu"]
|
|
6
6
|
license = "MIT"
|
|
@@ -16,6 +16,7 @@ python = "^3.11"
|
|
|
16
16
|
nltk = "^3.8"
|
|
17
17
|
colorama = "^0.4.6"
|
|
18
18
|
json_repair = "^0.30"
|
|
19
|
+
nest_asyncio = "^1.6.0"
|
|
19
20
|
|
|
20
21
|
|
|
21
22
|
[build-system]
|
|
@@ -1,9 +1,9 @@
|
|
|
1
1
|
from .data_types import LLMInformationExtractionFrame, LLMInformationExtractionDocument
|
|
2
|
-
from .engines import LlamaCppInferenceEngine, OllamaInferenceEngine, HuggingFaceHubInferenceEngine, OpenAIInferenceEngine, LiteLLMInferenceEngine
|
|
2
|
+
from .engines import LlamaCppInferenceEngine, OllamaInferenceEngine, HuggingFaceHubInferenceEngine, OpenAIInferenceEngine, AzureOpenAIInferenceEngine, LiteLLMInferenceEngine
|
|
3
3
|
from .extractors import BasicFrameExtractor, ReviewFrameExtractor, SentenceFrameExtractor, SentenceReviewFrameExtractor, SentenceCoTFrameExtractor, BinaryRelationExtractor, MultiClassRelationExtractor
|
|
4
4
|
from .prompt_editor import PromptEditor
|
|
5
5
|
|
|
6
6
|
__all__ = ["LLMInformationExtractionFrame", "LLMInformationExtractionDocument",
|
|
7
|
-
"LlamaCppInferenceEngine", "OllamaInferenceEngine", "HuggingFaceHubInferenceEngine", "OpenAIInferenceEngine", "LiteLLMInferenceEngine",
|
|
7
|
+
"LlamaCppInferenceEngine", "OllamaInferenceEngine", "HuggingFaceHubInferenceEngine", "OpenAIInferenceEngine", "AzureOpenAIInferenceEngine", "LiteLLMInferenceEngine",
|
|
8
8
|
"BasicFrameExtractor", "ReviewFrameExtractor", "SentenceFrameExtractor", "SentenceReviewFrameExtractor", "SentenceCoTFrameExtractor", "BinaryRelationExtractor", "MultiClassRelationExtractor",
|
|
9
9
|
"PromptEditor"]
|
|
@@ -290,9 +290,88 @@ class OpenAIInferenceEngine(InferenceEngine):
|
|
|
290
290
|
if stream:
|
|
291
291
|
res = ''
|
|
292
292
|
for chunk in response:
|
|
293
|
-
if chunk.choices
|
|
294
|
-
|
|
295
|
-
|
|
293
|
+
if len(chunk.choices) > 0:
|
|
294
|
+
if chunk.choices[0].delta.content is not None:
|
|
295
|
+
res += chunk.choices[0].delta.content
|
|
296
|
+
print(chunk.choices[0].delta.content, end="", flush=True)
|
|
297
|
+
return res
|
|
298
|
+
|
|
299
|
+
return response.choices[0].message.content
|
|
300
|
+
|
|
301
|
+
|
|
302
|
+
async def chat_async(self, messages:List[Dict[str,str]], max_new_tokens:int=2048, temperature:float=0.0, **kwrs) -> str:
|
|
303
|
+
"""
|
|
304
|
+
Async version of chat method. Streaming is not supported.
|
|
305
|
+
"""
|
|
306
|
+
response = await self.async_client.chat.completions.create(
|
|
307
|
+
model=self.model,
|
|
308
|
+
messages=messages,
|
|
309
|
+
max_tokens=max_new_tokens,
|
|
310
|
+
temperature=temperature,
|
|
311
|
+
stream=False,
|
|
312
|
+
**kwrs
|
|
313
|
+
)
|
|
314
|
+
|
|
315
|
+
return response.choices[0].message.content
|
|
316
|
+
|
|
317
|
+
|
|
318
|
+
class AzureOpenAIInferenceEngine(InferenceEngine):
|
|
319
|
+
def __init__(self, model:str, api_version:str, **kwrs):
|
|
320
|
+
"""
|
|
321
|
+
The Azure OpenAI API inference engine.
|
|
322
|
+
For parameters and documentation, refer to
|
|
323
|
+
- https://azure.microsoft.com/en-us/products/ai-services/openai-service
|
|
324
|
+
- https://learn.microsoft.com/en-us/azure/ai-services/openai/quickstart
|
|
325
|
+
|
|
326
|
+
Parameters:
|
|
327
|
+
----------
|
|
328
|
+
model : str
|
|
329
|
+
model name as described in https://platform.openai.com/docs/models
|
|
330
|
+
api_version : str
|
|
331
|
+
the Azure OpenAI API version
|
|
332
|
+
"""
|
|
333
|
+
if importlib.util.find_spec("openai") is None:
|
|
334
|
+
raise ImportError("OpenAI Python API library not found. Please install OpanAI (```pip install openai```).")
|
|
335
|
+
|
|
336
|
+
from openai import AzureOpenAI, AsyncAzureOpenAI
|
|
337
|
+
self.model = model
|
|
338
|
+
self.api_version = api_version
|
|
339
|
+
self.client = AzureOpenAI(api_version=self.api_version,
|
|
340
|
+
**kwrs)
|
|
341
|
+
self.async_client = AsyncAzureOpenAI(api_version=self.api_version,
|
|
342
|
+
**kwrs)
|
|
343
|
+
|
|
344
|
+
def chat(self, messages:List[Dict[str,str]], max_new_tokens:int=2048, temperature:float=0.0, stream:bool=False, **kwrs) -> str:
|
|
345
|
+
"""
|
|
346
|
+
This method inputs chat messages and outputs LLM generated text.
|
|
347
|
+
|
|
348
|
+
Parameters:
|
|
349
|
+
----------
|
|
350
|
+
messages : List[Dict[str,str]]
|
|
351
|
+
a list of dict with role and content. role must be one of {"system", "user", "assistant"}
|
|
352
|
+
max_new_tokens : str, Optional
|
|
353
|
+
the max number of new tokens LLM can generate.
|
|
354
|
+
temperature : float, Optional
|
|
355
|
+
the temperature for token sampling.
|
|
356
|
+
stream : bool, Optional
|
|
357
|
+
if True, LLM generated text will be printed in terminal in real-time.
|
|
358
|
+
"""
|
|
359
|
+
response = self.client.chat.completions.create(
|
|
360
|
+
model=self.model,
|
|
361
|
+
messages=messages,
|
|
362
|
+
max_tokens=max_new_tokens,
|
|
363
|
+
temperature=temperature,
|
|
364
|
+
stream=stream,
|
|
365
|
+
**kwrs
|
|
366
|
+
)
|
|
367
|
+
|
|
368
|
+
if stream:
|
|
369
|
+
res = ''
|
|
370
|
+
for chunk in response:
|
|
371
|
+
if len(chunk.choices) > 0:
|
|
372
|
+
if chunk.choices[0].delta.content is not None:
|
|
373
|
+
res += chunk.choices[0].delta.content
|
|
374
|
+
print(chunk.choices[0].delta.content, end="", flush=True)
|
|
296
375
|
return res
|
|
297
376
|
|
|
298
377
|
return response.choices[0].message.content
|
|
@@ -312,6 +391,7 @@ class OpenAIInferenceEngine(InferenceEngine):
|
|
|
312
391
|
)
|
|
313
392
|
|
|
314
393
|
return response.choices[0].message.content
|
|
394
|
+
|
|
315
395
|
|
|
316
396
|
class LiteLLMInferenceEngine(InferenceEngine):
|
|
317
397
|
def __init__(self, model:str=None, base_url:str=None, api_key:str=None):
|
|
@@ -67,7 +67,7 @@ class PromptEditor:
|
|
|
67
67
|
return prompt
|
|
68
68
|
|
|
69
69
|
|
|
70
|
-
def rewrite(self, draft:str) -> str:
|
|
70
|
+
def rewrite(self, draft:str, **kwrs) -> str:
|
|
71
71
|
"""
|
|
72
72
|
This method inputs a prompt draft and rewrites it following the extractor's guideline.
|
|
73
73
|
"""
|
|
@@ -79,10 +79,10 @@ class PromptEditor:
|
|
|
79
79
|
prompt_template=rewrite_prompt_template)
|
|
80
80
|
messages = [{"role": "system", "content": self.system_prompt},
|
|
81
81
|
{"role": "user", "content": prompt}]
|
|
82
|
-
res = self.inference_engine.chat(messages, stream=True)
|
|
82
|
+
res = self.inference_engine.chat(messages, stream=True, **kwrs)
|
|
83
83
|
return res
|
|
84
84
|
|
|
85
|
-
def comment(self, draft:str) -> str:
|
|
85
|
+
def comment(self, draft:str, **kwrs) -> str:
|
|
86
86
|
"""
|
|
87
87
|
This method inputs a prompt draft and comment following the extractor's guideline.
|
|
88
88
|
"""
|
|
@@ -94,11 +94,11 @@ class PromptEditor:
|
|
|
94
94
|
prompt_template=comment_prompt_template)
|
|
95
95
|
messages = [{"role": "system", "content": self.system_prompt},
|
|
96
96
|
{"role": "user", "content": prompt}]
|
|
97
|
-
res = self.inference_engine.chat(messages, stream=True)
|
|
97
|
+
res = self.inference_engine.chat(messages, stream=True, **kwrs)
|
|
98
98
|
return res
|
|
99
99
|
|
|
100
100
|
|
|
101
|
-
def _terminal_chat(self):
|
|
101
|
+
def _terminal_chat(self, **kwrs):
|
|
102
102
|
"""
|
|
103
103
|
This method runs an interactive chat session in the terminal to help users write prompt templates.
|
|
104
104
|
"""
|
|
@@ -126,11 +126,11 @@ class PromptEditor:
|
|
|
126
126
|
# Chat
|
|
127
127
|
messages.append({"role": "user", "content": user_input})
|
|
128
128
|
print(f"{Fore.BLUE}Assistant: {Style.RESET_ALL}", end="")
|
|
129
|
-
response = self.inference_engine.chat(messages, stream=True)
|
|
129
|
+
response = self.inference_engine.chat(messages, stream=True, **kwrs)
|
|
130
130
|
messages.append({"role": "assistant", "content": response})
|
|
131
131
|
|
|
132
132
|
|
|
133
|
-
def _IPython_chat(self):
|
|
133
|
+
def _IPython_chat(self, **kwrs):
|
|
134
134
|
"""
|
|
135
135
|
This method runs an interactive chat session in Jupyter/IPython using ipywidgets to help users write prompt templates.
|
|
136
136
|
"""
|
|
@@ -186,7 +186,7 @@ class PromptEditor:
|
|
|
186
186
|
|
|
187
187
|
# Get assistant's response and append it to conversation
|
|
188
188
|
print("Assistant: ", end="")
|
|
189
|
-
response = self.inference_engine.chat(messages, stream=True)
|
|
189
|
+
response = self.inference_engine.chat(messages, stream=True, **kwrs)
|
|
190
190
|
messages.append({"role": "assistant", "content": response})
|
|
191
191
|
|
|
192
192
|
# Display the assistant's response
|
|
@@ -200,11 +200,11 @@ class PromptEditor:
|
|
|
200
200
|
display(input_box)
|
|
201
201
|
display(output_area)
|
|
202
202
|
|
|
203
|
-
def chat(self):
|
|
203
|
+
def chat(self, **kwrs):
|
|
204
204
|
"""
|
|
205
205
|
External method that detects the environment and calls the appropriate chat method.
|
|
206
206
|
"""
|
|
207
207
|
if 'ipykernel' in sys.modules:
|
|
208
|
-
self._IPython_chat()
|
|
208
|
+
self._IPython_chat(**kwrs)
|
|
209
209
|
else:
|
|
210
|
-
self._terminal_chat()
|
|
210
|
+
self._terminal_chat(**kwrs)
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
{llm_ie-0.4.2 → llm_ie-0.4.4}/src/llm_ie/asset/prompt_guide/BasicFrameExtractor_prompt_guide.txt
RENAMED
|
File without changes
|
{llm_ie-0.4.2 → llm_ie-0.4.4}/src/llm_ie/asset/prompt_guide/BinaryRelationExtractor_prompt_guide.txt
RENAMED
|
File without changes
|
|
File without changes
|
{llm_ie-0.4.2 → llm_ie-0.4.4}/src/llm_ie/asset/prompt_guide/ReviewFrameExtractor_prompt_guide.txt
RENAMED
|
File without changes
|
|
File without changes
|
{llm_ie-0.4.2 → llm_ie-0.4.4}/src/llm_ie/asset/prompt_guide/SentenceFrameExtractor_prompt_guide.txt
RENAMED
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|