olca 0.2.7__tar.gz → 0.2.9__tar.gz
Sign up to get free protection for your applications and to get access to all the features.
- {olca-0.2.7 → olca-0.2.9}/PKG-INFO +1 -1
- {olca-0.2.7 → olca-0.2.9}/olca/olcacli.py +22 -14
- {olca-0.2.7 → olca-0.2.9}/olca.egg-info/PKG-INFO +1 -1
- {olca-0.2.7 → olca-0.2.9}/pyproject.toml +1 -1
- {olca-0.2.7 → olca-0.2.9}/setup.py +1 -1
- {olca-0.2.7 → olca-0.2.9}/LICENSE +0 -0
- {olca-0.2.7 → olca-0.2.9}/README.md +0 -0
- {olca-0.2.7 → olca-0.2.9}/olca/__init__.py +0 -0
- {olca-0.2.7 → olca-0.2.9}/olca.egg-info/SOURCES.txt +0 -0
- {olca-0.2.7 → olca-0.2.9}/olca.egg-info/dependency_links.txt +0 -0
- {olca-0.2.7 → olca-0.2.9}/olca.egg-info/entry_points.txt +0 -0
- {olca-0.2.7 → olca-0.2.9}/olca.egg-info/requires.txt +0 -0
- {olca-0.2.7 → olca-0.2.9}/olca.egg-info/top_level.txt +0 -0
- {olca-0.2.7 → olca-0.2.9}/setup.cfg +0 -0
@@ -273,19 +273,27 @@ def main():
|
|
273
273
|
print("For troubleshooting, visit: https://python.langchain.com/docs/troubleshooting/errors/GRAPH_RECURSION_LIMIT")
|
274
274
|
|
275
275
|
def generate_config_example():
|
276
|
-
|
277
|
-
|
278
|
-
|
279
|
-
|
280
|
-
|
281
|
-
|
282
|
-
|
283
|
-
|
284
|
-
|
285
|
-
|
286
|
-
|
287
|
-
|
288
|
-
|
276
|
+
try:
|
277
|
+
config = {
|
278
|
+
"api_keyname": input("api_keyname [OPENAI_API_KEY_olca]: ") or "OPENAI_API_KEY_olca",
|
279
|
+
"model_name": input("model_name [gpt-4o-mini]: ") or "gpt-4o-mini",
|
280
|
+
"recursion_limit": int(input("recursion_limit [12]: ") or 12),
|
281
|
+
"temperature": float(input("temperature [0]: ") or 0),
|
282
|
+
"human": input("human [true]: ").lower() in ["true", "yes", "y", "1", ""] or True,
|
283
|
+
"tracing": input("tracing [true]: ").lower() in ["true", "yes", "y", "1", ""] or True,
|
284
|
+
"system_instructions": input("system_instructions [Hello, I am a chatbot. How can I help you today?]: ") or "You are interacting using the human tool addressing carefully what the user is asking. You carefully go step by step in your interaction with the user. You are clear and concise in your communication wrapping up the conversation in a coherent manner for the interaction. You make sure to always present what you think is the end-result of the work before quitting the loop and exit the workflow.",
|
285
|
+
"user_input": input("user_input [Interact with me to write a story using the 3 act structure that we will save in ./story/]: ") or "Interact with me to write a story using the 3 act structure that we will save in ./story/"
|
286
|
+
}
|
287
|
+
with open('olca.yml', 'w') as file:
|
288
|
+
yaml.dump(config, file)
|
289
|
+
print("Configuration file 'olca.yml' created successfully.")
|
290
|
+
except KeyboardInterrupt:
|
291
|
+
print("\nConfiguration canceled by user.")
|
292
|
+
exit(0)
|
289
293
|
|
290
294
|
if __name__ == "__main__":
|
291
|
-
|
295
|
+
try:
|
296
|
+
main()
|
297
|
+
except KeyboardInterrupt:
|
298
|
+
print("\nExiting gracefully.")
|
299
|
+
exit(0)
|
@@ -2,7 +2,7 @@ from setuptools import setup, find_packages
|
|
2
2
|
|
3
3
|
setup(
|
4
4
|
name='olca',
|
5
|
-
version = "0.2.
|
5
|
+
version = "0.2.9",
|
6
6
|
author='Jean GUillaume ISabelle',
|
7
7
|
author_email='jgi@jgwill.com',
|
8
8
|
description='A Python package for experimenting with Langchain agent and interactivity in Terminal modalities.',
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|