lionagi 0.9.4__py3-none-any.whl → 0.9.6__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- lionagi/libs/schema/function_to_schema.py +84 -24
- lionagi/operations/ReAct/ReAct.py +35 -1
- lionagi/operations/ReAct/utils.py +5 -2
- lionagi/operations/chat/chat.py +7 -5
- lionagi/operations/communicate/communicate.py +2 -0
- lionagi/operations/interpret/interpret.py +2 -1
- lionagi/operations/operate/operate.py +2 -0
- lionagi/service/endpoints/base.py +47 -5
- lionagi/service/endpoints/match_endpoint.py +7 -0
- lionagi/service/imodel.py +34 -15
- lionagi/service/providers/ollama_/__init__.py +3 -0
- lionagi/service/providers/ollama_/chat_completions.py +134 -0
- lionagi/service/providers/openai_/chat_completions.py +2 -0
- lionagi/session/branch.py +16 -0
- lionagi/version.py +1 -1
- {lionagi-0.9.4.dist-info → lionagi-0.9.6.dist-info}/METADATA +1 -1
- {lionagi-0.9.4.dist-info → lionagi-0.9.6.dist-info}/RECORD +19 -17
- {lionagi-0.9.4.dist-info → lionagi-0.9.6.dist-info}/WHEEL +0 -0
- {lionagi-0.9.4.dist-info → lionagi-0.9.6.dist-info}/licenses/LICENSE +0 -0
@@ -5,7 +5,13 @@
|
|
5
5
|
import inspect
|
6
6
|
from typing import Any, Literal
|
7
7
|
|
8
|
-
from
|
8
|
+
from pydantic import Field, field_validator
|
9
|
+
|
10
|
+
from lionagi.libs.schema.extract_docstring import extract_docstring
|
11
|
+
from lionagi.libs.validate.common_field_validators import (
|
12
|
+
validate_model_to_type,
|
13
|
+
)
|
14
|
+
from lionagi.operatives.models.schema_model import SchemaModel
|
9
15
|
|
10
16
|
py_json_msp = {
|
11
17
|
"str": "string",
|
@@ -18,12 +24,60 @@ py_json_msp = {
|
|
18
24
|
}
|
19
25
|
|
20
26
|
|
27
|
+
class FunctionSchema(SchemaModel):
|
28
|
+
name: str
|
29
|
+
description: str | None = Field(
|
30
|
+
None,
|
31
|
+
description=(
|
32
|
+
"A description of what the function does, used by the "
|
33
|
+
"model to choose when and how to call the function."
|
34
|
+
),
|
35
|
+
)
|
36
|
+
parameters: dict[str, Any] | None = Field(
|
37
|
+
None,
|
38
|
+
description=(
|
39
|
+
"The parameters the functions accepts, described as a JSON Schema object. "
|
40
|
+
"See the guide (https://platform.openai.com/docs/guides/function-calling) "
|
41
|
+
"for examples, and the JSON Schema reference for documentation about the "
|
42
|
+
"format. Omitting parameters defines a function with an empty parameter list."
|
43
|
+
),
|
44
|
+
validation_alias="request_options",
|
45
|
+
)
|
46
|
+
strict: bool | None = Field(
|
47
|
+
None,
|
48
|
+
description=(
|
49
|
+
"Whether to enable strict schema adherence when generating the function call. "
|
50
|
+
"If set to true, the model will follow the exact schema defined in the parameters "
|
51
|
+
"field. Only a subset of JSON Schema is supported when strict is true."
|
52
|
+
),
|
53
|
+
)
|
54
|
+
|
55
|
+
@field_validator("parameters", mode="before")
|
56
|
+
def _validate_parameters(cls, v):
|
57
|
+
if v is None:
|
58
|
+
return None
|
59
|
+
if isinstance(v, dict):
|
60
|
+
return v
|
61
|
+
try:
|
62
|
+
model_type = validate_model_to_type(cls, v)
|
63
|
+
return model_type.model_json_schema()
|
64
|
+
except Exception:
|
65
|
+
raise ValueError(f"Invalid model type: {v}")
|
66
|
+
|
67
|
+
def to_dict(self):
|
68
|
+
dict_ = super().to_dict()
|
69
|
+
return {"type": "function", "function": dict_}
|
70
|
+
|
71
|
+
|
21
72
|
def function_to_schema(
|
22
73
|
f_,
|
23
74
|
style: Literal["google", "rest"] = "google",
|
24
75
|
*,
|
76
|
+
request_options: dict[str, Any] | None = None,
|
77
|
+
strict: bool = None,
|
25
78
|
func_description: str = None,
|
26
79
|
parametert_description: dict[str, str] = None,
|
80
|
+
return_obj: bool = False,
|
27
81
|
) -> dict:
|
28
82
|
"""
|
29
83
|
Generate a schema description for a given function. in openai format
|
@@ -78,27 +132,33 @@ def function_to_schema(
|
|
78
132
|
"required": [],
|
79
133
|
}
|
80
134
|
|
81
|
-
|
82
|
-
|
83
|
-
|
84
|
-
|
85
|
-
|
86
|
-
|
87
|
-
|
88
|
-
|
89
|
-
|
90
|
-
|
91
|
-
|
92
|
-
|
93
|
-
"
|
94
|
-
|
95
|
-
|
96
|
-
|
97
|
-
|
98
|
-
|
99
|
-
|
100
|
-
|
101
|
-
|
102
|
-
|
103
|
-
|
135
|
+
if not request_options:
|
136
|
+
for name, param in sig.parameters.items():
|
137
|
+
# Default type to string and update if type hint is available
|
138
|
+
param_type = "string"
|
139
|
+
if param.annotation is not inspect.Parameter.empty:
|
140
|
+
param_type = py_json_msp[param.annotation.__name__]
|
141
|
+
|
142
|
+
# Extract parameter description from docstring, if available
|
143
|
+
param_description = parametert_description.get(name)
|
144
|
+
|
145
|
+
# Assuming all parameters are required for simplicity
|
146
|
+
parameters["required"].append(name)
|
147
|
+
parameters["properties"][name] = {
|
148
|
+
"type": param_type,
|
149
|
+
"description": param_description,
|
150
|
+
}
|
151
|
+
else:
|
152
|
+
parameters = request_options
|
153
|
+
|
154
|
+
params = {
|
155
|
+
"name": func_name,
|
156
|
+
"description": func_description,
|
157
|
+
"parameters": parameters,
|
104
158
|
}
|
159
|
+
if strict:
|
160
|
+
params["strict"] = strict
|
161
|
+
|
162
|
+
if return_obj:
|
163
|
+
return FunctionSchema(**params)
|
164
|
+
return FunctionSchema(**params).to_dict()
|
@@ -47,6 +47,8 @@ async def ReAct(
|
|
47
47
|
analysis_model: iModel | None = None,
|
48
48
|
verbose_analysis: bool = False,
|
49
49
|
verbose_length: int = None,
|
50
|
+
include_token_usage_to_model: bool = True,
|
51
|
+
continue_after_failed_response: bool = False,
|
50
52
|
**kwargs,
|
51
53
|
):
|
52
54
|
outs = []
|
@@ -73,6 +75,8 @@ async def ReAct(
|
|
73
75
|
verbose_analysis=verbose_analysis,
|
74
76
|
display_as=display_as,
|
75
77
|
verbose_length=verbose_length,
|
78
|
+
include_token_usage_to_model=include_token_usage_to_model,
|
79
|
+
continue_after_failed_response=continue_after_failed_response,
|
76
80
|
**kwargs,
|
77
81
|
):
|
78
82
|
analysis, str_ = i
|
@@ -101,6 +105,8 @@ async def ReAct(
|
|
101
105
|
analysis_model=analysis_model,
|
102
106
|
display_as=display_as,
|
103
107
|
verbose_length=verbose_length,
|
108
|
+
include_token_usage_to_model=include_token_usage_to_model,
|
109
|
+
continue_after_failed_response=continue_after_failed_response,
|
104
110
|
**kwargs,
|
105
111
|
):
|
106
112
|
outs.append(i)
|
@@ -131,6 +137,8 @@ async def ReActStream(
|
|
131
137
|
verbose_analysis: bool = False,
|
132
138
|
display_as: Literal["json", "yaml"] = "yaml",
|
133
139
|
verbose_length: int = None,
|
140
|
+
include_token_usage_to_model: bool = True,
|
141
|
+
continue_after_failed_response: bool = False,
|
134
142
|
**kwargs,
|
135
143
|
) -> AsyncGenerator:
|
136
144
|
irfm: FieldModel | None = None
|
@@ -213,6 +221,9 @@ async def ReActStream(
|
|
213
221
|
kwargs_for_operate = copy(kwargs)
|
214
222
|
kwargs_for_operate["actions"] = True
|
215
223
|
kwargs_for_operate["reason"] = True
|
224
|
+
kwargs_for_operate["include_token_usage_to_model"] = (
|
225
|
+
include_token_usage_to_model
|
226
|
+
)
|
216
227
|
|
217
228
|
# Step 1: Generate initial ReAct analysis
|
218
229
|
analysis: ReActAnalysis = await branch.operate(
|
@@ -255,7 +266,7 @@ async def ReActStream(
|
|
255
266
|
if isinstance(analysis, dict)
|
256
267
|
else False
|
257
268
|
)
|
258
|
-
and (extensions if max_extensions else 0) > 0
|
269
|
+
and (extensions - 1 if max_extensions else 0) > 0
|
259
270
|
):
|
260
271
|
new_instruction = None
|
261
272
|
if extensions == max_extensions:
|
@@ -272,6 +283,9 @@ async def ReActStream(
|
|
272
283
|
operate_kwargs["reason"] = True
|
273
284
|
operate_kwargs["response_format"] = ReActAnalysis
|
274
285
|
operate_kwargs["action_strategy"] = analysis.action_strategy
|
286
|
+
operate_kwargs["include_token_usage_to_model"] = (
|
287
|
+
include_token_usage_to_model
|
288
|
+
)
|
275
289
|
if analysis.action_batch_size:
|
276
290
|
operate_kwargs["action_batch_size"] = analysis.action_batch_size
|
277
291
|
if irfm:
|
@@ -289,6 +303,7 @@ async def ReActStream(
|
|
289
303
|
operate_kwargs["guidance"] = guide + operate_kwargs.get(
|
290
304
|
"guidance", ""
|
291
305
|
)
|
306
|
+
operate_kwargs["reasoning_effort"] = reasoning_effort
|
292
307
|
|
293
308
|
analysis = await branch.operate(
|
294
309
|
instruction=new_instruction,
|
@@ -298,6 +313,16 @@ async def ReActStream(
|
|
298
313
|
)
|
299
314
|
round_count += 1
|
300
315
|
|
316
|
+
if isinstance(analysis, dict) and all(
|
317
|
+
i is None for i in analysis.values()
|
318
|
+
):
|
319
|
+
if not continue_after_failed_response:
|
320
|
+
raise ValueError(
|
321
|
+
"All values in the response are None. "
|
322
|
+
"This might be due to a failed response. "
|
323
|
+
"Set `continue_after_failed_response=True` to ignore this error."
|
324
|
+
)
|
325
|
+
|
301
326
|
# If verbose, show round analysis
|
302
327
|
if verbose_analysis:
|
303
328
|
str_ = f"\n### ReAct Round No.{round_count} Analysis:\n"
|
@@ -329,6 +354,15 @@ async def ReActStream(
|
|
329
354
|
response_format=response_format,
|
330
355
|
**(response_kwargs or {}),
|
331
356
|
)
|
357
|
+
if isinstance(analysis, dict) and all(
|
358
|
+
i is None for i in analysis.values()
|
359
|
+
):
|
360
|
+
if not continue_after_failed_response:
|
361
|
+
raise ValueError(
|
362
|
+
"All values in the response are None. "
|
363
|
+
"This might be due to a failed response. "
|
364
|
+
"Set `continue_after_failed_response=True` to ignore this error."
|
365
|
+
)
|
332
366
|
except Exception:
|
333
367
|
out = branch.msgs.last_response.response
|
334
368
|
|
@@ -30,6 +30,8 @@ class ReActAnalysis(BaseModel):
|
|
30
30
|
2) A list of planned actions to perform before finalizing,
|
31
31
|
3) Indication whether more expansions/rounds are needed,
|
32
32
|
4) Additional tuning knobs: how to handle validation, how to execute actions, etc.
|
33
|
+
Remember do not repeat yourself, and aim to use the most efficient way to achieve
|
34
|
+
the goal to user's satisfaction.
|
33
35
|
"""
|
34
36
|
|
35
37
|
# Standard ReAct strings for controlling expansions:
|
@@ -38,11 +40,12 @@ class ReActAnalysis(BaseModel):
|
|
38
40
|
"If you are not ready to finalize, set extension_needed to True. "
|
39
41
|
"hint: you should set extension_needed to True if the overall goal"
|
40
42
|
"is not yet achieved. Do not set it to False, if you are just providing"
|
41
|
-
"an interim answer. You have up to {extensions} expansions. Please
|
43
|
+
"an interim answer. You have up to {extensions} expansions. Please "
|
44
|
+
"strategize accordingly and continue."
|
42
45
|
)
|
43
46
|
CONTINUE_EXT_PROMPT: ClassVar[str] = (
|
44
47
|
"Another round is available. You may do multiple actions if needed. "
|
45
|
-
"You have up to {extensions} expansions. Please continue."
|
48
|
+
"You have up to {extensions} expansions. Please strategize accordingly and continue."
|
46
49
|
)
|
47
50
|
ANSWER_PROMPT: ClassVar[str] = (
|
48
51
|
"Given your reasoning and actions, please now provide the final answer "
|
lionagi/operations/chat/chat.py
CHANGED
@@ -36,6 +36,7 @@ async def chat(
|
|
36
36
|
image_detail: Literal["low", "high", "auto"] = None,
|
37
37
|
plain_content: str = None,
|
38
38
|
return_ins_res_message: bool = False,
|
39
|
+
include_token_usage_to_model: bool = False,
|
39
40
|
**kwargs,
|
40
41
|
) -> tuple[Instruction, AssistantResponse]:
|
41
42
|
ins: Instruction = branch.msgs.create_instruction(
|
@@ -151,11 +152,12 @@ async def chat(
|
|
151
152
|
kwargs["messages"] = [i.chat_msg for i in messages]
|
152
153
|
imodel = imodel or branch.chat_model
|
153
154
|
|
154
|
-
meth =
|
155
|
-
|
156
|
-
|
157
|
-
|
158
|
-
|
155
|
+
meth = imodel.invoke
|
156
|
+
if "stream" not in kwargs or not kwargs["stream"]:
|
157
|
+
kwargs["include_token_usage_to_model"] = include_token_usage_to_model
|
158
|
+
else:
|
159
|
+
meth = imodel.stream
|
160
|
+
|
159
161
|
api_call = await meth(**kwargs)
|
160
162
|
branch._log_manager.log(Log.create(api_call))
|
161
163
|
|
@@ -35,6 +35,7 @@ async def communicate(
|
|
35
35
|
fuzzy_match_kwargs=None,
|
36
36
|
clear_messages=False,
|
37
37
|
operative_model=None,
|
38
|
+
include_token_usage_to_model: bool = False,
|
38
39
|
**kwargs,
|
39
40
|
):
|
40
41
|
if operative_model:
|
@@ -80,6 +81,7 @@ async def communicate(
|
|
80
81
|
image_detail=image_detail,
|
81
82
|
plain_content=plain_content,
|
82
83
|
return_ins_res_message=True,
|
84
|
+
include_token_usage_to_model=include_token_usage_to_model,
|
83
85
|
**kwargs,
|
84
86
|
)
|
85
87
|
branch.msgs.add_message(instruction=ins)
|
@@ -20,7 +20,8 @@ async def interpret(
|
|
20
20
|
instruction = (
|
21
21
|
"You are given a user's raw instruction or question. Your task is to rewrite it into a clearer,"
|
22
22
|
"more structured prompt for an LLM or system, making any implicit or missing details explicit. "
|
23
|
-
"Return only the re-written prompt."
|
23
|
+
"Return only the re-written prompt. Do not assume any details not mentioned in the input, nor "
|
24
|
+
"give additional instruction than what is explicitly stated."
|
24
25
|
)
|
25
26
|
guidance = (
|
26
27
|
f"Domain hint: {domain or 'general'}. "
|
@@ -63,6 +63,7 @@ async def operate(
|
|
63
63
|
] = "return_value",
|
64
64
|
operative_model: type[BaseModel] = None,
|
65
65
|
request_model: type[BaseModel] = None,
|
66
|
+
include_token_usage_to_model: bool = False,
|
66
67
|
**kwargs,
|
67
68
|
) -> list | BaseModel | None | dict | str:
|
68
69
|
if operative_model:
|
@@ -138,6 +139,7 @@ async def operate(
|
|
138
139
|
image_detail=image_detail,
|
139
140
|
tool_schemas=tool_schemas,
|
140
141
|
return_ins_res_message=True,
|
142
|
+
include_token_usage_to_model=include_token_usage_to_model,
|
141
143
|
**kwargs,
|
142
144
|
)
|
143
145
|
branch.msgs.add_message(instruction=ins)
|
@@ -80,6 +80,7 @@ class EndpointConfig(BaseModel):
|
|
80
80
|
api_version: str | None = None
|
81
81
|
allowed_roles: list[str] | None = None
|
82
82
|
request_options: type | None = Field(None, exclude=True)
|
83
|
+
invoke_with_endpoint: bool | None = None
|
83
84
|
|
84
85
|
|
85
86
|
class EndPoint(ABC):
|
@@ -91,19 +92,28 @@ class EndPoint(ABC):
|
|
91
92
|
HTTP requests.
|
92
93
|
"""
|
93
94
|
|
94
|
-
def __init__(
|
95
|
+
def __init__(
|
96
|
+
self, config: dict | EndpointConfig | type[EndpointConfig], **kwargs
|
97
|
+
) -> None:
|
95
98
|
"""Initializes the EndPoint with a given configuration.
|
96
99
|
|
97
100
|
Args:
|
98
|
-
config (dict): Configuration data that matches the EndpointConfig
|
101
|
+
config (dict | EndpointConfig): Configuration data that matches the EndpointConfig
|
99
102
|
schema.
|
100
103
|
"""
|
101
|
-
|
104
|
+
if isinstance(config, dict):
|
105
|
+
self.config = EndpointConfig(**config)
|
106
|
+
if isinstance(config, EndpointConfig):
|
107
|
+
self.config = config
|
108
|
+
if isinstance(config, type) and issubclass(config, EndpointConfig):
|
109
|
+
self.config = config()
|
110
|
+
if kwargs:
|
111
|
+
self.update_config(**kwargs)
|
102
112
|
|
103
113
|
def update_config(self, **kwargs):
|
104
114
|
config = self.config.model_dump()
|
105
115
|
config.update(kwargs)
|
106
|
-
self.config =
|
116
|
+
self.config = self.config.model_validate(config)
|
107
117
|
|
108
118
|
@property
|
109
119
|
def name(self) -> str | None:
|
@@ -349,11 +359,38 @@ class APICalling(Event):
|
|
349
359
|
endpoint: EndPoint = Field(exclude=True)
|
350
360
|
is_cached: bool = Field(default=False, exclude=True)
|
351
361
|
should_invoke_endpoint: bool = Field(default=True, exclude=True)
|
362
|
+
include_token_usage_to_model: bool = Field(
|
363
|
+
default=False,
|
364
|
+
exclude=True,
|
365
|
+
description="Whether to include token usage information into instruction messages",
|
366
|
+
)
|
367
|
+
response_obj: BaseModel | None = Field(None, exclude=True)
|
352
368
|
|
353
369
|
@model_validator(mode="after")
|
354
370
|
def _validate_streaming(self) -> Self:
|
355
371
|
if self.payload.get("stream") is True:
|
356
372
|
self.streaming = True
|
373
|
+
|
374
|
+
if self.include_token_usage_to_model:
|
375
|
+
if isinstance(self.payload["messages"][-1], dict):
|
376
|
+
required_tokens = self.required_tokens
|
377
|
+
self.payload["messages"][-1][
|
378
|
+
"content"
|
379
|
+
] += f"\n\nEstimated Current Token Usage: {required_tokens}"
|
380
|
+
if "model" in self.payload:
|
381
|
+
if (
|
382
|
+
self.payload["model"].startswith("gpt-4")
|
383
|
+
or "o1mini" in self.payload["model"]
|
384
|
+
or "o1-preview" in self.payload["model"]
|
385
|
+
):
|
386
|
+
self.payload["messages"][-1]["content"] += "/128_000"
|
387
|
+
elif "o1" in self.payload["model"]:
|
388
|
+
self.payload["messages"][-1]["content"] += "/200_000"
|
389
|
+
elif "sonnet" in self.payload["model"]:
|
390
|
+
self.payload["messages"][-1]["content"] += "/200_000"
|
391
|
+
elif "haiku" in self.payload["model"]:
|
392
|
+
self.payload["messages"][-1]["content"] += "/200_000"
|
393
|
+
|
357
394
|
return self
|
358
395
|
|
359
396
|
@property
|
@@ -622,7 +659,12 @@ class APICalling(Event):
|
|
622
659
|
f"API call to {self.endpoint.full_url} failed: {e1}"
|
623
660
|
)
|
624
661
|
else:
|
625
|
-
self.
|
662
|
+
self.response_obj = response
|
663
|
+
self.execution.response = (
|
664
|
+
response.model_dump()
|
665
|
+
if isinstance(response, BaseModel)
|
666
|
+
else response
|
667
|
+
)
|
626
668
|
self.execution.status = EventStatus.COMPLETED
|
627
669
|
|
628
670
|
def __str__(self) -> str:
|
@@ -48,6 +48,13 @@ def match_endpoint(
|
|
48
48
|
|
49
49
|
return OpenRouterChatCompletionEndPoint()
|
50
50
|
|
51
|
+
if provider == "ollama":
|
52
|
+
from ..providers.ollama_.chat_completions import (
|
53
|
+
OllamaChatCompletionEndPoint,
|
54
|
+
)
|
55
|
+
|
56
|
+
return OllamaChatCompletionEndPoint()
|
57
|
+
|
51
58
|
return OpenAIChatCompletionEndPoint(
|
52
59
|
config={
|
53
60
|
"provider": provider,
|
lionagi/service/imodel.py
CHANGED
@@ -51,7 +51,7 @@ class iModel:
|
|
51
51
|
interval: float | None = None,
|
52
52
|
limit_requests: int = None,
|
53
53
|
limit_tokens: int = None,
|
54
|
-
invoke_with_endpoint: bool =
|
54
|
+
invoke_with_endpoint: bool = None,
|
55
55
|
concurrency_limit: int | None = None,
|
56
56
|
streaming_process_func: Callable = None,
|
57
57
|
requires_api_key: bool = True,
|
@@ -95,6 +95,16 @@ class iModel:
|
|
95
95
|
Additional keyword arguments, such as `model`, or any other
|
96
96
|
provider-specific fields.
|
97
97
|
"""
|
98
|
+
model = kwargs.get("model", None)
|
99
|
+
if model:
|
100
|
+
if not provider:
|
101
|
+
if "/" in model:
|
102
|
+
provider = model.split("/")[0]
|
103
|
+
model = model.replace(provider + "/", "")
|
104
|
+
kwargs["model"] = model
|
105
|
+
else:
|
106
|
+
raise ValueError("Provider must be provided")
|
107
|
+
|
98
108
|
if api_key is None:
|
99
109
|
provider = str(provider or "").strip().lower()
|
100
110
|
match provider:
|
@@ -110,6 +120,8 @@ class iModel:
|
|
110
120
|
api_key = "GROQ_API_KEY"
|
111
121
|
case "exa":
|
112
122
|
api_key = "EXA_API_KEY"
|
123
|
+
case "ollama":
|
124
|
+
api_key = "ollama"
|
113
125
|
case "":
|
114
126
|
if requires_api_key:
|
115
127
|
raise ValueError("API key must be provided")
|
@@ -121,16 +133,6 @@ class iModel:
|
|
121
133
|
api_key = os.getenv(api_key)
|
122
134
|
|
123
135
|
kwargs["api_key"] = api_key
|
124
|
-
model = kwargs.get("model", None)
|
125
|
-
if model:
|
126
|
-
if not provider:
|
127
|
-
if "/" in model:
|
128
|
-
provider = model.split("/")[0]
|
129
|
-
model = model.replace(provider + "/", "")
|
130
|
-
kwargs["model"] = model
|
131
|
-
else:
|
132
|
-
raise ValueError("Provider must be provided")
|
133
|
-
|
134
136
|
if isinstance(endpoint, EndPoint):
|
135
137
|
self.endpoint = endpoint
|
136
138
|
else:
|
@@ -145,7 +147,13 @@ class iModel:
|
|
145
147
|
if base_url:
|
146
148
|
self.endpoint.config.base_url = base_url
|
147
149
|
|
148
|
-
|
150
|
+
if (
|
151
|
+
invoke_with_endpoint is None
|
152
|
+
and self.endpoint.config.invoke_with_endpoint is True
|
153
|
+
):
|
154
|
+
invoke_with_endpoint = True
|
155
|
+
|
156
|
+
self.should_invoke_endpoint = invoke_with_endpoint or False
|
149
157
|
self.kwargs = kwargs
|
150
158
|
self.executor = RateLimitedAPIExecutor(
|
151
159
|
queue_capacity=queue_capacity,
|
@@ -162,7 +170,9 @@ class iModel:
|
|
162
170
|
else:
|
163
171
|
self.streaming_process_func = streaming_process_func
|
164
172
|
|
165
|
-
def create_api_calling(
|
173
|
+
def create_api_calling(
|
174
|
+
self, include_token_usage_to_model: bool = False, **kwargs
|
175
|
+
) -> APICalling:
|
166
176
|
"""Constructs an `APICalling` object from endpoint-specific payload.
|
167
177
|
|
168
178
|
Args:
|
@@ -183,6 +193,7 @@ class iModel:
|
|
183
193
|
endpoint=self.endpoint,
|
184
194
|
is_cached=payload.get("is_cached", False),
|
185
195
|
should_invoke_endpoint=self.should_invoke_endpoint,
|
196
|
+
include_token_usage_to_model=include_token_usage_to_model,
|
186
197
|
)
|
187
198
|
|
188
199
|
async def process_chunk(self, chunk) -> None:
|
@@ -200,7 +211,12 @@ class iModel:
|
|
200
211
|
return await self.streaming_process_func(chunk)
|
201
212
|
return self.streaming_process_func(chunk)
|
202
213
|
|
203
|
-
async def stream(
|
214
|
+
async def stream(
|
215
|
+
self,
|
216
|
+
api_call=None,
|
217
|
+
include_token_usage_to_model: bool = False,
|
218
|
+
**kwargs,
|
219
|
+
) -> AsyncGenerator:
|
204
220
|
"""Performs a streaming API call with the given arguments.
|
205
221
|
|
206
222
|
Args:
|
@@ -214,7 +230,10 @@ class iModel:
|
|
214
230
|
"""
|
215
231
|
if api_call is None:
|
216
232
|
kwargs["stream"] = True
|
217
|
-
api_call = self.create_api_calling(
|
233
|
+
api_call = self.create_api_calling(
|
234
|
+
include_token_usage_to_model=include_token_usage_to_model,
|
235
|
+
**kwargs,
|
236
|
+
)
|
218
237
|
await self.executor.append(api_call)
|
219
238
|
|
220
239
|
if (
|
@@ -0,0 +1,134 @@
|
|
1
|
+
# Copyright (c) 2023 - 2025, HaiyangLi <quantocean.li at gmail dot com>
|
2
|
+
#
|
3
|
+
# SPDX-License-Identifier: Apache-2.0
|
4
|
+
|
5
|
+
from collections.abc import AsyncGenerator
|
6
|
+
|
7
|
+
from lionagi.service.endpoints.chat_completion import ChatCompletionEndPoint
|
8
|
+
|
9
|
+
CHAT_COMPLETION_CONFIG = {
|
10
|
+
"provider": "ollama",
|
11
|
+
"base_url": "http://localhost:11434/v1",
|
12
|
+
"endpoint": "chat",
|
13
|
+
"method": "post",
|
14
|
+
"openai_compatible": True,
|
15
|
+
"is_invokeable": True,
|
16
|
+
"requires_tokens": True,
|
17
|
+
"is_streamable": True,
|
18
|
+
"required_kwargs": {
|
19
|
+
"messages",
|
20
|
+
"model",
|
21
|
+
},
|
22
|
+
"optional_kwargs": {
|
23
|
+
"frequency_penalty",
|
24
|
+
"presence_penalty",
|
25
|
+
"response_format",
|
26
|
+
"seed",
|
27
|
+
"stop",
|
28
|
+
"stream",
|
29
|
+
"stream_options",
|
30
|
+
"temperature",
|
31
|
+
"top_p",
|
32
|
+
"max_tokens",
|
33
|
+
"tools",
|
34
|
+
# "tool_choice",
|
35
|
+
# "logit_bias",
|
36
|
+
# "user",
|
37
|
+
# "n",
|
38
|
+
# "logprobs",
|
39
|
+
},
|
40
|
+
"allowed_roles": ["user", "assistant", "system"],
|
41
|
+
"invoke_with_endpoint": True,
|
42
|
+
}
|
43
|
+
|
44
|
+
|
45
|
+
class OllamaChatCompletionEndPoint(ChatCompletionEndPoint):
|
46
|
+
"""
|
47
|
+
Documentation: https://platform.openai.com/docs/api-reference/chat/create
|
48
|
+
"""
|
49
|
+
|
50
|
+
def __init__(self, config: dict = CHAT_COMPLETION_CONFIG):
|
51
|
+
from lionagi.libs.package.imports import check_import
|
52
|
+
|
53
|
+
check_import("openai")
|
54
|
+
check_import("ollama")
|
55
|
+
|
56
|
+
from ollama import list, pull
|
57
|
+
from openai import AsyncOpenAI
|
58
|
+
|
59
|
+
super().__init__(config)
|
60
|
+
self.client = AsyncOpenAI(
|
61
|
+
base_url=self.config.base_url,
|
62
|
+
api_key="ollama",
|
63
|
+
)
|
64
|
+
self._pull = pull
|
65
|
+
self._list = list
|
66
|
+
|
67
|
+
async def _invoke(
|
68
|
+
self,
|
69
|
+
payload: dict,
|
70
|
+
headers: dict,
|
71
|
+
**kwargs,
|
72
|
+
):
|
73
|
+
self._check_model(payload["model"])
|
74
|
+
params = {**payload, **kwargs}
|
75
|
+
headers.pop("Authorization", None)
|
76
|
+
params["extra_headers"] = headers
|
77
|
+
|
78
|
+
if "response_format" in payload:
|
79
|
+
return await self.client.beta.chat.completions.parse(**params)
|
80
|
+
params.pop("response_format", None)
|
81
|
+
return await self.client.chat.completions.create(**params)
|
82
|
+
|
83
|
+
async def _stream(
|
84
|
+
self,
|
85
|
+
payload: dict,
|
86
|
+
headers: dict,
|
87
|
+
**kwargs,
|
88
|
+
) -> AsyncGenerator:
|
89
|
+
|
90
|
+
self._check_model(payload["model"])
|
91
|
+
params = {**payload, **kwargs}
|
92
|
+
headers.pop("Authorization", None)
|
93
|
+
params["extra_headers"] = headers
|
94
|
+
|
95
|
+
async for chunk in self.client.beta.chat.completions.stream(**params):
|
96
|
+
yield chunk
|
97
|
+
|
98
|
+
@property
|
99
|
+
def allowed_roles(self):
|
100
|
+
return ["system", "user", "assistant"]
|
101
|
+
|
102
|
+
def _pull_model(self, model: str):
|
103
|
+
from tqdm import tqdm
|
104
|
+
|
105
|
+
current_digest, bars = "", {}
|
106
|
+
for progress in self._pull(model, stream=True):
|
107
|
+
digest = progress.get("digest", "")
|
108
|
+
if digest != current_digest and current_digest in bars:
|
109
|
+
bars[current_digest].close()
|
110
|
+
|
111
|
+
if not digest:
|
112
|
+
print(progress.get("status"))
|
113
|
+
continue
|
114
|
+
|
115
|
+
if digest not in bars and (total := progress.get("total")):
|
116
|
+
bars[digest] = tqdm(
|
117
|
+
total=total,
|
118
|
+
desc=f"pulling {digest[7:19]}",
|
119
|
+
unit="B",
|
120
|
+
unit_scale=True,
|
121
|
+
)
|
122
|
+
|
123
|
+
if completed := progress.get("completed"):
|
124
|
+
bars[digest].update(completed - bars[digest].n)
|
125
|
+
|
126
|
+
current_digest = digest
|
127
|
+
|
128
|
+
def _list_local_models(self) -> set:
|
129
|
+
response = self._list()
|
130
|
+
return {i.model for i in response.models}
|
131
|
+
|
132
|
+
def _check_model(self, model: str):
|
133
|
+
if model not in self._list_local_models():
|
134
|
+
self._pull_model(model)
|
@@ -89,6 +89,8 @@ class OpenAIChatCompletionEndPoint(ChatCompletionEndPoint):
|
|
89
89
|
payload.pop("top_p", None)
|
90
90
|
if payload["messages"][0].get("role") == "system":
|
91
91
|
payload["messages"][0]["role"] = "developer"
|
92
|
+
else:
|
93
|
+
payload.pop("reasoning_effort", None)
|
92
94
|
|
93
95
|
return {
|
94
96
|
"payload": payload,
|
lionagi/session/branch.py
CHANGED
@@ -941,6 +941,7 @@ class Branch(Element, Communicatable, Relational):
|
|
941
941
|
] = "return_value",
|
942
942
|
operative_model: type[BaseModel] = None,
|
943
943
|
request_model: type[BaseModel] = None,
|
944
|
+
include_token_usage_to_model: bool = False,
|
944
945
|
**kwargs,
|
945
946
|
) -> list | BaseModel | None | dict | str:
|
946
947
|
"""
|
@@ -1028,6 +1029,8 @@ class Branch(Element, Communicatable, Relational):
|
|
1028
1029
|
Alias for `response_format`.
|
1029
1030
|
request_model (type[BaseModel], optional):
|
1030
1031
|
Another alias for `response_format`.
|
1032
|
+
include_token_usage_to_model:
|
1033
|
+
If `True`, includes token usage in the model messages.
|
1031
1034
|
**kwargs:
|
1032
1035
|
Additional keyword arguments passed to the LLM via `branch.chat()`.
|
1033
1036
|
|
@@ -1080,6 +1083,7 @@ class Branch(Element, Communicatable, Relational):
|
|
1080
1083
|
operative_model=operative_model,
|
1081
1084
|
request_model=request_model,
|
1082
1085
|
imodel=imodel,
|
1086
|
+
include_token_usage_to_model=include_token_usage_to_model,
|
1083
1087
|
**kwargs,
|
1084
1088
|
)
|
1085
1089
|
|
@@ -1106,6 +1110,7 @@ class Branch(Element, Communicatable, Relational):
|
|
1106
1110
|
fuzzy_match_kwargs: dict = None,
|
1107
1111
|
clear_messages: bool = False,
|
1108
1112
|
operative_model: type[BaseModel] = None,
|
1113
|
+
include_token_usage_to_model: bool = False,
|
1109
1114
|
**kwargs,
|
1110
1115
|
):
|
1111
1116
|
"""
|
@@ -1190,6 +1195,7 @@ class Branch(Element, Communicatable, Relational):
|
|
1190
1195
|
fuzzy_match_kwargs=fuzzy_match_kwargs,
|
1191
1196
|
clear_messages=clear_messages,
|
1192
1197
|
operative_model=operative_model,
|
1198
|
+
include_token_usage_to_model=include_token_usage_to_model,
|
1193
1199
|
**kwargs,
|
1194
1200
|
)
|
1195
1201
|
|
@@ -1639,6 +1645,7 @@ class Branch(Element, Communicatable, Relational):
|
|
1639
1645
|
analysis_model: iModel | None = None,
|
1640
1646
|
verbose: bool = False,
|
1641
1647
|
verbose_length: int = None,
|
1648
|
+
include_token_usage_to_model: bool = True,
|
1642
1649
|
**kwargs,
|
1643
1650
|
):
|
1644
1651
|
"""
|
@@ -1688,6 +1695,12 @@ class Branch(Element, Communicatable, Relational):
|
|
1688
1695
|
analysis_model (iModel | None, optional):
|
1689
1696
|
A custom LLM model for generating the ReAct analysis steps. If `None`,
|
1690
1697
|
uses the branch's default `chat_model`.
|
1698
|
+
include_token_usage_to_model:
|
1699
|
+
If `True`, includes token usage in the model messages.
|
1700
|
+
verbose (bool):
|
1701
|
+
If `True`, logs detailed information about the process.
|
1702
|
+
verbose_length (int):
|
1703
|
+
If `verbose=True`, limits the length of logged strings to this value.
|
1691
1704
|
**kwargs:
|
1692
1705
|
Additional keyword arguments passed into the initial `branch.operate()` call.
|
1693
1706
|
|
@@ -1733,6 +1746,7 @@ class Branch(Element, Communicatable, Relational):
|
|
1733
1746
|
intermediate_listable=intermediate_listable,
|
1734
1747
|
reasoning_effort=reasoning_effort,
|
1735
1748
|
display_as=display_as,
|
1749
|
+
include_token_usage_to_model=include_token_usage_to_model,
|
1736
1750
|
**kwargs,
|
1737
1751
|
)
|
1738
1752
|
|
@@ -1758,6 +1772,7 @@ class Branch(Element, Communicatable, Relational):
|
|
1758
1772
|
verbose: bool = False,
|
1759
1773
|
display_as: Literal["json", "yaml"] = "yaml",
|
1760
1774
|
verbose_length: int = None,
|
1775
|
+
include_token_usage_to_model: bool = True,
|
1761
1776
|
**kwargs,
|
1762
1777
|
) -> AsyncGenerator:
|
1763
1778
|
from lionagi.operations.ReAct.ReAct import ReActStream
|
@@ -1784,6 +1799,7 @@ class Branch(Element, Communicatable, Relational):
|
|
1784
1799
|
verbose_analysis=True,
|
1785
1800
|
display_as=display_as,
|
1786
1801
|
verbose_length=verbose_length,
|
1802
|
+
include_token_usage_to_model=include_token_usage_to_model,
|
1787
1803
|
**kwargs,
|
1788
1804
|
):
|
1789
1805
|
analysis, str_ = result
|
lionagi/version.py
CHANGED
@@ -1 +1 @@
|
|
1
|
-
__version__ = "0.9.
|
1
|
+
__version__ = "0.9.6"
|
@@ -4,7 +4,7 @@ lionagi/_errors.py,sha256=JlBTFJnRWtVYcRxKb7fWFiJHLbykl1E19mSJ8sXYVxg,455
|
|
4
4
|
lionagi/_types.py,sha256=9g7iytvSj3UjZxD-jL06_fxuNfgZyWT3Qnp0XYp1wQU,63
|
5
5
|
lionagi/settings.py,sha256=W52mM34E6jXF3GyqCFzVREKZrmnUqtZm_BVDsUiDI_s,1627
|
6
6
|
lionagi/utils.py,sha256=K36D9AAGiMPR4eM9tYoiVgvH-NdPPSeMQPls09s7keQ,73223
|
7
|
-
lionagi/version.py,sha256=
|
7
|
+
lionagi/version.py,sha256=IgVHjr-TeioZYLJSkvpT80LLGi6U3ONzR1cfYfd5XNQ,22
|
8
8
|
lionagi/libs/__init__.py,sha256=5y5joOZzfFWERl75auAcNcKC3lImVJ5ZZGvvHZUFCJM,112
|
9
9
|
lionagi/libs/parse.py,sha256=JRS3bql0InHJqATnAatl-hQv4N--XXw4P77JHhTFnrc,1011
|
10
10
|
lionagi/libs/file/__init__.py,sha256=5y5joOZzfFWERl75auAcNcKC3lImVJ5ZZGvvHZUFCJM,112
|
@@ -32,7 +32,7 @@ lionagi/libs/schema/__init__.py,sha256=5y5joOZzfFWERl75auAcNcKC3lImVJ5ZZGvvHZUFC
|
|
32
32
|
lionagi/libs/schema/as_readable.py,sha256=W4fi98WVkP5rfZ6A-iWqP5YFJexYCjt9Hf-l0iNs-2Q,5916
|
33
33
|
lionagi/libs/schema/extract_code_block.py,sha256=PuJbJj1JnqR5fSZudowPcVPpEoKISLr0MjTOOVXSzwY,2394
|
34
34
|
lionagi/libs/schema/extract_docstring.py,sha256=aYyLSRlB8lTH9QF9-6a56uph3AAkNuTyZ0S_duf5-fw,5729
|
35
|
-
lionagi/libs/schema/function_to_schema.py,sha256=
|
35
|
+
lionagi/libs/schema/function_to_schema.py,sha256=qLsM-_1ERlLdP_zBwz7ttxMZQ8jr5eTJN8IX1QSkSig,5560
|
36
36
|
lionagi/libs/schema/json_schema.py,sha256=cuHcaMr748O9g6suNGmRx4tRXcidd5-c7AMGjTIZyHM,7670
|
37
37
|
lionagi/libs/token_transform/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
38
38
|
lionagi/libs/token_transform/llmlingua.py,sha256=DkeLUlrb7rGx3nZ04aADU9HXXu5mZTf_DBwT0xhzIv4,7
|
@@ -48,24 +48,24 @@ lionagi/operations/__init__.py,sha256=2HExKTx2J3iKWsvS9YaF6i5SZgqcpAJkVmWbi2H5A5
|
|
48
48
|
lionagi/operations/manager.py,sha256=H7UY86PIxvxKdzJY9YVsWyJcqlwLWhVyvm4sYePH_uY,565
|
49
49
|
lionagi/operations/types.py,sha256=fM8HphnbBifMzhoKKvdl3JxGCBHlEGPJEYkLWj9b7vE,704
|
50
50
|
lionagi/operations/utils.py,sha256=LrWr_JEyJmSw5RL03KZhWgDKYsjFk0-OS8SoaGU7Jhs,1220
|
51
|
-
lionagi/operations/ReAct/ReAct.py,sha256=
|
51
|
+
lionagi/operations/ReAct/ReAct.py,sha256=f3_Q3Sc9U7i_4IBzFN3KxohUnMTMa_yFJ2I9rSlERvE,13578
|
52
52
|
lionagi/operations/ReAct/__init__.py,sha256=5y5joOZzfFWERl75auAcNcKC3lImVJ5ZZGvvHZUFCJM,112
|
53
|
-
lionagi/operations/ReAct/utils.py,sha256=
|
53
|
+
lionagi/operations/ReAct/utils.py,sha256=84Giel5ToqfbN5F6Tm0uw8yZTTnxiM_jWuFEhnKOxM8,3800
|
54
54
|
lionagi/operations/_act/__init__.py,sha256=5y5joOZzfFWERl75auAcNcKC3lImVJ5ZZGvvHZUFCJM,112
|
55
55
|
lionagi/operations/_act/act.py,sha256=CunHTTZcS6xNUe0xKSDgtMJ7-ucSvHeW4BtmVjXnaxk,2958
|
56
56
|
lionagi/operations/brainstorm/__init__.py,sha256=5y5joOZzfFWERl75auAcNcKC3lImVJ5ZZGvvHZUFCJM,112
|
57
57
|
lionagi/operations/brainstorm/brainstorm.py,sha256=iRZUW_V-0Ncw7Av0_WBk3oNRWb4LqQU5i2gQDxulYWY,17222
|
58
58
|
lionagi/operations/brainstorm/prompt.py,sha256=Dqi4NNeztdI4iutggRqjnOrG4a4E2JtwIAtRnjZ_ghQ,610
|
59
59
|
lionagi/operations/chat/__init__.py,sha256=5y5joOZzfFWERl75auAcNcKC3lImVJ5ZZGvvHZUFCJM,112
|
60
|
-
lionagi/operations/chat/chat.py,sha256=
|
60
|
+
lionagi/operations/chat/chat.py,sha256=xJAH2H0zyVvxiL3XtW3MC6YrwCCB1uCkwcQIJ1YsIOk,5466
|
61
61
|
lionagi/operations/communicate/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
62
|
-
lionagi/operations/communicate/communicate.py,sha256=
|
62
|
+
lionagi/operations/communicate/communicate.py,sha256=dPaPqg898biY6j_FlgH4HEJxTK6T_87ixXWhD6kbk40,3077
|
63
63
|
lionagi/operations/instruct/__init__.py,sha256=5y5joOZzfFWERl75auAcNcKC3lImVJ5ZZGvvHZUFCJM,112
|
64
64
|
lionagi/operations/instruct/instruct.py,sha256=-HDdCgvRGCNB5vMCV0xM2KEa9VrjosApXSKtQQ9d0xQ,795
|
65
65
|
lionagi/operations/interpret/__init__.py,sha256=5y5joOZzfFWERl75auAcNcKC3lImVJ5ZZGvvHZUFCJM,112
|
66
|
-
lionagi/operations/interpret/interpret.py,sha256=
|
66
|
+
lionagi/operations/interpret/interpret.py,sha256=8_F3oYaoYK8MDcK4iCwksBP7sI0UlgBiZSrUusdlKNo,1528
|
67
67
|
lionagi/operations/operate/__init__.py,sha256=5y5joOZzfFWERl75auAcNcKC3lImVJ5ZZGvvHZUFCJM,112
|
68
|
-
lionagi/operations/operate/operate.py,sha256=
|
68
|
+
lionagi/operations/operate/operate.py,sha256=j5dGWhHlcWnO-aaLZ4Xe0Hb1M7FGp9BGm356-LOH6rA,7333
|
69
69
|
lionagi/operations/parse/__init__.py,sha256=5y5joOZzfFWERl75auAcNcKC3lImVJ5ZZGvvHZUFCJM,112
|
70
70
|
lionagi/operations/parse/parse.py,sha256=i2q6YGRwsp2WEu0aySfZ6io7iLNAqPAx1wWd2hUjpgM,3296
|
71
71
|
lionagi/operations/plan/__init__.py,sha256=yGBPll6lOqVjadbTvDLGrTlMx3FfBW-e00z7AMvg7Uo,156
|
@@ -161,13 +161,13 @@ lionagi/protocols/messages/templates/instruction_message.jinja2,sha256=L-ptw5OHx
|
|
161
161
|
lionagi/protocols/messages/templates/system_message.jinja2,sha256=JRKJ0aFpYfaXSFouKc_N4unZ35C3yZTOWhIrIdCB5qk,215
|
162
162
|
lionagi/protocols/messages/templates/tool_schemas.jinja2,sha256=ozIaSDCRjIAhLyA8VM6S-YqS0w2NcctALSwx4LjDwII,126
|
163
163
|
lionagi/service/__init__.py,sha256=DMGXIqPsmut9H5GT0ZeSzQIzYzzPwI-2gLXydpbwiV8,21
|
164
|
-
lionagi/service/imodel.py,sha256=
|
164
|
+
lionagi/service/imodel.py,sha256=GIb0v0gSa9WJA5fmif8nAe-y-j8qHlZNb8FEIuzgE9s,15157
|
165
165
|
lionagi/service/manager.py,sha256=FkuqAtLErqLmXNnDtuAdTUFo4uuE_VL660BBGBhzInU,1435
|
166
166
|
lionagi/service/types.py,sha256=CHPi8Bxl_yJ1pl2jYZBOrTHbT8_oO9sK75d4LMB651g,486
|
167
167
|
lionagi/service/endpoints/__init__.py,sha256=5y5joOZzfFWERl75auAcNcKC3lImVJ5ZZGvvHZUFCJM,112
|
168
|
-
lionagi/service/endpoints/base.py,sha256=
|
168
|
+
lionagi/service/endpoints/base.py,sha256=sArF7dpSfer3OsBLMFqHGslAd4dgJUkCLtYkBWRmxOM,25441
|
169
169
|
lionagi/service/endpoints/chat_completion.py,sha256=nihV7kCYm7ixdm8dH0JW7vKjqH9yIom4QDXGeDwuO6E,2964
|
170
|
-
lionagi/service/endpoints/match_endpoint.py,sha256=
|
170
|
+
lionagi/service/endpoints/match_endpoint.py,sha256=x2T-ftzdqCrdITRLkH8UNRDY2Pm359DnX2RDXTBnbpc,2082
|
171
171
|
lionagi/service/endpoints/rate_limited_processor.py,sha256=P0CsMyhuG8OHCPYe2qez92Bm7v2ZRq4L5I6LOiAoGYs,5199
|
172
172
|
lionagi/service/endpoints/token_calculator.py,sha256=-AKwDvV7C8k8MTmd62ymT0ETSUPWBJ_DQKLZUutlyfg,6161
|
173
173
|
lionagi/service/providers/__init__.py,sha256=5y5joOZzfFWERl75auAcNcKC3lImVJ5ZZGvvHZUFCJM,112
|
@@ -180,15 +180,17 @@ lionagi/service/providers/exa_/search.py,sha256=Z3pyJH8KiWiquJSJw8Rd6D7x43BwTFHb
|
|
180
180
|
lionagi/service/providers/exa_/types.py,sha256=8ODjXpFajBE9-DGqBJNS--GObwmLSDi667xS84z_AgA,139
|
181
181
|
lionagi/service/providers/groq_/__init__.py,sha256=5y5joOZzfFWERl75auAcNcKC3lImVJ5ZZGvvHZUFCJM,112
|
182
182
|
lionagi/service/providers/groq_/chat_completions.py,sha256=q1p_1qus4vduWWBzs9V_KbNrqU2Tu2o8TZm6Fh09fw4,1343
|
183
|
+
lionagi/service/providers/ollama_/__init__.py,sha256=5y5joOZzfFWERl75auAcNcKC3lImVJ5ZZGvvHZUFCJM,112
|
184
|
+
lionagi/service/providers/ollama_/chat_completions.py,sha256=gPemTJO9dPd68l60kDYppO29uYqmLole4JGkv9Fz1Us,3764
|
183
185
|
lionagi/service/providers/openai_/__init__.py,sha256=5y5joOZzfFWERl75auAcNcKC3lImVJ5ZZGvvHZUFCJM,112
|
184
|
-
lionagi/service/providers/openai_/chat_completions.py,sha256=
|
186
|
+
lionagi/service/providers/openai_/chat_completions.py,sha256=y3RAgI5WQH5EwT1wZxp5ttnkCxUJEcOM3Cta6u9cpQo,2867
|
185
187
|
lionagi/service/providers/openrouter_/__init__.py,sha256=5y5joOZzfFWERl75auAcNcKC3lImVJ5ZZGvvHZUFCJM,112
|
186
188
|
lionagi/service/providers/openrouter_/chat_completions.py,sha256=0pdXjJCXmCPPbKKVubrnqofaodTOxWTJam8fd3NgrNk,1525
|
187
189
|
lionagi/service/providers/perplexity_/__init__.py,sha256=5y5joOZzfFWERl75auAcNcKC3lImVJ5ZZGvvHZUFCJM,112
|
188
190
|
lionagi/service/providers/perplexity_/chat_completions.py,sha256=O4MIS_3xIINGjkAZdlw0Bu_jAfBDR4VZA1F8JW2EU1M,1197
|
189
191
|
lionagi/service/providers/perplexity_/models.py,sha256=gXH4XGkhZ4aFxvMSDTlHq9Rz1mhu3aTENXAtE-BIr6U,4866
|
190
192
|
lionagi/session/__init__.py,sha256=5y5joOZzfFWERl75auAcNcKC3lImVJ5ZZGvvHZUFCJM,112
|
191
|
-
lionagi/session/branch.py,sha256=
|
193
|
+
lionagi/session/branch.py,sha256=dKlaM6hh_q7OoXkz4E5S3aS4ksqC2yzdhjzI7xe6pzU,72439
|
192
194
|
lionagi/session/prompts.py,sha256=AhuHL19s0TijVZX3tMKUKMi6l88xeVdpkuEn2vJSRyU,3236
|
193
195
|
lionagi/session/session.py,sha256=8SuNMiJX6IAW6Ou8aDK0LsVG7zcD5yd22sakMyrd3pw,8987
|
194
196
|
lionagi/tools/__init__.py,sha256=5y5joOZzfFWERl75auAcNcKC3lImVJ5ZZGvvHZUFCJM,112
|
@@ -211,7 +213,7 @@ lionagi/tools/file/writer.py,sha256=5y5joOZzfFWERl75auAcNcKC3lImVJ5ZZGvvHZUFCJM,
|
|
211
213
|
lionagi/tools/file/providers/__init__.py,sha256=5y5joOZzfFWERl75auAcNcKC3lImVJ5ZZGvvHZUFCJM,112
|
212
214
|
lionagi/tools/file/providers/docling_.py,sha256=5y5joOZzfFWERl75auAcNcKC3lImVJ5ZZGvvHZUFCJM,112
|
213
215
|
lionagi/tools/query/__init__.py,sha256=5y5joOZzfFWERl75auAcNcKC3lImVJ5ZZGvvHZUFCJM,112
|
214
|
-
lionagi-0.9.
|
215
|
-
lionagi-0.9.
|
216
|
-
lionagi-0.9.
|
217
|
-
lionagi-0.9.
|
216
|
+
lionagi-0.9.6.dist-info/METADATA,sha256=WGUAokWhfSlKKpws48WJRgBmwQJ1MTBAQd3o_EVdPdg,18053
|
217
|
+
lionagi-0.9.6.dist-info/WHEEL,sha256=qtCwoSJWgHk21S1Kb4ihdzI2rlJ1ZKaIurTj_ngOhyQ,87
|
218
|
+
lionagi-0.9.6.dist-info/licenses/LICENSE,sha256=VXFWsdoN5AAknBCgFqQNgPWYx7OPp-PFEP961zGdOjc,11288
|
219
|
+
lionagi-0.9.6.dist-info/RECORD,,
|
File without changes
|
File without changes
|