openlit 1.33.10__py3-none-any.whl → 1.33.12__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- openlit/__helpers.py +125 -88
- openlit/__init__.py +38 -11
- openlit/instrumentation/ag2/__init__.py +19 -20
- openlit/instrumentation/ag2/ag2.py +134 -69
- openlit/instrumentation/ai21/__init__.py +22 -21
- openlit/instrumentation/ai21/ai21.py +82 -546
- openlit/instrumentation/ai21/async_ai21.py +82 -546
- openlit/instrumentation/ai21/utils.py +409 -0
- openlit/instrumentation/anthropic/__init__.py +16 -16
- openlit/instrumentation/anthropic/anthropic.py +61 -353
- openlit/instrumentation/anthropic/async_anthropic.py +62 -354
- openlit/instrumentation/anthropic/utils.py +251 -0
- openlit/instrumentation/assemblyai/__init__.py +2 -2
- openlit/instrumentation/assemblyai/assemblyai.py +3 -3
- openlit/instrumentation/astra/__init__.py +25 -25
- openlit/instrumentation/astra/astra.py +2 -2
- openlit/instrumentation/astra/async_astra.py +2 -2
- openlit/instrumentation/azure_ai_inference/__init__.py +5 -5
- openlit/instrumentation/azure_ai_inference/async_azure_ai_inference.py +8 -8
- openlit/instrumentation/azure_ai_inference/azure_ai_inference.py +8 -8
- openlit/instrumentation/bedrock/__init__.py +2 -2
- openlit/instrumentation/bedrock/bedrock.py +3 -3
- openlit/instrumentation/chroma/__init__.py +9 -9
- openlit/instrumentation/chroma/chroma.py +2 -2
- openlit/instrumentation/cohere/__init__.py +7 -7
- openlit/instrumentation/cohere/async_cohere.py +9 -9
- openlit/instrumentation/cohere/cohere.py +9 -9
- openlit/instrumentation/controlflow/__init__.py +4 -4
- openlit/instrumentation/controlflow/controlflow.py +2 -2
- openlit/instrumentation/crawl4ai/__init__.py +3 -3
- openlit/instrumentation/crawl4ai/async_crawl4ai.py +2 -2
- openlit/instrumentation/crawl4ai/crawl4ai.py +2 -2
- openlit/instrumentation/crewai/__init__.py +3 -3
- openlit/instrumentation/crewai/crewai.py +2 -2
- openlit/instrumentation/dynamiq/__init__.py +5 -5
- openlit/instrumentation/dynamiq/dynamiq.py +2 -2
- openlit/instrumentation/elevenlabs/__init__.py +5 -5
- openlit/instrumentation/elevenlabs/async_elevenlabs.py +3 -3
- openlit/instrumentation/elevenlabs/elevenlabs.py +3 -3
- openlit/instrumentation/embedchain/__init__.py +2 -2
- openlit/instrumentation/embedchain/embedchain.py +4 -4
- openlit/instrumentation/firecrawl/__init__.py +3 -3
- openlit/instrumentation/firecrawl/firecrawl.py +2 -2
- openlit/instrumentation/google_ai_studio/__init__.py +3 -3
- openlit/instrumentation/google_ai_studio/async_google_ai_studio.py +3 -3
- openlit/instrumentation/google_ai_studio/google_ai_studio.py +3 -3
- openlit/instrumentation/gpt4all/__init__.py +3 -3
- openlit/instrumentation/gpt4all/gpt4all.py +7 -7
- openlit/instrumentation/groq/__init__.py +3 -3
- openlit/instrumentation/groq/async_groq.py +5 -5
- openlit/instrumentation/groq/groq.py +5 -5
- openlit/instrumentation/haystack/__init__.py +2 -2
- openlit/instrumentation/haystack/haystack.py +2 -2
- openlit/instrumentation/julep/__init__.py +7 -7
- openlit/instrumentation/julep/async_julep.py +3 -3
- openlit/instrumentation/julep/julep.py +3 -3
- openlit/instrumentation/langchain/__init__.py +2 -2
- openlit/instrumentation/langchain/async_langchain.py +13 -9
- openlit/instrumentation/langchain/langchain.py +13 -8
- openlit/instrumentation/letta/__init__.py +7 -7
- openlit/instrumentation/letta/letta.py +5 -5
- openlit/instrumentation/litellm/__init__.py +5 -5
- openlit/instrumentation/litellm/async_litellm.py +8 -8
- openlit/instrumentation/litellm/litellm.py +8 -8
- openlit/instrumentation/llamaindex/__init__.py +2 -2
- openlit/instrumentation/llamaindex/llamaindex.py +2 -2
- openlit/instrumentation/mem0/__init__.py +2 -2
- openlit/instrumentation/mem0/mem0.py +2 -2
- openlit/instrumentation/milvus/__init__.py +2 -2
- openlit/instrumentation/milvus/milvus.py +2 -2
- openlit/instrumentation/mistral/__init__.py +7 -7
- openlit/instrumentation/mistral/async_mistral.py +10 -10
- openlit/instrumentation/mistral/mistral.py +10 -10
- openlit/instrumentation/multion/__init__.py +7 -7
- openlit/instrumentation/multion/async_multion.py +5 -5
- openlit/instrumentation/multion/multion.py +5 -5
- openlit/instrumentation/ollama/__init__.py +11 -9
- openlit/instrumentation/ollama/async_ollama.py +71 -465
- openlit/instrumentation/ollama/ollama.py +71 -465
- openlit/instrumentation/ollama/utils.py +332 -0
- openlit/instrumentation/openai/__init__.py +11 -11
- openlit/instrumentation/openai/async_openai.py +18 -18
- openlit/instrumentation/openai/openai.py +18 -18
- openlit/instrumentation/phidata/__init__.py +2 -2
- openlit/instrumentation/phidata/phidata.py +2 -2
- openlit/instrumentation/pinecone/__init__.py +6 -6
- openlit/instrumentation/pinecone/pinecone.py +2 -2
- openlit/instrumentation/premai/__init__.py +3 -3
- openlit/instrumentation/premai/premai.py +7 -7
- openlit/instrumentation/qdrant/__init__.py +2 -2
- openlit/instrumentation/qdrant/async_qdrant.py +2 -2
- openlit/instrumentation/qdrant/qdrant.py +2 -2
- openlit/instrumentation/reka/__init__.py +3 -3
- openlit/instrumentation/reka/async_reka.py +3 -3
- openlit/instrumentation/reka/reka.py +3 -3
- openlit/instrumentation/together/__init__.py +5 -5
- openlit/instrumentation/together/async_together.py +8 -8
- openlit/instrumentation/together/together.py +8 -8
- openlit/instrumentation/transformers/__init__.py +2 -2
- openlit/instrumentation/transformers/transformers.py +4 -4
- openlit/instrumentation/vertexai/__init__.py +9 -9
- openlit/instrumentation/vertexai/async_vertexai.py +4 -4
- openlit/instrumentation/vertexai/vertexai.py +4 -4
- openlit/instrumentation/vllm/__init__.py +2 -2
- openlit/instrumentation/vllm/vllm.py +3 -3
- openlit/otel/events.py +85 -0
- openlit/otel/tracing.py +3 -13
- openlit/semcov/__init__.py +13 -1
- {openlit-1.33.10.dist-info → openlit-1.33.12.dist-info}/METADATA +2 -2
- openlit-1.33.12.dist-info/RECORD +126 -0
- openlit-1.33.10.dist-info/RECORD +0 -122
- {openlit-1.33.10.dist-info → openlit-1.33.12.dist-info}/LICENSE +0 -0
- {openlit-1.33.10.dist-info → openlit-1.33.12.dist-info}/WHEEL +0 -0
@@ -29,7 +29,7 @@ class ElevenLabsInstrumentor(BaseInstrumentor):
|
|
29
29
|
tracer = kwargs.get("tracer")
|
30
30
|
metrics = kwargs.get("metrics_dict")
|
31
31
|
pricing_info = kwargs.get("pricing_info", {})
|
32
|
-
|
32
|
+
capture_message_content = kwargs.get("capture_message_content", False)
|
33
33
|
disable_metrics = kwargs.get("disable_metrics")
|
34
34
|
version = importlib.metadata.version("elevenlabs")
|
35
35
|
|
@@ -38,7 +38,7 @@ class ElevenLabsInstrumentor(BaseInstrumentor):
|
|
38
38
|
"elevenlabs.client",
|
39
39
|
"ElevenLabs.generate",
|
40
40
|
generate("elevenlabs.generate", version, environment, application_name,
|
41
|
-
tracer, pricing_info,
|
41
|
+
tracer, pricing_info, capture_message_content, metrics, disable_metrics),
|
42
42
|
)
|
43
43
|
|
44
44
|
# sync text_to_speech.convert
|
@@ -46,7 +46,7 @@ class ElevenLabsInstrumentor(BaseInstrumentor):
|
|
46
46
|
"elevenlabs.text_to_speech.client",
|
47
47
|
"TextToSpeechClient.convert",
|
48
48
|
generate("elevenlabs.text_to_speech", version, environment, application_name,
|
49
|
-
tracer, pricing_info,
|
49
|
+
tracer, pricing_info, capture_message_content, metrics, disable_metrics),
|
50
50
|
)
|
51
51
|
|
52
52
|
# async generate
|
@@ -54,7 +54,7 @@ class ElevenLabsInstrumentor(BaseInstrumentor):
|
|
54
54
|
"elevenlabs.client",
|
55
55
|
"AsyncElevenLabs.generate",
|
56
56
|
async_generate("elevenlabs.generate", version, environment, application_name,
|
57
|
-
tracer, pricing_info,
|
57
|
+
tracer, pricing_info, capture_message_content, metrics, disable_metrics),
|
58
58
|
)
|
59
59
|
|
60
60
|
# sync text_to_speech.convert
|
@@ -62,7 +62,7 @@ class ElevenLabsInstrumentor(BaseInstrumentor):
|
|
62
62
|
"elevenlabs.text_to_speech.client",
|
63
63
|
"AsyncTextToSpeechClient.convert",
|
64
64
|
generate("elevenlabs.text_to_speech", version, environment, application_name,
|
65
|
-
tracer, pricing_info,
|
65
|
+
tracer, pricing_info, capture_message_content, metrics, disable_metrics),
|
66
66
|
)
|
67
67
|
|
68
68
|
def _uninstrument(self, **kwargs):
|
@@ -18,7 +18,7 @@ from openlit.semcov import SemanticConvetion
|
|
18
18
|
logger = logging.getLogger(__name__)
|
19
19
|
|
20
20
|
def async_generate(gen_ai_endpoint, version, environment, application_name,
|
21
|
-
tracer, pricing_info,
|
21
|
+
tracer, pricing_info, capture_message_content, metrics, disable_metrics):
|
22
22
|
"""
|
23
23
|
Generates a telemetry wrapper for creating speech audio to collect metrics.
|
24
24
|
|
@@ -28,7 +28,7 @@ def async_generate(gen_ai_endpoint, version, environment, application_name,
|
|
28
28
|
application_name: Name of the application using the ElevenLabs API.
|
29
29
|
tracer: OpenTelemetry tracer for creating spans.
|
30
30
|
pricing_info: Information used for calculating the cost of generating speech audio.
|
31
|
-
|
31
|
+
capture_message_content: Flag indicating whether to trace the input text and generated audio.
|
32
32
|
|
33
33
|
Returns:
|
34
34
|
A function that wraps the speech audio creation method to add telemetry.
|
@@ -105,7 +105,7 @@ def async_generate(gen_ai_endpoint, version, environment, application_name,
|
|
105
105
|
cost)
|
106
106
|
span.set_attribute(SemanticConvetion.GEN_AI_SDK_VERSION,
|
107
107
|
version)
|
108
|
-
if
|
108
|
+
if capture_message_content:
|
109
109
|
span.add_event(
|
110
110
|
name=SemanticConvetion.GEN_AI_CONTENT_PROMPT_EVENT,
|
111
111
|
attributes={
|
@@ -18,7 +18,7 @@ from openlit.semcov import SemanticConvetion
|
|
18
18
|
logger = logging.getLogger(__name__)
|
19
19
|
|
20
20
|
def generate(gen_ai_endpoint, version, environment, application_name,
|
21
|
-
tracer, pricing_info,
|
21
|
+
tracer, pricing_info, capture_message_content, metrics, disable_metrics):
|
22
22
|
"""
|
23
23
|
Generates a telemetry wrapper for creating speech audio to collect metrics.
|
24
24
|
|
@@ -28,7 +28,7 @@ def generate(gen_ai_endpoint, version, environment, application_name,
|
|
28
28
|
application_name: Name of the application using the ElevenLabs API.
|
29
29
|
tracer: OpenTelemetry tracer for creating spans.
|
30
30
|
pricing_info: Information used for calculating the cost of generating speech audio.
|
31
|
-
|
31
|
+
capture_message_content: Flag indicating whether to trace the input text and generated audio.
|
32
32
|
|
33
33
|
Returns:
|
34
34
|
A function that wraps the speech audio creation method to add telemetry.
|
@@ -106,7 +106,7 @@ def generate(gen_ai_endpoint, version, environment, application_name,
|
|
106
106
|
cost)
|
107
107
|
span.set_attribute(SemanticConvetion.GEN_AI_SDK_VERSION,
|
108
108
|
version)
|
109
|
-
if
|
109
|
+
if capture_message_content:
|
110
110
|
span.add_event(
|
111
111
|
name=SemanticConvetion.GEN_AI_CONTENT_PROMPT_EVENT,
|
112
112
|
attributes={
|
@@ -35,7 +35,7 @@ class EmbedChainInstrumentor(BaseInstrumentor):
|
|
35
35
|
environment = kwargs.get("environment")
|
36
36
|
tracer = kwargs.get("tracer")
|
37
37
|
pricing_info = kwargs.get("pricing_info")
|
38
|
-
|
38
|
+
capture_message_content = kwargs.get("capture_message_content")
|
39
39
|
version = importlib.metadata.version("embedchain")
|
40
40
|
|
41
41
|
for wrapped_method in WRAPPED_METHODS:
|
@@ -47,7 +47,7 @@ class EmbedChainInstrumentor(BaseInstrumentor):
|
|
47
47
|
wrap_package,
|
48
48
|
wrap_object,
|
49
49
|
wrapper(gen_ai_endpoint, version, environment, application_name,
|
50
|
-
tracer, pricing_info,
|
50
|
+
tracer, pricing_info, capture_message_content),
|
51
51
|
)
|
52
52
|
|
53
53
|
@staticmethod
|
@@ -13,7 +13,7 @@ from openlit.semcov import SemanticConvetion
|
|
13
13
|
logger = logging.getLogger(__name__)
|
14
14
|
|
15
15
|
def evaluate(gen_ai_endpoint, version, environment, application_name,
|
16
|
-
tracer, pricing_info,
|
16
|
+
tracer, pricing_info, capture_message_content):
|
17
17
|
"""
|
18
18
|
Creates a wrapper around a function call to trace and log its execution metrics.
|
19
19
|
|
@@ -27,7 +27,7 @@ def evaluate(gen_ai_endpoint, version, environment, application_name,
|
|
27
27
|
- application_name (str): Name of the EmbedChain application.
|
28
28
|
- tracer (opentelemetry.trace.Tracer): The tracer object used for OpenTelemetry tracing.
|
29
29
|
- pricing_info (dict): Information about the pricing for internal metrics (currently not used).
|
30
|
-
-
|
30
|
+
- capture_message_content (bool): Flag indicating whether to trace the content of the response.
|
31
31
|
|
32
32
|
Returns:
|
33
33
|
- function: A higher-order function that takes a function 'wrapped' and returns
|
@@ -91,7 +91,7 @@ def evaluate(gen_ai_endpoint, version, environment, application_name,
|
|
91
91
|
return wrapper
|
92
92
|
|
93
93
|
def get_data_sources(gen_ai_endpoint, version, environment, application_name,
|
94
|
-
tracer, pricing_info,
|
94
|
+
tracer, pricing_info, capture_message_content):
|
95
95
|
"""
|
96
96
|
Creates a wrapper around a function call to trace and log its execution metrics.
|
97
97
|
|
@@ -105,7 +105,7 @@ def get_data_sources(gen_ai_endpoint, version, environment, application_name,
|
|
105
105
|
- application_name (str): Name of the EmbedChain application.
|
106
106
|
- tracer (opentelemetry.trace.Tracer): The tracer object used for OpenTelemetry tracing.
|
107
107
|
- pricing_info (dict): Information about the pricing for internal metrics (currently not used).
|
108
|
-
-
|
108
|
+
- capture_message_content (bool): Flag indicating whether to trace the content of the response.
|
109
109
|
|
110
110
|
Returns:
|
111
111
|
- function: A higher-order function that takes a function 'wrapped' and returns
|
@@ -26,7 +26,7 @@ class FireCrawlInstrumentor(BaseInstrumentor):
|
|
26
26
|
tracer = kwargs.get("tracer")
|
27
27
|
metrics = kwargs.get("metrics_dict")
|
28
28
|
pricing_info = kwargs.get("pricing_info", {})
|
29
|
-
|
29
|
+
capture_message_content = kwargs.get("capture_message_content", False)
|
30
30
|
disable_metrics = kwargs.get("disable_metrics")
|
31
31
|
version = importlib.metadata.version("firecrawl-py")
|
32
32
|
|
@@ -34,14 +34,14 @@ class FireCrawlInstrumentor(BaseInstrumentor):
|
|
34
34
|
"firecrawl.firecrawl",
|
35
35
|
"FirecrawlApp.scrape_url",
|
36
36
|
wrap_crawl("firecrawl.scrape_url", version, environment, application_name,
|
37
|
-
tracer, pricing_info,
|
37
|
+
tracer, pricing_info, capture_message_content, metrics, disable_metrics),
|
38
38
|
)
|
39
39
|
|
40
40
|
wrap_function_wrapper(
|
41
41
|
"firecrawl.firecrawl",
|
42
42
|
"FirecrawlApp.crawl_url",
|
43
43
|
wrap_crawl("firecrawl.crawl_url", version, environment, application_name,
|
44
|
-
tracer, pricing_info,
|
44
|
+
tracer, pricing_info, capture_message_content, metrics, disable_metrics),
|
45
45
|
)
|
46
46
|
|
47
47
|
def _uninstrument(self, **kwargs):
|
@@ -15,7 +15,7 @@ from openlit.semcov import SemanticConvetion
|
|
15
15
|
logger = logging.getLogger(__name__)
|
16
16
|
|
17
17
|
def wrap_crawl(gen_ai_endpoint, version, environment, application_name,
|
18
|
-
tracer, pricing_info,
|
18
|
+
tracer, pricing_info, capture_message_content, metrics, disable_metrics):
|
19
19
|
"""
|
20
20
|
Generates a telemetry wrapper for chat completions to collect metrics.
|
21
21
|
|
@@ -26,7 +26,7 @@ def wrap_crawl(gen_ai_endpoint, version, environment, application_name,
|
|
26
26
|
application_name: Name of the application using the FireCrawl Agent.
|
27
27
|
tracer: OpenTelemetry tracer for creating spans.
|
28
28
|
pricing_info: Information used for calculating the cost of FireCrawl usage.
|
29
|
-
|
29
|
+
capture_message_content: Flag indicating whether to trace the actual content.
|
30
30
|
|
31
31
|
Returns:
|
32
32
|
A function that wraps the chat completions method to add telemetry.
|
@@ -30,7 +30,7 @@ class GoogleAIStudioInstrumentor(BaseInstrumentor):
|
|
30
30
|
tracer = kwargs.get("tracer")
|
31
31
|
metrics = kwargs.get("metrics_dict")
|
32
32
|
pricing_info = kwargs.get("pricing_info", {})
|
33
|
-
|
33
|
+
capture_message_content = kwargs.get("capture_message_content", False)
|
34
34
|
disable_metrics = kwargs.get("disable_metrics")
|
35
35
|
version = importlib.metadata.version("google-genai")
|
36
36
|
|
@@ -39,7 +39,7 @@ class GoogleAIStudioInstrumentor(BaseInstrumentor):
|
|
39
39
|
"google.genai.models",
|
40
40
|
"Models.generate_content",
|
41
41
|
generate(version, environment, application_name,
|
42
|
-
tracer, pricing_info,
|
42
|
+
tracer, pricing_info, capture_message_content, metrics, disable_metrics),
|
43
43
|
)
|
44
44
|
|
45
45
|
# async generate
|
@@ -47,7 +47,7 @@ class GoogleAIStudioInstrumentor(BaseInstrumentor):
|
|
47
47
|
"google.genai.models",
|
48
48
|
"AsyncModels.generate_content",
|
49
49
|
async_generate(version, environment,
|
50
|
-
application_name, tracer, pricing_info,
|
50
|
+
application_name, tracer, pricing_info, capture_message_content, metrics,
|
51
51
|
disable_metrics),
|
52
52
|
)
|
53
53
|
|
@@ -19,7 +19,7 @@ from openlit.semcov import SemanticConvetion
|
|
19
19
|
logger = logging.getLogger(__name__)
|
20
20
|
|
21
21
|
def async_generate(version, environment, application_name,
|
22
|
-
tracer, pricing_info,
|
22
|
+
tracer, pricing_info, capture_message_content, metrics, disable_metrics):
|
23
23
|
"""
|
24
24
|
Generates a telemetry wrapper for chat to collect metrics.
|
25
25
|
|
@@ -30,7 +30,7 @@ def async_generate(version, environment, application_name,
|
|
30
30
|
application_name: Name of the application using the Google AI Studio API.
|
31
31
|
tracer: OpenTelemetry tracer for creating spans.
|
32
32
|
pricing_info: Information used for calculating the cost of Google AI Studio usage.
|
33
|
-
|
33
|
+
capture_message_content: Flag indicating whether to trace the actual content.
|
34
34
|
|
35
35
|
Returns:
|
36
36
|
A function that wraps the chat method to add telemetry.
|
@@ -165,7 +165,7 @@ def async_generate(version, environment, application_name,
|
|
165
165
|
end_time - start_time)
|
166
166
|
span.set_attribute(SemanticConvetion.GEN_AI_SDK_VERSION,
|
167
167
|
version)
|
168
|
-
if
|
168
|
+
if capture_message_content:
|
169
169
|
span.add_event(
|
170
170
|
name=SemanticConvetion.GEN_AI_CONTENT_PROMPT_EVENT,
|
171
171
|
attributes={
|
@@ -19,7 +19,7 @@ from openlit.semcov import SemanticConvetion
|
|
19
19
|
logger = logging.getLogger(__name__)
|
20
20
|
|
21
21
|
def generate(version, environment, application_name,
|
22
|
-
tracer, pricing_info,
|
22
|
+
tracer, pricing_info, capture_message_content, metrics, disable_metrics):
|
23
23
|
"""
|
24
24
|
Generates a telemetry wrapper for chat to collect metrics.
|
25
25
|
|
@@ -30,7 +30,7 @@ def generate(version, environment, application_name,
|
|
30
30
|
application_name: Name of the application using the Google AI Studio API.
|
31
31
|
tracer: OpenTelemetry tracer for creating spans.
|
32
32
|
pricing_info: Information used for calculating the cost of Google AI Studio usage.
|
33
|
-
|
33
|
+
capture_message_content: Flag indicating whether to trace the actual content.
|
34
34
|
|
35
35
|
Returns:
|
36
36
|
A function that wraps the chat method to add telemetry.
|
@@ -165,7 +165,7 @@ def generate(version, environment, application_name,
|
|
165
165
|
end_time - start_time)
|
166
166
|
span.set_attribute(SemanticConvetion.GEN_AI_SDK_VERSION,
|
167
167
|
version)
|
168
|
-
if
|
168
|
+
if capture_message_content:
|
169
169
|
span.add_event(
|
170
170
|
name=SemanticConvetion.GEN_AI_CONTENT_PROMPT_EVENT,
|
171
171
|
attributes={
|
@@ -26,7 +26,7 @@ class GPT4AllInstrumentor(BaseInstrumentor):
|
|
26
26
|
tracer = kwargs.get("tracer")
|
27
27
|
metrics = kwargs.get("metrics_dict")
|
28
28
|
pricing_info = kwargs.get("pricing_info", {})
|
29
|
-
|
29
|
+
capture_message_content = kwargs.get("capture_message_content", False)
|
30
30
|
disable_metrics = kwargs.get("disable_metrics")
|
31
31
|
version = importlib.metadata.version("gpt4all")
|
32
32
|
|
@@ -35,7 +35,7 @@ class GPT4AllInstrumentor(BaseInstrumentor):
|
|
35
35
|
"gpt4all",
|
36
36
|
"GPT4All.generate",
|
37
37
|
generate(version, environment, application_name,
|
38
|
-
tracer, pricing_info,
|
38
|
+
tracer, pricing_info, capture_message_content, metrics, disable_metrics),
|
39
39
|
)
|
40
40
|
|
41
41
|
# embed
|
@@ -43,7 +43,7 @@ class GPT4AllInstrumentor(BaseInstrumentor):
|
|
43
43
|
"gpt4all",
|
44
44
|
"Embed4All.embed",
|
45
45
|
embed(version, environment, application_name,
|
46
|
-
tracer, pricing_info,
|
46
|
+
tracer, pricing_info, capture_message_content, metrics, disable_metrics),
|
47
47
|
)
|
48
48
|
|
49
49
|
|
@@ -20,7 +20,7 @@ from openlit.semcov import SemanticConvetion
|
|
20
20
|
logger = logging.getLogger(__name__)
|
21
21
|
|
22
22
|
def generate(version, environment, application_name,
|
23
|
-
tracer, pricing_info,
|
23
|
+
tracer, pricing_info, capture_message_content, metrics, disable_metrics):
|
24
24
|
"""
|
25
25
|
Generates a telemetry wrapper for chat completions to collect metrics.
|
26
26
|
|
@@ -30,7 +30,7 @@ def generate(version, environment, application_name,
|
|
30
30
|
application_name: Name of the application using the GPT4All API.
|
31
31
|
tracer: OpenTelemetry tracer for creating spans.
|
32
32
|
pricing_info: Information used for calculating GPT4All usage.
|
33
|
-
|
33
|
+
capture_message_content: Flag indicating whether to trace the actual content.
|
34
34
|
|
35
35
|
Returns:
|
36
36
|
A function that wraps the chat completions method to add telemetry.
|
@@ -168,7 +168,7 @@ def generate(version, environment, application_name,
|
|
168
168
|
version)
|
169
169
|
self._span.set_attribute(SemanticConvetion.GEN_AI_USAGE_COST,
|
170
170
|
0)
|
171
|
-
if
|
171
|
+
if capture_message_content:
|
172
172
|
self._span.add_event(
|
173
173
|
name=SemanticConvetion.GEN_AI_CONTENT_PROMPT_EVENT,
|
174
174
|
attributes={
|
@@ -319,7 +319,7 @@ def generate(version, environment, application_name,
|
|
319
319
|
version)
|
320
320
|
span.set_attribute(SemanticConvetion.GEN_AI_USAGE_COST,
|
321
321
|
0)
|
322
|
-
if
|
322
|
+
if capture_message_content:
|
323
323
|
span.add_event(
|
324
324
|
name=SemanticConvetion.GEN_AI_CONTENT_PROMPT_EVENT,
|
325
325
|
attributes={
|
@@ -374,7 +374,7 @@ def generate(version, environment, application_name,
|
|
374
374
|
return wrapper
|
375
375
|
|
376
376
|
def embed(version, environment, application_name,
|
377
|
-
tracer, pricing_info,
|
377
|
+
tracer, pricing_info, capture_message_content, metrics, disable_metrics):
|
378
378
|
"""
|
379
379
|
Generates a telemetry wrapper for embeddings to collect metrics.
|
380
380
|
|
@@ -384,7 +384,7 @@ def embed(version, environment, application_name,
|
|
384
384
|
application_name: Name of the application using the GPT4All API.
|
385
385
|
tracer: OpenTelemetry tracer for creating spans.
|
386
386
|
pricing_info: Information used for calculating GPT4All usage.
|
387
|
-
|
387
|
+
capture_message_content: Flag indicating whether to trace the actual content.
|
388
388
|
|
389
389
|
Returns:
|
390
390
|
A function that wraps the embeddings method to add telemetry.
|
@@ -452,7 +452,7 @@ def embed(version, environment, application_name,
|
|
452
452
|
span.set_attribute(SemanticConvetion.GEN_AI_USAGE_COST,
|
453
453
|
0)
|
454
454
|
|
455
|
-
if
|
455
|
+
if capture_message_content:
|
456
456
|
span.add_event(
|
457
457
|
name=SemanticConvetion.GEN_AI_CONTENT_PROMPT_EVENT,
|
458
458
|
attributes={
|
@@ -25,7 +25,7 @@ class GroqInstrumentor(BaseInstrumentor):
|
|
25
25
|
tracer = kwargs.get("tracer")
|
26
26
|
metrics = kwargs.get("metrics_dict")
|
27
27
|
pricing_info = kwargs.get("pricing_info", {})
|
28
|
-
|
28
|
+
capture_message_content = kwargs.get("capture_message_content", False)
|
29
29
|
disable_metrics = kwargs.get("disable_metrics")
|
30
30
|
version = importlib.metadata.version("groq")
|
31
31
|
|
@@ -34,7 +34,7 @@ class GroqInstrumentor(BaseInstrumentor):
|
|
34
34
|
"groq.resources.chat.completions",
|
35
35
|
"Completions.create",
|
36
36
|
chat(version, environment, application_name,
|
37
|
-
tracer, pricing_info,
|
37
|
+
tracer, pricing_info, capture_message_content, metrics, disable_metrics),
|
38
38
|
)
|
39
39
|
|
40
40
|
#async
|
@@ -42,7 +42,7 @@ class GroqInstrumentor(BaseInstrumentor):
|
|
42
42
|
"groq.resources.chat.completions",
|
43
43
|
"AsyncCompletions.create",
|
44
44
|
async_chat(version, environment, application_name,
|
45
|
-
tracer, pricing_info,
|
45
|
+
tracer, pricing_info, capture_message_content, metrics, disable_metrics),
|
46
46
|
)
|
47
47
|
|
48
48
|
def _uninstrument(self, **kwargs):
|
@@ -21,7 +21,7 @@ from openlit.semcov import SemanticConvetion
|
|
21
21
|
logger = logging.getLogger(__name__)
|
22
22
|
|
23
23
|
def async_chat(version, environment, application_name,
|
24
|
-
tracer, pricing_info,
|
24
|
+
tracer, pricing_info, capture_message_content, metrics, disable_metrics):
|
25
25
|
"""
|
26
26
|
Generates a telemetry wrapper for chat completions to collect metrics.
|
27
27
|
|
@@ -31,7 +31,7 @@ def async_chat(version, environment, application_name,
|
|
31
31
|
application_name: Name of the application using the Groq API.
|
32
32
|
tracer: OpenTelemetry tracer for creating spans.
|
33
33
|
pricing_info: Information used for calculating the cost of Groq usage.
|
34
|
-
|
34
|
+
capture_message_content: Flag indicating whether to trace the actual content.
|
35
35
|
|
36
36
|
Returns:
|
37
37
|
A function that wraps the chat completions method to add telemetry.
|
@@ -221,7 +221,7 @@ def async_chat(version, environment, application_name,
|
|
221
221
|
self._ttft)
|
222
222
|
self._span.set_attribute(SemanticConvetion.GEN_AI_SDK_VERSION,
|
223
223
|
version)
|
224
|
-
if
|
224
|
+
if capture_message_content:
|
225
225
|
self._span.add_event(
|
226
226
|
name=SemanticConvetion.GEN_AI_CONTENT_PROMPT_EVENT,
|
227
227
|
attributes={
|
@@ -396,7 +396,7 @@ def async_chat(version, environment, application_name,
|
|
396
396
|
end_time - start_time)
|
397
397
|
span.set_attribute(SemanticConvetion.GEN_AI_SDK_VERSION,
|
398
398
|
version)
|
399
|
-
if
|
399
|
+
if capture_message_content:
|
400
400
|
span.add_event(
|
401
401
|
name=SemanticConvetion.GEN_AI_CONTENT_PROMPT_EVENT,
|
402
402
|
attributes={
|
@@ -407,7 +407,7 @@ def async_chat(version, environment, application_name,
|
|
407
407
|
for i in range(kwargs.get('n',1)):
|
408
408
|
span.set_attribute(SemanticConvetion.GEN_AI_RESPONSE_FINISH_REASON,
|
409
409
|
[response_dict.get('choices')[i].get('finish_reason')])
|
410
|
-
if
|
410
|
+
if capture_message_content:
|
411
411
|
span.add_event(
|
412
412
|
name=SemanticConvetion.GEN_AI_CONTENT_COMPLETION_EVENT,
|
413
413
|
attributes={
|
@@ -21,7 +21,7 @@ from openlit.semcov import SemanticConvetion
|
|
21
21
|
logger = logging.getLogger(__name__)
|
22
22
|
|
23
23
|
def chat(version, environment, application_name,
|
24
|
-
tracer, pricing_info,
|
24
|
+
tracer, pricing_info, capture_message_content, metrics, disable_metrics):
|
25
25
|
"""
|
26
26
|
Generates a telemetry wrapper for chat completions to collect metrics.
|
27
27
|
|
@@ -31,7 +31,7 @@ def chat(version, environment, application_name,
|
|
31
31
|
application_name: Name of the application using the Groq API.
|
32
32
|
tracer: OpenTelemetry tracer for creating spans.
|
33
33
|
pricing_info: Information used for calculating the cost of Groq usage.
|
34
|
-
|
34
|
+
capture_message_content: Flag indicating whether to trace the actual content.
|
35
35
|
|
36
36
|
Returns:
|
37
37
|
A function that wraps the chat completions method to add telemetry.
|
@@ -221,7 +221,7 @@ def chat(version, environment, application_name,
|
|
221
221
|
self._ttft)
|
222
222
|
self._span.set_attribute(SemanticConvetion.GEN_AI_SDK_VERSION,
|
223
223
|
version)
|
224
|
-
if
|
224
|
+
if capture_message_content:
|
225
225
|
self._span.add_event(
|
226
226
|
name=SemanticConvetion.GEN_AI_CONTENT_PROMPT_EVENT,
|
227
227
|
attributes={
|
@@ -396,7 +396,7 @@ def chat(version, environment, application_name,
|
|
396
396
|
end_time - start_time)
|
397
397
|
span.set_attribute(SemanticConvetion.GEN_AI_SDK_VERSION,
|
398
398
|
version)
|
399
|
-
if
|
399
|
+
if capture_message_content:
|
400
400
|
span.add_event(
|
401
401
|
name=SemanticConvetion.GEN_AI_CONTENT_PROMPT_EVENT,
|
402
402
|
attributes={
|
@@ -407,7 +407,7 @@ def chat(version, environment, application_name,
|
|
407
407
|
for i in range(kwargs.get('n',1)):
|
408
408
|
span.set_attribute(SemanticConvetion.GEN_AI_RESPONSE_FINISH_REASON,
|
409
409
|
[response_dict.get('choices')[i].get('finish_reason')])
|
410
|
-
if
|
410
|
+
if capture_message_content:
|
411
411
|
span.add_event(
|
412
412
|
name=SemanticConvetion.GEN_AI_CONTENT_COMPLETION_EVENT,
|
413
413
|
attributes={
|
@@ -29,7 +29,7 @@ class HaystackInstrumentor(BaseInstrumentor):
|
|
29
29
|
environment = kwargs.get("environment")
|
30
30
|
tracer = kwargs.get("tracer")
|
31
31
|
pricing_info = kwargs.get("pricing_info")
|
32
|
-
|
32
|
+
capture_message_content = kwargs.get("capture_message_content")
|
33
33
|
version = importlib.metadata.version("haystack-ai")
|
34
34
|
|
35
35
|
for wrapped_method in WRAPPED_METHODS:
|
@@ -41,7 +41,7 @@ class HaystackInstrumentor(BaseInstrumentor):
|
|
41
41
|
wrap_package,
|
42
42
|
wrap_object,
|
43
43
|
wrapper(gen_ai_endpoint, version, environment, application_name,
|
44
|
-
tracer, pricing_info,
|
44
|
+
tracer, pricing_info, capture_message_content),
|
45
45
|
)
|
46
46
|
|
47
47
|
@staticmethod
|
@@ -13,7 +13,7 @@ from openlit.semcov import SemanticConvetion
|
|
13
13
|
logger = logging.getLogger(__name__)
|
14
14
|
|
15
15
|
def join_data(gen_ai_endpoint, version, environment, application_name,
|
16
|
-
tracer, pricing_info,
|
16
|
+
tracer, pricing_info, capture_message_content):
|
17
17
|
"""
|
18
18
|
Creates a wrapper around a function call to trace and log its execution metrics.
|
19
19
|
|
@@ -27,7 +27,7 @@ def join_data(gen_ai_endpoint, version, environment, application_name,
|
|
27
27
|
- application_name (str): Name of the Haystack application.
|
28
28
|
- tracer (opentelemetry.trace.Tracer): The tracer object used for OpenTelemetry tracing.
|
29
29
|
- pricing_info (dict): Information about the pricing for internal metrics (currently not used).
|
30
|
-
-
|
30
|
+
- capture_message_content (bool): Flag indicating whether to trace the content of the response.
|
31
31
|
|
32
32
|
Returns:
|
33
33
|
- function: A higher-order function that takes a function 'wrapped' and returns
|
@@ -30,7 +30,7 @@ class JulepInstrumentor(BaseInstrumentor):
|
|
30
30
|
tracer = kwargs.get("tracer")
|
31
31
|
metrics = kwargs.get("metrics_dict")
|
32
32
|
pricing_info = kwargs.get("pricing_info", {})
|
33
|
-
|
33
|
+
capture_message_content = kwargs.get("capture_message_content", False)
|
34
34
|
disable_metrics = kwargs.get("disable_metrics")
|
35
35
|
version = importlib.metadata.version("julep")
|
36
36
|
|
@@ -39,19 +39,19 @@ class JulepInstrumentor(BaseInstrumentor):
|
|
39
39
|
"julep.resources.agents.agents",
|
40
40
|
"AgentsResource.create",
|
41
41
|
wrap_julep("julep.agents_create", version, environment, application_name,
|
42
|
-
tracer, pricing_info,
|
42
|
+
tracer, pricing_info, capture_message_content, metrics, disable_metrics),
|
43
43
|
)
|
44
44
|
wrap_function_wrapper(
|
45
45
|
"julep.resources.tasks",
|
46
46
|
"TasksResource.create",
|
47
47
|
wrap_julep("julep.task_create", version, environment, application_name,
|
48
|
-
tracer, pricing_info,
|
48
|
+
tracer, pricing_info, capture_message_content, metrics, disable_metrics),
|
49
49
|
)
|
50
50
|
wrap_function_wrapper(
|
51
51
|
"julep.resources.executions.executions",
|
52
52
|
"ExecutionsResource.create",
|
53
53
|
wrap_julep("julep.execution_create", version, environment, application_name,
|
54
|
-
tracer, pricing_info,
|
54
|
+
tracer, pricing_info, capture_message_content, metrics, disable_metrics),
|
55
55
|
)
|
56
56
|
|
57
57
|
# async
|
@@ -59,19 +59,19 @@ class JulepInstrumentor(BaseInstrumentor):
|
|
59
59
|
"julep.resources.agents.agents",
|
60
60
|
"AsyncAgentsResource.create",
|
61
61
|
async_wrap_julep("julep.agents_create", version, environment, application_name,
|
62
|
-
tracer, pricing_info,
|
62
|
+
tracer, pricing_info, capture_message_content, metrics, disable_metrics),
|
63
63
|
)
|
64
64
|
wrap_function_wrapper(
|
65
65
|
"julep.resources.tasks",
|
66
66
|
"AsyncTasksResource.create",
|
67
67
|
async_wrap_julep("julep.task_create", version, environment, application_name,
|
68
|
-
tracer, pricing_info,
|
68
|
+
tracer, pricing_info, capture_message_content, metrics, disable_metrics),
|
69
69
|
)
|
70
70
|
wrap_function_wrapper(
|
71
71
|
"julep.resources.executions.executions",
|
72
72
|
"AsyncExecutionsResource.create",
|
73
73
|
async_wrap_julep("julep.execution_create", version, environment, application_name,
|
74
|
-
tracer, pricing_info,
|
74
|
+
tracer, pricing_info, capture_message_content, metrics, disable_metrics),
|
75
75
|
)
|
76
76
|
|
77
77
|
|
@@ -13,7 +13,7 @@ from openlit.semcov import SemanticConvetion
|
|
13
13
|
logger = logging.getLogger(__name__)
|
14
14
|
|
15
15
|
def async_wrap_julep(gen_ai_endpoint, version, environment, application_name,
|
16
|
-
tracer, pricing_info,
|
16
|
+
tracer, pricing_info, capture_message_content, metrics, disable_metrics):
|
17
17
|
"""
|
18
18
|
Creates a wrapper around a function call to trace and log its execution metrics.
|
19
19
|
|
@@ -27,7 +27,7 @@ def async_wrap_julep(gen_ai_endpoint, version, environment, application_name,
|
|
27
27
|
- application_name (str): Name of the application.
|
28
28
|
- tracer (opentelemetry.trace.Tracer): The tracer object used for OpenTelemetry tracing.
|
29
29
|
- pricing_info (dict): Information about the pricing for internal metrics (currently not used).
|
30
|
-
-
|
30
|
+
- capture_message_content (bool): Flag indicating whether to trace the content of the response.
|
31
31
|
|
32
32
|
Returns:
|
33
33
|
- function: A higher-order function that takes a function 'wrapped' and returns
|
@@ -88,7 +88,7 @@ def async_wrap_julep(gen_ai_endpoint, version, environment, application_name,
|
|
88
88
|
elif gen_ai_endpoint == "julep.execution_create":
|
89
89
|
span.set_attribute(SemanticConvetion.GEN_AI_AGENT_TASK_ID,
|
90
90
|
kwargs.get("task_id", ""))
|
91
|
-
if
|
91
|
+
if capture_message_content:
|
92
92
|
span.add_event(
|
93
93
|
name=SemanticConvetion.GEN_AI_CONTENT_PROMPT_EVENT,
|
94
94
|
attributes={
|
@@ -13,7 +13,7 @@ from openlit.semcov import SemanticConvetion
|
|
13
13
|
logger = logging.getLogger(__name__)
|
14
14
|
|
15
15
|
def wrap_julep(gen_ai_endpoint, version, environment, application_name,
|
16
|
-
tracer, pricing_info,
|
16
|
+
tracer, pricing_info, capture_message_content, metrics, disable_metrics):
|
17
17
|
"""
|
18
18
|
Creates a wrapper around a function call to trace and log its execution metrics.
|
19
19
|
|
@@ -27,7 +27,7 @@ def wrap_julep(gen_ai_endpoint, version, environment, application_name,
|
|
27
27
|
- application_name (str): Name of the Langchain application.
|
28
28
|
- tracer (opentelemetry.trace.Tracer): The tracer object used for OpenTelemetry tracing.
|
29
29
|
- pricing_info (dict): Information about the pricing for internal metrics (currently not used).
|
30
|
-
-
|
30
|
+
- capture_message_content (bool): Flag indicating whether to trace the content of the response.
|
31
31
|
|
32
32
|
Returns:
|
33
33
|
- function: A higher-order function that takes a function 'wrapped' and returns
|
@@ -88,7 +88,7 @@ def wrap_julep(gen_ai_endpoint, version, environment, application_name,
|
|
88
88
|
elif gen_ai_endpoint == "julep.execution_create":
|
89
89
|
span.set_attribute(SemanticConvetion.GEN_AI_AGENT_TASK_ID,
|
90
90
|
kwargs.get("task_id", ""))
|
91
|
-
if
|
91
|
+
if capture_message_content:
|
92
92
|
span.add_event(
|
93
93
|
name=SemanticConvetion.GEN_AI_CONTENT_PROMPT_EVENT,
|
94
94
|
attributes={
|