openlit 1.16.2__py3-none-any.whl → 1.17.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- openlit/instrumentation/anthropic/anthropic.py +28 -10
- openlit/instrumentation/anthropic/async_anthropic.py +27 -10
- openlit/instrumentation/bedrock/bedrock.py +13 -5
- openlit/instrumentation/cohere/cohere.py +33 -12
- openlit/instrumentation/elevenlabs/async_elevenlabs.py +6 -2
- openlit/instrumentation/elevenlabs/elevenlabs.py +6 -2
- openlit/instrumentation/gpt4all/gpt4all.py +30 -10
- openlit/instrumentation/groq/async_groq.py +31 -11
- openlit/instrumentation/groq/groq.py +31 -11
- openlit/instrumentation/mistral/async_mistral.py +33 -12
- openlit/instrumentation/mistral/mistral.py +33 -12
- openlit/instrumentation/ollama/async_ollama.py +57 -20
- openlit/instrumentation/ollama/ollama.py +57 -20
- openlit/instrumentation/openai/async_azure_openai.py +94 -35
- openlit/instrumentation/openai/async_openai.py +68 -27
- openlit/instrumentation/openai/azure_openai.py +89 -31
- openlit/instrumentation/openai/openai.py +68 -29
- openlit/instrumentation/transformers/transformers.py +20 -16
- openlit/instrumentation/vertexai/async_vertexai.py +104 -35
- openlit/instrumentation/vertexai/vertexai.py +104 -35
- openlit/semcov/__init__.py +3 -1
- {openlit-1.16.2.dist-info → openlit-1.17.0.dist-info}/METADATA +1 -1
- {openlit-1.16.2.dist-info → openlit-1.17.0.dist-info}/RECORD +25 -25
- {openlit-1.16.2.dist-info → openlit-1.17.0.dist-info}/LICENSE +0 -0
- {openlit-1.16.2.dist-info → openlit-1.17.0.dist-info}/WHEEL +0 -0
@@ -103,10 +103,18 @@ def generate_content_async(gen_ai_endpoint, version, environment, application_na
|
|
103
103
|
span.set_attribute(SemanticConvetion.GEN_AI_USAGE_COST,
|
104
104
|
cost)
|
105
105
|
if trace_content:
|
106
|
-
span.
|
107
|
-
|
108
|
-
|
109
|
-
|
106
|
+
span.add_event(
|
107
|
+
name=SemanticConvetion.GEN_AI_CONTENT_PROMPT_EVENT,
|
108
|
+
attributes={
|
109
|
+
SemanticConvetion.GEN_AI_CONTENT_PROMPT: prompt,
|
110
|
+
},
|
111
|
+
)
|
112
|
+
span.add_event(
|
113
|
+
name=SemanticConvetion.GEN_AI_CONTENT_COMPLETION_EVENT,
|
114
|
+
attributes={
|
115
|
+
SemanticConvetion.GEN_AI_CONTENT_COMPLETION: llmresponse,
|
116
|
+
},
|
117
|
+
)
|
110
118
|
|
111
119
|
span.set_status(Status(StatusCode.OK))
|
112
120
|
|
@@ -181,10 +189,19 @@ def generate_content_async(gen_ai_endpoint, version, environment, application_na
|
|
181
189
|
span.set_attribute(SemanticConvetion.GEN_AI_USAGE_COST,
|
182
190
|
cost)
|
183
191
|
if trace_content:
|
184
|
-
span.
|
185
|
-
|
186
|
-
|
187
|
-
|
192
|
+
span.add_event(
|
193
|
+
name=SemanticConvetion.GEN_AI_CONTENT_PROMPT_EVENT,
|
194
|
+
attributes={
|
195
|
+
SemanticConvetion.GEN_AI_CONTENT_PROMPT: prompt,
|
196
|
+
},
|
197
|
+
)
|
198
|
+
span.add_event(
|
199
|
+
name=SemanticConvetion.GEN_AI_CONTENT_COMPLETION_EVENT,
|
200
|
+
attributes={
|
201
|
+
# pylint: disable=line-too-long
|
202
|
+
SemanticConvetion.GEN_AI_CONTENT_COMPLETION: response.candidates[0].content.parts[0].text,
|
203
|
+
},
|
204
|
+
)
|
188
205
|
|
189
206
|
span.set_status(Status(StatusCode.OK))
|
190
207
|
|
@@ -316,10 +333,18 @@ def send_message_async(gen_ai_endpoint, version, environment, application_name,
|
|
316
333
|
span.set_attribute(SemanticConvetion.GEN_AI_USAGE_COST,
|
317
334
|
cost)
|
318
335
|
if trace_content:
|
319
|
-
span.
|
320
|
-
|
321
|
-
|
322
|
-
|
336
|
+
span.add_event(
|
337
|
+
name=SemanticConvetion.GEN_AI_CONTENT_PROMPT_EVENT,
|
338
|
+
attributes={
|
339
|
+
SemanticConvetion.GEN_AI_CONTENT_PROMPT: prompt,
|
340
|
+
},
|
341
|
+
)
|
342
|
+
span.add_event(
|
343
|
+
name=SemanticConvetion.GEN_AI_CONTENT_COMPLETION_EVENT,
|
344
|
+
attributes={
|
345
|
+
SemanticConvetion.GEN_AI_CONTENT_COMPLETION: llmresponse,
|
346
|
+
},
|
347
|
+
)
|
323
348
|
|
324
349
|
span.set_status(Status(StatusCode.OK))
|
325
350
|
|
@@ -394,11 +419,19 @@ def send_message_async(gen_ai_endpoint, version, environment, application_name,
|
|
394
419
|
span.set_attribute(SemanticConvetion.GEN_AI_USAGE_COST,
|
395
420
|
cost)
|
396
421
|
if trace_content:
|
397
|
-
span.
|
398
|
-
|
399
|
-
|
400
|
-
|
401
|
-
|
422
|
+
span.add_event(
|
423
|
+
name=SemanticConvetion.GEN_AI_CONTENT_PROMPT_EVENT,
|
424
|
+
attributes={
|
425
|
+
SemanticConvetion.GEN_AI_CONTENT_PROMPT: prompt,
|
426
|
+
},
|
427
|
+
)
|
428
|
+
span.add_event(
|
429
|
+
name=SemanticConvetion.GEN_AI_CONTENT_COMPLETION_EVENT,
|
430
|
+
attributes={
|
431
|
+
# pylint: disable=line-too-long
|
432
|
+
SemanticConvetion.GEN_AI_CONTENT_COMPLETION: response.candidates[0].content.parts[0].text,
|
433
|
+
},
|
434
|
+
)
|
402
435
|
|
403
436
|
span.set_status(Status(StatusCode.OK))
|
404
437
|
|
@@ -516,10 +549,18 @@ def predict_async(gen_ai_endpoint, version, environment, application_name, trace
|
|
516
549
|
span.set_attribute(SemanticConvetion.GEN_AI_USAGE_COST,
|
517
550
|
cost)
|
518
551
|
if trace_content:
|
519
|
-
span.
|
520
|
-
|
521
|
-
|
522
|
-
|
552
|
+
span.add_event(
|
553
|
+
name=SemanticConvetion.GEN_AI_CONTENT_PROMPT_EVENT,
|
554
|
+
attributes={
|
555
|
+
SemanticConvetion.GEN_AI_CONTENT_PROMPT: prompt,
|
556
|
+
},
|
557
|
+
)
|
558
|
+
span.add_event(
|
559
|
+
name=SemanticConvetion.GEN_AI_CONTENT_COMPLETION_EVENT,
|
560
|
+
attributes={
|
561
|
+
SemanticConvetion.GEN_AI_CONTENT_COMPLETION: response.text,
|
562
|
+
},
|
563
|
+
)
|
523
564
|
|
524
565
|
span.set_status(Status(StatusCode.OK))
|
525
566
|
|
@@ -648,10 +689,18 @@ def predict_streaming_async(gen_ai_endpoint, version, environment, application_n
|
|
648
689
|
span.set_attribute(SemanticConvetion.GEN_AI_USAGE_COST,
|
649
690
|
cost)
|
650
691
|
if trace_content:
|
651
|
-
span.
|
652
|
-
|
653
|
-
|
654
|
-
|
692
|
+
span.add_event(
|
693
|
+
name=SemanticConvetion.GEN_AI_CONTENT_PROMPT_EVENT,
|
694
|
+
attributes={
|
695
|
+
SemanticConvetion.GEN_AI_CONTENT_PROMPT: prompt,
|
696
|
+
},
|
697
|
+
)
|
698
|
+
span.add_event(
|
699
|
+
name=SemanticConvetion.GEN_AI_CONTENT_COMPLETION_EVENT,
|
700
|
+
attributes={
|
701
|
+
SemanticConvetion.GEN_AI_CONTENT_COMPLETION: llmresponse,
|
702
|
+
},
|
703
|
+
)
|
655
704
|
|
656
705
|
span.set_status(Status(StatusCode.OK))
|
657
706
|
|
@@ -765,10 +814,18 @@ def start_chat_async(gen_ai_endpoint, version, environment, application_name, tr
|
|
765
814
|
span.set_attribute(SemanticConvetion.GEN_AI_USAGE_COST,
|
766
815
|
cost)
|
767
816
|
if trace_content:
|
768
|
-
span.
|
769
|
-
|
770
|
-
|
771
|
-
|
817
|
+
span.add_event(
|
818
|
+
name=SemanticConvetion.GEN_AI_CONTENT_PROMPT_EVENT,
|
819
|
+
attributes={
|
820
|
+
SemanticConvetion.GEN_AI_CONTENT_PROMPT: prompt,
|
821
|
+
},
|
822
|
+
)
|
823
|
+
span.add_event(
|
824
|
+
name=SemanticConvetion.GEN_AI_CONTENT_COMPLETION_EVENT,
|
825
|
+
attributes={
|
826
|
+
SemanticConvetion.GEN_AI_CONTENT_COMPLETION: response.text,
|
827
|
+
},
|
828
|
+
)
|
772
829
|
|
773
830
|
span.set_status(Status(StatusCode.OK))
|
774
831
|
|
@@ -895,10 +952,18 @@ def start_chat_streaming_async(gen_ai_endpoint, version, environment, applicatio
|
|
895
952
|
span.set_attribute(SemanticConvetion.GEN_AI_USAGE_COST,
|
896
953
|
cost)
|
897
954
|
if trace_content:
|
898
|
-
span.
|
899
|
-
|
900
|
-
|
901
|
-
|
955
|
+
span.add_event(
|
956
|
+
name=SemanticConvetion.GEN_AI_CONTENT_PROMPT_EVENT,
|
957
|
+
attributes={
|
958
|
+
SemanticConvetion.GEN_AI_CONTENT_PROMPT: prompt,
|
959
|
+
},
|
960
|
+
)
|
961
|
+
span.add_event(
|
962
|
+
name=SemanticConvetion.GEN_AI_CONTENT_COMPLETION_EVENT,
|
963
|
+
attributes={
|
964
|
+
SemanticConvetion.GEN_AI_CONTENT_COMPLETION: llmresponse,
|
965
|
+
},
|
966
|
+
)
|
902
967
|
|
903
968
|
span.set_status(Status(StatusCode.OK))
|
904
969
|
|
@@ -1006,8 +1071,12 @@ def embeddings_async(gen_ai_endpoint, version, environment, application_name, tr
|
|
1006
1071
|
span.set_attribute(SemanticConvetion.GEN_AI_USAGE_COST,
|
1007
1072
|
cost)
|
1008
1073
|
if trace_content:
|
1009
|
-
span.
|
1010
|
-
|
1074
|
+
span.add_event(
|
1075
|
+
name=SemanticConvetion.GEN_AI_CONTENT_PROMPT_EVENT,
|
1076
|
+
attributes={
|
1077
|
+
SemanticConvetion.GEN_AI_CONTENT_PROMPT: prompt,
|
1078
|
+
},
|
1079
|
+
)
|
1011
1080
|
|
1012
1081
|
span.set_status(Status(StatusCode.OK))
|
1013
1082
|
|
@@ -103,10 +103,18 @@ def generate_content(gen_ai_endpoint, version, environment, application_name, tr
|
|
103
103
|
span.set_attribute(SemanticConvetion.GEN_AI_USAGE_COST,
|
104
104
|
cost)
|
105
105
|
if trace_content:
|
106
|
-
span.
|
107
|
-
|
108
|
-
|
109
|
-
|
106
|
+
span.add_event(
|
107
|
+
name=SemanticConvetion.GEN_AI_CONTENT_PROMPT_EVENT,
|
108
|
+
attributes={
|
109
|
+
SemanticConvetion.GEN_AI_CONTENT_PROMPT: prompt,
|
110
|
+
},
|
111
|
+
)
|
112
|
+
span.add_event(
|
113
|
+
name=SemanticConvetion.GEN_AI_CONTENT_COMPLETION_EVENT,
|
114
|
+
attributes={
|
115
|
+
SemanticConvetion.GEN_AI_CONTENT_COMPLETION: llmresponse,
|
116
|
+
},
|
117
|
+
)
|
110
118
|
|
111
119
|
span.set_status(Status(StatusCode.OK))
|
112
120
|
|
@@ -181,10 +189,19 @@ def generate_content(gen_ai_endpoint, version, environment, application_name, tr
|
|
181
189
|
span.set_attribute(SemanticConvetion.GEN_AI_USAGE_COST,
|
182
190
|
cost)
|
183
191
|
if trace_content:
|
184
|
-
span.
|
185
|
-
|
186
|
-
|
187
|
-
|
192
|
+
span.add_event(
|
193
|
+
name=SemanticConvetion.GEN_AI_CONTENT_PROMPT_EVENT,
|
194
|
+
attributes={
|
195
|
+
SemanticConvetion.GEN_AI_CONTENT_PROMPT: prompt,
|
196
|
+
},
|
197
|
+
)
|
198
|
+
span.add_event(
|
199
|
+
name=SemanticConvetion.GEN_AI_CONTENT_COMPLETION_EVENT,
|
200
|
+
attributes={
|
201
|
+
# pylint: disable=line-too-long
|
202
|
+
SemanticConvetion.GEN_AI_CONTENT_COMPLETION: response.candidates[0].content.parts[0].text,
|
203
|
+
},
|
204
|
+
)
|
188
205
|
|
189
206
|
span.set_status(Status(StatusCode.OK))
|
190
207
|
|
@@ -316,10 +333,18 @@ def send_message(gen_ai_endpoint, version, environment, application_name, tracer
|
|
316
333
|
span.set_attribute(SemanticConvetion.GEN_AI_USAGE_COST,
|
317
334
|
cost)
|
318
335
|
if trace_content:
|
319
|
-
span.
|
320
|
-
|
321
|
-
|
322
|
-
|
336
|
+
span.add_event(
|
337
|
+
name=SemanticConvetion.GEN_AI_CONTENT_PROMPT_EVENT,
|
338
|
+
attributes={
|
339
|
+
SemanticConvetion.GEN_AI_CONTENT_PROMPT: prompt,
|
340
|
+
},
|
341
|
+
)
|
342
|
+
span.add_event(
|
343
|
+
name=SemanticConvetion.GEN_AI_CONTENT_COMPLETION_EVENT,
|
344
|
+
attributes={
|
345
|
+
SemanticConvetion.GEN_AI_CONTENT_COMPLETION: llmresponse,
|
346
|
+
},
|
347
|
+
)
|
323
348
|
|
324
349
|
span.set_status(Status(StatusCode.OK))
|
325
350
|
|
@@ -394,11 +419,19 @@ def send_message(gen_ai_endpoint, version, environment, application_name, tracer
|
|
394
419
|
span.set_attribute(SemanticConvetion.GEN_AI_USAGE_COST,
|
395
420
|
cost)
|
396
421
|
if trace_content:
|
397
|
-
span.
|
398
|
-
|
399
|
-
|
400
|
-
|
401
|
-
|
422
|
+
span.add_event(
|
423
|
+
name=SemanticConvetion.GEN_AI_CONTENT_PROMPT_EVENT,
|
424
|
+
attributes={
|
425
|
+
SemanticConvetion.GEN_AI_CONTENT_PROMPT: prompt,
|
426
|
+
},
|
427
|
+
)
|
428
|
+
span.add_event(
|
429
|
+
name=SemanticConvetion.GEN_AI_CONTENT_COMPLETION_EVENT,
|
430
|
+
attributes={
|
431
|
+
# pylint: disable=line-too-long
|
432
|
+
SemanticConvetion.GEN_AI_CONTENT_COMPLETION: response.candidates[0].content.parts[0].text,
|
433
|
+
},
|
434
|
+
)
|
402
435
|
|
403
436
|
span.set_status(Status(StatusCode.OK))
|
404
437
|
|
@@ -516,10 +549,18 @@ def predict(gen_ai_endpoint, version, environment, application_name, tracer,
|
|
516
549
|
span.set_attribute(SemanticConvetion.GEN_AI_USAGE_COST,
|
517
550
|
cost)
|
518
551
|
if trace_content:
|
519
|
-
span.
|
520
|
-
|
521
|
-
|
522
|
-
|
552
|
+
span.add_event(
|
553
|
+
name=SemanticConvetion.GEN_AI_CONTENT_PROMPT_EVENT,
|
554
|
+
attributes={
|
555
|
+
SemanticConvetion.GEN_AI_CONTENT_PROMPT: prompt,
|
556
|
+
},
|
557
|
+
)
|
558
|
+
span.add_event(
|
559
|
+
name=SemanticConvetion.GEN_AI_CONTENT_COMPLETION_EVENT,
|
560
|
+
attributes={
|
561
|
+
SemanticConvetion.GEN_AI_CONTENT_COMPLETION: response.text,
|
562
|
+
},
|
563
|
+
)
|
523
564
|
|
524
565
|
span.set_status(Status(StatusCode.OK))
|
525
566
|
|
@@ -648,10 +689,18 @@ def predict_streaming(gen_ai_endpoint, version, environment, application_name, t
|
|
648
689
|
span.set_attribute(SemanticConvetion.GEN_AI_USAGE_COST,
|
649
690
|
cost)
|
650
691
|
if trace_content:
|
651
|
-
span.
|
652
|
-
|
653
|
-
|
654
|
-
|
692
|
+
span.add_event(
|
693
|
+
name=SemanticConvetion.GEN_AI_CONTENT_PROMPT_EVENT,
|
694
|
+
attributes={
|
695
|
+
SemanticConvetion.GEN_AI_CONTENT_PROMPT: prompt,
|
696
|
+
},
|
697
|
+
)
|
698
|
+
span.add_event(
|
699
|
+
name=SemanticConvetion.GEN_AI_CONTENT_COMPLETION_EVENT,
|
700
|
+
attributes={
|
701
|
+
SemanticConvetion.GEN_AI_CONTENT_COMPLETION: llmresponse,
|
702
|
+
},
|
703
|
+
)
|
655
704
|
|
656
705
|
span.set_status(Status(StatusCode.OK))
|
657
706
|
|
@@ -765,10 +814,18 @@ def start_chat(gen_ai_endpoint, version, environment, application_name, tracer,
|
|
765
814
|
span.set_attribute(SemanticConvetion.GEN_AI_USAGE_COST,
|
766
815
|
cost)
|
767
816
|
if trace_content:
|
768
|
-
span.
|
769
|
-
|
770
|
-
|
771
|
-
|
817
|
+
span.add_event(
|
818
|
+
name=SemanticConvetion.GEN_AI_CONTENT_PROMPT_EVENT,
|
819
|
+
attributes={
|
820
|
+
SemanticConvetion.GEN_AI_CONTENT_PROMPT: prompt,
|
821
|
+
},
|
822
|
+
)
|
823
|
+
span.add_event(
|
824
|
+
name=SemanticConvetion.GEN_AI_CONTENT_COMPLETION_EVENT,
|
825
|
+
attributes={
|
826
|
+
SemanticConvetion.GEN_AI_CONTENT_COMPLETION: response.text,
|
827
|
+
},
|
828
|
+
)
|
772
829
|
|
773
830
|
span.set_status(Status(StatusCode.OK))
|
774
831
|
|
@@ -895,10 +952,18 @@ def start_chat_streaming(gen_ai_endpoint, version, environment, application_name
|
|
895
952
|
span.set_attribute(SemanticConvetion.GEN_AI_USAGE_COST,
|
896
953
|
cost)
|
897
954
|
if trace_content:
|
898
|
-
span.
|
899
|
-
|
900
|
-
|
901
|
-
|
955
|
+
span.add_event(
|
956
|
+
name=SemanticConvetion.GEN_AI_CONTENT_PROMPT_EVENT,
|
957
|
+
attributes={
|
958
|
+
SemanticConvetion.GEN_AI_CONTENT_PROMPT: prompt,
|
959
|
+
},
|
960
|
+
)
|
961
|
+
span.add_event(
|
962
|
+
name=SemanticConvetion.GEN_AI_CONTENT_COMPLETION_EVENT,
|
963
|
+
attributes={
|
964
|
+
SemanticConvetion.GEN_AI_CONTENT_COMPLETION: llmresponse,
|
965
|
+
},
|
966
|
+
)
|
902
967
|
|
903
968
|
span.set_status(Status(StatusCode.OK))
|
904
969
|
|
@@ -1006,8 +1071,12 @@ def embeddings(gen_ai_endpoint, version, environment, application_name, tracer,
|
|
1006
1071
|
span.set_attribute(SemanticConvetion.GEN_AI_USAGE_COST,
|
1007
1072
|
cost)
|
1008
1073
|
if trace_content:
|
1009
|
-
span.
|
1010
|
-
|
1074
|
+
span.add_event(
|
1075
|
+
name=SemanticConvetion.GEN_AI_CONTENT_PROMPT_EVENT,
|
1076
|
+
attributes={
|
1077
|
+
SemanticConvetion.GEN_AI_CONTENT_PROMPT: prompt,
|
1078
|
+
},
|
1079
|
+
)
|
1011
1080
|
|
1012
1081
|
span.set_status(Status(StatusCode.OK))
|
1013
1082
|
|
openlit/semcov/__init__.py
CHANGED
@@ -70,7 +70,9 @@ class SemanticConvetion:
|
|
70
70
|
GEN_AI_RESPONSE_IMAGE = "gen_ai.response.image" # Not used directly in code yet
|
71
71
|
|
72
72
|
# GenAI Content
|
73
|
+
GEN_AI_CONTENT_PROMPT_EVENT = "gen_ai.content.prompt"
|
73
74
|
GEN_AI_CONTENT_PROMPT = "gen_ai.prompt"
|
75
|
+
GEN_AI_CONTENT_COMPLETION_EVENT = "gen_ai.content.completion"
|
74
76
|
GEN_AI_CONTENT_COMPLETION = "gen_ai.completion"
|
75
77
|
GEN_AI_CONTENT_REVISED_PROMPT = "gen_ai.content.revised_prompt"
|
76
78
|
|
@@ -94,7 +96,7 @@ class SemanticConvetion:
|
|
94
96
|
GEN_AI_SYSTEM_COHERE = "cohere"
|
95
97
|
GEN_AI_SYSTEM_MISTRAL = "mistral"
|
96
98
|
GEN_AI_SYSTEM_BEDROCK = "bedrock"
|
97
|
-
GEN_AI_SYSTEM_VERTEXAI = "
|
99
|
+
GEN_AI_SYSTEM_VERTEXAI = "vertex_ai"
|
98
100
|
GEN_AI_SYSTEM_GROQ = "groq"
|
99
101
|
GEN_AI_SYSTEM_OLLAMA = "ollama"
|
100
102
|
GEN_AI_SYSTEM_GPT4ALL = "gpt4all"
|
@@ -1,6 +1,6 @@
|
|
1
1
|
Metadata-Version: 2.1
|
2
2
|
Name: openlit
|
3
|
-
Version: 1.
|
3
|
+
Version: 1.17.0
|
4
4
|
Summary: OpenTelemetry-native Auto instrumentation library for monitoring LLM Applications, facilitating the integration of observability into your GenAI-driven projects
|
5
5
|
Home-page: https://github.com/openlit/openlit/tree/main/openlit/python
|
6
6
|
Keywords: OpenTelemetry,otel,otlp,llm,tracing,openai,anthropic,claude,cohere,llm monitoring,observability,monitoring,gpt,Generative AI,chatGPT
|
@@ -1,25 +1,25 @@
|
|
1
1
|
openlit/__helpers.py,sha256=lrn4PBs9owDudiCY2NBoVbAi7AU_HtUpyOj0oqPBsPY,5545
|
2
2
|
openlit/__init__.py,sha256=eJKH1Op7wzBsuoBYuM_C022Jo7cCtRQBJxf2lpDfe_o,14981
|
3
3
|
openlit/instrumentation/anthropic/__init__.py,sha256=oaU53BOPyfUKbEzYvLr1DPymDluurSnwo4Hernf2XdU,1955
|
4
|
-
openlit/instrumentation/anthropic/anthropic.py,sha256=
|
5
|
-
openlit/instrumentation/anthropic/async_anthropic.py,sha256=
|
4
|
+
openlit/instrumentation/anthropic/anthropic.py,sha256=y7CEGhKOGHWt8G_5Phr4qPJTfPGRJIAr9Yk6nM3CcvM,16775
|
5
|
+
openlit/instrumentation/anthropic/async_anthropic.py,sha256=Zz1KRKIG9wGn0quOoLvjORC-49IvHQpJ6GBdB-4PfCQ,16816
|
6
6
|
openlit/instrumentation/bedrock/__init__.py,sha256=DLLYio4S4gUzRElqNRT8WMKzM79HZwOBVjXfJI4BfaA,1545
|
7
|
-
openlit/instrumentation/bedrock/bedrock.py,sha256=
|
7
|
+
openlit/instrumentation/bedrock/bedrock.py,sha256=HqRZeiAFeNdlhlnt4DSLda8qkMP3nPKq_zhdxDssXmY,9498
|
8
8
|
openlit/instrumentation/chroma/__init__.py,sha256=61lFpHlUEQUobsUJZHXdvOViKwsOH8AOvSfc4VgCmiM,3253
|
9
9
|
openlit/instrumentation/chroma/chroma.py,sha256=E80j_41UeZi8RzTsHbpvi1izOA_n-0-3_VdrA68AJPA,10531
|
10
10
|
openlit/instrumentation/cohere/__init__.py,sha256=PC5T1qIg9pwLNocBP_WjG5B_6p_z019s8quk_fNLAMs,1920
|
11
|
-
openlit/instrumentation/cohere/cohere.py,sha256=
|
11
|
+
openlit/instrumentation/cohere/cohere.py,sha256=62-P2K39v6pIJme6vTVViLJ9PP8q_UWkTv2l3Wa2gHA,21217
|
12
12
|
openlit/instrumentation/elevenlabs/__init__.py,sha256=BZjAe-kzFJpKxT0tKksXVfZgirvgEp8qM3SfegWU5co,2631
|
13
|
-
openlit/instrumentation/elevenlabs/async_elevenlabs.py,sha256=
|
14
|
-
openlit/instrumentation/elevenlabs/elevenlabs.py,sha256=
|
13
|
+
openlit/instrumentation/elevenlabs/async_elevenlabs.py,sha256=yMYACh95SFr5EYklKnXw2DrPFa3iIgM4qQMWjO1itMU,5690
|
14
|
+
openlit/instrumentation/elevenlabs/elevenlabs.py,sha256=mFnD7sgT47OxaXJz0Vc1nrNjXEpcGQDj5run3gA48Lw,6089
|
15
15
|
openlit/instrumentation/embedchain/__init__.py,sha256=8TYk1OEbz46yF19dr-gB_x80VZMagU3kJ8-QihPXTeA,1929
|
16
16
|
openlit/instrumentation/embedchain/embedchain.py,sha256=SLlr7qieT3kp4M6OYSRy8FaVCXQ2t3oPyIiE99ioNE4,7892
|
17
17
|
openlit/instrumentation/gpt4all/__init__.py,sha256=-59CP2B3-HGZJ_vC-fI9Dt-0BuQXRhSCWCjnaGeU15Q,1802
|
18
|
-
openlit/instrumentation/gpt4all/gpt4all.py,sha256=
|
18
|
+
openlit/instrumentation/gpt4all/gpt4all.py,sha256=dbxqZeuTrv_y6wyDOIEmC8-Dc4iCGgLpj3l5JiodLMI,18787
|
19
19
|
openlit/instrumentation/gpu/__init__.py,sha256=Dj2MLar0DB20-t6W3pfR-3jfR_mwg4SYwhzIrH_n9sU,5596
|
20
20
|
openlit/instrumentation/groq/__init__.py,sha256=uW_0G6HSanQyK2dIXYhzR604pDiyPQfybzc37DsfSew,1911
|
21
|
-
openlit/instrumentation/groq/async_groq.py,sha256=
|
22
|
-
openlit/instrumentation/groq/groq.py,sha256=
|
21
|
+
openlit/instrumentation/groq/async_groq.py,sha256=myob-d9V66YiNmkFd9rtmMaXjlLiSMVHi_US16L-rZw,20011
|
22
|
+
openlit/instrumentation/groq/groq.py,sha256=m4gFPbYzjUUIgjXZ0Alu2Zy1HcO5takCFA2XFnkcGVo,19975
|
23
23
|
openlit/instrumentation/haystack/__init__.py,sha256=QK6XxxZUHX8vMv2Crk7rNBOc64iOOBLhJGL_lPlAZ8s,1758
|
24
24
|
openlit/instrumentation/haystack/haystack.py,sha256=oQIZiDhdp3gnJnhYQ1OouJMc9YT0pQ-_31cmNuopa68,3891
|
25
25
|
openlit/instrumentation/langchain/__init__.py,sha256=19C7YGSF-6u5VlvKkThNS4zZqvxw-fQfRsKufZ9onfk,2881
|
@@ -29,29 +29,29 @@ openlit/instrumentation/llamaindex/llamaindex.py,sha256=uiIigbwhonSbJWA7LpgOVI1R
|
|
29
29
|
openlit/instrumentation/milvus/__init__.py,sha256=qi1yfmMrvkDtnrN_6toW8qC9BRL78bq7ayWpObJ8Bq4,2961
|
30
30
|
openlit/instrumentation/milvus/milvus.py,sha256=qhKIoggBAJhRctRrBYz69AcvXH-eh7oBn_l9WfxpAjI,9121
|
31
31
|
openlit/instrumentation/mistral/__init__.py,sha256=zJCIpFWRbsYrvooOJYuqwyuKeSOQLWbyXWCObL-Snks,3156
|
32
|
-
openlit/instrumentation/mistral/async_mistral.py,sha256=
|
33
|
-
openlit/instrumentation/mistral/mistral.py,sha256
|
32
|
+
openlit/instrumentation/mistral/async_mistral.py,sha256=WXU46nbe61IglfGehrwdprMVB6hNiuqqmJHE3XvvP0E,22192
|
33
|
+
openlit/instrumentation/mistral/mistral.py,sha256=DC-wLIypokPEEAbVSKX5sytv94DY2QDnk12401e5vq8,22039
|
34
34
|
openlit/instrumentation/ollama/__init__.py,sha256=cOax8PiypDuo_FC4WvDCYBRo7lH5nV9xU92h7k-eZbg,3812
|
35
|
-
openlit/instrumentation/ollama/async_ollama.py,sha256=
|
36
|
-
openlit/instrumentation/ollama/ollama.py,sha256=
|
35
|
+
openlit/instrumentation/ollama/async_ollama.py,sha256=UBwl-Jv4VPyPuhKIesL0VYIQ2u1AZ0U8c9MZQzEPn6c,30594
|
36
|
+
openlit/instrumentation/ollama/ollama.py,sha256=bZPgkJJUD3RgJTUUt98cklfZ8Y1lsSgQGOUXXHk53BI,30504
|
37
37
|
openlit/instrumentation/openai/__init__.py,sha256=AZ2cPr3TMKkgGdMl_yXMeSi7bWhtmMqOW1iHdzHHGHA,16265
|
38
|
-
openlit/instrumentation/openai/async_azure_openai.py,sha256=
|
39
|
-
openlit/instrumentation/openai/async_openai.py,sha256=
|
40
|
-
openlit/instrumentation/openai/azure_openai.py,sha256=
|
41
|
-
openlit/instrumentation/openai/openai.py,sha256=
|
38
|
+
openlit/instrumentation/openai/async_azure_openai.py,sha256=XbST1UE_zXzNL6RX2XwCsK_a6IhG9PHVTMKBjGrUcB0,48961
|
39
|
+
openlit/instrumentation/openai/async_openai.py,sha256=RGNpKLsHYfJXjj1ImuWRJToVSs0wdvMNp2kyTBrBaDw,47578
|
40
|
+
openlit/instrumentation/openai/azure_openai.py,sha256=dZUc5MtCwg_sZJWiruG6exYGhPAm-339sqs3sKZNRPU,48761
|
41
|
+
openlit/instrumentation/openai/openai.py,sha256=pOeehRgd0GMITGMYUU6PUsjvPDUWmF-L_ER9ry1hL_c,48214
|
42
42
|
openlit/instrumentation/pinecone/__init__.py,sha256=Mv9bElqNs07_JQkYyNnO0wOM3hdbprmw7sttdMeKC7g,2526
|
43
43
|
openlit/instrumentation/pinecone/pinecone.py,sha256=0EhLmtOuvwWVvAKh3e56wyd8wzQq1oaLOmF15SVHxVE,8765
|
44
44
|
openlit/instrumentation/qdrant/__init__.py,sha256=OJIg17-IGmBEvBYVKjCHcJ0hFXuEL7XV_jzUTqkolN8,4799
|
45
45
|
openlit/instrumentation/qdrant/qdrant.py,sha256=4uHKYGvWQtRAEVLUWo3o4joJw7hFm2NxVuBu5YKZKiI,14456
|
46
46
|
openlit/instrumentation/transformers/__init__.py,sha256=4GBtjzcJU4XiPexIUYEqF3pNZMeQw4Gm5B-cyumaFjs,1468
|
47
|
-
openlit/instrumentation/transformers/transformers.py,sha256=
|
47
|
+
openlit/instrumentation/transformers/transformers.py,sha256=KNAT2ROjziW6OAP6Y0Ec4oS2T2jx9y2mzpBgR_e78bI,7625
|
48
48
|
openlit/instrumentation/vertexai/__init__.py,sha256=N3E9HtzefD-zC0fvmfGYiDmSqssoavp_i59wfuYLyMw,6079
|
49
|
-
openlit/instrumentation/vertexai/async_vertexai.py,sha256=
|
50
|
-
openlit/instrumentation/vertexai/vertexai.py,sha256=
|
49
|
+
openlit/instrumentation/vertexai/async_vertexai.py,sha256=8JwSwLPPA4lAatf4w_5kJ5_YZDLwl5yG8N59cTD-EZM,55198
|
50
|
+
openlit/instrumentation/vertexai/vertexai.py,sha256=R6dDQfC3YFoZDygxU2fkflcMsqIv8AVoU3XOwWSvpwA,54951
|
51
51
|
openlit/otel/metrics.py,sha256=O7NoaDz0bY19mqpE4-0PcKwEe-B-iJFRgOCaanAuZAc,4291
|
52
52
|
openlit/otel/tracing.py,sha256=vL1ifMbARPBpqK--yXYsCM6y5dSu5LFIKqkhZXtYmUc,3712
|
53
|
-
openlit/semcov/__init__.py,sha256=
|
54
|
-
openlit-1.
|
55
|
-
openlit-1.
|
56
|
-
openlit-1.
|
57
|
-
openlit-1.
|
53
|
+
openlit/semcov/__init__.py,sha256=7E8qfoNRQbu6QAYBOkSUlESOQ8-VzvLTIKyj75BrHsY,7466
|
54
|
+
openlit-1.17.0.dist-info/LICENSE,sha256=xx0jnfkXJvxRnG63LTGOxlggYnIysveWIZ6H3PNdCrQ,11357
|
55
|
+
openlit-1.17.0.dist-info/METADATA,sha256=mY_exhD2wObnFejJJJjXxlMwdMZ34uZBKlObiwhMc1Y,14120
|
56
|
+
openlit-1.17.0.dist-info/WHEEL,sha256=sP946D7jFCHeNz5Iq4fL4Lu-PrWrFsgfLXbbkciIZwg,88
|
57
|
+
openlit-1.17.0.dist-info/RECORD,,
|
File without changes
|
File without changes
|