sentry-sdk 3.0.0a4__py2.py3-none-any.whl → 3.0.0a5__py2.py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of sentry-sdk might be problematic. Click here for more details.
- sentry_sdk/__init__.py +1 -0
- sentry_sdk/api.py +57 -0
- sentry_sdk/consts.py +113 -4
- sentry_sdk/integrations/starlite.py +1 -1
- sentry_sdk/opentelemetry/scope.py +3 -1
- sentry_sdk/opentelemetry/span_processor.py +1 -0
- sentry_sdk/tracing.py +6 -1
- {sentry_sdk-3.0.0a4.dist-info → sentry_sdk-3.0.0a5.dist-info}/METADATA +1 -1
- {sentry_sdk-3.0.0a4.dist-info → sentry_sdk-3.0.0a5.dist-info}/RECORD +13 -13
- {sentry_sdk-3.0.0a4.dist-info → sentry_sdk-3.0.0a5.dist-info}/WHEEL +0 -0
- {sentry_sdk-3.0.0a4.dist-info → sentry_sdk-3.0.0a5.dist-info}/entry_points.txt +0 -0
- {sentry_sdk-3.0.0a4.dist-info → sentry_sdk-3.0.0a5.dist-info}/licenses/LICENSE +0 -0
- {sentry_sdk-3.0.0a4.dist-info → sentry_sdk-3.0.0a5.dist-info}/top_level.txt +0 -0
sentry_sdk/__init__.py
CHANGED
sentry_sdk/api.py
CHANGED
|
@@ -76,6 +76,7 @@ __all__ = [
|
|
|
76
76
|
"start_session",
|
|
77
77
|
"end_session",
|
|
78
78
|
"set_transaction_name",
|
|
79
|
+
"update_current_span",
|
|
79
80
|
]
|
|
80
81
|
|
|
81
82
|
|
|
@@ -341,3 +342,59 @@ def end_session() -> None:
|
|
|
341
342
|
@scopemethod
|
|
342
343
|
def set_transaction_name(name: str, source: Optional[str] = None) -> None:
|
|
343
344
|
return get_current_scope().set_transaction_name(name, source)
|
|
345
|
+
|
|
346
|
+
|
|
347
|
+
def update_current_span(op=None, name=None, attributes=None):
|
|
348
|
+
# type: (Optional[str], Optional[str], Optional[dict[str, Union[str, int, float, bool]]]) -> None
|
|
349
|
+
"""
|
|
350
|
+
Update the current active span with the provided parameters.
|
|
351
|
+
|
|
352
|
+
This function allows you to modify properties of the currently active span.
|
|
353
|
+
If no span is currently active, this function will do nothing.
|
|
354
|
+
|
|
355
|
+
:param op: The operation name for the span. This is a high-level description
|
|
356
|
+
of what the span represents (e.g., "http.client", "db.query").
|
|
357
|
+
You can use predefined constants from :py:class:`sentry_sdk.consts.OP`
|
|
358
|
+
or provide your own string. If not provided, the span's operation will
|
|
359
|
+
remain unchanged.
|
|
360
|
+
:type op: str or None
|
|
361
|
+
|
|
362
|
+
:param name: The human-readable name/description for the span. This provides
|
|
363
|
+
more specific details about what the span represents (e.g., "GET /api/users",
|
|
364
|
+
"SELECT * FROM users"). If not provided, the span's name will remain unchanged.
|
|
365
|
+
:type name: str or None
|
|
366
|
+
|
|
367
|
+
:param attributes: A dictionary of key-value pairs to add as attributes to the span.
|
|
368
|
+
Attribute values must be strings, integers, floats, or booleans. These
|
|
369
|
+
attributes will be merged with any existing span data. If not provided,
|
|
370
|
+
no attributes will be added.
|
|
371
|
+
:type attributes: dict[str, Union[str, int, float, bool]] or None
|
|
372
|
+
|
|
373
|
+
:returns: None
|
|
374
|
+
|
|
375
|
+
.. versionadded:: 2.35.0
|
|
376
|
+
|
|
377
|
+
Example::
|
|
378
|
+
|
|
379
|
+
import sentry_sdk
|
|
380
|
+
from sentry_sdk.consts import OP
|
|
381
|
+
|
|
382
|
+
sentry_sdk.update_current_span(
|
|
383
|
+
op=OP.FUNCTION,
|
|
384
|
+
name="process_user_data",
|
|
385
|
+
attributes={"user_id": 123, "batch_size": 50}
|
|
386
|
+
)
|
|
387
|
+
"""
|
|
388
|
+
current_span = get_current_span()
|
|
389
|
+
|
|
390
|
+
if current_span is None:
|
|
391
|
+
return
|
|
392
|
+
|
|
393
|
+
if op is not None:
|
|
394
|
+
current_span.op = op
|
|
395
|
+
|
|
396
|
+
if name is not None:
|
|
397
|
+
current_span.name = name
|
|
398
|
+
|
|
399
|
+
if attributes is not None:
|
|
400
|
+
current_span.set_attributes(attributes)
|
sentry_sdk/consts.py
CHANGED
|
@@ -103,6 +103,9 @@ class SPANDATA:
|
|
|
103
103
|
|
|
104
104
|
AI_CITATIONS = "ai.citations"
|
|
105
105
|
"""
|
|
106
|
+
.. deprecated::
|
|
107
|
+
This attribute is deprecated. Use GEN_AI_* attributes instead.
|
|
108
|
+
|
|
106
109
|
References or sources cited by the AI model in its response.
|
|
107
110
|
Example: ["Smith et al. 2020", "Jones 2019"]
|
|
108
111
|
"""
|
|
@@ -115,65 +118,97 @@ class SPANDATA:
|
|
|
115
118
|
|
|
116
119
|
AI_DOCUMENTS = "ai.documents"
|
|
117
120
|
"""
|
|
121
|
+
.. deprecated::
|
|
122
|
+
This attribute is deprecated. Use GEN_AI_* attributes instead.
|
|
123
|
+
|
|
118
124
|
Documents or content chunks used as context for the AI model.
|
|
119
125
|
Example: ["doc1.txt", "doc2.pdf"]
|
|
120
126
|
"""
|
|
121
127
|
|
|
122
128
|
AI_FINISH_REASON = "ai.finish_reason"
|
|
123
129
|
"""
|
|
130
|
+
.. deprecated::
|
|
131
|
+
This attribute is deprecated. Use GEN_AI_RESPONSE_FINISH_REASONS instead.
|
|
132
|
+
|
|
124
133
|
The reason why the model stopped generating.
|
|
125
134
|
Example: "length"
|
|
126
135
|
"""
|
|
127
136
|
|
|
128
137
|
AI_FREQUENCY_PENALTY = "ai.frequency_penalty"
|
|
129
138
|
"""
|
|
139
|
+
.. deprecated::
|
|
140
|
+
This attribute is deprecated. Use GEN_AI_REQUEST_FREQUENCY_PENALTY instead.
|
|
141
|
+
|
|
130
142
|
Used to reduce repetitiveness of generated tokens.
|
|
131
143
|
Example: 0.5
|
|
132
144
|
"""
|
|
133
145
|
|
|
134
146
|
AI_FUNCTION_CALL = "ai.function_call"
|
|
135
147
|
"""
|
|
148
|
+
.. deprecated::
|
|
149
|
+
This attribute is deprecated. Use GEN_AI_RESPONSE_TOOL_CALLS instead.
|
|
150
|
+
|
|
136
151
|
For an AI model call, the function that was called. This is deprecated for OpenAI, and replaced by tool_calls
|
|
137
152
|
"""
|
|
138
153
|
|
|
139
154
|
AI_GENERATION_ID = "ai.generation_id"
|
|
140
155
|
"""
|
|
156
|
+
.. deprecated::
|
|
157
|
+
This attribute is deprecated. Use GEN_AI_RESPONSE_ID instead.
|
|
158
|
+
|
|
141
159
|
Unique identifier for the completion.
|
|
142
160
|
Example: "gen_123abc"
|
|
143
161
|
"""
|
|
144
162
|
|
|
145
163
|
AI_INPUT_MESSAGES = "ai.input_messages"
|
|
146
164
|
"""
|
|
165
|
+
.. deprecated::
|
|
166
|
+
This attribute is deprecated. Use GEN_AI_REQUEST_MESSAGES instead.
|
|
167
|
+
|
|
147
168
|
The input messages to an LLM call.
|
|
148
169
|
Example: [{"role": "user", "message": "hello"}]
|
|
149
170
|
"""
|
|
150
171
|
|
|
151
172
|
AI_LOGIT_BIAS = "ai.logit_bias"
|
|
152
173
|
"""
|
|
174
|
+
.. deprecated::
|
|
175
|
+
This attribute is deprecated. Use GEN_AI_* attributes instead.
|
|
176
|
+
|
|
153
177
|
For an AI model call, the logit bias
|
|
154
178
|
"""
|
|
155
179
|
|
|
156
180
|
AI_METADATA = "ai.metadata"
|
|
157
181
|
"""
|
|
182
|
+
.. deprecated::
|
|
183
|
+
This attribute is deprecated. Use GEN_AI_* attributes instead.
|
|
184
|
+
|
|
158
185
|
Extra metadata passed to an AI pipeline step.
|
|
159
186
|
Example: {"executed_function": "add_integers"}
|
|
160
187
|
"""
|
|
161
188
|
|
|
162
189
|
AI_MODEL_ID = "ai.model_id"
|
|
163
190
|
"""
|
|
164
|
-
|
|
191
|
+
.. deprecated::
|
|
192
|
+
This attribute is deprecated. Use GEN_AI_REQUEST_MODEL or GEN_AI_RESPONSE_MODEL instead.
|
|
193
|
+
|
|
194
|
+
The unique descriptor of the model being executed.
|
|
165
195
|
Example: gpt-4
|
|
166
196
|
"""
|
|
167
197
|
|
|
168
198
|
AI_PIPELINE_NAME = "ai.pipeline.name"
|
|
169
199
|
"""
|
|
200
|
+
.. deprecated::
|
|
201
|
+
This attribute is deprecated. Use GEN_AI_PIPELINE_NAME instead.
|
|
202
|
+
|
|
170
203
|
Name of the AI pipeline or chain being executed.
|
|
171
|
-
DEPRECATED: Use GEN_AI_PIPELINE_NAME instead.
|
|
172
204
|
Example: "qa-pipeline"
|
|
173
205
|
"""
|
|
174
206
|
|
|
175
207
|
AI_PREAMBLE = "ai.preamble"
|
|
176
208
|
"""
|
|
209
|
+
.. deprecated::
|
|
210
|
+
This attribute is deprecated. Use GEN_AI_* attributes instead.
|
|
211
|
+
|
|
177
212
|
For an AI model call, the preamble parameter.
|
|
178
213
|
Preambles are a part of the prompt used to adjust the model's overall behavior and conversation style.
|
|
179
214
|
Example: "You are now a clown."
|
|
@@ -181,6 +216,9 @@ class SPANDATA:
|
|
|
181
216
|
|
|
182
217
|
AI_PRESENCE_PENALTY = "ai.presence_penalty"
|
|
183
218
|
"""
|
|
219
|
+
.. deprecated::
|
|
220
|
+
This attribute is deprecated. Use GEN_AI_REQUEST_PRESENCE_PENALTY instead.
|
|
221
|
+
|
|
184
222
|
Used to reduce repetitiveness of generated tokens.
|
|
185
223
|
Example: 0.5
|
|
186
224
|
"""
|
|
@@ -193,89 +231,133 @@ class SPANDATA:
|
|
|
193
231
|
|
|
194
232
|
AI_RAW_PROMPTING = "ai.raw_prompting"
|
|
195
233
|
"""
|
|
234
|
+
.. deprecated::
|
|
235
|
+
This attribute is deprecated. Use GEN_AI_* attributes instead.
|
|
236
|
+
|
|
196
237
|
Minimize pre-processing done to the prompt sent to the LLM.
|
|
197
238
|
Example: true
|
|
198
239
|
"""
|
|
199
240
|
|
|
200
241
|
AI_RESPONSE_FORMAT = "ai.response_format"
|
|
201
242
|
"""
|
|
243
|
+
.. deprecated::
|
|
244
|
+
This attribute is deprecated. Use GEN_AI_* attributes instead.
|
|
245
|
+
|
|
202
246
|
For an AI model call, the format of the response
|
|
203
247
|
"""
|
|
204
248
|
|
|
205
249
|
AI_RESPONSES = "ai.responses"
|
|
206
250
|
"""
|
|
251
|
+
.. deprecated::
|
|
252
|
+
This attribute is deprecated. Use GEN_AI_RESPONSE_TEXT instead.
|
|
253
|
+
|
|
207
254
|
The responses to an AI model call. Always as a list.
|
|
208
255
|
Example: ["hello", "world"]
|
|
209
256
|
"""
|
|
210
257
|
|
|
211
258
|
AI_SEARCH_QUERIES = "ai.search_queries"
|
|
212
259
|
"""
|
|
260
|
+
.. deprecated::
|
|
261
|
+
This attribute is deprecated. Use GEN_AI_* attributes instead.
|
|
262
|
+
|
|
213
263
|
Queries used to search for relevant context or documents.
|
|
214
264
|
Example: ["climate change effects", "renewable energy"]
|
|
215
265
|
"""
|
|
216
266
|
|
|
217
267
|
AI_SEARCH_REQUIRED = "ai.is_search_required"
|
|
218
268
|
"""
|
|
269
|
+
.. deprecated::
|
|
270
|
+
This attribute is deprecated. Use GEN_AI_* attributes instead.
|
|
271
|
+
|
|
219
272
|
Boolean indicating if the model needs to perform a search.
|
|
220
273
|
Example: true
|
|
221
274
|
"""
|
|
222
275
|
|
|
223
276
|
AI_SEARCH_RESULTS = "ai.search_results"
|
|
224
277
|
"""
|
|
278
|
+
.. deprecated::
|
|
279
|
+
This attribute is deprecated. Use GEN_AI_* attributes instead.
|
|
280
|
+
|
|
225
281
|
Results returned from search queries for context.
|
|
226
282
|
Example: ["Result 1", "Result 2"]
|
|
227
283
|
"""
|
|
228
284
|
|
|
229
285
|
AI_SEED = "ai.seed"
|
|
230
286
|
"""
|
|
287
|
+
.. deprecated::
|
|
288
|
+
This attribute is deprecated. Use GEN_AI_REQUEST_SEED instead.
|
|
289
|
+
|
|
231
290
|
The seed, ideally models given the same seed and same other parameters will produce the exact same output.
|
|
232
291
|
Example: 123.45
|
|
233
292
|
"""
|
|
234
293
|
|
|
235
294
|
AI_STREAMING = "ai.streaming"
|
|
236
295
|
"""
|
|
296
|
+
.. deprecated::
|
|
297
|
+
This attribute is deprecated. Use GEN_AI_RESPONSE_STREAMING instead.
|
|
298
|
+
|
|
237
299
|
Whether or not the AI model call's response was streamed back asynchronously
|
|
238
|
-
DEPRECATED: Use GEN_AI_RESPONSE_STREAMING instead.
|
|
239
300
|
Example: true
|
|
240
301
|
"""
|
|
241
302
|
|
|
242
303
|
AI_TAGS = "ai.tags"
|
|
243
304
|
"""
|
|
305
|
+
.. deprecated::
|
|
306
|
+
This attribute is deprecated. Use GEN_AI_* attributes instead.
|
|
307
|
+
|
|
244
308
|
Tags that describe an AI pipeline step.
|
|
245
309
|
Example: {"executed_function": "add_integers"}
|
|
246
310
|
"""
|
|
247
311
|
|
|
248
312
|
AI_TEMPERATURE = "ai.temperature"
|
|
249
313
|
"""
|
|
314
|
+
.. deprecated::
|
|
315
|
+
This attribute is deprecated. Use GEN_AI_REQUEST_TEMPERATURE instead.
|
|
316
|
+
|
|
250
317
|
For an AI model call, the temperature parameter. Temperature essentially means how random the output will be.
|
|
251
318
|
Example: 0.5
|
|
252
319
|
"""
|
|
253
320
|
|
|
254
321
|
AI_TEXTS = "ai.texts"
|
|
255
322
|
"""
|
|
323
|
+
.. deprecated::
|
|
324
|
+
This attribute is deprecated. Use GEN_AI_* attributes instead.
|
|
325
|
+
|
|
256
326
|
Raw text inputs provided to the model.
|
|
257
327
|
Example: ["What is machine learning?"]
|
|
258
328
|
"""
|
|
259
329
|
|
|
260
330
|
AI_TOP_K = "ai.top_k"
|
|
261
331
|
"""
|
|
332
|
+
.. deprecated::
|
|
333
|
+
This attribute is deprecated. Use GEN_AI_REQUEST_TOP_K instead.
|
|
334
|
+
|
|
262
335
|
For an AI model call, the top_k parameter. Top_k essentially controls how random the output will be.
|
|
263
336
|
Example: 35
|
|
264
337
|
"""
|
|
265
338
|
|
|
266
339
|
AI_TOP_P = "ai.top_p"
|
|
267
340
|
"""
|
|
341
|
+
.. deprecated::
|
|
342
|
+
This attribute is deprecated. Use GEN_AI_REQUEST_TOP_P instead.
|
|
343
|
+
|
|
268
344
|
For an AI model call, the top_p parameter. Top_p essentially controls how random the output will be.
|
|
269
345
|
Example: 0.5
|
|
270
346
|
"""
|
|
271
347
|
|
|
272
348
|
AI_TOOL_CALLS = "ai.tool_calls"
|
|
273
349
|
"""
|
|
350
|
+
.. deprecated::
|
|
351
|
+
This attribute is deprecated. Use GEN_AI_RESPONSE_TOOL_CALLS instead.
|
|
352
|
+
|
|
274
353
|
For an AI model call, the function that was called. This is deprecated for OpenAI, and replaced by tool_calls
|
|
275
354
|
"""
|
|
276
355
|
|
|
277
356
|
AI_TOOLS = "ai.tools"
|
|
278
357
|
"""
|
|
358
|
+
.. deprecated::
|
|
359
|
+
This attribute is deprecated. Use GEN_AI_REQUEST_AVAILABLE_TOOLS instead.
|
|
360
|
+
|
|
279
361
|
For an AI model call, the functions that are available
|
|
280
362
|
"""
|
|
281
363
|
|
|
@@ -287,6 +369,9 @@ class SPANDATA:
|
|
|
287
369
|
|
|
288
370
|
AI_WARNINGS = "ai.warnings"
|
|
289
371
|
"""
|
|
372
|
+
.. deprecated::
|
|
373
|
+
This attribute is deprecated. Use GEN_AI_* attributes instead.
|
|
374
|
+
|
|
290
375
|
Warning messages generated during model execution.
|
|
291
376
|
Example: ["Token limit exceeded"]
|
|
292
377
|
"""
|
|
@@ -391,6 +476,18 @@ class SPANDATA:
|
|
|
391
476
|
Example: "qa-pipeline"
|
|
392
477
|
"""
|
|
393
478
|
|
|
479
|
+
GEN_AI_RESPONSE_FINISH_REASONS = "gen_ai.response.finish_reasons"
|
|
480
|
+
"""
|
|
481
|
+
The reason why the model stopped generating.
|
|
482
|
+
Example: "COMPLETE"
|
|
483
|
+
"""
|
|
484
|
+
|
|
485
|
+
GEN_AI_RESPONSE_ID = "gen_ai.response.id"
|
|
486
|
+
"""
|
|
487
|
+
Unique identifier for the completion.
|
|
488
|
+
Example: "gen_123abc"
|
|
489
|
+
"""
|
|
490
|
+
|
|
394
491
|
GEN_AI_RESPONSE_MODEL = "gen_ai.response.model"
|
|
395
492
|
"""
|
|
396
493
|
Exact model identifier used to generate the response
|
|
@@ -451,12 +548,24 @@ class SPANDATA:
|
|
|
451
548
|
Example: 0.1
|
|
452
549
|
"""
|
|
453
550
|
|
|
551
|
+
GEN_AI_REQUEST_SEED = "gen_ai.request.seed"
|
|
552
|
+
"""
|
|
553
|
+
The seed, ideally models given the same seed and same other parameters will produce the exact same output.
|
|
554
|
+
Example: "1234567890"
|
|
555
|
+
"""
|
|
556
|
+
|
|
454
557
|
GEN_AI_REQUEST_TEMPERATURE = "gen_ai.request.temperature"
|
|
455
558
|
"""
|
|
456
559
|
The temperature parameter used to control randomness in the output.
|
|
457
560
|
Example: 0.7
|
|
458
561
|
"""
|
|
459
562
|
|
|
563
|
+
GEN_AI_REQUEST_TOP_K = "gen_ai.request.top_k"
|
|
564
|
+
"""
|
|
565
|
+
Limits the model to only consider the K most likely next tokens, where K is an integer (e.g., top_k=20 means only the 20 highest probability tokens are considered).
|
|
566
|
+
Example: 35
|
|
567
|
+
"""
|
|
568
|
+
|
|
460
569
|
GEN_AI_REQUEST_TOP_P = "gen_ai.request.top_p"
|
|
461
570
|
"""
|
|
462
571
|
The top_p parameter used to control diversity via nucleus sampling.
|
|
@@ -1272,4 +1381,4 @@ DEFAULT_OPTIONS = _get_default_options()
|
|
|
1272
1381
|
del _get_default_options
|
|
1273
1382
|
|
|
1274
1383
|
|
|
1275
|
-
VERSION = "3.0.
|
|
1384
|
+
VERSION = "3.0.0a5"
|
|
@@ -17,7 +17,7 @@ try:
|
|
|
17
17
|
from starlite.plugins.base import get_plugin_for_value # type: ignore
|
|
18
18
|
from starlite.routes.http import HTTPRoute # type: ignore
|
|
19
19
|
from starlite.utils import ConnectionDataExtractor, is_async_callable, Ref # type: ignore
|
|
20
|
-
from pydantic import BaseModel
|
|
20
|
+
from pydantic import BaseModel
|
|
21
21
|
except ImportError:
|
|
22
22
|
raise DidNotEnable("Starlite is not installed")
|
|
23
23
|
|
|
@@ -85,7 +85,9 @@ class PotelScope(Scope):
|
|
|
85
85
|
|
|
86
86
|
span_context = self._incoming_otel_span_context()
|
|
87
87
|
if span_context is None:
|
|
88
|
-
|
|
88
|
+
# force a new trace since no incoming stuff
|
|
89
|
+
with use_span(INVALID_SPAN):
|
|
90
|
+
yield
|
|
89
91
|
else:
|
|
90
92
|
with use_span(NonRecordingSpan(span_context)):
|
|
91
93
|
yield
|
sentry_sdk/tracing.py
CHANGED
|
@@ -352,7 +352,8 @@ class Span:
|
|
|
352
352
|
return self.get_attribute(SentrySpanAttribute.NAME)
|
|
353
353
|
|
|
354
354
|
@name.setter
|
|
355
|
-
def name(self, value:
|
|
355
|
+
def name(self, value: str) -> None:
|
|
356
|
+
self._otel_span.update_name(value)
|
|
356
357
|
self.set_attribute(SentrySpanAttribute.NAME, value)
|
|
357
358
|
|
|
358
359
|
@property
|
|
@@ -450,6 +451,10 @@ class Span:
|
|
|
450
451
|
|
|
451
452
|
self._otel_span.set_attribute(key, serialized_value)
|
|
452
453
|
|
|
454
|
+
def set_attributes(self, attributes: dict[str, Any]) -> None:
|
|
455
|
+
for key, value in attributes.items():
|
|
456
|
+
self.set_attribute(key, value)
|
|
457
|
+
|
|
453
458
|
@property
|
|
454
459
|
def status(self) -> Optional[str]:
|
|
455
460
|
"""
|
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
sentry_sdk/__init__.py,sha256=
|
|
1
|
+
sentry_sdk/__init__.py,sha256=C3U0psDzNPIoYrHqjV1w5TXU6xJWUGHe1e0y58ODF3c,1309
|
|
2
2
|
sentry_sdk/_compat.py,sha256=wP2ZOiqdpaVaC2IPh3O-GwDK1H-3-momIn5bmMNrCGY,2969
|
|
3
3
|
sentry_sdk/_init_implementation.py,sha256=PJrU7NZTWOE7hH2n_odcs1E3Nc_2TK2ilLwn5trJXns,1521
|
|
4
4
|
sentry_sdk/_log_batcher.py,sha256=Ugg89Q1hnNgNfBV2-SUl_8qJN9Qy-Ux8qWw7VQR0KXo,4840
|
|
@@ -6,10 +6,10 @@ sentry_sdk/_lru_cache.py,sha256=H_BCecDD-7qSmcI0c6Q8i1nzBJMRBLFqXuax-ofh6wg,1189
|
|
|
6
6
|
sentry_sdk/_queue.py,sha256=AWE9LpaPhNm4VEUiY3TJWIXfAw44D-dgFXuhZdBHF-k,11119
|
|
7
7
|
sentry_sdk/_types.py,sha256=owny4XuTQM31ljk5SEGS2x-bfanue26OE5qkgWE5e5Y,8647
|
|
8
8
|
sentry_sdk/_werkzeug.py,sha256=qEPz7ZYBeBkILq3snw0GnaZQbt5mItMQO9JQ3YDCCgg,3702
|
|
9
|
-
sentry_sdk/api.py,sha256=
|
|
9
|
+
sentry_sdk/api.py,sha256=muzQWcZ0O_9Gw-62NA-GC827E9zkxhRhvD6GPK5W0Vo,10824
|
|
10
10
|
sentry_sdk/attachments.py,sha256=la0cbz1yc8cV4PQyM3kgArvWjGWtjxq945wgiKdaoCs,3046
|
|
11
11
|
sentry_sdk/client.py,sha256=SYX-kvGdrNzmr0drB0J7gNkF5tbBVWpMMYopJwLR3ys,35284
|
|
12
|
-
sentry_sdk/consts.py,sha256=
|
|
12
|
+
sentry_sdk/consts.py,sha256=6Ua20kG1lXeuq7GzKmwnT-oDcOC7zwBClBW1LRSfQn8,50993
|
|
13
13
|
sentry_sdk/debug.py,sha256=ac50G-ZSRUTLhu6VVaJnc4sGsQM1lIN6qWR1AS4oTH4,773
|
|
14
14
|
sentry_sdk/envelope.py,sha256=2Bmjlm7bctrtkYeWVPNGBHBAL6pMFtULUXCONeq6QY8,9118
|
|
15
15
|
sentry_sdk/feature_flags.py,sha256=uLIi4SjN1EuFP59AduFS5dIzqUYOidPZUCZjuqrAbhg,2144
|
|
@@ -22,7 +22,7 @@ sentry_sdk/serializer.py,sha256=twylwvlYrP67Mq4AXVQEuNeeOVAkyIdnyvBj8yLwq7U,1236
|
|
|
22
22
|
sentry_sdk/session.py,sha256=UunkzHmyV7rZu2a6kWMB5BZNJnQc9WiEkJnPu5aHjto,5097
|
|
23
23
|
sentry_sdk/sessions.py,sha256=0puTgebSwYPNH6zq004QLe1q4VM7XOK8iCzsFfcGLzs,6069
|
|
24
24
|
sentry_sdk/spotlight.py,sha256=mHFbVED17LBAaVdRJbNfE9c9NmknWTZ4Y3lkYkJkJuI,8492
|
|
25
|
-
sentry_sdk/tracing.py,sha256=
|
|
25
|
+
sentry_sdk/tracing.py,sha256=NdAKEAHg6Eycs9-dhVW2gCk0WWNlMXrtY0qrgz2WS5U,17869
|
|
26
26
|
sentry_sdk/tracing_utils.py,sha256=osw7NwUO3vJWB1cFQgrJEao4qYWF6mRvx_-u09zxhu4,27000
|
|
27
27
|
sentry_sdk/transport.py,sha256=MBUPJR8ZV-AtYdDRX342I_OSIhgo9h89qofTVGgajHI,28353
|
|
28
28
|
sentry_sdk/types.py,sha256=NLbnRzww2K3_oGz2GzcC8TdX5L2DXYso1-H1uCv2Hwc,1222
|
|
@@ -89,7 +89,7 @@ sentry_sdk/integrations/serverless.py,sha256=aY3ZZjlF3Sl0uKDfIEvAxfFYIOppV9nKArH
|
|
|
89
89
|
sentry_sdk/integrations/socket.py,sha256=B0SYwJFCo_3UbG9l5YhIa_E8GBWsJYg8ddxRBajwU9I,3454
|
|
90
90
|
sentry_sdk/integrations/sqlalchemy.py,sha256=JSc45HmsKPWZlRu9y8ctA3pAjcTBIOjCi6tDn4G1oKA,4339
|
|
91
91
|
sentry_sdk/integrations/starlette.py,sha256=EMukyRLrUFQhtYlOiB446qk_PynOqOy6ZGeiWDUy4bw,25244
|
|
92
|
-
sentry_sdk/integrations/starlite.py,sha256=
|
|
92
|
+
sentry_sdk/integrations/starlite.py,sha256=HE08ZcvXUIE_u642cZtY3dRaMvUYqlMURUTcoySENv8,10521
|
|
93
93
|
sentry_sdk/integrations/statsig.py,sha256=3Abtq8S5EqHUYVaMf61_MiIx5oZfLz09d61Cz97_TSY,1241
|
|
94
94
|
sentry_sdk/integrations/stdlib.py,sha256=HNc2haRxGZHU9ObFvknBdYGhoAeylIG_XfYY1lMUDPM,10345
|
|
95
95
|
sentry_sdk/integrations/strawberry.py,sha256=fwolVXMtauXXzqzcYtKw34tixsw-I_KlBPWZIU0lmLs,13506
|
|
@@ -152,17 +152,17 @@ sentry_sdk/opentelemetry/consts.py,sha256=k1lpmm-a6K4TNaOzhW4iIQH48Ztep2jzof39Rp
|
|
|
152
152
|
sentry_sdk/opentelemetry/contextvars_context.py,sha256=qoZIhur-ZYnpV7StKvw9f2i6DkoJwLe69vIyMFPJyf8,3036
|
|
153
153
|
sentry_sdk/opentelemetry/propagator.py,sha256=z5dQPtD0yC1ToKAHa5vO8tBiuIt9Ry_f4GWYO2jFO8s,3638
|
|
154
154
|
sentry_sdk/opentelemetry/sampler.py,sha256=AC-nrSrweeIzhxz2Sixlkst98juIDu30poiVaCYrP6Y,12113
|
|
155
|
-
sentry_sdk/opentelemetry/scope.py,sha256=
|
|
156
|
-
sentry_sdk/opentelemetry/span_processor.py,sha256=
|
|
155
|
+
sentry_sdk/opentelemetry/scope.py,sha256=Hx6OiR5zGfXC19n5KhMhsMUz85_IIYSjdA5sES1iZRw,6947
|
|
156
|
+
sentry_sdk/opentelemetry/span_processor.py,sha256=j7Fv19RzUocqtP81NVa5SKwr9LvQ69mT-YVe0nh4kFI,11185
|
|
157
157
|
sentry_sdk/opentelemetry/tracing.py,sha256=b8f2SFQE2A8GBE17_GcqLq0VN2gJ7SzySLlsxZ4Yzt4,2757
|
|
158
158
|
sentry_sdk/opentelemetry/utils.py,sha256=tlTPBFbHp0pTm6-z-p38978wg5ca0A7NfuolFE4AeZc,14422
|
|
159
159
|
sentry_sdk/profiler/__init__.py,sha256=bYeDkmLQliS2KkNSOGa8Sx4zN3pjTc3WmYIvbkcGkmQ,153
|
|
160
160
|
sentry_sdk/profiler/continuous_profiler.py,sha256=AXwhI-NI6hx82EnKEy_oNQPPl9nwLQJKcj_QRX6KKAg,21574
|
|
161
161
|
sentry_sdk/profiler/transaction_profiler.py,sha256=gbe9RedrGqPakPi3uY9fUDXp0IcM9djHuVcYUcbDRfg,25381
|
|
162
162
|
sentry_sdk/profiler/utils.py,sha256=cdjpSQ9IAchtMPLWNFGDjZHIXArTbcPtFPEngB9Tb9c,6475
|
|
163
|
-
sentry_sdk-3.0.
|
|
164
|
-
sentry_sdk-3.0.
|
|
165
|
-
sentry_sdk-3.0.
|
|
166
|
-
sentry_sdk-3.0.
|
|
167
|
-
sentry_sdk-3.0.
|
|
168
|
-
sentry_sdk-3.0.
|
|
163
|
+
sentry_sdk-3.0.0a5.dist-info/licenses/LICENSE,sha256=KhQNZg9GKBL6KQvHQNBGMxJsXsRdhLebVp4Sew7t3Qs,1093
|
|
164
|
+
sentry_sdk-3.0.0a5.dist-info/METADATA,sha256=W5PgWwtJXxOaRFw_G6kh1mbZ7cZp96mYiZoQ0TJbV6A,10151
|
|
165
|
+
sentry_sdk-3.0.0a5.dist-info/WHEEL,sha256=JNWh1Fm1UdwIQV075glCn4MVuCRs0sotJIq-J6rbxCU,109
|
|
166
|
+
sentry_sdk-3.0.0a5.dist-info/entry_points.txt,sha256=-FP10-IbDq7-9RSn7JcaVV6-nDwVN2kwvA46zNTNwtk,78
|
|
167
|
+
sentry_sdk-3.0.0a5.dist-info/top_level.txt,sha256=XrQz30XE9FKXSY_yGLrd9bsv2Rk390GTDJOSujYaMxI,11
|
|
168
|
+
sentry_sdk-3.0.0a5.dist-info/RECORD,,
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|