|
|
import os |
|
|
import uuid |
|
|
import random |
|
|
import socket |
|
|
from collections.abc import Mapping |
|
|
from datetime import datetime, timezone |
|
|
from importlib import import_module |
|
|
from typing import cast, overload |
|
|
|
|
|
from sentry_sdk._compat import PY37, check_uwsgi_thread_support |
|
|
from sentry_sdk.utils import ( |
|
|
ContextVar, |
|
|
capture_internal_exceptions, |
|
|
current_stacktrace, |
|
|
env_to_bool, |
|
|
format_timestamp, |
|
|
get_sdk_name, |
|
|
get_type_name, |
|
|
get_default_release, |
|
|
handle_in_app, |
|
|
is_gevent, |
|
|
logger, |
|
|
) |
|
|
from sentry_sdk.serializer import serialize |
|
|
from sentry_sdk.tracing import trace |
|
|
from sentry_sdk.transport import BaseHttpTransport, make_transport |
|
|
from sentry_sdk.consts import ( |
|
|
DEFAULT_MAX_VALUE_LENGTH, |
|
|
DEFAULT_OPTIONS, |
|
|
INSTRUMENTER, |
|
|
VERSION, |
|
|
ClientConstructor, |
|
|
) |
|
|
from sentry_sdk.integrations import _DEFAULT_INTEGRATIONS, setup_integrations |
|
|
from sentry_sdk.sessions import SessionFlusher |
|
|
from sentry_sdk.envelope import Envelope |
|
|
from sentry_sdk.profiler.continuous_profiler import setup_continuous_profiler |
|
|
from sentry_sdk.profiler.transaction_profiler import ( |
|
|
has_profiling_enabled, |
|
|
Profile, |
|
|
setup_profiler, |
|
|
) |
|
|
from sentry_sdk.scrubber import EventScrubber |
|
|
from sentry_sdk.monitor import Monitor |
|
|
from sentry_sdk.spotlight import setup_spotlight |
|
|
|
|
|
from typing import TYPE_CHECKING |
|
|
|
|
|
if TYPE_CHECKING: |
|
|
from typing import Any |
|
|
from typing import Callable |
|
|
from typing import Dict |
|
|
from typing import Optional |
|
|
from typing import Sequence |
|
|
from typing import Type |
|
|
from typing import Union |
|
|
from typing import TypeVar |
|
|
|
|
|
from sentry_sdk._types import Event, Hint, SDKInfo |
|
|
from sentry_sdk.integrations import Integration |
|
|
from sentry_sdk.metrics import MetricsAggregator |
|
|
from sentry_sdk.scope import Scope |
|
|
from sentry_sdk.session import Session |
|
|
from sentry_sdk.spotlight import SpotlightClient |
|
|
from sentry_sdk.transport import Transport |
|
|
|
|
|
I = TypeVar("I", bound=Integration) |
|
|
|
|
|
_client_init_debug = ContextVar("client_init_debug") |
|
|
|
|
|
|
|
|
SDK_INFO = { |
|
|
"name": "sentry.python", |
|
|
"version": VERSION, |
|
|
"packages": [{"name": "pypi:sentry-sdk", "version": VERSION}], |
|
|
} |
|
|
|
|
|
|
|
|
def _get_options(*args, **kwargs): |
|
|
|
|
|
if args and (isinstance(args[0], (bytes, str)) or args[0] is None): |
|
|
dsn = args[0] |
|
|
args = args[1:] |
|
|
else: |
|
|
dsn = None |
|
|
|
|
|
if len(args) > 1: |
|
|
raise TypeError("Only single positional argument is expected") |
|
|
|
|
|
rv = dict(DEFAULT_OPTIONS) |
|
|
options = dict(*args, **kwargs) |
|
|
if dsn is not None and options.get("dsn") is None: |
|
|
options["dsn"] = dsn |
|
|
|
|
|
for key, value in options.items(): |
|
|
if key not in rv: |
|
|
raise TypeError("Unknown option %r" % (key,)) |
|
|
|
|
|
rv[key] = value |
|
|
|
|
|
if rv["dsn"] is None: |
|
|
rv["dsn"] = os.environ.get("SENTRY_DSN") |
|
|
|
|
|
if rv["release"] is None: |
|
|
rv["release"] = get_default_release() |
|
|
|
|
|
if rv["environment"] is None: |
|
|
rv["environment"] = os.environ.get("SENTRY_ENVIRONMENT") or "production" |
|
|
|
|
|
if rv["debug"] is None: |
|
|
rv["debug"] = env_to_bool(os.environ.get("SENTRY_DEBUG", "False"), strict=True) |
|
|
|
|
|
if rv["server_name"] is None and hasattr(socket, "gethostname"): |
|
|
rv["server_name"] = socket.gethostname() |
|
|
|
|
|
if rv["instrumenter"] is None: |
|
|
rv["instrumenter"] = INSTRUMENTER.SENTRY |
|
|
|
|
|
if rv["project_root"] is None: |
|
|
try: |
|
|
project_root = os.getcwd() |
|
|
except Exception: |
|
|
project_root = None |
|
|
|
|
|
rv["project_root"] = project_root |
|
|
|
|
|
if rv["enable_tracing"] is True and rv["traces_sample_rate"] is None: |
|
|
rv["traces_sample_rate"] = 1.0 |
|
|
|
|
|
if rv["event_scrubber"] is None: |
|
|
rv["event_scrubber"] = EventScrubber( |
|
|
send_default_pii=( |
|
|
False if rv["send_default_pii"] is None else rv["send_default_pii"] |
|
|
) |
|
|
) |
|
|
|
|
|
if rv["socket_options"] and not isinstance(rv["socket_options"], list): |
|
|
logger.warning( |
|
|
"Ignoring socket_options because of unexpected format. See urllib3.HTTPConnection.socket_options for the expected format." |
|
|
) |
|
|
rv["socket_options"] = None |
|
|
|
|
|
return rv |
|
|
|
|
|
|
|
|
try: |
|
|
|
|
|
module_not_found_error = ModuleNotFoundError |
|
|
except Exception: |
|
|
|
|
|
module_not_found_error = ImportError |
|
|
|
|
|
|
|
|
class BaseClient: |
|
|
""" |
|
|
.. versionadded:: 2.0.0 |
|
|
|
|
|
The basic definition of a client that is used for sending data to Sentry. |
|
|
""" |
|
|
|
|
|
spotlight = None |
|
|
|
|
|
def __init__(self, options=None): |
|
|
|
|
|
self.options = ( |
|
|
options if options is not None else DEFAULT_OPTIONS |
|
|
) |
|
|
|
|
|
self.transport = None |
|
|
self.monitor = None |
|
|
self.metrics_aggregator = None |
|
|
|
|
|
def __getstate__(self, *args, **kwargs): |
|
|
|
|
|
return {"options": {}} |
|
|
|
|
|
def __setstate__(self, *args, **kwargs): |
|
|
|
|
|
pass |
|
|
|
|
|
@property |
|
|
def dsn(self): |
|
|
|
|
|
return None |
|
|
|
|
|
def should_send_default_pii(self): |
|
|
|
|
|
return False |
|
|
|
|
|
def is_active(self): |
|
|
|
|
|
""" |
|
|
.. versionadded:: 2.0.0 |
|
|
|
|
|
Returns whether the client is active (able to send data to Sentry) |
|
|
""" |
|
|
return False |
|
|
|
|
|
def capture_event(self, *args, **kwargs): |
|
|
|
|
|
return None |
|
|
|
|
|
def capture_session(self, *args, **kwargs): |
|
|
|
|
|
return None |
|
|
|
|
|
if TYPE_CHECKING: |
|
|
|
|
|
@overload |
|
|
def get_integration(self, name_or_class): |
|
|
|
|
|
... |
|
|
|
|
|
@overload |
|
|
def get_integration(self, name_or_class): |
|
|
|
|
|
... |
|
|
|
|
|
def get_integration(self, name_or_class): |
|
|
|
|
|
return None |
|
|
|
|
|
def close(self, *args, **kwargs): |
|
|
|
|
|
return None |
|
|
|
|
|
def flush(self, *args, **kwargs): |
|
|
|
|
|
return None |
|
|
|
|
|
def __enter__(self): |
|
|
|
|
|
return self |
|
|
|
|
|
def __exit__(self, exc_type, exc_value, tb): |
|
|
|
|
|
return None |
|
|
|
|
|
|
|
|
class NonRecordingClient(BaseClient): |
|
|
""" |
|
|
.. versionadded:: 2.0.0 |
|
|
|
|
|
A client that does not send any events to Sentry. This is used as a fallback when the Sentry SDK is not yet initialized. |
|
|
""" |
|
|
|
|
|
pass |
|
|
|
|
|
|
|
|
class _Client(BaseClient): |
|
|
""" |
|
|
The client is internally responsible for capturing the events and |
|
|
forwarding them to sentry through the configured transport. It takes |
|
|
the client options as keyword arguments and optionally the DSN as first |
|
|
argument. |
|
|
|
|
|
Alias of :py:class:`sentry_sdk.Client`. (Was created for better intelisense support) |
|
|
""" |
|
|
|
|
|
def __init__(self, *args, **kwargs): |
|
|
|
|
|
super(_Client, self).__init__(options=get_options(*args, **kwargs)) |
|
|
self._init_impl() |
|
|
|
|
|
def __getstate__(self): |
|
|
|
|
|
return {"options": self.options} |
|
|
|
|
|
def __setstate__(self, state): |
|
|
|
|
|
self.options = state["options"] |
|
|
self._init_impl() |
|
|
|
|
|
def _setup_instrumentation(self, functions_to_trace): |
|
|
|
|
|
""" |
|
|
Instruments the functions given in the list `functions_to_trace` with the `@sentry_sdk.tracing.trace` decorator. |
|
|
""" |
|
|
for function in functions_to_trace: |
|
|
class_name = None |
|
|
function_qualname = function["qualified_name"] |
|
|
module_name, function_name = function_qualname.rsplit(".", 1) |
|
|
|
|
|
try: |
|
|
|
|
|
|
|
|
|
|
|
module_obj = import_module(module_name) |
|
|
function_obj = getattr(module_obj, function_name) |
|
|
setattr(module_obj, function_name, trace(function_obj)) |
|
|
logger.debug("Enabled tracing for %s", function_qualname) |
|
|
except module_not_found_error: |
|
|
try: |
|
|
|
|
|
|
|
|
|
|
|
module_name, class_name = module_name.rsplit(".", 1) |
|
|
module_obj = import_module(module_name) |
|
|
class_obj = getattr(module_obj, class_name) |
|
|
function_obj = getattr(class_obj, function_name) |
|
|
function_type = type(class_obj.__dict__[function_name]) |
|
|
traced_function = trace(function_obj) |
|
|
|
|
|
if function_type in (staticmethod, classmethod): |
|
|
traced_function = staticmethod(traced_function) |
|
|
|
|
|
setattr(class_obj, function_name, traced_function) |
|
|
setattr(module_obj, class_name, class_obj) |
|
|
logger.debug("Enabled tracing for %s", function_qualname) |
|
|
|
|
|
except Exception as e: |
|
|
logger.warning( |
|
|
"Can not enable tracing for '%s'. (%s) Please check your `functions_to_trace` parameter.", |
|
|
function_qualname, |
|
|
e, |
|
|
) |
|
|
|
|
|
except Exception as e: |
|
|
logger.warning( |
|
|
"Can not enable tracing for '%s'. (%s) Please check your `functions_to_trace` parameter.", |
|
|
function_qualname, |
|
|
e, |
|
|
) |
|
|
|
|
|
def _init_impl(self): |
|
|
|
|
|
old_debug = _client_init_debug.get(False) |
|
|
|
|
|
def _capture_envelope(envelope): |
|
|
|
|
|
if self.transport is not None: |
|
|
self.transport.capture_envelope(envelope) |
|
|
|
|
|
try: |
|
|
_client_init_debug.set(self.options["debug"]) |
|
|
self.transport = make_transport(self.options) |
|
|
|
|
|
self.monitor = None |
|
|
if self.transport: |
|
|
if self.options["enable_backpressure_handling"]: |
|
|
self.monitor = Monitor(self.transport) |
|
|
|
|
|
self.session_flusher = SessionFlusher(capture_func=_capture_envelope) |
|
|
|
|
|
self.metrics_aggregator = None |
|
|
experiments = self.options.get("_experiments", {}) |
|
|
if experiments.get("enable_metrics", True): |
|
|
|
|
|
|
|
|
metrics_supported = not is_gevent() or PY37 |
|
|
if metrics_supported: |
|
|
from sentry_sdk.metrics import MetricsAggregator |
|
|
|
|
|
self.metrics_aggregator = MetricsAggregator( |
|
|
capture_func=_capture_envelope, |
|
|
enable_code_locations=bool( |
|
|
experiments.get("metric_code_locations", True) |
|
|
), |
|
|
) |
|
|
else: |
|
|
logger.info( |
|
|
"Metrics not supported on Python 3.6 and lower with gevent." |
|
|
) |
|
|
|
|
|
max_request_body_size = ("always", "never", "small", "medium") |
|
|
if self.options["max_request_body_size"] not in max_request_body_size: |
|
|
raise ValueError( |
|
|
"Invalid value for max_request_body_size. Must be one of {}".format( |
|
|
max_request_body_size |
|
|
) |
|
|
) |
|
|
|
|
|
if self.options["_experiments"].get("otel_powered_performance", False): |
|
|
logger.debug( |
|
|
"[OTel] Enabling experimental OTel-powered performance monitoring." |
|
|
) |
|
|
self.options["instrumenter"] = INSTRUMENTER.OTEL |
|
|
if ( |
|
|
"sentry_sdk.integrations.opentelemetry.integration.OpenTelemetryIntegration" |
|
|
not in _DEFAULT_INTEGRATIONS |
|
|
): |
|
|
_DEFAULT_INTEGRATIONS.append( |
|
|
"sentry_sdk.integrations.opentelemetry.integration.OpenTelemetryIntegration", |
|
|
) |
|
|
|
|
|
self.integrations = setup_integrations( |
|
|
self.options["integrations"], |
|
|
with_defaults=self.options["default_integrations"], |
|
|
with_auto_enabling_integrations=self.options[ |
|
|
"auto_enabling_integrations" |
|
|
], |
|
|
disabled_integrations=self.options["disabled_integrations"], |
|
|
) |
|
|
|
|
|
spotlight_config = self.options.get("spotlight") |
|
|
if spotlight_config is None and "SENTRY_SPOTLIGHT" in os.environ: |
|
|
spotlight_env_value = os.environ["SENTRY_SPOTLIGHT"] |
|
|
spotlight_config = env_to_bool(spotlight_env_value, strict=True) |
|
|
self.options["spotlight"] = ( |
|
|
spotlight_config |
|
|
if spotlight_config is not None |
|
|
else spotlight_env_value |
|
|
) |
|
|
|
|
|
if self.options.get("spotlight"): |
|
|
self.spotlight = setup_spotlight(self.options) |
|
|
|
|
|
sdk_name = get_sdk_name(list(self.integrations.keys())) |
|
|
SDK_INFO["name"] = sdk_name |
|
|
logger.debug("Setting SDK name to '%s'", sdk_name) |
|
|
|
|
|
if has_profiling_enabled(self.options): |
|
|
try: |
|
|
setup_profiler(self.options) |
|
|
except Exception as e: |
|
|
logger.debug("Can not set up profiler. (%s)", e) |
|
|
else: |
|
|
try: |
|
|
setup_continuous_profiler( |
|
|
self.options, |
|
|
sdk_info=SDK_INFO, |
|
|
capture_func=_capture_envelope, |
|
|
) |
|
|
except Exception as e: |
|
|
logger.debug("Can not set up continuous profiler. (%s)", e) |
|
|
|
|
|
finally: |
|
|
_client_init_debug.set(old_debug) |
|
|
|
|
|
self._setup_instrumentation(self.options.get("functions_to_trace", [])) |
|
|
|
|
|
if ( |
|
|
self.monitor |
|
|
or self.metrics_aggregator |
|
|
or has_profiling_enabled(self.options) |
|
|
or isinstance(self.transport, BaseHttpTransport) |
|
|
): |
|
|
|
|
|
|
|
|
check_uwsgi_thread_support() |
|
|
|
|
|
def is_active(self): |
|
|
|
|
|
""" |
|
|
.. versionadded:: 2.0.0 |
|
|
|
|
|
Returns whether the client is active (able to send data to Sentry) |
|
|
""" |
|
|
return True |
|
|
|
|
|
def should_send_default_pii(self): |
|
|
|
|
|
""" |
|
|
.. versionadded:: 2.0.0 |
|
|
|
|
|
Returns whether the client should send default PII (Personally Identifiable Information) data to Sentry. |
|
|
""" |
|
|
result = self.options.get("send_default_pii") |
|
|
if result is None: |
|
|
result = not self.options["dsn"] and self.spotlight is not None |
|
|
|
|
|
return result |
|
|
|
|
|
@property |
|
|
def dsn(self): |
|
|
|
|
|
"""Returns the configured DSN as string.""" |
|
|
return self.options["dsn"] |
|
|
|
|
|
def _prepare_event( |
|
|
self, |
|
|
event, |
|
|
hint, |
|
|
scope, |
|
|
): |
|
|
|
|
|
|
|
|
if event.get("timestamp") is None: |
|
|
event["timestamp"] = datetime.now(timezone.utc) |
|
|
|
|
|
if scope is not None: |
|
|
is_transaction = event.get("type") == "transaction" |
|
|
spans_before = len(event.get("spans", [])) |
|
|
event_ = scope.apply_to_event(event, hint, self.options) |
|
|
|
|
|
|
|
|
if event_ is None: |
|
|
if self.transport: |
|
|
self.transport.record_lost_event( |
|
|
"event_processor", |
|
|
data_category=("transaction" if is_transaction else "error"), |
|
|
) |
|
|
if is_transaction: |
|
|
self.transport.record_lost_event( |
|
|
"event_processor", |
|
|
data_category="span", |
|
|
quantity=spans_before + 1, |
|
|
) |
|
|
return None |
|
|
|
|
|
event = event_ |
|
|
|
|
|
spans_delta = spans_before - len(event.get("spans", [])) |
|
|
if is_transaction and spans_delta > 0 and self.transport is not None: |
|
|
self.transport.record_lost_event( |
|
|
"event_processor", data_category="span", quantity=spans_delta |
|
|
) |
|
|
|
|
|
if ( |
|
|
self.options["attach_stacktrace"] |
|
|
and "exception" not in event |
|
|
and "stacktrace" not in event |
|
|
and "threads" not in event |
|
|
): |
|
|
with capture_internal_exceptions(): |
|
|
event["threads"] = { |
|
|
"values": [ |
|
|
{ |
|
|
"stacktrace": current_stacktrace( |
|
|
include_local_variables=self.options.get( |
|
|
"include_local_variables", True |
|
|
), |
|
|
max_value_length=self.options.get( |
|
|
"max_value_length", DEFAULT_MAX_VALUE_LENGTH |
|
|
), |
|
|
), |
|
|
"crashed": False, |
|
|
"current": True, |
|
|
} |
|
|
] |
|
|
} |
|
|
|
|
|
for key in "release", "environment", "server_name", "dist": |
|
|
if event.get(key) is None and self.options[key] is not None: |
|
|
event[key] = str(self.options[key]).strip() |
|
|
if event.get("sdk") is None: |
|
|
sdk_info = dict(SDK_INFO) |
|
|
sdk_info["integrations"] = sorted(self.integrations.keys()) |
|
|
event["sdk"] = sdk_info |
|
|
|
|
|
if event.get("platform") is None: |
|
|
event["platform"] = "python" |
|
|
|
|
|
event = handle_in_app( |
|
|
event, |
|
|
self.options["in_app_exclude"], |
|
|
self.options["in_app_include"], |
|
|
self.options["project_root"], |
|
|
) |
|
|
|
|
|
if event is not None: |
|
|
event_scrubber = self.options["event_scrubber"] |
|
|
if event_scrubber: |
|
|
event_scrubber.scrub_event(event) |
|
|
|
|
|
|
|
|
|
|
|
if event is not None: |
|
|
event = cast( |
|
|
"Event", |
|
|
serialize( |
|
|
cast("Dict[str, Any]", event), |
|
|
max_request_body_size=self.options.get("max_request_body_size"), |
|
|
max_value_length=self.options.get("max_value_length"), |
|
|
custom_repr=self.options.get("custom_repr"), |
|
|
), |
|
|
) |
|
|
|
|
|
before_send = self.options["before_send"] |
|
|
if ( |
|
|
before_send is not None |
|
|
and event is not None |
|
|
and event.get("type") != "transaction" |
|
|
): |
|
|
new_event = None |
|
|
with capture_internal_exceptions(): |
|
|
new_event = before_send(event, hint or {}) |
|
|
if new_event is None: |
|
|
logger.info("before send dropped event") |
|
|
if self.transport: |
|
|
self.transport.record_lost_event( |
|
|
"before_send", data_category="error" |
|
|
) |
|
|
event = new_event |
|
|
|
|
|
before_send_transaction = self.options["before_send_transaction"] |
|
|
if ( |
|
|
before_send_transaction is not None |
|
|
and event is not None |
|
|
and event.get("type") == "transaction" |
|
|
): |
|
|
new_event = None |
|
|
spans_before = len(event.get("spans", [])) |
|
|
with capture_internal_exceptions(): |
|
|
new_event = before_send_transaction(event, hint or {}) |
|
|
if new_event is None: |
|
|
logger.info("before send transaction dropped event") |
|
|
if self.transport: |
|
|
self.transport.record_lost_event( |
|
|
reason="before_send", data_category="transaction" |
|
|
) |
|
|
self.transport.record_lost_event( |
|
|
reason="before_send", |
|
|
data_category="span", |
|
|
quantity=spans_before + 1, |
|
|
) |
|
|
else: |
|
|
spans_delta = spans_before - len(new_event.get("spans", [])) |
|
|
if spans_delta > 0 and self.transport is not None: |
|
|
self.transport.record_lost_event( |
|
|
reason="before_send", data_category="span", quantity=spans_delta |
|
|
) |
|
|
|
|
|
event = new_event |
|
|
|
|
|
return event |
|
|
|
|
|
def _is_ignored_error(self, event, hint): |
|
|
|
|
|
exc_info = hint.get("exc_info") |
|
|
if exc_info is None: |
|
|
return False |
|
|
|
|
|
error = exc_info[0] |
|
|
error_type_name = get_type_name(exc_info[0]) |
|
|
error_full_name = "%s.%s" % (exc_info[0].__module__, error_type_name) |
|
|
|
|
|
for ignored_error in self.options["ignore_errors"]: |
|
|
|
|
|
|
|
|
if isinstance(ignored_error, str): |
|
|
if ignored_error == error_full_name or ignored_error == error_type_name: |
|
|
return True |
|
|
else: |
|
|
if issubclass(error, ignored_error): |
|
|
return True |
|
|
|
|
|
return False |
|
|
|
|
|
def _should_capture( |
|
|
self, |
|
|
event, |
|
|
hint, |
|
|
scope=None, |
|
|
): |
|
|
|
|
|
|
|
|
is_transaction = event.get("type") == "transaction" |
|
|
if is_transaction: |
|
|
return True |
|
|
|
|
|
ignoring_prevents_recursion = scope is not None and not scope._should_capture |
|
|
if ignoring_prevents_recursion: |
|
|
return False |
|
|
|
|
|
ignored_by_config_option = self._is_ignored_error(event, hint) |
|
|
if ignored_by_config_option: |
|
|
return False |
|
|
|
|
|
return True |
|
|
|
|
|
def _should_sample_error( |
|
|
self, |
|
|
event, |
|
|
hint, |
|
|
): |
|
|
|
|
|
error_sampler = self.options.get("error_sampler", None) |
|
|
|
|
|
if callable(error_sampler): |
|
|
with capture_internal_exceptions(): |
|
|
sample_rate = error_sampler(event, hint) |
|
|
else: |
|
|
sample_rate = self.options["sample_rate"] |
|
|
|
|
|
try: |
|
|
not_in_sample_rate = sample_rate < 1.0 and random.random() >= sample_rate |
|
|
except NameError: |
|
|
logger.warning( |
|
|
"The provided error_sampler raised an error. Defaulting to sampling the event." |
|
|
) |
|
|
|
|
|
|
|
|
|
|
|
not_in_sample_rate = False |
|
|
except TypeError: |
|
|
parameter, verb = ( |
|
|
("error_sampler", "returned") |
|
|
if callable(error_sampler) |
|
|
else ("sample_rate", "contains") |
|
|
) |
|
|
logger.warning( |
|
|
"The provided %s %s an invalid value of %s. The value should be a float or a bool. Defaulting to sampling the event." |
|
|
% (parameter, verb, repr(sample_rate)) |
|
|
) |
|
|
|
|
|
|
|
|
|
|
|
not_in_sample_rate = False |
|
|
|
|
|
if not_in_sample_rate: |
|
|
|
|
|
if self.transport: |
|
|
self.transport.record_lost_event("sample_rate", data_category="error") |
|
|
|
|
|
return False |
|
|
|
|
|
return True |
|
|
|
|
|
def _update_session_from_event( |
|
|
self, |
|
|
session, |
|
|
event, |
|
|
): |
|
|
|
|
|
|
|
|
crashed = False |
|
|
errored = False |
|
|
user_agent = None |
|
|
|
|
|
exceptions = (event.get("exception") or {}).get("values") |
|
|
if exceptions: |
|
|
errored = True |
|
|
for error in exceptions: |
|
|
mechanism = error.get("mechanism") |
|
|
if isinstance(mechanism, Mapping) and mechanism.get("handled") is False: |
|
|
crashed = True |
|
|
break |
|
|
|
|
|
user = event.get("user") |
|
|
|
|
|
if session.user_agent is None: |
|
|
headers = (event.get("request") or {}).get("headers") |
|
|
headers_dict = headers if isinstance(headers, dict) else {} |
|
|
for k, v in headers_dict.items(): |
|
|
if k.lower() == "user-agent": |
|
|
user_agent = v |
|
|
break |
|
|
|
|
|
session.update( |
|
|
status="crashed" if crashed else None, |
|
|
user=user, |
|
|
user_agent=user_agent, |
|
|
errors=session.errors + (errored or crashed), |
|
|
) |
|
|
|
|
|
def capture_event( |
|
|
self, |
|
|
event, |
|
|
hint=None, |
|
|
scope=None, |
|
|
): |
|
|
|
|
|
"""Captures an event. |
|
|
|
|
|
:param event: A ready-made event that can be directly sent to Sentry. |
|
|
|
|
|
:param hint: Contains metadata about the event that can be read from `before_send`, such as the original exception object or a HTTP request object. |
|
|
|
|
|
:param scope: An optional :py:class:`sentry_sdk.Scope` to apply to events. |
|
|
|
|
|
:returns: An event ID. May be `None` if there is no DSN set or of if the SDK decided to discard the event for other reasons. In such situations setting `debug=True` on `init()` may help. |
|
|
""" |
|
|
hint = dict(hint or ()) |
|
|
|
|
|
if not self._should_capture(event, hint, scope): |
|
|
return None |
|
|
|
|
|
profile = event.pop("profile", None) |
|
|
|
|
|
event_id = event.get("event_id") |
|
|
if event_id is None: |
|
|
event["event_id"] = event_id = uuid.uuid4().hex |
|
|
event_opt = self._prepare_event(event, hint, scope) |
|
|
if event_opt is None: |
|
|
return None |
|
|
|
|
|
|
|
|
|
|
|
session = scope._session if scope else None |
|
|
if session: |
|
|
self._update_session_from_event(session, event) |
|
|
|
|
|
is_transaction = event_opt.get("type") == "transaction" |
|
|
is_checkin = event_opt.get("type") == "check_in" |
|
|
|
|
|
if ( |
|
|
not is_transaction |
|
|
and not is_checkin |
|
|
and not self._should_sample_error(event, hint) |
|
|
): |
|
|
return None |
|
|
|
|
|
attachments = hint.get("attachments") |
|
|
|
|
|
trace_context = event_opt.get("contexts", {}).get("trace") or {} |
|
|
dynamic_sampling_context = trace_context.pop("dynamic_sampling_context", {}) |
|
|
|
|
|
headers = { |
|
|
"event_id": event_opt["event_id"], |
|
|
"sent_at": format_timestamp(datetime.now(timezone.utc)), |
|
|
} |
|
|
|
|
|
if dynamic_sampling_context: |
|
|
headers["trace"] = dynamic_sampling_context |
|
|
|
|
|
envelope = Envelope(headers=headers) |
|
|
|
|
|
if is_transaction: |
|
|
if isinstance(profile, Profile): |
|
|
envelope.add_profile(profile.to_json(event_opt, self.options)) |
|
|
envelope.add_transaction(event_opt) |
|
|
elif is_checkin: |
|
|
envelope.add_checkin(event_opt) |
|
|
else: |
|
|
envelope.add_event(event_opt) |
|
|
|
|
|
for attachment in attachments or (): |
|
|
envelope.add_item(attachment.to_envelope_item()) |
|
|
|
|
|
return_value = None |
|
|
if self.spotlight: |
|
|
self.spotlight.capture_envelope(envelope) |
|
|
return_value = event_id |
|
|
|
|
|
if self.transport is not None: |
|
|
self.transport.capture_envelope(envelope) |
|
|
return_value = event_id |
|
|
|
|
|
return return_value |
|
|
|
|
|
def capture_session( |
|
|
self, session |
|
|
): |
|
|
|
|
|
if not session.release: |
|
|
logger.info("Discarded session update because of missing release") |
|
|
else: |
|
|
self.session_flusher.add_session(session) |
|
|
|
|
|
if TYPE_CHECKING: |
|
|
|
|
|
@overload |
|
|
def get_integration(self, name_or_class): |
|
|
|
|
|
... |
|
|
|
|
|
@overload |
|
|
def get_integration(self, name_or_class): |
|
|
|
|
|
... |
|
|
|
|
|
def get_integration( |
|
|
self, name_or_class |
|
|
): |
|
|
|
|
|
"""Returns the integration for this client by name or class. |
|
|
If the client does not have that integration then `None` is returned. |
|
|
""" |
|
|
if isinstance(name_or_class, str): |
|
|
integration_name = name_or_class |
|
|
elif name_or_class.identifier is not None: |
|
|
integration_name = name_or_class.identifier |
|
|
else: |
|
|
raise ValueError("Integration has no name") |
|
|
|
|
|
return self.integrations.get(integration_name) |
|
|
|
|
|
def close( |
|
|
self, |
|
|
timeout=None, |
|
|
callback=None, |
|
|
): |
|
|
|
|
|
""" |
|
|
Close the client and shut down the transport. Arguments have the same |
|
|
semantics as :py:meth:`Client.flush`. |
|
|
""" |
|
|
if self.transport is not None: |
|
|
self.flush(timeout=timeout, callback=callback) |
|
|
self.session_flusher.kill() |
|
|
if self.metrics_aggregator is not None: |
|
|
self.metrics_aggregator.kill() |
|
|
if self.monitor: |
|
|
self.monitor.kill() |
|
|
self.transport.kill() |
|
|
self.transport = None |
|
|
|
|
|
def flush( |
|
|
self, |
|
|
timeout=None, |
|
|
callback=None, |
|
|
): |
|
|
|
|
|
""" |
|
|
Wait for the current events to be sent. |
|
|
|
|
|
:param timeout: Wait for at most `timeout` seconds. If no `timeout` is provided, the `shutdown_timeout` option value is used. |
|
|
|
|
|
:param callback: Is invoked with the number of pending events and the configured timeout. |
|
|
""" |
|
|
if self.transport is not None: |
|
|
if timeout is None: |
|
|
timeout = self.options["shutdown_timeout"] |
|
|
self.session_flusher.flush() |
|
|
if self.metrics_aggregator is not None: |
|
|
self.metrics_aggregator.flush() |
|
|
self.transport.flush(timeout=timeout, callback=callback) |
|
|
|
|
|
def __enter__(self): |
|
|
|
|
|
return self |
|
|
|
|
|
def __exit__(self, exc_type, exc_value, tb): |
|
|
|
|
|
self.close() |
|
|
|
|
|
|
|
|
from typing import TYPE_CHECKING |
|
|
|
|
|
if TYPE_CHECKING: |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
class get_options(ClientConstructor, Dict[str, Any]): |
|
|
pass |
|
|
|
|
|
class Client(ClientConstructor, _Client): |
|
|
pass |
|
|
|
|
|
else: |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
get_options = (lambda: _get_options)() |
|
|
Client = (lambda: _Client)() |
|
|
|