elasticapm/conf/__init__.py (629 lines of code) (raw):
# BSD 3-Clause License
#
# Copyright (c) 2012, the Sentry Team, see AUTHORS for more details
# Copyright (c) 2019, Elasticsearch BV
# All rights reserved.
#
# Redistribution and use in source and binary forms, with or without
# modification, are permitted provided that the following conditions are met:
#
# * Redistributions of source code must retain the above copyright notice, this
# list of conditions and the following disclaimer.
#
# * Redistributions in binary form must reproduce the above copyright notice,
# this list of conditions and the following disclaimer in the documentation
# and/or other materials provided with the distribution.
#
# * Neither the name of the copyright holder nor the names of its
# contributors may be used to endorse or promote products derived from
# this software without specific prior written permission.
#
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
# AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
# IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
# DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE
# FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
# DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
# SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
# CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
# OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
import logging
import logging.handlers
import math
import os
import re
import threading
from datetime import timedelta
import _hashlib
from elasticapm.conf.constants import BASE_SANITIZE_FIELD_NAMES, TRACE_CONTINUATION_STRATEGY
from elasticapm.utils import compat, starmatch_to_regex
from elasticapm.utils.logging import get_logger
from elasticapm.utils.threading import IntervalTimer, ThreadManager
__all__ = ("setup_logging", "Config")
logger = get_logger("elasticapm.conf")
log_levels_map = {
"trace": 5,
"debug": logging.DEBUG,
"info": logging.INFO,
"warning": logging.WARNING,
"warn": logging.WARNING,
"error": logging.ERROR,
"critical": logging.CRITICAL,
"off": 1000,
}
logfile_set_up = False
class ConfigurationError(ValueError):
def __init__(self, msg, field_name) -> None:
self.field_name = field_name
super(ValueError, self).__init__(msg)
class _ConfigValue(object):
"""
Base class for configuration values
dict_key
String representing the key used for this config value in dict configs.
env_key
String representing the key used in environment variables for this
config value. If not specified, will be set to `"ELASTIC_APM_" + dict_key`.
type
Type of value stored in this config value.
validators
List of validator classes. Must be callables, which will be called with
a value and the dict_key for the config value. The validator either
returns the validated value or raises a ConfigurationError if validation
fails.
callbacks
List of functions which will be called when the config value is updated.
The callbacks must match this signature:
callback(dict_key, old_value, new_value, config_instance)
Note that callbacks wait until the end of any given `update()` operation
and are called at this point. This, coupled with the fact that callbacks
receive the config instance, means that callbacks can utilize multiple
configuration values (such as is the case for logging). This is
complicated if more than one of the involved config values are
dynamic, as both would need callbacks and the callback would need to
be idempotent.
callbacks_on_default
Whether the callback should be called on config initialization if the
default value is used. Default: True
default
The default for this config value if not user-configured.
required
Whether this config value is required. If a default is specified,
this is a redundant option (except to ensure that this config value
is specified if a default were ever to be removed).
Note that _ConfigValues and any inheriting classes must implement __set__
and __get__. The calling instance will always be a _ConfigBase descendant
and the __set__ and __get__ calls will access `instance._values[self.dict_key]`
to get and set values.
"""
def __init__(
self,
dict_key,
env_key=None,
type=str,
validators=None,
callbacks=None,
callbacks_on_default=True,
default=None,
required=False,
) -> None:
self.type = type
self.dict_key = dict_key
self.validators = validators
self.callbacks = callbacks
self.default = default
self.required = required
if env_key is None:
env_key = "ELASTIC_APM_" + dict_key
self.env_key = env_key
self.callbacks_on_default = callbacks_on_default
def __get__(self, instance, owner):
if instance:
return instance._values.get(self.dict_key, self.default)
else:
return self.default
def __set__(self, config_instance, value) -> None:
value = self._validate(config_instance, value)
self._callback_if_changed(config_instance, value)
config_instance._values[self.dict_key] = value
def _validate(self, instance, value):
if value is None and self.required:
raise ConfigurationError(
"Configuration error: value for {} is required.".format(self.dict_key), self.dict_key
)
if self.validators and value is not None:
for validator in self.validators:
value = validator(value, self.dict_key)
if self.type and value is not None:
try:
value = self.type(value)
except ValueError as e:
raise ConfigurationError("{}: {}".format(self.dict_key, str(e)), self.dict_key)
instance._errors.pop(self.dict_key, None)
return value
def _callback_if_changed(self, instance, new_value) -> None:
"""
If the value changed (checked against instance._values[self.dict_key]),
then run the callback function (if defined)
"""
old_value = instance._values.get(self.dict_key, self.default)
if old_value != new_value:
instance.callbacks_queue.append((self.dict_key, old_value, new_value))
def call_callbacks(self, old_value, new_value, config_instance):
if not self.callbacks:
return
for callback in self.callbacks:
try:
callback(self.dict_key, old_value, new_value, config_instance)
except Exception as e:
raise ConfigurationError(
"Callback {} raised an exception when setting {} to {}: {}".format(
callback, self.dict_key, new_value, e
),
self.dict_key,
)
class _ListConfigValue(_ConfigValue):
def __init__(self, dict_key, list_separator=",", **kwargs) -> None:
self.list_separator = list_separator
super(_ListConfigValue, self).__init__(dict_key, **kwargs)
def __set__(self, instance, value) -> None:
if isinstance(value, str):
value = value.split(self.list_separator)
elif value is not None:
value = list(value)
if value:
value = [self.type(item) for item in value]
self._callback_if_changed(instance, value)
instance._values[self.dict_key] = value
class _DictConfigValue(_ConfigValue):
def __init__(self, dict_key, item_separator=",", keyval_separator="=", **kwargs) -> None:
self.item_separator = item_separator
self.keyval_separator = keyval_separator
super(_DictConfigValue, self).__init__(dict_key, **kwargs)
def __set__(self, instance, value) -> None:
if isinstance(value, str):
items = (item.split(self.keyval_separator) for item in value.split(self.item_separator))
value = {key.strip(): self.type(val.strip()) for key, val in items}
elif not isinstance(value, dict):
# TODO: better error handling
value = None
self._callback_if_changed(instance, value)
instance._values[self.dict_key] = value
class _BoolConfigValue(_ConfigValue):
def __init__(self, dict_key, true_string="true", false_string="false", **kwargs) -> None:
self.true_string = true_string
self.false_string = false_string
# this is necessary to have the bool type preserved in _validate
kwargs["type"] = bool
super(_BoolConfigValue, self).__init__(dict_key, **kwargs)
def __set__(self, instance, value) -> None:
if isinstance(value, str):
if value.lower() == self.true_string:
value = True
elif value.lower() == self.false_string:
value = False
value = self._validate(instance, value)
self._callback_if_changed(instance, value)
instance._values[self.dict_key] = bool(value)
class _DurationConfigValue(_ConfigValue):
units = (
("us", 0.000001),
("ms", 0.001),
("s", 1),
("m", 60),
)
def __init__(self, dict_key, allow_microseconds=False, unitless_factor=None, **kwargs) -> None:
self.type = None # no type coercion
used_units = self.units if allow_microseconds else self.units[1:]
pattern = "|".join(unit[0] for unit in used_units)
unit_multipliers = dict(used_units)
unit_required = ""
if unitless_factor:
unit_multipliers[None] = unitless_factor
unit_required = "?"
regex = f"((?:-)?\\d+)({pattern}){unit_required}$"
verbose_pattern = f"\\d+({pattern}){unit_required}$"
duration_validator = UnitValidator(
regex=regex, verbose_pattern=verbose_pattern, unit_multipliers=unit_multipliers
)
validators = kwargs.pop("validators", [])
validators.insert(0, duration_validator)
super().__init__(dict_key, validators=validators, **kwargs)
def __set__(self, config_instance, value) -> None:
value = self._validate(config_instance, value)
value = timedelta(seconds=float(value))
self._callback_if_changed(config_instance, value)
config_instance._values[self.dict_key] = value
class RegexValidator(object):
def __init__(self, regex, verbose_pattern=None) -> None:
self.regex = regex
self.verbose_pattern = verbose_pattern or regex
def __call__(self, value, field_name):
value = str(value)
match = re.match(self.regex, value)
if match:
return value
raise ConfigurationError(
"{}={} does not match pattern {}".format(
field_name,
value,
self.verbose_pattern,
),
field_name,
)
class UnitValidator(object):
def __init__(self, regex, verbose_pattern, unit_multipliers) -> None:
self.regex = regex
self.verbose_pattern = verbose_pattern
self.unit_multipliers = unit_multipliers
def __call__(self, value, field_name):
value = str(value)
match = re.match(self.regex, value, re.IGNORECASE)
if not match:
raise ConfigurationError(
"{}={} does not match pattern {}".format(
field_name,
value,
self.verbose_pattern,
),
field_name,
)
val, unit = match.groups()
try:
val = int(val) * self.unit_multipliers[unit]
except KeyError:
raise ConfigurationError("{}={} is not a supported unit".format(field_name, unit), field_name)
return val
class PrecisionValidator(object):
"""
Forces a float value to `precision` digits of precision.
Rounds half away from zero.
If `minimum` is provided, and the value rounds to 0 (but was not zero to
begin with), use the minimum instead.
"""
def __init__(self, precision=0, minimum=None) -> None:
self.precision = precision
self.minimum = minimum
def __call__(self, value, field_name):
try:
value = float(value)
except ValueError:
raise ConfigurationError("{}={} is not a float".format(field_name, value), field_name)
multiplier = 10**self.precision
rounded = math.floor(value * multiplier + 0.5) / multiplier
if rounded == 0 and self.minimum and value != 0:
rounded = self.minimum
return rounded
size_validator = UnitValidator(
r"^(\d+)(b|kb|mb|gb)$", r"\d+(b|KB|MB|GB)", {"b": 1, "kb": 1024, "mb": 1024 * 1024, "gb": 1024 * 1024 * 1024}
)
class ExcludeRangeValidator(object):
def __init__(self, range_start, range_end, range_desc) -> None:
self.range_start = range_start
self.range_end = range_end
self.range_desc = range_desc
def __call__(self, value, field_name):
if self.range_start <= value <= self.range_end:
raise ConfigurationError(
"{}={} cannot be in range: {}".format(
field_name,
value,
self.range_desc.format(**{"range_start": self.range_start, "range_end": self.range_end}),
),
field_name,
)
return value
class FileIsReadableValidator(object):
def __call__(self, value, field_name):
value = os.path.normpath(value)
if not os.path.exists(value):
raise ConfigurationError("{}={} does not exist".format(field_name, value), field_name)
elif not os.path.isfile(value):
raise ConfigurationError("{}={} is not a file".format(field_name, value), field_name)
elif not os.access(value, os.R_OK):
raise ConfigurationError("{}={} is not readable".format(field_name, value), field_name)
return value
def _in_fips_mode():
try:
return _hashlib.get_fips_mode() == 1
except AttributeError:
# versions older of Python3.9 do not have the helper
return False
class SupportedValueInFipsModeValidator(object):
"""If FIPS mode is enabled only supported_value is accepted"""
def __init__(self, supported_value) -> None:
self.supported_value = supported_value
def __call__(self, value, field_name):
if _in_fips_mode():
if value != self.supported_value:
raise ConfigurationError(
"{}={} must be set to {} if FIPS mode is enabled".format(field_name, value, self.supported_value),
field_name,
)
return value
class EnumerationValidator(object):
"""
Validator which ensures that a given config value is chosen from a list
of valid string options.
"""
def __init__(self, valid_values, case_sensitive=False) -> None:
"""
valid_values
List of valid string values for the config value
case_sensitive
Whether to compare case when comparing a value to the valid list.
Defaults to False (case-insensitive)
"""
self.case_sensitive = case_sensitive
if case_sensitive:
self.valid_values = {s: s for s in valid_values}
else:
self.valid_values = {s.lower(): s for s in valid_values}
def __call__(self, value, field_name):
if self.case_sensitive:
ret = self.valid_values.get(value)
else:
ret = self.valid_values.get(value.lower())
if ret is None:
raise ConfigurationError(
"{}={} is not in the list of valid values: {}".format(
field_name,
value,
list(self.valid_values.values()),
),
field_name,
)
return ret
def _log_level_callback(dict_key, old_value, new_value, config_instance) -> None:
elasticapm_logger = logging.getLogger("elasticapm")
elasticapm_logger.setLevel(log_levels_map.get(new_value, 100))
global logfile_set_up
if not logfile_set_up and config_instance.log_file:
logfile_set_up = True
filehandler = logging.handlers.RotatingFileHandler(
config_instance.log_file, maxBytes=config_instance.log_file_size, backupCount=1
)
try:
import ecs_logging
filehandler.setFormatter(ecs_logging.StdlibFormatter())
except ImportError:
pass
elasticapm_logger.addHandler(filehandler)
def _log_ecs_reformatting_callback(dict_key, old_value, new_value, config_instance) -> None:
"""
If ecs_logging is installed and log_ecs_reformatting is set to "override", we should
set the ecs_logging.StdlibFormatter as the formatted for every handler in
the root logger, and set the default processor for structlog to the
ecs_logging.StructlogFormatter.
"""
if new_value.lower() == "override":
try:
import ecs_logging
except ImportError:
return
# Stdlib
root_logger = logging.getLogger()
formatter = ecs_logging.StdlibFormatter()
for handler in root_logger.handlers:
handler.setFormatter(formatter)
# Structlog
try:
import structlog
structlog.configure(processors=[ecs_logging.StructlogFormatter()])
except ImportError:
pass
class _ConfigBase(object):
_NO_VALUE = object() # sentinel object
def __init__(self, config_dict=None, env_dict=None, inline_dict=None, copy=False) -> None:
"""
config_dict
Configuration dict as is common for frameworks such as flask and django.
Keys match the _ConfigValue.dict_key (usually all caps)
env_dict
Environment variables dict. Keys match the _ConfigValue.env_key
(usually "ELASTIC_APM_" + dict_key)
inline_dict
Any config passed in as kwargs to the Client object. Typically
the keys match the names of the _ConfigValue variables in the Config
object.
copy
Whether this object is being created to copy an existing Config
object. If True, don't run the initial `update` (which would call
callbacks if present)
"""
self._values = {}
self._errors = {}
self._dict_key_lookup = {}
self.callbacks_queue = []
for config_value in self.__class__.__dict__.values():
if not isinstance(config_value, _ConfigValue):
continue
self._dict_key_lookup[config_value.dict_key] = config_value
if not copy:
self.update(config_dict, env_dict, inline_dict, initial=True)
def update(self, config_dict=None, env_dict=None, inline_dict=None, initial=False) -> None:
if config_dict is None:
config_dict = {}
if env_dict is None:
env_dict = os.environ
if inline_dict is None:
inline_dict = {}
for field, config_value in self.__class__.__dict__.items():
if not isinstance(config_value, _ConfigValue):
continue
new_value = self._NO_VALUE
# first check environment
if config_value.env_key and config_value.env_key in env_dict:
new_value = env_dict[config_value.env_key]
# check the inline config
elif field in inline_dict:
new_value = inline_dict[field]
# finally, check config dictionary
elif config_value.dict_key in config_dict:
new_value = config_dict[config_value.dict_key]
# only set if new_value changed. We'll fall back to the field default if not.
if new_value is not self._NO_VALUE:
try:
setattr(self, field, new_value)
except ConfigurationError as e:
self._errors[e.field_name] = str(e)
# handle initial callbacks
if (
initial
and config_value.callbacks_on_default
and getattr(self, field) is not None
and getattr(self, field) == config_value.default
):
self.callbacks_queue.append((config_value.dict_key, self._NO_VALUE, config_value.default))
# if a field has not been provided by any config source, we have to check separately if it is required
if config_value.required and getattr(self, field) is None:
self._errors[config_value.dict_key] = "Configuration error: value for {} is required.".format(
config_value.dict_key
)
self.call_pending_callbacks()
def call_pending_callbacks(self) -> None:
"""
Call callbacks for config options matching list of tuples:
(dict_key, old_value, new_value)
"""
for dict_key, old_value, new_value in self.callbacks_queue:
self._dict_key_lookup[dict_key].call_callbacks(old_value, new_value, self)
self.callbacks_queue = []
@property
def values(self):
return self._values
@values.setter
def values(self, values) -> None:
self._values = values
@property
def errors(self):
return self._errors
def copy(self):
c = self.__class__(copy=True)
c._errors = {}
c.values = self.values.copy()
return c
class WrapperConfig(_ConfigBase):
enabled = _BoolConfigValue("ENABLED", default=True)
instrument = _BoolConfigValue("INSTRUMENT", default=True)
class Config(_ConfigBase):
service_name = _ConfigValue(
"SERVICE_NAME",
validators=[RegexValidator("^[a-zA-Z0-9 _-]+$")],
default="unknown-python-service",
required=True,
)
service_node_name = _ConfigValue("SERVICE_NODE_NAME")
environment = _ConfigValue("ENVIRONMENT")
transport_json_serializer = _ConfigValue("TRANSPORT_JSON_SERIALIZER")
secret_token = _ConfigValue("SECRET_TOKEN")
api_key = _ConfigValue("API_KEY")
debug = _BoolConfigValue("DEBUG", default=False)
server_url = _ConfigValue("SERVER_URL", default="http://127.0.0.1:8200", required=True)
server_cert = _ConfigValue("SERVER_CERT", validators=[FileIsReadableValidator()])
server_ca_cert_file = _ConfigValue("SERVER_CA_CERT_FILE", validators=[FileIsReadableValidator()])
verify_server_cert = _BoolConfigValue(
"VERIFY_SERVER_CERT", default=True, validators=[SupportedValueInFipsModeValidator(supported_value=True)]
)
use_certifi = _BoolConfigValue("USE_CERTIFI", default=True)
include_paths = _ListConfigValue("INCLUDE_PATHS")
exclude_paths = _ListConfigValue("EXCLUDE_PATHS", default=compat.get_default_library_patters())
filter_exception_types = _ListConfigValue("FILTER_EXCEPTION_TYPES")
server_timeout = _ConfigValue(
"SERVER_TIMEOUT",
type=float,
validators=[
UnitValidator(r"^((?:-)?\d+)(ms|s|m)?$", r"\d+(ms|s|m)", {"ms": 0.001, "s": 1, "m": 60, None: 1000})
],
default=5,
)
hostname = _ConfigValue("HOSTNAME", default=None)
auto_log_stacks = _BoolConfigValue("AUTO_LOG_STACKS", default=True)
transport_class = _ConfigValue("TRANSPORT_CLASS", default="elasticapm.transport.http.Transport", required=True)
processors = _ListConfigValue(
"PROCESSORS",
default=[
"elasticapm.processors.sanitize_stacktrace_locals",
"elasticapm.processors.sanitize_http_request_cookies",
"elasticapm.processors.sanitize_http_response_cookies",
"elasticapm.processors.sanitize_http_headers",
"elasticapm.processors.sanitize_http_wsgi_env",
"elasticapm.processors.sanitize_http_request_body",
],
)
sanitize_field_names = _ListConfigValue(
"SANITIZE_FIELD_NAMES", type=starmatch_to_regex, default=BASE_SANITIZE_FIELD_NAMES
)
metrics_sets = _ListConfigValue(
"METRICS_SETS",
default=[
"elasticapm.metrics.sets.cpu.CPUMetricSet",
],
)
metrics_interval = _DurationConfigValue(
"METRICS_INTERVAL",
validators=[ExcludeRangeValidator(0.001, 0.999, "{range_start} - {range_end} s")],
default=timedelta(seconds=30),
)
breakdown_metrics = _BoolConfigValue("BREAKDOWN_METRICS", default=True)
prometheus_metrics = _BoolConfigValue("PROMETHEUS_METRICS", default=False)
prometheus_metrics_prefix = _ConfigValue("PROMETHEUS_METRICS_PREFIX", default="prometheus.metrics.")
disable_metrics = _ListConfigValue("DISABLE_METRICS", type=starmatch_to_regex, default=[])
central_config = _BoolConfigValue("CENTRAL_CONFIG", default=True)
api_request_size = _ConfigValue("API_REQUEST_SIZE", type=int, validators=[size_validator], default=768 * 1024)
api_request_time = _DurationConfigValue("API_REQUEST_TIME", default=timedelta(seconds=10))
transaction_sample_rate = _ConfigValue(
"TRANSACTION_SAMPLE_RATE", type=float, validators=[PrecisionValidator(4, 0.0001)], default=1.0
)
transaction_max_spans = _ConfigValue("TRANSACTION_MAX_SPANS", type=int, default=500)
stack_trace_limit = _ConfigValue("STACK_TRACE_LIMIT", type=int, default=50)
span_frames_min_duration = _DurationConfigValue(
"SPAN_FRAMES_MIN_DURATION", default=timedelta(seconds=0.005), unitless_factor=0.001
)
span_stack_trace_min_duration = _DurationConfigValue(
"SPAN_STACK_TRACE_MIN_DURATION", default=timedelta(seconds=0.005), unitless_factor=0.001
)
span_compression_enabled = _BoolConfigValue("SPAN_COMPRESSION_ENABLED", default=True)
span_compression_exact_match_max_duration = _DurationConfigValue(
"SPAN_COMPRESSION_EXACT_MATCH_MAX_DURATION",
default=timedelta(seconds=0.05),
)
span_compression_same_kind_max_duration = _DurationConfigValue(
"SPAN_COMPRESSION_SAME_KIND_MAX_DURATION",
default=timedelta(seconds=0),
)
exit_span_min_duration = _DurationConfigValue(
"EXIT_SPAN_MIN_DURATION",
default=timedelta(seconds=0),
)
collect_local_variables = _ConfigValue("COLLECT_LOCAL_VARIABLES", default="errors")
source_lines_error_app_frames = _ConfigValue("SOURCE_LINES_ERROR_APP_FRAMES", type=int, default=5)
source_lines_error_library_frames = _ConfigValue("SOURCE_LINES_ERROR_LIBRARY_FRAMES", type=int, default=5)
source_lines_span_app_frames = _ConfigValue("SOURCE_LINES_SPAN_APP_FRAMES", type=int, default=0)
source_lines_span_library_frames = _ConfigValue("SOURCE_LINES_SPAN_LIBRARY_FRAMES", type=int, default=0)
local_var_max_length = _ConfigValue("LOCAL_VAR_MAX_LENGTH", type=int, default=200)
local_var_list_max_length = _ConfigValue("LOCAL_VAR_LIST_MAX_LENGTH", type=int, default=10)
local_var_dict_max_length = _ConfigValue("LOCAL_VAR_DICT_MAX_LENGTH", type=int, default=10)
capture_body = _ConfigValue(
"CAPTURE_BODY",
default="off",
validators=[lambda val, _: {"errors": "error", "transactions": "transaction"}.get(val, val)],
)
async_mode = _BoolConfigValue("ASYNC_MODE", default=True)
instrument_django_middleware = _BoolConfigValue("INSTRUMENT_DJANGO_MIDDLEWARE", default=True)
autoinsert_django_middleware = _BoolConfigValue("AUTOINSERT_DJANGO_MIDDLEWARE", default=True)
transactions_ignore_patterns = _ListConfigValue("TRANSACTIONS_IGNORE_PATTERNS", default=[])
transaction_ignore_urls = _ListConfigValue("TRANSACTION_IGNORE_URLS", type=starmatch_to_regex, default=[])
ignore_message_queues = _ListConfigValue("IGNORE_MESSAGE_QUEUES", type=starmatch_to_regex, default=[])
service_version = _ConfigValue("SERVICE_VERSION")
framework_name = _ConfigValue("FRAMEWORK_NAME")
framework_version = _ConfigValue("FRAMEWORK_VERSION")
global_labels = _DictConfigValue("GLOBAL_LABELS")
disable_send = _BoolConfigValue("DISABLE_SEND", default=False)
enabled = _BoolConfigValue("ENABLED", default=True)
recording = _BoolConfigValue("RECORDING", default=True)
instrument = _BoolConfigValue("INSTRUMENT", default=True)
enable_distributed_tracing = _BoolConfigValue("ENABLE_DISTRIBUTED_TRACING", default=True)
capture_headers = _BoolConfigValue("CAPTURE_HEADERS", default=True)
django_transaction_name_from_route = _BoolConfigValue("DJANGO_TRANSACTION_NAME_FROM_ROUTE", default=False)
disable_log_record_factory = _BoolConfigValue("DISABLE_LOG_RECORD_FACTORY", default=False)
use_elastic_traceparent_header = _BoolConfigValue("USE_ELASTIC_TRACEPARENT_HEADER", default=True)
use_elastic_excepthook = _BoolConfigValue("USE_ELASTIC_EXCEPTHOOK", default=False)
cloud_provider = _ConfigValue("CLOUD_PROVIDER", default=True)
log_level = _ConfigValue(
"LOG_LEVEL",
validators=[EnumerationValidator(["trace", "debug", "info", "warning", "warn", "error", "critical", "off"])],
callbacks=[_log_level_callback],
)
log_file = _ConfigValue("LOG_FILE", default="")
log_file_size = _ConfigValue("LOG_FILE_SIZE", validators=[size_validator], type=int, default=50 * 1024 * 1024)
log_ecs_reformatting = _ConfigValue(
"LOG_ECS_REFORMATTING",
validators=[EnumerationValidator(["off", "override"])],
callbacks=[_log_ecs_reformatting_callback],
default="off",
)
trace_continuation_strategy = _ConfigValue(
"TRACE_CONTINUATION_STRATEGY",
validators=[
EnumerationValidator(
[
TRACE_CONTINUATION_STRATEGY.CONTINUE,
TRACE_CONTINUATION_STRATEGY.RESTART,
TRACE_CONTINUATION_STRATEGY.RESTART_EXTERNAL,
]
)
],
default=TRACE_CONTINUATION_STRATEGY.CONTINUE,
)
include_process_args = _BoolConfigValue("INCLUDE_PROCESS_ARGS", default=False)
@property
def is_recording(self):
if not self.enabled:
return False
else:
return self.recording
class VersionedConfig(ThreadManager):
"""
A thin layer around Config that provides versioning
"""
__slots__ = (
"_config",
"_version",
"_first_config",
"_first_version",
"_lock",
"transport",
"_update_thread",
"pid",
"start_stop_order",
)
def __init__(self, config_object, version, transport=None) -> None:
"""
Create a new VersionedConfig with an initial Config object
:param config_object: the initial Config object
:param version: a version identifier for the configuration
"""
self._config = self._first_config = config_object
self._version = self._first_version = version
self.transport = transport
self._lock = threading.Lock()
self._update_thread = None
super(VersionedConfig, self).__init__()
def update(self, version: str, **config):
"""
Update the configuration version
:param version: version identifier for the new configuration
:param config: a key/value map of new configuration
:return: configuration errors, if any
"""
new_config = self._config.copy()
# pass an empty env dict to ensure the environment doesn't get precedence
new_config.update(inline_dict=config, env_dict={})
if not new_config.errors:
with self._lock:
self._version = version
self._config = new_config
else:
return new_config.errors
def reset(self) -> None:
"""
Reset state to the original configuration
Note that because ConfigurationValues can have callbacks, we need to
note any differences between the original configuration and the most
recent configuration and run any callbacks that might exist for those
values.
"""
callbacks = []
for key in self._config.values.keys():
if key in self._first_config.values and self._config.values[key] != self._first_config.values[key]:
callbacks.append((key, self._config.values[key], self._first_config.values[key]))
with self._lock:
self._version = self._first_version
self._config = self._first_config
self._config.callbacks_queue.extend(callbacks)
self._config.call_pending_callbacks()
@property
def changed(self) -> bool:
return self._config != self._first_config
def __getattr__(self, item):
return getattr(self._config, item)
def __setattr__(self, name, value) -> None:
if name not in self.__slots__:
setattr(self._config, name, value)
else:
super(VersionedConfig, self).__setattr__(name, value)
@property
def config_version(self):
return self._version
def update_config(self):
if not self.transport:
logger.warning("No transport set for config updates, skipping")
return
logger.debug("Checking for new config...")
keys = {"service": {"name": self.service_name}}
if self.environment:
keys["service"]["environment"] = self.environment
new_version, new_config, next_run = self.transport.get_config(self.config_version, keys)
if new_version and new_config:
errors = self.update(new_version, **new_config)
if errors:
logger.error("Error applying new configuration: %s", repr(errors))
else:
logger.info(
"Applied new remote configuration: %s",
"; ".join("%s=%s" % (str(k), str(v)) for k, v in new_config.items()),
)
elif new_version == self.config_version:
logger.debug("Remote config unchanged")
elif not new_config and self.changed:
logger.debug("Remote config disappeared, resetting to original")
self.reset()
return next_run
def start_thread(self, pid=None) -> None:
self._update_thread = IntervalTimer(
self.update_config, 1, "eapm conf updater", daemon=True, evaluate_function_interval=True
)
self._update_thread.start()
super(VersionedConfig, self).start_thread(pid=pid)
def stop_thread(self) -> None:
if self._update_thread:
self._update_thread.cancel()
self._update_thread = None
def setup_logging(handler):
"""
Configures logging to pipe to Elastic APM.
For a typical Python install:
>>> from elasticapm.handlers.logging import LoggingHandler
>>> client = ElasticAPM(...)
>>> setup_logging(LoggingHandler(client))
Within Django:
>>> from elasticapm.contrib.django.handlers import LoggingHandler
>>> setup_logging(LoggingHandler())
Returns a boolean based on if logging was configured or not.
"""
# TODO We should probably revisit this. Does it make more sense as
# a method within the Client class? The Client object could easily
# pass itself into LoggingHandler and we could eliminate args altogether.
logger = logging.getLogger()
if handler.__class__ in map(type, logger.handlers):
return False
logger.addHandler(handler)
return True