elasticapm/utils/stacks.py (213 lines of code) (raw):

# BSD 3-Clause License # # Copyright (c) 2012, the Sentry Team, see AUTHORS for more details # Copyright (c) 2019, Elasticsearch BV # All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are met: # # * Redistributions of source code must retain the above copyright notice, this # list of conditions and the following disclaimer. # # * Redistributions in binary form must reproduce the above copyright notice, # this list of conditions and the following disclaimer in the documentation # and/or other materials provided with the distribution. # # * Neither the name of the copyright holder nor the names of its # contributors may be used to endorse or promote products derived from # this software without specific prior written permission. # # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" # AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE # IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE # DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE # FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL # DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR # SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER # CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, # OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE import fnmatch import inspect import itertools import os import re import sys from functools import lru_cache from elasticapm.utils.encoding import transform _coding_re = re.compile(r"coding[:=]\s*([-\w.]+)") @lru_cache(512) def get_lines_from_file(filename, lineno, context_lines, loader=None, module_name=None): """ Returns context_lines before and after lineno from file. Returns (pre_context_lineno, pre_context, context_line, post_context). """ lineno = lineno - 1 lower_bound = max(0, lineno - context_lines) upper_bound = lineno + context_lines source = None if loader is not None and hasattr(loader, "get_source"): result = get_source_lines_from_loader(loader, module_name, lineno, lower_bound, upper_bound) if result is not None: return result if source is None: try: with open(filename, "rb") as file_obj: encoding = "utf8" # try to find encoding of source file by "coding" header # if none is found, utf8 is used as a fallback for line in itertools.islice(file_obj, 0, 2): match = _coding_re.search(line.decode("utf8")) if match: encoding = match.group(1) break file_obj.seek(0) lines = [ str(line, encoding, "replace") for line in itertools.islice(file_obj, lower_bound, upper_bound + 1) ] offset = lineno - lower_bound return ( [line.strip("\r\n") for line in lines[0:offset]], lines[offset].strip("\r\n"), [line.strip("\r\n") for line in lines[offset + 1 :]] if len(lines) > offset else [], ) except (OSError, IOError, IndexError): pass return None, None, None def get_source_lines_from_loader(loader, module_name, lineno, lower_bound, upper_bound): try: source = loader.get_source(module_name) except ImportError: # ImportError: Loader for module cProfile cannot handle module __main__ return None if source is not None: source = source.splitlines() else: return None, None, None try: pre_context = [line.strip("\r\n") for line in source[lower_bound:lineno]] context_line = source[lineno].strip("\r\n") post_context = [line.strip("\r\n") for line in source[(lineno + 1) : upper_bound + 1]] except IndexError: # the file may have changed since it was loaded into memory return None, None, None return pre_context, context_line, post_context def get_culprit(frames, include_paths=None, exclude_paths=None): # We iterate through each frame looking for a deterministic culprit # When one is found, we mark it as last "best guess" (best_guess) and then # check it against ``exclude_paths``. If it isnt listed, then we # use this option. If nothing is found, we use the "best guess". if include_paths is None: include_paths = [] if exclude_paths is None: exclude_paths = [] best_guess = None culprit = None for frame in frames: try: culprit = ".".join((f or "<unknown>" for f in [frame.get("module"), frame.get("function")])) except KeyError: continue if any((culprit.startswith(k) for k in include_paths)): if not (best_guess and any((culprit.startswith(k) for k in exclude_paths))): best_guess = culprit elif best_guess: break # Return either the best guess or the last frames call return best_guess or culprit def _getitem_from_frame(f_locals, key, default=None): """ f_locals is not guaranteed to have .get(), but it will always support __getitem__. Even if it doesnt, we return ``default``. """ try: return f_locals[key] except Exception: return default def to_dict(dictish): """ Given something that closely resembles a dictionary, we attempt to coerce it into a propery dictionary. """ if hasattr(dictish, "iterkeys"): m = dictish.iterkeys elif hasattr(dictish, "keys"): m = dictish.keys else: raise ValueError(dictish) return dict((k, dictish[k]) for k in m()) def iter_traceback_frames(tb, config=None): """ Given a traceback object, it will iterate over all frames that do not contain the ``__traceback_hide__`` local variable. """ max_frames = config.stack_trace_limit if config else -1 if not max_frames: return frames = [] while tb: # support for __traceback_hide__ which is used by a few libraries # to hide internal frames. frame = tb.tb_frame f_locals = getattr(frame, "f_locals", None) if f_locals is None or not _getitem_from_frame(f_locals, "__traceback_hide__"): frames.append((frame, getattr(tb, "tb_lineno", None))) tb = tb.tb_next if max_frames != -1: frames = frames[-max_frames:] for frame in frames: yield frame def iter_stack_frames(frames=None, start_frame=None, skip=0, skip_top_modules=(), config=None): """ Given an optional list of frames (defaults to current stack), iterates over all frames that do not contain the ``__traceback_hide__`` local variable. Frames can be skipped by either providing a number, or a tuple of module names. If the module of a frame matches one of the names (using `.startswith`, that frame will be skipped. This matching will only be done until the first frame doesn't match. This is useful to filter out frames that are caused by frame collection itself. :param frames: a list of frames, or None :param start_frame: a Frame object or None :param skip: number of frames to skip from the beginning :param skip_top_modules: tuple of strings :param config: agent configuration """ if not frames: frame = start_frame if start_frame is not None else inspect.currentframe().f_back frames = _walk_stack(frame) max_frames = config.stack_trace_limit if config else -1 stop_ignoring = False frames_count = 0 # can't use i, as we don't want to count skipped and ignored frames for i, frame in enumerate(frames): if max_frames != -1 and frames_count == max_frames: break if i < skip: continue f_globals = getattr(frame, "f_globals", {}) if not stop_ignoring and f_globals.get("__name__", "").startswith(skip_top_modules): continue stop_ignoring = True f_locals = getattr(frame, "f_locals", {}) if not _getitem_from_frame(f_locals, "__traceback_hide__"): frames_count += 1 yield frame, frame.f_lineno def get_frame_info( frame, lineno, with_locals=True, library_frame_context_lines=None, in_app_frame_context_lines=None, include_paths_re=None, exclude_paths_re=None, locals_processor_func=None, ): # Support hidden frames f_locals = getattr(frame, "f_locals", {}) if _getitem_from_frame(f_locals, "__traceback_hide__"): return None f_globals = getattr(frame, "f_globals", {}) loader = f_globals.get("__loader__") module_name = f_globals.get("__name__") f_code = getattr(frame, "f_code", None) if f_code: abs_path = frame.f_code.co_filename function = frame.f_code.co_name else: abs_path = None function = None # Try to pull a relative file path # This changes /foo/site-packages/baz/bar.py into baz/bar.py try: base_filename = sys.modules[module_name.split(".", 1)[0]].__file__ filename = abs_path.split(base_filename.rsplit(os.path.sep, 2)[0], 1)[-1].lstrip(os.path.sep) except Exception: filename = abs_path if not filename: filename = abs_path frame_result = { "abs_path": abs_path, "filename": filename, "module": module_name, "function": function, "lineno": lineno, "library_frame": is_library_frame(abs_path, include_paths_re, exclude_paths_re), } context_lines = library_frame_context_lines if frame_result["library_frame"] else in_app_frame_context_lines if context_lines and lineno is not None and abs_path: # context_metadata will be processed by elasticapm.processors.add_context_lines_to_frames. # This ensures that blocking operations (reading from source files) happens on the background # processing thread. frame_result["context_metadata"] = (abs_path, lineno, int(context_lines / 2), loader, module_name) if with_locals: if f_locals is not None and not isinstance(f_locals, dict): # XXX: Genshi (and maybe others) have broken implementations of # f_locals that are not actually dictionaries try: f_locals = to_dict(f_locals) except Exception: f_locals = "<invalid local scope>" if locals_processor_func: f_locals = {varname: locals_processor_func(var) for varname, var in f_locals.items()} frame_result["vars"] = transform(f_locals) return frame_result def get_stack_info( frames, with_locals=True, library_frame_context_lines=None, in_app_frame_context_lines=None, include_paths_re=None, exclude_paths_re=None, locals_processor_func=None, ): """ Given a list of frames, returns a list of stack information dictionary objects that are JSON-ready. We have to be careful here as certain implementations of the _Frame class do not contain the necessary data to lookup all of the information we want. :param frames: a list of (Frame, lineno) tuples :param with_locals: boolean to indicate if local variables should be collected :param include_paths_re: a regex to determine if a frame is not a library frame :param exclude_paths_re: a regex to exclude frames from not being library frames :param locals_processor_func: a function to call on all local variables :return: """ results = [] for frame, lineno in frames: result = get_frame_info( frame, lineno, library_frame_context_lines=library_frame_context_lines, in_app_frame_context_lines=in_app_frame_context_lines, with_locals=with_locals, include_paths_re=include_paths_re, exclude_paths_re=exclude_paths_re, locals_processor_func=locals_processor_func, ) if result: results.append(result) return results def _walk_stack(frame): while frame: yield frame frame = frame.f_back @lru_cache(512) def is_library_frame(abs_file_path, include_paths_re, exclude_paths_re): if not abs_file_path: return True if include_paths_re and include_paths_re.match(abs_file_path): # frame is in-app, return False return False elif exclude_paths_re and exclude_paths_re.match(abs_file_path): return True # if neither excluded nor included, assume it's an in-app frame return False def get_path_regex(paths): """ compiles a list of path globs into a single pattern that matches any of the given paths :param paths: a list of strings representing fnmatch path globs :return: a compiled regex """ return re.compile("|".join(map(fnmatch.translate, paths)))