mirror of
https://github.com/python/cpython
synced 2024-11-02 10:13:47 +00:00
04fabe22dd
Co-authored-by: Irit Katriel <1055913+iritkatriel@users.noreply.github.com>
1595 lines
63 KiB
Python
1595 lines
63 KiB
Python
"""Extract, format and print information about Python stack traces."""
|
|
|
|
import os
|
|
import io
|
|
import collections.abc
|
|
import itertools
|
|
import linecache
|
|
import sys
|
|
import textwrap
|
|
import warnings
|
|
from contextlib import suppress
|
|
|
|
__all__ = ['extract_stack', 'extract_tb', 'format_exception',
|
|
'format_exception_only', 'format_list', 'format_stack',
|
|
'format_tb', 'print_exc', 'format_exc', 'print_exception',
|
|
'print_last', 'print_stack', 'print_tb', 'clear_frames',
|
|
'FrameSummary', 'StackSummary', 'TracebackException',
|
|
'walk_stack', 'walk_tb']
|
|
|
|
#
|
|
# Formatting and printing lists of traceback lines.
|
|
#
|
|
|
|
_COLORIZE = True
|
|
|
|
def print_list(extracted_list, file=None):
|
|
"""Print the list of tuples as returned by extract_tb() or
|
|
extract_stack() as a formatted stack trace to the given file."""
|
|
if file is None:
|
|
file = sys.stderr
|
|
for item in StackSummary.from_list(extracted_list).format():
|
|
print(item, file=file, end="")
|
|
|
|
def format_list(extracted_list):
|
|
"""Format a list of tuples or FrameSummary objects for printing.
|
|
|
|
Given a list of tuples or FrameSummary objects as returned by
|
|
extract_tb() or extract_stack(), return a list of strings ready
|
|
for printing.
|
|
|
|
Each string in the resulting list corresponds to the item with the
|
|
same index in the argument list. Each string ends in a newline;
|
|
the strings may contain internal newlines as well, for those items
|
|
whose source text line is not None.
|
|
"""
|
|
return StackSummary.from_list(extracted_list).format()
|
|
|
|
#
|
|
# Printing and Extracting Tracebacks.
|
|
#
|
|
|
|
def print_tb(tb, limit=None, file=None):
|
|
"""Print up to 'limit' stack trace entries from the traceback 'tb'.
|
|
|
|
If 'limit' is omitted or None, all entries are printed. If 'file'
|
|
is omitted or None, the output goes to sys.stderr; otherwise
|
|
'file' should be an open file or file-like object with a write()
|
|
method.
|
|
"""
|
|
print_list(extract_tb(tb, limit=limit), file=file)
|
|
|
|
def format_tb(tb, limit=None):
|
|
"""A shorthand for 'format_list(extract_tb(tb, limit))'."""
|
|
return extract_tb(tb, limit=limit).format()
|
|
|
|
def extract_tb(tb, limit=None):
|
|
"""
|
|
Return a StackSummary object representing a list of
|
|
pre-processed entries from traceback.
|
|
|
|
This is useful for alternate formatting of stack traces. If
|
|
'limit' is omitted or None, all entries are extracted. A
|
|
pre-processed stack trace entry is a FrameSummary object
|
|
containing attributes filename, lineno, name, and line
|
|
representing the information that is usually printed for a stack
|
|
trace. The line is a string with leading and trailing
|
|
whitespace stripped; if the source is not available it is None.
|
|
"""
|
|
return StackSummary._extract_from_extended_frame_gen(
|
|
_walk_tb_with_full_positions(tb), limit=limit)
|
|
|
|
#
|
|
# Exception formatting and output.
|
|
#
|
|
|
|
_cause_message = (
|
|
"\nThe above exception was the direct cause "
|
|
"of the following exception:\n\n")
|
|
|
|
_context_message = (
|
|
"\nDuring handling of the above exception, "
|
|
"another exception occurred:\n\n")
|
|
|
|
|
|
class _Sentinel:
|
|
def __repr__(self):
|
|
return "<implicit>"
|
|
|
|
_sentinel = _Sentinel()
|
|
|
|
def _parse_value_tb(exc, value, tb):
|
|
if (value is _sentinel) != (tb is _sentinel):
|
|
raise ValueError("Both or neither of value and tb must be given")
|
|
if value is tb is _sentinel:
|
|
if exc is not None:
|
|
if isinstance(exc, BaseException):
|
|
return exc, exc.__traceback__
|
|
|
|
raise TypeError(f'Exception expected for value, '
|
|
f'{type(exc).__name__} found')
|
|
else:
|
|
return None, None
|
|
return value, tb
|
|
|
|
|
|
def print_exception(exc, /, value=_sentinel, tb=_sentinel, limit=None, \
|
|
file=None, chain=True, **kwargs):
|
|
"""Print exception up to 'limit' stack trace entries from 'tb' to 'file'.
|
|
|
|
This differs from print_tb() in the following ways: (1) if
|
|
traceback is not None, it prints a header "Traceback (most recent
|
|
call last):"; (2) it prints the exception type and value after the
|
|
stack trace; (3) if type is SyntaxError and value has the
|
|
appropriate format, it prints the line where the syntax error
|
|
occurred with a caret on the next line indicating the approximate
|
|
position of the error.
|
|
"""
|
|
colorize = kwargs.get("colorize", False)
|
|
value, tb = _parse_value_tb(exc, value, tb)
|
|
te = TracebackException(type(value), value, tb, limit=limit, compact=True)
|
|
te.print(file=file, chain=chain, colorize=colorize)
|
|
|
|
|
|
BUILTIN_EXCEPTION_LIMIT = object()
|
|
|
|
def _can_colorize():
|
|
if sys.platform == "win32":
|
|
try:
|
|
import nt
|
|
if not nt._supports_virtual_terminal():
|
|
return False
|
|
except (ImportError, AttributeError):
|
|
return False
|
|
|
|
if os.environ.get("PYTHON_COLORS") == "0":
|
|
return False
|
|
if os.environ.get("PYTHON_COLORS") == "1":
|
|
return True
|
|
if "NO_COLOR" in os.environ:
|
|
return False
|
|
if not _COLORIZE:
|
|
return False
|
|
if "FORCE_COLOR" in os.environ:
|
|
return True
|
|
if os.environ.get("TERM") == "dumb":
|
|
return False
|
|
try:
|
|
return os.isatty(sys.stderr.fileno())
|
|
except io.UnsupportedOperation:
|
|
return sys.stderr.isatty()
|
|
|
|
def _print_exception_bltin(exc, /):
|
|
file = sys.stderr if sys.stderr is not None else sys.__stderr__
|
|
colorize = _can_colorize()
|
|
return print_exception(exc, limit=BUILTIN_EXCEPTION_LIMIT, file=file, colorize=colorize)
|
|
|
|
|
|
def format_exception(exc, /, value=_sentinel, tb=_sentinel, limit=None, \
|
|
chain=True):
|
|
"""Format a stack trace and the exception information.
|
|
|
|
The arguments have the same meaning as the corresponding arguments
|
|
to print_exception(). The return value is a list of strings, each
|
|
ending in a newline and some containing internal newlines. When
|
|
these lines are concatenated and printed, exactly the same text is
|
|
printed as does print_exception().
|
|
"""
|
|
value, tb = _parse_value_tb(exc, value, tb)
|
|
te = TracebackException(type(value), value, tb, limit=limit, compact=True)
|
|
return list(te.format(chain=chain))
|
|
|
|
|
|
def format_exception_only(exc, /, value=_sentinel, *, show_group=False):
|
|
"""Format the exception part of a traceback.
|
|
|
|
The return value is a list of strings, each ending in a newline.
|
|
|
|
The list contains the exception's message, which is
|
|
normally a single string; however, for :exc:`SyntaxError` exceptions, it
|
|
contains several lines that (when printed) display detailed information
|
|
about where the syntax error occurred. Following the message, the list
|
|
contains the exception's ``__notes__``.
|
|
|
|
When *show_group* is ``True``, and the exception is an instance of
|
|
:exc:`BaseExceptionGroup`, the nested exceptions are included as
|
|
well, recursively, with indentation relative to their nesting depth.
|
|
"""
|
|
if value is _sentinel:
|
|
value = exc
|
|
te = TracebackException(type(value), value, None, compact=True)
|
|
return list(te.format_exception_only(show_group=show_group))
|
|
|
|
|
|
# -- not official API but folk probably use these two functions.
|
|
|
|
def _format_final_exc_line(etype, value, *, insert_final_newline=True, colorize=False):
|
|
valuestr = _safe_string(value, 'exception')
|
|
end_char = "\n" if insert_final_newline else ""
|
|
if colorize:
|
|
if value is None or not valuestr:
|
|
line = f"{_ANSIColors.BOLD_MAGENTA}{etype}{_ANSIColors.RESET}{end_char}"
|
|
else:
|
|
line = f"{_ANSIColors.BOLD_MAGENTA}{etype}{_ANSIColors.RESET}: {_ANSIColors.MAGENTA}{valuestr}{_ANSIColors.RESET}{end_char}"
|
|
else:
|
|
if value is None or not valuestr:
|
|
line = f"{etype}{end_char}"
|
|
else:
|
|
line = f"{etype}: {valuestr}{end_char}"
|
|
return line
|
|
|
|
def _safe_string(value, what, func=str):
|
|
try:
|
|
return func(value)
|
|
except:
|
|
return f'<{what} {func.__name__}() failed>'
|
|
|
|
# --
|
|
|
|
def print_exc(limit=None, file=None, chain=True):
|
|
"""Shorthand for 'print_exception(sys.exception(), limit, file, chain)'."""
|
|
print_exception(sys.exception(), limit=limit, file=file, chain=chain)
|
|
|
|
def format_exc(limit=None, chain=True):
|
|
"""Like print_exc() but return a string."""
|
|
return "".join(format_exception(sys.exception(), limit=limit, chain=chain))
|
|
|
|
def print_last(limit=None, file=None, chain=True):
|
|
"""This is a shorthand for 'print_exception(sys.last_exc, limit, file, chain)'."""
|
|
if not hasattr(sys, "last_exc") and not hasattr(sys, "last_type"):
|
|
raise ValueError("no last exception")
|
|
|
|
if hasattr(sys, "last_exc"):
|
|
print_exception(sys.last_exc, limit, file, chain)
|
|
else:
|
|
print_exception(sys.last_type, sys.last_value, sys.last_traceback,
|
|
limit, file, chain)
|
|
|
|
|
|
#
|
|
# Printing and Extracting Stacks.
|
|
#
|
|
|
|
def print_stack(f=None, limit=None, file=None):
|
|
"""Print a stack trace from its invocation point.
|
|
|
|
The optional 'f' argument can be used to specify an alternate
|
|
stack frame at which to start. The optional 'limit' and 'file'
|
|
arguments have the same meaning as for print_exception().
|
|
"""
|
|
if f is None:
|
|
f = sys._getframe().f_back
|
|
print_list(extract_stack(f, limit=limit), file=file)
|
|
|
|
|
|
def format_stack(f=None, limit=None):
|
|
"""Shorthand for 'format_list(extract_stack(f, limit))'."""
|
|
if f is None:
|
|
f = sys._getframe().f_back
|
|
return format_list(extract_stack(f, limit=limit))
|
|
|
|
|
|
def extract_stack(f=None, limit=None):
|
|
"""Extract the raw traceback from the current stack frame.
|
|
|
|
The return value has the same format as for extract_tb(). The
|
|
optional 'f' and 'limit' arguments have the same meaning as for
|
|
print_stack(). Each item in the list is a quadruple (filename,
|
|
line number, function name, text), and the entries are in order
|
|
from oldest to newest stack frame.
|
|
"""
|
|
if f is None:
|
|
f = sys._getframe().f_back
|
|
stack = StackSummary.extract(walk_stack(f), limit=limit)
|
|
stack.reverse()
|
|
return stack
|
|
|
|
|
|
def clear_frames(tb):
|
|
"Clear all references to local variables in the frames of a traceback."
|
|
while tb is not None:
|
|
try:
|
|
tb.tb_frame.clear()
|
|
except RuntimeError:
|
|
# Ignore the exception raised if the frame is still executing.
|
|
pass
|
|
tb = tb.tb_next
|
|
|
|
|
|
class FrameSummary:
|
|
"""Information about a single frame from a traceback.
|
|
|
|
- :attr:`filename` The filename for the frame.
|
|
- :attr:`lineno` The line within filename for the frame that was
|
|
active when the frame was captured.
|
|
- :attr:`name` The name of the function or method that was executing
|
|
when the frame was captured.
|
|
- :attr:`line` The text from the linecache module for the
|
|
of code that was running when the frame was captured.
|
|
- :attr:`locals` Either None if locals were not supplied, or a dict
|
|
mapping the name to the repr() of the variable.
|
|
"""
|
|
|
|
__slots__ = ('filename', 'lineno', 'end_lineno', 'colno', 'end_colno',
|
|
'name', '_lines', '_lines_dedented', 'locals')
|
|
|
|
def __init__(self, filename, lineno, name, *, lookup_line=True,
|
|
locals=None, line=None,
|
|
end_lineno=None, colno=None, end_colno=None):
|
|
"""Construct a FrameSummary.
|
|
|
|
:param lookup_line: If True, `linecache` is consulted for the source
|
|
code line. Otherwise, the line will be looked up when first needed.
|
|
:param locals: If supplied the frame locals, which will be captured as
|
|
object representations.
|
|
:param line: If provided, use this instead of looking up the line in
|
|
the linecache.
|
|
"""
|
|
self.filename = filename
|
|
self.lineno = lineno
|
|
self.end_lineno = lineno if end_lineno is None else end_lineno
|
|
self.colno = colno
|
|
self.end_colno = end_colno
|
|
self.name = name
|
|
self._lines = line
|
|
self._lines_dedented = None
|
|
if lookup_line:
|
|
self.line
|
|
self.locals = {k: _safe_string(v, 'local', func=repr)
|
|
for k, v in locals.items()} if locals else None
|
|
|
|
def __eq__(self, other):
|
|
if isinstance(other, FrameSummary):
|
|
return (self.filename == other.filename and
|
|
self.lineno == other.lineno and
|
|
self.name == other.name and
|
|
self.locals == other.locals)
|
|
if isinstance(other, tuple):
|
|
return (self.filename, self.lineno, self.name, self.line) == other
|
|
return NotImplemented
|
|
|
|
def __getitem__(self, pos):
|
|
return (self.filename, self.lineno, self.name, self.line)[pos]
|
|
|
|
def __iter__(self):
|
|
return iter([self.filename, self.lineno, self.name, self.line])
|
|
|
|
def __repr__(self):
|
|
return "<FrameSummary file {filename}, line {lineno} in {name}>".format(
|
|
filename=self.filename, lineno=self.lineno, name=self.name)
|
|
|
|
def __len__(self):
|
|
return 4
|
|
|
|
def _set_lines(self):
|
|
if (
|
|
self._lines is None
|
|
and self.lineno is not None
|
|
and self.end_lineno is not None
|
|
):
|
|
lines = []
|
|
for lineno in range(self.lineno, self.end_lineno + 1):
|
|
# treat errors (empty string) and empty lines (newline) as the same
|
|
lines.append(linecache.getline(self.filename, lineno).rstrip())
|
|
self._lines = "\n".join(lines) + "\n"
|
|
|
|
@property
|
|
def _original_lines(self):
|
|
# Returns the line as-is from the source, without modifying whitespace.
|
|
self._set_lines()
|
|
return self._lines
|
|
|
|
@property
|
|
def _dedented_lines(self):
|
|
# Returns _original_lines, but dedented
|
|
self._set_lines()
|
|
if self._lines_dedented is None and self._lines is not None:
|
|
self._lines_dedented = textwrap.dedent(self._lines)
|
|
return self._lines_dedented
|
|
|
|
@property
|
|
def line(self):
|
|
self._set_lines()
|
|
if self._lines is None:
|
|
return None
|
|
# return only the first line, stripped
|
|
return self._lines.partition("\n")[0].strip()
|
|
|
|
|
|
def walk_stack(f):
|
|
"""Walk a stack yielding the frame and line number for each frame.
|
|
|
|
This will follow f.f_back from the given frame. If no frame is given, the
|
|
current stack is used. Usually used with StackSummary.extract.
|
|
"""
|
|
if f is None:
|
|
f = sys._getframe().f_back.f_back.f_back.f_back
|
|
while f is not None:
|
|
yield f, f.f_lineno
|
|
f = f.f_back
|
|
|
|
|
|
def walk_tb(tb):
|
|
"""Walk a traceback yielding the frame and line number for each frame.
|
|
|
|
This will follow tb.tb_next (and thus is in the opposite order to
|
|
walk_stack). Usually used with StackSummary.extract.
|
|
"""
|
|
while tb is not None:
|
|
yield tb.tb_frame, tb.tb_lineno
|
|
tb = tb.tb_next
|
|
|
|
|
|
def _walk_tb_with_full_positions(tb):
|
|
# Internal version of walk_tb that yields full code positions including
|
|
# end line and column information.
|
|
while tb is not None:
|
|
positions = _get_code_position(tb.tb_frame.f_code, tb.tb_lasti)
|
|
# Yield tb_lineno when co_positions does not have a line number to
|
|
# maintain behavior with walk_tb.
|
|
if positions[0] is None:
|
|
yield tb.tb_frame, (tb.tb_lineno, ) + positions[1:]
|
|
else:
|
|
yield tb.tb_frame, positions
|
|
tb = tb.tb_next
|
|
|
|
|
|
def _get_code_position(code, instruction_index):
|
|
if instruction_index < 0:
|
|
return (None, None, None, None)
|
|
positions_gen = code.co_positions()
|
|
return next(itertools.islice(positions_gen, instruction_index // 2, None))
|
|
|
|
|
|
_RECURSIVE_CUTOFF = 3 # Also hardcoded in traceback.c.
|
|
|
|
class _ANSIColors:
|
|
RED = '\x1b[31m'
|
|
BOLD_RED = '\x1b[1;31m'
|
|
MAGENTA = '\x1b[35m'
|
|
BOLD_MAGENTA = '\x1b[1;35m'
|
|
GREY = '\x1b[90m'
|
|
RESET = '\x1b[0m'
|
|
|
|
class StackSummary(list):
|
|
"""A list of FrameSummary objects, representing a stack of frames."""
|
|
|
|
@classmethod
|
|
def extract(klass, frame_gen, *, limit=None, lookup_lines=True,
|
|
capture_locals=False):
|
|
"""Create a StackSummary from a traceback or stack object.
|
|
|
|
:param frame_gen: A generator that yields (frame, lineno) tuples
|
|
whose summaries are to be included in the stack.
|
|
:param limit: None to include all frames or the number of frames to
|
|
include.
|
|
:param lookup_lines: If True, lookup lines for each frame immediately,
|
|
otherwise lookup is deferred until the frame is rendered.
|
|
:param capture_locals: If True, the local variables from each frame will
|
|
be captured as object representations into the FrameSummary.
|
|
"""
|
|
def extended_frame_gen():
|
|
for f, lineno in frame_gen:
|
|
yield f, (lineno, None, None, None)
|
|
|
|
return klass._extract_from_extended_frame_gen(
|
|
extended_frame_gen(), limit=limit, lookup_lines=lookup_lines,
|
|
capture_locals=capture_locals)
|
|
|
|
@classmethod
|
|
def _extract_from_extended_frame_gen(klass, frame_gen, *, limit=None,
|
|
lookup_lines=True, capture_locals=False):
|
|
# Same as extract but operates on a frame generator that yields
|
|
# (frame, (lineno, end_lineno, colno, end_colno)) in the stack.
|
|
# Only lineno is required, the remaining fields can be None if the
|
|
# information is not available.
|
|
builtin_limit = limit is BUILTIN_EXCEPTION_LIMIT
|
|
if limit is None or builtin_limit:
|
|
limit = getattr(sys, 'tracebacklimit', None)
|
|
if limit is not None and limit < 0:
|
|
limit = 0
|
|
if limit is not None:
|
|
if builtin_limit:
|
|
frame_gen = tuple(frame_gen)
|
|
frame_gen = frame_gen[len(frame_gen) - limit:]
|
|
elif limit >= 0:
|
|
frame_gen = itertools.islice(frame_gen, limit)
|
|
else:
|
|
frame_gen = collections.deque(frame_gen, maxlen=-limit)
|
|
|
|
result = klass()
|
|
fnames = set()
|
|
for f, (lineno, end_lineno, colno, end_colno) in frame_gen:
|
|
co = f.f_code
|
|
filename = co.co_filename
|
|
name = co.co_name
|
|
fnames.add(filename)
|
|
linecache.lazycache(filename, f.f_globals)
|
|
# Must defer line lookups until we have called checkcache.
|
|
if capture_locals:
|
|
f_locals = f.f_locals
|
|
else:
|
|
f_locals = None
|
|
result.append(FrameSummary(
|
|
filename, lineno, name, lookup_line=False, locals=f_locals,
|
|
end_lineno=end_lineno, colno=colno, end_colno=end_colno))
|
|
for filename in fnames:
|
|
linecache.checkcache(filename)
|
|
|
|
# If immediate lookup was desired, trigger lookups now.
|
|
if lookup_lines:
|
|
for f in result:
|
|
f.line
|
|
return result
|
|
|
|
@classmethod
|
|
def from_list(klass, a_list):
|
|
"""
|
|
Create a StackSummary object from a supplied list of
|
|
FrameSummary objects or old-style list of tuples.
|
|
"""
|
|
# While doing a fast-path check for isinstance(a_list, StackSummary) is
|
|
# appealing, idlelib.run.cleanup_traceback and other similar code may
|
|
# break this by making arbitrary frames plain tuples, so we need to
|
|
# check on a frame by frame basis.
|
|
result = StackSummary()
|
|
for frame in a_list:
|
|
if isinstance(frame, FrameSummary):
|
|
result.append(frame)
|
|
else:
|
|
filename, lineno, name, line = frame
|
|
result.append(FrameSummary(filename, lineno, name, line=line))
|
|
return result
|
|
|
|
def format_frame_summary(self, frame_summary, **kwargs):
|
|
"""Format the lines for a single FrameSummary.
|
|
|
|
Returns a string representing one frame involved in the stack. This
|
|
gets called for every frame to be printed in the stack summary.
|
|
"""
|
|
colorize = kwargs.get("colorize", False)
|
|
row = []
|
|
filename = frame_summary.filename
|
|
if frame_summary.filename.startswith("<stdin>-"):
|
|
filename = "<stdin>"
|
|
if colorize:
|
|
row.append(' File {}"{}"{}, line {}{}{}, in {}{}{}\n'.format(
|
|
_ANSIColors.MAGENTA,
|
|
filename,
|
|
_ANSIColors.RESET,
|
|
_ANSIColors.MAGENTA,
|
|
frame_summary.lineno,
|
|
_ANSIColors.RESET,
|
|
_ANSIColors.MAGENTA,
|
|
frame_summary.name,
|
|
_ANSIColors.RESET,
|
|
)
|
|
)
|
|
else:
|
|
row.append(' File "{}", line {}, in {}\n'.format(
|
|
filename, frame_summary.lineno, frame_summary.name))
|
|
if frame_summary._dedented_lines and frame_summary._dedented_lines.strip():
|
|
if (
|
|
frame_summary.colno is None or
|
|
frame_summary.end_colno is None
|
|
):
|
|
# only output first line if column information is missing
|
|
row.append(textwrap.indent(frame_summary.line, ' ') + "\n")
|
|
else:
|
|
# get first and last line
|
|
all_lines_original = frame_summary._original_lines.splitlines()
|
|
first_line = all_lines_original[0]
|
|
# assume all_lines_original has enough lines (since we constructed it)
|
|
last_line = all_lines_original[frame_summary.end_lineno - frame_summary.lineno]
|
|
|
|
# character index of the start/end of the instruction
|
|
start_offset = _byte_offset_to_character_offset(first_line, frame_summary.colno)
|
|
end_offset = _byte_offset_to_character_offset(last_line, frame_summary.end_colno)
|
|
|
|
all_lines = frame_summary._dedented_lines.splitlines()[
|
|
:frame_summary.end_lineno - frame_summary.lineno + 1
|
|
]
|
|
|
|
# adjust start/end offset based on dedent
|
|
dedent_characters = len(first_line) - len(all_lines[0])
|
|
start_offset = max(0, start_offset - dedent_characters)
|
|
end_offset = max(0, end_offset - dedent_characters)
|
|
|
|
# When showing this on a terminal, some of the non-ASCII characters
|
|
# might be rendered as double-width characters, so we need to take
|
|
# that into account when calculating the length of the line.
|
|
dp_start_offset = _display_width(all_lines[0], offset=start_offset)
|
|
dp_end_offset = _display_width(all_lines[-1], offset=end_offset)
|
|
|
|
# get exact code segment corresponding to the instruction
|
|
segment = "\n".join(all_lines)
|
|
segment = segment[start_offset:len(segment) - (len(all_lines[-1]) - end_offset)]
|
|
|
|
# attempt to parse for anchors
|
|
anchors = None
|
|
with suppress(Exception):
|
|
anchors = _extract_caret_anchors_from_line_segment(segment)
|
|
|
|
# only use carets if there are anchors or the carets do not span all lines
|
|
show_carets = False
|
|
if anchors or all_lines[0][:start_offset].lstrip() or all_lines[-1][end_offset:].rstrip():
|
|
show_carets = True
|
|
|
|
result = []
|
|
|
|
# only display first line, last line, and lines around anchor start/end
|
|
significant_lines = {0, len(all_lines) - 1}
|
|
|
|
anchors_left_end_offset = 0
|
|
anchors_right_start_offset = 0
|
|
primary_char = "^"
|
|
secondary_char = "^"
|
|
if anchors:
|
|
anchors_left_end_offset = anchors.left_end_offset
|
|
anchors_right_start_offset = anchors.right_start_offset
|
|
# computed anchor positions do not take start_offset into account,
|
|
# so account for it here
|
|
if anchors.left_end_lineno == 0:
|
|
anchors_left_end_offset += start_offset
|
|
if anchors.right_start_lineno == 0:
|
|
anchors_right_start_offset += start_offset
|
|
|
|
# account for display width
|
|
anchors_left_end_offset = _display_width(
|
|
all_lines[anchors.left_end_lineno], offset=anchors_left_end_offset
|
|
)
|
|
anchors_right_start_offset = _display_width(
|
|
all_lines[anchors.right_start_lineno], offset=anchors_right_start_offset
|
|
)
|
|
|
|
primary_char = anchors.primary_char
|
|
secondary_char = anchors.secondary_char
|
|
significant_lines.update(
|
|
range(anchors.left_end_lineno - 1, anchors.left_end_lineno + 2)
|
|
)
|
|
significant_lines.update(
|
|
range(anchors.right_start_lineno - 1, anchors.right_start_lineno + 2)
|
|
)
|
|
|
|
# remove bad line numbers
|
|
significant_lines.discard(-1)
|
|
significant_lines.discard(len(all_lines))
|
|
|
|
def output_line(lineno):
|
|
"""output all_lines[lineno] along with carets"""
|
|
result.append(all_lines[lineno] + "\n")
|
|
if not show_carets:
|
|
return
|
|
num_spaces = len(all_lines[lineno]) - len(all_lines[lineno].lstrip())
|
|
carets = []
|
|
num_carets = dp_end_offset if lineno == len(all_lines) - 1 else _display_width(all_lines[lineno])
|
|
# compute caret character for each position
|
|
for col in range(num_carets):
|
|
if col < num_spaces or (lineno == 0 and col < dp_start_offset):
|
|
# before first non-ws char of the line, or before start of instruction
|
|
carets.append(' ')
|
|
elif anchors and (
|
|
lineno > anchors.left_end_lineno or
|
|
(lineno == anchors.left_end_lineno and col >= anchors_left_end_offset)
|
|
) and (
|
|
lineno < anchors.right_start_lineno or
|
|
(lineno == anchors.right_start_lineno and col < anchors_right_start_offset)
|
|
):
|
|
# within anchors
|
|
carets.append(secondary_char)
|
|
else:
|
|
carets.append(primary_char)
|
|
if colorize:
|
|
# Replace the previous line with a red version of it only in the parts covered
|
|
# by the carets.
|
|
line = result[-1]
|
|
colorized_line_parts = []
|
|
colorized_carets_parts = []
|
|
|
|
for color, group in itertools.groupby(itertools.zip_longest(line, carets, fillvalue=""), key=lambda x: x[1]):
|
|
caret_group = list(group)
|
|
if color == "^":
|
|
colorized_line_parts.append(_ANSIColors.BOLD_RED + "".join(char for char, _ in caret_group) + _ANSIColors.RESET)
|
|
colorized_carets_parts.append(_ANSIColors.BOLD_RED + "".join(caret for _, caret in caret_group) + _ANSIColors.RESET)
|
|
elif color == "~":
|
|
colorized_line_parts.append(_ANSIColors.RED + "".join(char for char, _ in caret_group) + _ANSIColors.RESET)
|
|
colorized_carets_parts.append(_ANSIColors.RED + "".join(caret for _, caret in caret_group) + _ANSIColors.RESET)
|
|
else:
|
|
colorized_line_parts.append("".join(char for char, _ in caret_group))
|
|
colorized_carets_parts.append("".join(caret for _, caret in caret_group))
|
|
|
|
colorized_line = "".join(colorized_line_parts)
|
|
colorized_carets = "".join(colorized_carets_parts)
|
|
result[-1] = colorized_line
|
|
result.append(colorized_carets + "\n")
|
|
else:
|
|
result.append("".join(carets) + "\n")
|
|
|
|
# display significant lines
|
|
sig_lines_list = sorted(significant_lines)
|
|
for i, lineno in enumerate(sig_lines_list):
|
|
if i:
|
|
linediff = lineno - sig_lines_list[i - 1]
|
|
if linediff == 2:
|
|
# 1 line in between - just output it
|
|
output_line(lineno - 1)
|
|
elif linediff > 2:
|
|
# > 1 line in between - abbreviate
|
|
result.append(f"...<{linediff - 1} lines>...\n")
|
|
output_line(lineno)
|
|
|
|
row.append(
|
|
textwrap.indent(textwrap.dedent("".join(result)), ' ', lambda line: True)
|
|
)
|
|
if frame_summary.locals:
|
|
for name, value in sorted(frame_summary.locals.items()):
|
|
row.append(' {name} = {value}\n'.format(name=name, value=value))
|
|
|
|
return ''.join(row)
|
|
|
|
def format(self, **kwargs):
|
|
"""Format the stack ready for printing.
|
|
|
|
Returns a list of strings ready for printing. Each string in the
|
|
resulting list corresponds to a single frame from the stack.
|
|
Each string ends in a newline; the strings may contain internal
|
|
newlines as well, for those items with source text lines.
|
|
|
|
For long sequences of the same frame and line, the first few
|
|
repetitions are shown, followed by a summary line stating the exact
|
|
number of further repetitions.
|
|
"""
|
|
colorize = kwargs.get("colorize", False)
|
|
result = []
|
|
last_file = None
|
|
last_line = None
|
|
last_name = None
|
|
count = 0
|
|
for frame_summary in self:
|
|
formatted_frame = self.format_frame_summary(frame_summary, colorize=colorize)
|
|
if formatted_frame is None:
|
|
continue
|
|
if (last_file is None or last_file != frame_summary.filename or
|
|
last_line is None or last_line != frame_summary.lineno or
|
|
last_name is None or last_name != frame_summary.name):
|
|
if count > _RECURSIVE_CUTOFF:
|
|
count -= _RECURSIVE_CUTOFF
|
|
result.append(
|
|
f' [Previous line repeated {count} more '
|
|
f'time{"s" if count > 1 else ""}]\n'
|
|
)
|
|
last_file = frame_summary.filename
|
|
last_line = frame_summary.lineno
|
|
last_name = frame_summary.name
|
|
count = 0
|
|
count += 1
|
|
if count > _RECURSIVE_CUTOFF:
|
|
continue
|
|
result.append(formatted_frame)
|
|
|
|
if count > _RECURSIVE_CUTOFF:
|
|
count -= _RECURSIVE_CUTOFF
|
|
result.append(
|
|
f' [Previous line repeated {count} more '
|
|
f'time{"s" if count > 1 else ""}]\n'
|
|
)
|
|
return result
|
|
|
|
|
|
def _byte_offset_to_character_offset(str, offset):
|
|
as_utf8 = str.encode('utf-8')
|
|
return len(as_utf8[:offset].decode("utf-8", errors="replace"))
|
|
|
|
|
|
_Anchors = collections.namedtuple(
|
|
"_Anchors",
|
|
[
|
|
"left_end_lineno",
|
|
"left_end_offset",
|
|
"right_start_lineno",
|
|
"right_start_offset",
|
|
"primary_char",
|
|
"secondary_char",
|
|
],
|
|
defaults=["~", "^"]
|
|
)
|
|
|
|
def _extract_caret_anchors_from_line_segment(segment):
|
|
"""
|
|
Given source code `segment` corresponding to a FrameSummary, determine:
|
|
- for binary ops, the location of the binary op
|
|
- for indexing and function calls, the location of the brackets.
|
|
`segment` is expected to be a valid Python expression.
|
|
"""
|
|
import ast
|
|
|
|
try:
|
|
# Without parentheses, `segment` is parsed as a statement.
|
|
# Binary ops, subscripts, and calls are expressions, so
|
|
# we can wrap them with parentheses to parse them as
|
|
# (possibly multi-line) expressions.
|
|
# e.g. if we try to highlight the addition in
|
|
# x = (
|
|
# a +
|
|
# b
|
|
# )
|
|
# then we would ast.parse
|
|
# a +
|
|
# b
|
|
# which is not a valid statement because of the newline.
|
|
# Adding brackets makes it a valid expression.
|
|
# (
|
|
# a +
|
|
# b
|
|
# )
|
|
# Line locations will be different than the original,
|
|
# which is taken into account later on.
|
|
tree = ast.parse(f"(\n{segment}\n)")
|
|
except SyntaxError:
|
|
return None
|
|
|
|
if len(tree.body) != 1:
|
|
return None
|
|
|
|
lines = segment.splitlines()
|
|
|
|
def normalize(lineno, offset):
|
|
"""Get character index given byte offset"""
|
|
return _byte_offset_to_character_offset(lines[lineno], offset)
|
|
|
|
def next_valid_char(lineno, col):
|
|
"""Gets the next valid character index in `lines`, if
|
|
the current location is not valid. Handles empty lines.
|
|
"""
|
|
while lineno < len(lines) and col >= len(lines[lineno]):
|
|
col = 0
|
|
lineno += 1
|
|
assert lineno < len(lines) and col < len(lines[lineno])
|
|
return lineno, col
|
|
|
|
def increment(lineno, col):
|
|
"""Get the next valid character index in `lines`."""
|
|
col += 1
|
|
lineno, col = next_valid_char(lineno, col)
|
|
return lineno, col
|
|
|
|
def nextline(lineno, col):
|
|
"""Get the next valid character at least on the next line"""
|
|
col = 0
|
|
lineno += 1
|
|
lineno, col = next_valid_char(lineno, col)
|
|
return lineno, col
|
|
|
|
def increment_until(lineno, col, stop):
|
|
"""Get the next valid non-"\\#" character that satisfies the `stop` predicate"""
|
|
while True:
|
|
ch = lines[lineno][col]
|
|
if ch in "\\#":
|
|
lineno, col = nextline(lineno, col)
|
|
elif not stop(ch):
|
|
lineno, col = increment(lineno, col)
|
|
else:
|
|
break
|
|
return lineno, col
|
|
|
|
def setup_positions(expr, force_valid=True):
|
|
"""Get the lineno/col position of the end of `expr`. If `force_valid` is True,
|
|
forces the position to be a valid character (e.g. if the position is beyond the
|
|
end of the line, move to the next line)
|
|
"""
|
|
# -2 since end_lineno is 1-indexed and because we added an extra
|
|
# bracket + newline to `segment` when calling ast.parse
|
|
lineno = expr.end_lineno - 2
|
|
col = normalize(lineno, expr.end_col_offset)
|
|
return next_valid_char(lineno, col) if force_valid else (lineno, col)
|
|
|
|
statement = tree.body[0]
|
|
match statement:
|
|
case ast.Expr(expr):
|
|
match expr:
|
|
case ast.BinOp():
|
|
# ast gives these locations for BinOp subexpressions
|
|
# ( left_expr ) + ( right_expr )
|
|
# left^^^^^ right^^^^^
|
|
lineno, col = setup_positions(expr.left)
|
|
|
|
# First operator character is the first non-space/')' character
|
|
lineno, col = increment_until(lineno, col, lambda x: not x.isspace() and x != ')')
|
|
|
|
# binary op is 1 or 2 characters long, on the same line,
|
|
# before the right subexpression
|
|
right_col = col + 1
|
|
if (
|
|
right_col < len(lines[lineno])
|
|
and (
|
|
# operator char should not be in the right subexpression
|
|
expr.right.lineno - 2 > lineno or
|
|
right_col < normalize(expr.right.lineno - 2, expr.right.col_offset)
|
|
)
|
|
and not (ch := lines[lineno][right_col]).isspace()
|
|
and ch not in "\\#"
|
|
):
|
|
right_col += 1
|
|
|
|
# right_col can be invalid since it is exclusive
|
|
return _Anchors(lineno, col, lineno, right_col)
|
|
case ast.Subscript():
|
|
# ast gives these locations for value and slice subexpressions
|
|
# ( value_expr ) [ slice_expr ]
|
|
# value^^^^^ slice^^^^^
|
|
# subscript^^^^^^^^^^^^^^^^^^^^
|
|
|
|
# find left bracket
|
|
left_lineno, left_col = setup_positions(expr.value)
|
|
left_lineno, left_col = increment_until(left_lineno, left_col, lambda x: x == '[')
|
|
# find right bracket (final character of expression)
|
|
right_lineno, right_col = setup_positions(expr, force_valid=False)
|
|
return _Anchors(left_lineno, left_col, right_lineno, right_col)
|
|
case ast.Call():
|
|
# ast gives these locations for function call expressions
|
|
# ( func_expr ) (args, kwargs)
|
|
# func^^^^^
|
|
# call^^^^^^^^^^^^^^^^^^^^^^^^
|
|
|
|
# find left bracket
|
|
left_lineno, left_col = setup_positions(expr.func)
|
|
left_lineno, left_col = increment_until(left_lineno, left_col, lambda x: x == '(')
|
|
# find right bracket (final character of expression)
|
|
right_lineno, right_col = setup_positions(expr, force_valid=False)
|
|
return _Anchors(left_lineno, left_col, right_lineno, right_col)
|
|
|
|
return None
|
|
|
|
_WIDE_CHAR_SPECIFIERS = "WF"
|
|
|
|
def _display_width(line, offset=None):
|
|
"""Calculate the extra amount of width space the given source
|
|
code segment might take if it were to be displayed on a fixed
|
|
width output device. Supports wide unicode characters and emojis."""
|
|
|
|
if offset is None:
|
|
offset = len(line)
|
|
|
|
# Fast track for ASCII-only strings
|
|
if line.isascii():
|
|
return offset
|
|
|
|
import unicodedata
|
|
|
|
return sum(
|
|
2 if unicodedata.east_asian_width(char) in _WIDE_CHAR_SPECIFIERS else 1
|
|
for char in line[:offset]
|
|
)
|
|
|
|
|
|
|
|
class _ExceptionPrintContext:
|
|
def __init__(self):
|
|
self.seen = set()
|
|
self.exception_group_depth = 0
|
|
self.need_close = False
|
|
|
|
def indent(self):
|
|
return ' ' * (2 * self.exception_group_depth)
|
|
|
|
def emit(self, text_gen, margin_char=None):
|
|
if margin_char is None:
|
|
margin_char = '|'
|
|
indent_str = self.indent()
|
|
if self.exception_group_depth:
|
|
indent_str += margin_char + ' '
|
|
|
|
if isinstance(text_gen, str):
|
|
yield textwrap.indent(text_gen, indent_str, lambda line: True)
|
|
else:
|
|
for text in text_gen:
|
|
yield textwrap.indent(text, indent_str, lambda line: True)
|
|
|
|
|
|
class TracebackException:
|
|
"""An exception ready for rendering.
|
|
|
|
The traceback module captures enough attributes from the original exception
|
|
to this intermediary form to ensure that no references are held, while
|
|
still being able to fully print or format it.
|
|
|
|
max_group_width and max_group_depth control the formatting of exception
|
|
groups. The depth refers to the nesting level of the group, and the width
|
|
refers to the size of a single exception group's exceptions array. The
|
|
formatted output is truncated when either limit is exceeded.
|
|
|
|
Use `from_exception` to create TracebackException instances from exception
|
|
objects, or the constructor to create TracebackException instances from
|
|
individual components.
|
|
|
|
- :attr:`__cause__` A TracebackException of the original *__cause__*.
|
|
- :attr:`__context__` A TracebackException of the original *__context__*.
|
|
- :attr:`exceptions` For exception groups - a list of TracebackException
|
|
instances for the nested *exceptions*. ``None`` for other exceptions.
|
|
- :attr:`__suppress_context__` The *__suppress_context__* value from the
|
|
original exception.
|
|
- :attr:`stack` A `StackSummary` representing the traceback.
|
|
- :attr:`exc_type` (deprecated) The class of the original traceback.
|
|
- :attr:`exc_type_str` String display of exc_type
|
|
- :attr:`filename` For syntax errors - the filename where the error
|
|
occurred.
|
|
- :attr:`lineno` For syntax errors - the linenumber where the error
|
|
occurred.
|
|
- :attr:`end_lineno` For syntax errors - the end linenumber where the error
|
|
occurred. Can be `None` if not present.
|
|
- :attr:`text` For syntax errors - the text where the error
|
|
occurred.
|
|
- :attr:`offset` For syntax errors - the offset into the text where the
|
|
error occurred.
|
|
- :attr:`end_offset` For syntax errors - the end offset into the text where
|
|
the error occurred. Can be `None` if not present.
|
|
- :attr:`msg` For syntax errors - the compiler error message.
|
|
"""
|
|
|
|
def __init__(self, exc_type, exc_value, exc_traceback, *, limit=None,
|
|
lookup_lines=True, capture_locals=False, compact=False,
|
|
max_group_width=15, max_group_depth=10, save_exc_type=True, _seen=None):
|
|
# NB: we need to accept exc_traceback, exc_value, exc_traceback to
|
|
# permit backwards compat with the existing API, otherwise we
|
|
# need stub thunk objects just to glue it together.
|
|
# Handle loops in __cause__ or __context__.
|
|
is_recursive_call = _seen is not None
|
|
if _seen is None:
|
|
_seen = set()
|
|
_seen.add(id(exc_value))
|
|
|
|
self.max_group_width = max_group_width
|
|
self.max_group_depth = max_group_depth
|
|
|
|
self.stack = StackSummary._extract_from_extended_frame_gen(
|
|
_walk_tb_with_full_positions(exc_traceback),
|
|
limit=limit, lookup_lines=lookup_lines,
|
|
capture_locals=capture_locals)
|
|
|
|
self._exc_type = exc_type if save_exc_type else None
|
|
|
|
# Capture now to permit freeing resources: only complication is in the
|
|
# unofficial API _format_final_exc_line
|
|
self._str = _safe_string(exc_value, 'exception')
|
|
try:
|
|
self.__notes__ = getattr(exc_value, '__notes__', None)
|
|
except Exception as e:
|
|
self.__notes__ = [
|
|
f'Ignored error getting __notes__: {_safe_string(e, '__notes__', repr)}']
|
|
|
|
self._is_syntax_error = False
|
|
self._have_exc_type = exc_type is not None
|
|
if exc_type is not None:
|
|
self.exc_type_qualname = exc_type.__qualname__
|
|
self.exc_type_module = exc_type.__module__
|
|
else:
|
|
self.exc_type_qualname = None
|
|
self.exc_type_module = None
|
|
|
|
if exc_type and issubclass(exc_type, SyntaxError):
|
|
# Handle SyntaxError's specially
|
|
self.filename = exc_value.filename
|
|
lno = exc_value.lineno
|
|
self.lineno = str(lno) if lno is not None else None
|
|
end_lno = exc_value.end_lineno
|
|
self.end_lineno = str(end_lno) if end_lno is not None else None
|
|
self.text = exc_value.text
|
|
self.offset = exc_value.offset
|
|
self.end_offset = exc_value.end_offset
|
|
self.msg = exc_value.msg
|
|
self._is_syntax_error = True
|
|
elif exc_type and issubclass(exc_type, ImportError) and \
|
|
getattr(exc_value, "name_from", None) is not None:
|
|
wrong_name = getattr(exc_value, "name_from", None)
|
|
suggestion = _compute_suggestion_error(exc_value, exc_traceback, wrong_name)
|
|
if suggestion:
|
|
self._str += f". Did you mean: '{suggestion}'?"
|
|
elif exc_type and issubclass(exc_type, (NameError, AttributeError)) and \
|
|
getattr(exc_value, "name", None) is not None:
|
|
wrong_name = getattr(exc_value, "name", None)
|
|
suggestion = _compute_suggestion_error(exc_value, exc_traceback, wrong_name)
|
|
if suggestion:
|
|
self._str += f". Did you mean: '{suggestion}'?"
|
|
if issubclass(exc_type, NameError):
|
|
wrong_name = getattr(exc_value, "name", None)
|
|
if wrong_name is not None and wrong_name in sys.stdlib_module_names:
|
|
if suggestion:
|
|
self._str += f" Or did you forget to import '{wrong_name}'?"
|
|
else:
|
|
self._str += f". Did you forget to import '{wrong_name}'?"
|
|
if lookup_lines:
|
|
self._load_lines()
|
|
self.__suppress_context__ = \
|
|
exc_value.__suppress_context__ if exc_value is not None else False
|
|
|
|
# Convert __cause__ and __context__ to `TracebackExceptions`s, use a
|
|
# queue to avoid recursion (only the top-level call gets _seen == None)
|
|
if not is_recursive_call:
|
|
queue = [(self, exc_value)]
|
|
while queue:
|
|
te, e = queue.pop()
|
|
if (e and e.__cause__ is not None
|
|
and id(e.__cause__) not in _seen):
|
|
cause = TracebackException(
|
|
type(e.__cause__),
|
|
e.__cause__,
|
|
e.__cause__.__traceback__,
|
|
limit=limit,
|
|
lookup_lines=lookup_lines,
|
|
capture_locals=capture_locals,
|
|
max_group_width=max_group_width,
|
|
max_group_depth=max_group_depth,
|
|
_seen=_seen)
|
|
else:
|
|
cause = None
|
|
|
|
if compact:
|
|
need_context = (cause is None and
|
|
e is not None and
|
|
not e.__suppress_context__)
|
|
else:
|
|
need_context = True
|
|
if (e and e.__context__ is not None
|
|
and need_context and id(e.__context__) not in _seen):
|
|
context = TracebackException(
|
|
type(e.__context__),
|
|
e.__context__,
|
|
e.__context__.__traceback__,
|
|
limit=limit,
|
|
lookup_lines=lookup_lines,
|
|
capture_locals=capture_locals,
|
|
max_group_width=max_group_width,
|
|
max_group_depth=max_group_depth,
|
|
_seen=_seen)
|
|
else:
|
|
context = None
|
|
|
|
if e and isinstance(e, BaseExceptionGroup):
|
|
exceptions = []
|
|
for exc in e.exceptions:
|
|
texc = TracebackException(
|
|
type(exc),
|
|
exc,
|
|
exc.__traceback__,
|
|
limit=limit,
|
|
lookup_lines=lookup_lines,
|
|
capture_locals=capture_locals,
|
|
max_group_width=max_group_width,
|
|
max_group_depth=max_group_depth,
|
|
_seen=_seen)
|
|
exceptions.append(texc)
|
|
else:
|
|
exceptions = None
|
|
|
|
te.__cause__ = cause
|
|
te.__context__ = context
|
|
te.exceptions = exceptions
|
|
if cause:
|
|
queue.append((te.__cause__, e.__cause__))
|
|
if context:
|
|
queue.append((te.__context__, e.__context__))
|
|
if exceptions:
|
|
queue.extend(zip(te.exceptions, e.exceptions))
|
|
|
|
@classmethod
|
|
def from_exception(cls, exc, *args, **kwargs):
|
|
"""Create a TracebackException from an exception."""
|
|
return cls(type(exc), exc, exc.__traceback__, *args, **kwargs)
|
|
|
|
@property
|
|
def exc_type(self):
|
|
warnings.warn('Deprecated in 3.13. Use exc_type_str instead.',
|
|
DeprecationWarning, stacklevel=2)
|
|
return self._exc_type
|
|
|
|
@property
|
|
def exc_type_str(self):
|
|
if not self._have_exc_type:
|
|
return None
|
|
stype = self.exc_type_qualname
|
|
smod = self.exc_type_module
|
|
if smod not in ("__main__", "builtins"):
|
|
if not isinstance(smod, str):
|
|
smod = "<unknown>"
|
|
stype = smod + '.' + stype
|
|
return stype
|
|
|
|
def _load_lines(self):
|
|
"""Private API. force all lines in the stack to be loaded."""
|
|
for frame in self.stack:
|
|
frame.line
|
|
|
|
def __eq__(self, other):
|
|
if isinstance(other, TracebackException):
|
|
return self.__dict__ == other.__dict__
|
|
return NotImplemented
|
|
|
|
def __str__(self):
|
|
return self._str
|
|
|
|
def format_exception_only(self, *, show_group=False, _depth=0, **kwargs):
|
|
"""Format the exception part of the traceback.
|
|
|
|
The return value is a generator of strings, each ending in a newline.
|
|
|
|
Generator yields the exception message.
|
|
For :exc:`SyntaxError` exceptions, it
|
|
also yields (before the exception message)
|
|
several lines that (when printed)
|
|
display detailed information about where the syntax error occurred.
|
|
Following the message, generator also yields
|
|
all the exception's ``__notes__``.
|
|
|
|
When *show_group* is ``True``, and the exception is an instance of
|
|
:exc:`BaseExceptionGroup`, the nested exceptions are included as
|
|
well, recursively, with indentation relative to their nesting depth.
|
|
"""
|
|
colorize = kwargs.get("colorize", False)
|
|
|
|
indent = 3 * _depth * ' '
|
|
if not self._have_exc_type:
|
|
yield indent + _format_final_exc_line(None, self._str, colorize=colorize)
|
|
return
|
|
|
|
stype = self.exc_type_str
|
|
if not self._is_syntax_error:
|
|
if _depth > 0:
|
|
# Nested exceptions needs correct handling of multiline messages.
|
|
formatted = _format_final_exc_line(
|
|
stype, self._str, insert_final_newline=False, colorize=colorize
|
|
).split('\n')
|
|
yield from [
|
|
indent + l + '\n'
|
|
for l in formatted
|
|
]
|
|
else:
|
|
yield _format_final_exc_line(stype, self._str, colorize=colorize)
|
|
else:
|
|
yield from [indent + l for l in self._format_syntax_error(stype, colorize=colorize)]
|
|
|
|
if (
|
|
isinstance(self.__notes__, collections.abc.Sequence)
|
|
and not isinstance(self.__notes__, (str, bytes))
|
|
):
|
|
for note in self.__notes__:
|
|
note = _safe_string(note, 'note')
|
|
yield from [indent + l + '\n' for l in note.split('\n')]
|
|
elif self.__notes__ is not None:
|
|
yield indent + "{}\n".format(_safe_string(self.__notes__, '__notes__', func=repr))
|
|
|
|
if self.exceptions and show_group:
|
|
for ex in self.exceptions:
|
|
yield from ex.format_exception_only(show_group=show_group, _depth=_depth+1, colorize=colorize)
|
|
|
|
def _format_syntax_error(self, stype, **kwargs):
|
|
"""Format SyntaxError exceptions (internal helper)."""
|
|
# Show exactly where the problem was found.
|
|
colorize = kwargs.get("colorize", False)
|
|
filename_suffix = ''
|
|
if self.lineno is not None:
|
|
if colorize:
|
|
yield ' File {}"{}"{}, line {}{}{}\n'.format(
|
|
_ANSIColors.MAGENTA,
|
|
self.filename or "<string>",
|
|
_ANSIColors.RESET,
|
|
_ANSIColors.MAGENTA,
|
|
self.lineno,
|
|
_ANSIColors.RESET,
|
|
)
|
|
else:
|
|
yield ' File "{}", line {}\n'.format(
|
|
self.filename or "<string>", self.lineno)
|
|
elif self.filename is not None:
|
|
filename_suffix = ' ({})'.format(self.filename)
|
|
|
|
text = self.text
|
|
if text is not None:
|
|
# text = " foo\n"
|
|
# rtext = " foo"
|
|
# ltext = "foo"
|
|
rtext = text.rstrip('\n')
|
|
ltext = rtext.lstrip(' \n\f')
|
|
spaces = len(rtext) - len(ltext)
|
|
if self.offset is None:
|
|
yield ' {}\n'.format(ltext)
|
|
else:
|
|
offset = self.offset
|
|
end_offset = self.end_offset if self.end_offset not in {None, 0} else offset
|
|
if self.text and offset > len(self.text):
|
|
offset = len(self.text) + 1
|
|
if self.text and end_offset > len(self.text):
|
|
end_offset = len(self.text) + 1
|
|
if offset >= end_offset or end_offset < 0:
|
|
end_offset = offset + 1
|
|
|
|
# Convert 1-based column offset to 0-based index into stripped text
|
|
colno = offset - 1 - spaces
|
|
end_colno = end_offset - 1 - spaces
|
|
caretspace = ' '
|
|
if colno >= 0:
|
|
# non-space whitespace (likes tabs) must be kept for alignment
|
|
caretspace = ((c if c.isspace() else ' ') for c in ltext[:colno])
|
|
start_color = end_color = ""
|
|
if colorize:
|
|
# colorize from colno to end_colno
|
|
ltext = (
|
|
ltext[:colno] +
|
|
_ANSIColors.BOLD_RED + ltext[colno:end_colno] + _ANSIColors.RESET +
|
|
ltext[end_colno:]
|
|
)
|
|
start_color = _ANSIColors.BOLD_RED
|
|
end_color = _ANSIColors.RESET
|
|
yield ' {}\n'.format(ltext)
|
|
yield ' {}{}{}{}\n'.format(
|
|
"".join(caretspace),
|
|
start_color,
|
|
('^' * (end_colno - colno)),
|
|
end_color,
|
|
)
|
|
else:
|
|
yield ' {}\n'.format(ltext)
|
|
msg = self.msg or "<no detail available>"
|
|
if colorize:
|
|
yield "{}{}{}: {}{}{}{}\n".format(
|
|
_ANSIColors.BOLD_MAGENTA,
|
|
stype,
|
|
_ANSIColors.RESET,
|
|
_ANSIColors.MAGENTA,
|
|
msg,
|
|
_ANSIColors.RESET,
|
|
filename_suffix)
|
|
else:
|
|
yield "{}: {}{}\n".format(stype, msg, filename_suffix)
|
|
|
|
def format(self, *, chain=True, _ctx=None, **kwargs):
|
|
"""Format the exception.
|
|
|
|
If chain is not *True*, *__cause__* and *__context__* will not be formatted.
|
|
|
|
The return value is a generator of strings, each ending in a newline and
|
|
some containing internal newlines. `print_exception` is a wrapper around
|
|
this method which just prints the lines to a file.
|
|
|
|
The message indicating which exception occurred is always the last
|
|
string in the output.
|
|
"""
|
|
colorize = kwargs.get("colorize", False)
|
|
if _ctx is None:
|
|
_ctx = _ExceptionPrintContext()
|
|
|
|
output = []
|
|
exc = self
|
|
if chain:
|
|
while exc:
|
|
if exc.__cause__ is not None:
|
|
chained_msg = _cause_message
|
|
chained_exc = exc.__cause__
|
|
elif (exc.__context__ is not None and
|
|
not exc.__suppress_context__):
|
|
chained_msg = _context_message
|
|
chained_exc = exc.__context__
|
|
else:
|
|
chained_msg = None
|
|
chained_exc = None
|
|
|
|
output.append((chained_msg, exc))
|
|
exc = chained_exc
|
|
else:
|
|
output.append((None, exc))
|
|
|
|
for msg, exc in reversed(output):
|
|
if msg is not None:
|
|
yield from _ctx.emit(msg)
|
|
if exc.exceptions is None:
|
|
if exc.stack:
|
|
yield from _ctx.emit('Traceback (most recent call last):\n')
|
|
yield from _ctx.emit(exc.stack.format(colorize=colorize))
|
|
yield from _ctx.emit(exc.format_exception_only(colorize=colorize))
|
|
elif _ctx.exception_group_depth > self.max_group_depth:
|
|
# exception group, but depth exceeds limit
|
|
yield from _ctx.emit(
|
|
f"... (max_group_depth is {self.max_group_depth})\n")
|
|
else:
|
|
# format exception group
|
|
is_toplevel = (_ctx.exception_group_depth == 0)
|
|
if is_toplevel:
|
|
_ctx.exception_group_depth += 1
|
|
|
|
if exc.stack:
|
|
yield from _ctx.emit(
|
|
'Exception Group Traceback (most recent call last):\n',
|
|
margin_char = '+' if is_toplevel else None)
|
|
yield from _ctx.emit(exc.stack.format(colorize=colorize))
|
|
|
|
yield from _ctx.emit(exc.format_exception_only(colorize=colorize))
|
|
num_excs = len(exc.exceptions)
|
|
if num_excs <= self.max_group_width:
|
|
n = num_excs
|
|
else:
|
|
n = self.max_group_width + 1
|
|
_ctx.need_close = False
|
|
for i in range(n):
|
|
last_exc = (i == n-1)
|
|
if last_exc:
|
|
# The closing frame may be added by a recursive call
|
|
_ctx.need_close = True
|
|
|
|
if self.max_group_width is not None:
|
|
truncated = (i >= self.max_group_width)
|
|
else:
|
|
truncated = False
|
|
title = f'{i+1}' if not truncated else '...'
|
|
yield (_ctx.indent() +
|
|
('+-' if i==0 else ' ') +
|
|
f'+---------------- {title} ----------------\n')
|
|
_ctx.exception_group_depth += 1
|
|
if not truncated:
|
|
yield from exc.exceptions[i].format(chain=chain, _ctx=_ctx)
|
|
else:
|
|
remaining = num_excs - self.max_group_width
|
|
plural = 's' if remaining > 1 else ''
|
|
yield from _ctx.emit(
|
|
f"and {remaining} more exception{plural}\n")
|
|
|
|
if last_exc and _ctx.need_close:
|
|
yield (_ctx.indent() +
|
|
"+------------------------------------\n")
|
|
_ctx.need_close = False
|
|
_ctx.exception_group_depth -= 1
|
|
|
|
if is_toplevel:
|
|
assert _ctx.exception_group_depth == 1
|
|
_ctx.exception_group_depth = 0
|
|
|
|
|
|
def print(self, *, file=None, chain=True, **kwargs):
|
|
"""Print the result of self.format(chain=chain) to 'file'."""
|
|
colorize = kwargs.get("colorize", False)
|
|
if file is None:
|
|
file = sys.stderr
|
|
for line in self.format(chain=chain, colorize=colorize):
|
|
print(line, file=file, end="")
|
|
|
|
|
|
_MAX_CANDIDATE_ITEMS = 750
|
|
_MAX_STRING_SIZE = 40
|
|
_MOVE_COST = 2
|
|
_CASE_COST = 1
|
|
|
|
|
|
def _substitution_cost(ch_a, ch_b):
|
|
if ch_a == ch_b:
|
|
return 0
|
|
if ch_a.lower() == ch_b.lower():
|
|
return _CASE_COST
|
|
return _MOVE_COST
|
|
|
|
|
|
def _compute_suggestion_error(exc_value, tb, wrong_name):
|
|
if wrong_name is None or not isinstance(wrong_name, str):
|
|
return None
|
|
if isinstance(exc_value, AttributeError):
|
|
obj = exc_value.obj
|
|
try:
|
|
d = dir(obj)
|
|
except Exception:
|
|
return None
|
|
elif isinstance(exc_value, ImportError):
|
|
try:
|
|
mod = __import__(exc_value.name)
|
|
d = dir(mod)
|
|
except Exception:
|
|
return None
|
|
else:
|
|
assert isinstance(exc_value, NameError)
|
|
# find most recent frame
|
|
if tb is None:
|
|
return None
|
|
while tb.tb_next is not None:
|
|
tb = tb.tb_next
|
|
frame = tb.tb_frame
|
|
d = (
|
|
list(frame.f_locals)
|
|
+ list(frame.f_globals)
|
|
+ list(frame.f_builtins)
|
|
)
|
|
|
|
# Check first if we are in a method and the instance
|
|
# has the wrong name as attribute
|
|
if 'self' in frame.f_locals:
|
|
self = frame.f_locals['self']
|
|
if hasattr(self, wrong_name):
|
|
return f"self.{wrong_name}"
|
|
|
|
try:
|
|
import _suggestions
|
|
except ImportError:
|
|
pass
|
|
else:
|
|
return _suggestions._generate_suggestions(d, wrong_name)
|
|
|
|
# Compute closest match
|
|
|
|
if len(d) > _MAX_CANDIDATE_ITEMS:
|
|
return None
|
|
wrong_name_len = len(wrong_name)
|
|
if wrong_name_len > _MAX_STRING_SIZE:
|
|
return None
|
|
best_distance = wrong_name_len
|
|
suggestion = None
|
|
for possible_name in d:
|
|
if possible_name == wrong_name:
|
|
# A missing attribute is "found". Don't suggest it (see GH-88821).
|
|
continue
|
|
# No more than 1/3 of the involved characters should need changed.
|
|
max_distance = (len(possible_name) + wrong_name_len + 3) * _MOVE_COST // 6
|
|
# Don't take matches we've already beaten.
|
|
max_distance = min(max_distance, best_distance - 1)
|
|
current_distance = _levenshtein_distance(wrong_name, possible_name, max_distance)
|
|
if current_distance > max_distance:
|
|
continue
|
|
if not suggestion or current_distance < best_distance:
|
|
suggestion = possible_name
|
|
best_distance = current_distance
|
|
return suggestion
|
|
|
|
|
|
def _levenshtein_distance(a, b, max_cost):
|
|
# A Python implementation of Python/suggestions.c:levenshtein_distance.
|
|
|
|
# Both strings are the same
|
|
if a == b:
|
|
return 0
|
|
|
|
# Trim away common affixes
|
|
pre = 0
|
|
while a[pre:] and b[pre:] and a[pre] == b[pre]:
|
|
pre += 1
|
|
a = a[pre:]
|
|
b = b[pre:]
|
|
post = 0
|
|
while a[:post or None] and b[:post or None] and a[post-1] == b[post-1]:
|
|
post -= 1
|
|
a = a[:post or None]
|
|
b = b[:post or None]
|
|
if not a or not b:
|
|
return _MOVE_COST * (len(a) + len(b))
|
|
if len(a) > _MAX_STRING_SIZE or len(b) > _MAX_STRING_SIZE:
|
|
return max_cost + 1
|
|
|
|
# Prefer shorter buffer
|
|
if len(b) < len(a):
|
|
a, b = b, a
|
|
|
|
# Quick fail when a match is impossible
|
|
if (len(b) - len(a)) * _MOVE_COST > max_cost:
|
|
return max_cost + 1
|
|
|
|
# Instead of producing the whole traditional len(a)-by-len(b)
|
|
# matrix, we can update just one row in place.
|
|
# Initialize the buffer row
|
|
row = list(range(_MOVE_COST, _MOVE_COST * (len(a) + 1), _MOVE_COST))
|
|
|
|
result = 0
|
|
for bindex in range(len(b)):
|
|
bchar = b[bindex]
|
|
distance = result = bindex * _MOVE_COST
|
|
minimum = sys.maxsize
|
|
for index in range(len(a)):
|
|
# 1) Previous distance in this row is cost(b[:b_index], a[:index])
|
|
substitute = distance + _substitution_cost(bchar, a[index])
|
|
# 2) cost(b[:b_index], a[:index+1]) from previous row
|
|
distance = row[index]
|
|
# 3) existing result is cost(b[:b_index+1], a[index])
|
|
|
|
insert_delete = min(result, distance) + _MOVE_COST
|
|
result = min(insert_delete, substitute)
|
|
|
|
# cost(b[:b_index+1], a[:index+1])
|
|
row[index] = result
|
|
if result < minimum:
|
|
minimum = result
|
|
if minimum > max_cost:
|
|
# Everything in this row is too big, so bail early.
|
|
return max_cost + 1
|
|
return result
|