summaryrefslogtreecommitdiff
path: root/venv/lib/python3.11/site-packages/markdown_it/tree.py
diff options
context:
space:
mode:
Diffstat (limited to 'venv/lib/python3.11/site-packages/markdown_it/tree.py')
-rw-r--r--venv/lib/python3.11/site-packages/markdown_it/tree.py345
1 files changed, 0 insertions, 345 deletions
diff --git a/venv/lib/python3.11/site-packages/markdown_it/tree.py b/venv/lib/python3.11/site-packages/markdown_it/tree.py
deleted file mode 100644
index 6641e5a..0000000
--- a/venv/lib/python3.11/site-packages/markdown_it/tree.py
+++ /dev/null
@@ -1,345 +0,0 @@
-"""A tree representation of a linear markdown-it token stream.
-
-This module is not part of upstream JavaScript markdown-it.
-"""
-from __future__ import annotations
-
-from collections.abc import Generator, Sequence
-import textwrap
-from typing import Any, NamedTuple, TypeVar, overload
-
-from .token import Token
-
-
-class _NesterTokens(NamedTuple):
- opening: Token
- closing: Token
-
-
-_NodeType = TypeVar("_NodeType", bound="SyntaxTreeNode")
-
-
-class SyntaxTreeNode:
- """A Markdown syntax tree node.
-
- A class that can be used to construct a tree representation of a linear
- `markdown-it-py` token stream.
-
- Each node in the tree represents either:
- - root of the Markdown document
- - a single unnested `Token`
- - a `Token` "_open" and "_close" token pair, and the tokens nested in
- between
- """
-
- def __init__(
- self, tokens: Sequence[Token] = (), *, create_root: bool = True
- ) -> None:
- """Initialize a `SyntaxTreeNode` from a token stream.
-
- If `create_root` is True, create a root node for the document.
- """
- # Only nodes representing an unnested token have self.token
- self.token: Token | None = None
-
- # Only containers have nester tokens
- self.nester_tokens: _NesterTokens | None = None
-
- # Root node does not have self.parent
- self._parent: Any = None
-
- # Empty list unless a non-empty container, or unnested token that has
- # children (i.e. inline or img)
- self._children: list[Any] = []
-
- if create_root:
- self._set_children_from_tokens(tokens)
- return
-
- if not tokens:
- raise ValueError(
- "Can only create root from empty token sequence."
- " Set `create_root=True`."
- )
- elif len(tokens) == 1:
- inline_token = tokens[0]
- if inline_token.nesting:
- raise ValueError(
- "Unequal nesting level at the start and end of token stream."
- )
- self.token = inline_token
- if inline_token.children:
- self._set_children_from_tokens(inline_token.children)
- else:
- self.nester_tokens = _NesterTokens(tokens[0], tokens[-1])
- self._set_children_from_tokens(tokens[1:-1])
-
- def __repr__(self) -> str:
- return f"{type(self).__name__}({self.type})"
-
- @overload
- def __getitem__(self: _NodeType, item: int) -> _NodeType:
- ...
-
- @overload
- def __getitem__(self: _NodeType, item: slice) -> list[_NodeType]:
- ...
-
- def __getitem__(self: _NodeType, item: int | slice) -> _NodeType | list[_NodeType]:
- return self.children[item]
-
- def to_tokens(self: _NodeType) -> list[Token]:
- """Recover the linear token stream."""
-
- def recursive_collect_tokens(node: _NodeType, token_list: list[Token]) -> None:
- if node.type == "root":
- for child in node.children:
- recursive_collect_tokens(child, token_list)
- elif node.token:
- token_list.append(node.token)
- else:
- assert node.nester_tokens
- token_list.append(node.nester_tokens.opening)
- for child in node.children:
- recursive_collect_tokens(child, token_list)
- token_list.append(node.nester_tokens.closing)
-
- tokens: list[Token] = []
- recursive_collect_tokens(self, tokens)
- return tokens
-
- @property
- def children(self: _NodeType) -> list[_NodeType]:
- return self._children
-
- @children.setter
- def children(self: _NodeType, value: list[_NodeType]) -> None:
- self._children = value
-
- @property
- def parent(self: _NodeType) -> _NodeType | None:
- return self._parent # type: ignore
-
- @parent.setter
- def parent(self: _NodeType, value: _NodeType | None) -> None:
- self._parent = value
-
- @property
- def is_root(self) -> bool:
- """Is the node a special root node?"""
- return not (self.token or self.nester_tokens)
-
- @property
- def is_nested(self) -> bool:
- """Is this node nested?.
-
- Returns `True` if the node represents a `Token` pair and tokens in the
- sequence between them, where `Token.nesting` of the first `Token` in
- the pair is 1 and nesting of the other `Token` is -1.
- """
- return bool(self.nester_tokens)
-
- @property
- def siblings(self: _NodeType) -> Sequence[_NodeType]:
- """Get siblings of the node.
-
- Gets the whole group of siblings, including self.
- """
- if not self.parent:
- return [self]
- return self.parent.children
-
- @property
- def type(self) -> str:
- """Get a string type of the represented syntax.
-
- - "root" for root nodes
- - `Token.type` if the node represents an unnested token
- - `Token.type` of the opening token, with "_open" suffix stripped, if
- the node represents a nester token pair
- """
- if self.is_root:
- return "root"
- if self.token:
- return self.token.type
- assert self.nester_tokens
- return _removesuffix(self.nester_tokens.opening.type, "_open")
-
- @property
- def next_sibling(self: _NodeType) -> _NodeType | None:
- """Get the next node in the sequence of siblings.
-
- Returns `None` if this is the last sibling.
- """
- self_index = self.siblings.index(self)
- if self_index + 1 < len(self.siblings):
- return self.siblings[self_index + 1]
- return None
-
- @property
- def previous_sibling(self: _NodeType) -> _NodeType | None:
- """Get the previous node in the sequence of siblings.
-
- Returns `None` if this is the first sibling.
- """
- self_index = self.siblings.index(self)
- if self_index - 1 >= 0:
- return self.siblings[self_index - 1]
- return None
-
- def _add_child(
- self,
- tokens: Sequence[Token],
- ) -> None:
- """Make a child node for `self`."""
- child = type(self)(tokens, create_root=False)
- child.parent = self
- self.children.append(child)
-
- def _set_children_from_tokens(self, tokens: Sequence[Token]) -> None:
- """Convert the token stream to a tree structure and set the resulting
- nodes as children of `self`."""
- reversed_tokens = list(reversed(tokens))
- while reversed_tokens:
- token = reversed_tokens.pop()
-
- if not token.nesting:
- self._add_child([token])
- continue
- if token.nesting != 1:
- raise ValueError("Invalid token nesting")
-
- nested_tokens = [token]
- nesting = 1
- while reversed_tokens and nesting:
- token = reversed_tokens.pop()
- nested_tokens.append(token)
- nesting += token.nesting
- if nesting:
- raise ValueError(f"unclosed tokens starting {nested_tokens[0]}")
-
- self._add_child(nested_tokens)
-
- def pretty(
- self, *, indent: int = 2, show_text: bool = False, _current: int = 0
- ) -> str:
- """Create an XML style string of the tree."""
- prefix = " " * _current
- text = prefix + f"<{self.type}"
- if not self.is_root and self.attrs:
- text += " " + " ".join(f"{k}={v!r}" for k, v in self.attrs.items())
- text += ">"
- if (
- show_text
- and not self.is_root
- and self.type in ("text", "text_special")
- and self.content
- ):
- text += "\n" + textwrap.indent(self.content, prefix + " " * indent)
- for child in self.children:
- text += "\n" + child.pretty(
- indent=indent, show_text=show_text, _current=_current + indent
- )
- return text
-
- def walk(
- self: _NodeType, *, include_self: bool = True
- ) -> Generator[_NodeType, None, None]:
- """Recursively yield all descendant nodes in the tree starting at self.
-
- The order mimics the order of the underlying linear token
- stream (i.e. depth first).
- """
- if include_self:
- yield self
- for child in self.children:
- yield from child.walk(include_self=True)
-
- # NOTE:
- # The values of the properties defined below directly map to properties
- # of the underlying `Token`s. A root node does not translate to a `Token`
- # object, so calling these property getters on a root node will raise an
- # `AttributeError`.
- #
- # There is no mapping for `Token.nesting` because the `is_nested` property
- # provides that data, and can be called on any node type, including root.
-
- def _attribute_token(self) -> Token:
- """Return the `Token` that is used as the data source for the
- properties defined below."""
- if self.token:
- return self.token
- if self.nester_tokens:
- return self.nester_tokens.opening
- raise AttributeError("Root node does not have the accessed attribute")
-
- @property
- def tag(self) -> str:
- """html tag name, e.g. \"p\" """
- return self._attribute_token().tag
-
- @property
- def attrs(self) -> dict[str, str | int | float]:
- """Html attributes."""
- return self._attribute_token().attrs
-
- def attrGet(self, name: str) -> None | str | int | float:
- """Get the value of attribute `name`, or null if it does not exist."""
- return self._attribute_token().attrGet(name)
-
- @property
- def map(self) -> tuple[int, int] | None:
- """Source map info. Format: `tuple[ line_begin, line_end ]`"""
- map_ = self._attribute_token().map
- if map_:
- # Type ignore because `Token`s attribute types are not perfect
- return tuple(map_) # type: ignore
- return None
-
- @property
- def level(self) -> int:
- """nesting level, the same as `state.level`"""
- return self._attribute_token().level
-
- @property
- def content(self) -> str:
- """In a case of self-closing tag (code, html, fence, etc.), it
- has contents of this tag."""
- return self._attribute_token().content
-
- @property
- def markup(self) -> str:
- """'*' or '_' for emphasis, fence string for fence, etc."""
- return self._attribute_token().markup
-
- @property
- def info(self) -> str:
- """fence infostring"""
- return self._attribute_token().info
-
- @property
- def meta(self) -> dict[Any, Any]:
- """A place for plugins to store an arbitrary data."""
- return self._attribute_token().meta
-
- @property
- def block(self) -> bool:
- """True for block-level tokens, false for inline tokens."""
- return self._attribute_token().block
-
- @property
- def hidden(self) -> bool:
- """If it's true, ignore this element when rendering.
- Used for tight lists to hide paragraphs."""
- return self._attribute_token().hidden
-
-
-def _removesuffix(string: str, suffix: str) -> str:
- """Remove a suffix from a string.
-
- Replace this with str.removesuffix() from stdlib when minimum Python
- version is 3.9.
- """
- if suffix and string.endswith(suffix):
- return string[: -len(suffix)]
- return string