diff options
Diffstat (limited to 'venv/lib/python3.11/site-packages/markdown_it/common')
12 files changed, 0 insertions, 511 deletions
| diff --git a/venv/lib/python3.11/site-packages/markdown_it/common/__init__.py b/venv/lib/python3.11/site-packages/markdown_it/common/__init__.py deleted file mode 100644 index e69de29..0000000 --- a/venv/lib/python3.11/site-packages/markdown_it/common/__init__.py +++ /dev/null diff --git a/venv/lib/python3.11/site-packages/markdown_it/common/__pycache__/__init__.cpython-311.pyc b/venv/lib/python3.11/site-packages/markdown_it/common/__pycache__/__init__.cpython-311.pycBinary files differ deleted file mode 100644 index eefb1d9..0000000 --- a/venv/lib/python3.11/site-packages/markdown_it/common/__pycache__/__init__.cpython-311.pyc +++ /dev/null diff --git a/venv/lib/python3.11/site-packages/markdown_it/common/__pycache__/entities.cpython-311.pyc b/venv/lib/python3.11/site-packages/markdown_it/common/__pycache__/entities.cpython-311.pycBinary files differ deleted file mode 100644 index 0f727f7..0000000 --- a/venv/lib/python3.11/site-packages/markdown_it/common/__pycache__/entities.cpython-311.pyc +++ /dev/null diff --git a/venv/lib/python3.11/site-packages/markdown_it/common/__pycache__/html_blocks.cpython-311.pyc b/venv/lib/python3.11/site-packages/markdown_it/common/__pycache__/html_blocks.cpython-311.pycBinary files differ deleted file mode 100644 index a59fe25..0000000 --- a/venv/lib/python3.11/site-packages/markdown_it/common/__pycache__/html_blocks.cpython-311.pyc +++ /dev/null diff --git a/venv/lib/python3.11/site-packages/markdown_it/common/__pycache__/html_re.cpython-311.pyc b/venv/lib/python3.11/site-packages/markdown_it/common/__pycache__/html_re.cpython-311.pycBinary files differ deleted file mode 100644 index a71e80d..0000000 --- a/venv/lib/python3.11/site-packages/markdown_it/common/__pycache__/html_re.cpython-311.pyc +++ /dev/null diff --git a/venv/lib/python3.11/site-packages/markdown_it/common/__pycache__/normalize_url.cpython-311.pyc b/venv/lib/python3.11/site-packages/markdown_it/common/__pycache__/normalize_url.cpython-311.pycBinary files differ deleted file mode 100644 index d084e07..0000000 --- a/venv/lib/python3.11/site-packages/markdown_it/common/__pycache__/normalize_url.cpython-311.pyc +++ /dev/null diff --git a/venv/lib/python3.11/site-packages/markdown_it/common/__pycache__/utils.cpython-311.pyc b/venv/lib/python3.11/site-packages/markdown_it/common/__pycache__/utils.cpython-311.pycBinary files differ deleted file mode 100644 index 24e8539..0000000 --- a/venv/lib/python3.11/site-packages/markdown_it/common/__pycache__/utils.cpython-311.pyc +++ /dev/null diff --git a/venv/lib/python3.11/site-packages/markdown_it/common/entities.py b/venv/lib/python3.11/site-packages/markdown_it/common/entities.py deleted file mode 100644 index 6bb2d34..0000000 --- a/venv/lib/python3.11/site-packages/markdown_it/common/entities.py +++ /dev/null @@ -1,4 +0,0 @@ -"""HTML5 entities map: { name -> characters }.""" -import html.entities - -entities = {name.rstrip(";"): chars for name, chars in html.entities.html5.items()} diff --git a/venv/lib/python3.11/site-packages/markdown_it/common/html_blocks.py b/venv/lib/python3.11/site-packages/markdown_it/common/html_blocks.py deleted file mode 100644 index 8b199af..0000000 --- a/venv/lib/python3.11/site-packages/markdown_it/common/html_blocks.py +++ /dev/null @@ -1,68 +0,0 @@ -"""List of valid html blocks names, according to commonmark spec -http://jgm.github.io/CommonMark/spec.html#html-blocks -""" - -block_names = [ -    "address", -    "article", -    "aside", -    "base", -    "basefont", -    "blockquote", -    "body", -    "caption", -    "center", -    "col", -    "colgroup", -    "dd", -    "details", -    "dialog", -    "dir", -    "div", -    "dl", -    "dt", -    "fieldset", -    "figcaption", -    "figure", -    "footer", -    "form", -    "frame", -    "frameset", -    "h1", -    "h2", -    "h3", -    "h4", -    "h5", -    "h6", -    "head", -    "header", -    "hr", -    "html", -    "iframe", -    "legend", -    "li", -    "link", -    "main", -    "menu", -    "menuitem", -    "nav", -    "noframes", -    "ol", -    "optgroup", -    "option", -    "p", -    "param", -    "section", -    "source", -    "summary", -    "table", -    "tbody", -    "td", -    "tfoot", -    "th", -    "thead", -    "title", -    "tr", -    "track", -    "ul", -] diff --git a/venv/lib/python3.11/site-packages/markdown_it/common/html_re.py b/venv/lib/python3.11/site-packages/markdown_it/common/html_re.py deleted file mode 100644 index f0c336d..0000000 --- a/venv/lib/python3.11/site-packages/markdown_it/common/html_re.py +++ /dev/null @@ -1,40 +0,0 @@ -"""Regexps to match html elements -""" - -import re - -attr_name = "[a-zA-Z_:][a-zA-Z0-9:._-]*" - -unquoted = "[^\"'=<>`\\x00-\\x20]+" -single_quoted = "'[^']*'" -double_quoted = '"[^"]*"' - -attr_value = "(?:" + unquoted + "|" + single_quoted + "|" + double_quoted + ")" - -attribute = "(?:\\s+" + attr_name + "(?:\\s*=\\s*" + attr_value + ")?)" - -open_tag = "<[A-Za-z][A-Za-z0-9\\-]*" + attribute + "*\\s*\\/?>" - -close_tag = "<\\/[A-Za-z][A-Za-z0-9\\-]*\\s*>" -comment = "<!---->|<!--(?:-?[^>-])(?:-?[^-])*-->" -processing = "<[?][\\s\\S]*?[?]>" -declaration = "<![A-Z]+\\s+[^>]*>" -cdata = "<!\\[CDATA\\[[\\s\\S]*?\\]\\]>" - -HTML_TAG_RE = re.compile( -    "^(?:" -    + open_tag -    + "|" -    + close_tag -    + "|" -    + comment -    + "|" -    + processing -    + "|" -    + declaration -    + "|" -    + cdata -    + ")" -) -HTML_OPEN_CLOSE_TAG_STR = "^(?:" + open_tag + "|" + close_tag + ")" -HTML_OPEN_CLOSE_TAG_RE = re.compile(HTML_OPEN_CLOSE_TAG_STR) diff --git a/venv/lib/python3.11/site-packages/markdown_it/common/normalize_url.py b/venv/lib/python3.11/site-packages/markdown_it/common/normalize_url.py deleted file mode 100644 index 92720b3..0000000 --- a/venv/lib/python3.11/site-packages/markdown_it/common/normalize_url.py +++ /dev/null @@ -1,81 +0,0 @@ -from __future__ import annotations - -from collections.abc import Callable -from contextlib import suppress -import re -from urllib.parse import quote, unquote, urlparse, urlunparse  # noqa: F401 - -import mdurl - -from .. import _punycode - -RECODE_HOSTNAME_FOR = ("http:", "https:", "mailto:") - - -def normalizeLink(url: str) -> str: -    """Normalize destination URLs in links - -    :: - -        [label]:   destination   'title' -                ^^^^^^^^^^^ -    """ -    parsed = mdurl.parse(url, slashes_denote_host=True) - -    # Encode hostnames in urls like: -    # `http://host/`, `https://host/`, `mailto:user@host`, `//host/` -    # -    # We don't encode unknown schemas, because it's likely that we encode -    # something we shouldn't (e.g. `skype:name` treated as `skype:host`) -    # -    if parsed.hostname and ( -        not parsed.protocol or parsed.protocol in RECODE_HOSTNAME_FOR -    ): -        with suppress(Exception): -            parsed = parsed._replace(hostname=_punycode.to_ascii(parsed.hostname)) - -    return mdurl.encode(mdurl.format(parsed)) - - -def normalizeLinkText(url: str) -> str: -    """Normalize autolink content - -    :: - -        <destination> -         ~~~~~~~~~~~ -    """ -    parsed = mdurl.parse(url, slashes_denote_host=True) - -    # Encode hostnames in urls like: -    # `http://host/`, `https://host/`, `mailto:user@host`, `//host/` -    # -    # We don't encode unknown schemas, because it's likely that we encode -    # something we shouldn't (e.g. `skype:name` treated as `skype:host`) -    # -    if parsed.hostname and ( -        not parsed.protocol or parsed.protocol in RECODE_HOSTNAME_FOR -    ): -        with suppress(Exception): -            parsed = parsed._replace(hostname=_punycode.to_unicode(parsed.hostname)) - -    # add '%' to exclude list because of https://github.com/markdown-it/markdown-it/issues/720 -    return mdurl.decode(mdurl.format(parsed), mdurl.DECODE_DEFAULT_CHARS + "%") - - -BAD_PROTO_RE = re.compile(r"^(vbscript|javascript|file|data):") -GOOD_DATA_RE = re.compile(r"^data:image\/(gif|png|jpeg|webp);") - - -def validateLink(url: str, validator: Callable[[str], bool] | None = None) -> bool: -    """Validate URL link is allowed in output. - -    This validator can prohibit more than really needed to prevent XSS. -    It's a tradeoff to keep code simple and to be secure by default. - -    Note: url should be normalized at this point, and existing entities decoded. -    """ -    if validator is not None: -        return validator(url) -    url = url.strip().lower() -    return bool(GOOD_DATA_RE.search(url)) if BAD_PROTO_RE.search(url) else True diff --git a/venv/lib/python3.11/site-packages/markdown_it/common/utils.py b/venv/lib/python3.11/site-packages/markdown_it/common/utils.py deleted file mode 100644 index 0d11e3e..0000000 --- a/venv/lib/python3.11/site-packages/markdown_it/common/utils.py +++ /dev/null @@ -1,318 +0,0 @@ -"""Utilities for parsing source text -""" -from __future__ import annotations - -import re -from typing import Match, TypeVar - -from .entities import entities - - -def charCodeAt(src: str, pos: int) -> int | None: -    """ -    Returns the Unicode value of the character at the specified location. - -    @param - index The zero-based index of the desired character. -    If there is no character at the specified index, NaN is returned. - -    This was added for compatibility with python -    """ -    try: -        return ord(src[pos]) -    except IndexError: -        return None - - -def charStrAt(src: str, pos: int) -> str | None: -    """ -    Returns the Unicode value of the character at the specified location. - -    @param - index The zero-based index of the desired character. -    If there is no character at the specified index, NaN is returned. - -    This was added for compatibility with python -    """ -    try: -        return src[pos] -    except IndexError: -        return None - - -_ItemTV = TypeVar("_ItemTV") - - -def arrayReplaceAt( -    src: list[_ItemTV], pos: int, newElements: list[_ItemTV] -) -> list[_ItemTV]: -    """ -    Remove element from array and put another array at those position. -    Useful for some operations with tokens -    """ -    return src[:pos] + newElements + src[pos + 1 :] - - -def isValidEntityCode(c: int) -> bool: -    # broken sequence -    if c >= 0xD800 and c <= 0xDFFF: -        return False -    # never used -    if c >= 0xFDD0 and c <= 0xFDEF: -        return False -    if ((c & 0xFFFF) == 0xFFFF) or ((c & 0xFFFF) == 0xFFFE): -        return False -    # control codes -    if c >= 0x00 and c <= 0x08: -        return False -    if c == 0x0B: -        return False -    if c >= 0x0E and c <= 0x1F: -        return False -    if c >= 0x7F and c <= 0x9F: -        return False -    # out of range -    if c > 0x10FFFF: -        return False -    return True - - -def fromCodePoint(c: int) -> str: -    """Convert ordinal to unicode. - -    Note, in the original Javascript two string characters were required, -    for codepoints larger than `0xFFFF`. -    But Python 3 can represent any unicode codepoint in one character. -    """ -    return chr(c) - - -# UNESCAPE_MD_RE = re.compile(r'\\([!"#$%&\'()*+,\-.\/:;<=>?@[\\\]^_`{|}~])') -# ENTITY_RE_g       = re.compile(r'&([a-z#][a-z0-9]{1,31})', re.IGNORECASE) -UNESCAPE_ALL_RE = re.compile( -    r'\\([!"#$%&\'()*+,\-.\/:;<=>?@[\\\]^_`{|}~])' + "|" + r"&([a-z#][a-z0-9]{1,31});", -    re.IGNORECASE, -) -DIGITAL_ENTITY_BASE10_RE = re.compile(r"#([0-9]{1,8})") -DIGITAL_ENTITY_BASE16_RE = re.compile(r"#x([a-f0-9]{1,8})", re.IGNORECASE) - - -def replaceEntityPattern(match: str, name: str) -> str: -    """Convert HTML entity patterns, -    see https://spec.commonmark.org/0.30/#entity-references -    """ -    if name in entities: -        return entities[name] - -    code: None | int = None -    if pat := DIGITAL_ENTITY_BASE10_RE.fullmatch(name): -        code = int(pat.group(1), 10) -    elif pat := DIGITAL_ENTITY_BASE16_RE.fullmatch(name): -        code = int(pat.group(1), 16) - -    if code is not None and isValidEntityCode(code): -        return fromCodePoint(code) - -    return match - - -def unescapeAll(string: str) -> str: -    def replacer_func(match: Match[str]) -> str: -        escaped = match.group(1) -        if escaped: -            return escaped -        entity = match.group(2) -        return replaceEntityPattern(match.group(), entity) - -    if "\\" not in string and "&" not in string: -        return string -    return UNESCAPE_ALL_RE.sub(replacer_func, string) - - -ESCAPABLE = r"""\\!"#$%&'()*+,./:;<=>?@\[\]^`{}|_~-""" -ESCAPE_CHAR = re.compile(r"\\([" + ESCAPABLE + r"])") - - -def stripEscape(string: str) -> str: -    """Strip escape \\ characters""" -    return ESCAPE_CHAR.sub(r"\1", string) - - -def escapeHtml(raw: str) -> str: -    """Replace special characters "&", "<", ">" and '"' to HTML-safe sequences.""" -    # like html.escape, but without escaping single quotes -    raw = raw.replace("&", "&")  # Must be done first! -    raw = raw.replace("<", "<") -    raw = raw.replace(">", ">") -    raw = raw.replace('"', """) -    return raw - - -# ////////////////////////////////////////////////////////////////////////////// - -REGEXP_ESCAPE_RE = re.compile(r"[.?*+^$[\]\\(){}|-]") - - -def escapeRE(string: str) -> str: -    string = REGEXP_ESCAPE_RE.sub("\\$&", string) -    return string - - -# ////////////////////////////////////////////////////////////////////////////// - - -def isSpace(code: int | None) -> bool: -    """Check if character code is a whitespace.""" -    return code in (0x09, 0x20) - - -def isStrSpace(ch: str | None) -> bool: -    """Check if character is a whitespace.""" -    return ch in ("\t", " ") - - -MD_WHITESPACE = { -    0x09,  # \t -    0x0A,  # \n -    0x0B,  # \v -    0x0C,  # \f -    0x0D,  # \r -    0x20,  # space -    0xA0, -    0x1680, -    0x202F, -    0x205F, -    0x3000, -} - - -def isWhiteSpace(code: int) -> bool: -    r"""Zs (unicode class) || [\t\f\v\r\n]""" -    if code >= 0x2000 and code <= 0x200A: -        return True -    return code in MD_WHITESPACE - - -# ////////////////////////////////////////////////////////////////////////////// - -UNICODE_PUNCT_RE = re.compile( -    r"[!-#%-\*,-\/:;\?@\[-\]_\{\}\xA1\xA7\xAB\xB6\xB7\xBB\xBF\u037E\u0387\u055A-\u055F\u0589\u058A\u05BE\u05C0\u05C3\u05C6\u05F3\u05F4\u0609\u060A\u060C\u060D\u061B\u061E\u061F\u066A-\u066D\u06D4\u0700-\u070D\u07F7-\u07F9\u0830-\u083E\u085E\u0964\u0965\u0970\u09FD\u0A76\u0AF0\u0C84\u0DF4\u0E4F\u0E5A\u0E5B\u0F04-\u0F12\u0F14\u0F3A-\u0F3D\u0F85\u0FD0-\u0FD4\u0FD9\u0FDA\u104A-\u104F\u10FB\u1360-\u1368\u1400\u166D\u166E\u169B\u169C\u16EB-\u16ED\u1735\u1736\u17D4-\u17D6\u17D8-\u17DA\u1800-\u180A\u1944\u1945\u1A1E\u1A1F\u1AA0-\u1AA6\u1AA8-\u1AAD\u1B5A-\u1B60\u1BFC-\u1BFF\u1C3B-\u1C3F\u1C7E\u1C7F\u1CC0-\u1CC7\u1CD3\u2010-\u2027\u2030-\u2043\u2045-\u2051\u2053-\u205E\u207D\u207E\u208D\u208E\u2308-\u230B\u2329\u232A\u2768-\u2775\u27C5\u27C6\u27E6-\u27EF\u2983-\u2998\u29D8-\u29DB\u29FC\u29FD\u2CF9-\u2CFC\u2CFE\u2CFF\u2D70\u2E00-\u2E2E\u2E30-\u2E4E\u3001-\u3003\u3008-\u3011\u3014-\u301F\u3030\u303D\u30A0\u30FB\uA4FE\uA4FF\uA60D-\uA60F\uA673\uA67E\uA6F2-\uA6F7\uA874-\uA877\uA8CE\uA8CF\uA8F8-\uA8FA\uA8FC\uA92E\uA92F\uA95F\uA9C1-\uA9CD\uA9DE\uA9DF\uAA5C-\uAA5F\uAADE\uAADF\uAAF0\uAAF1\uABEB\uFD3E\uFD3F\uFE10-\uFE19\uFE30-\uFE52\uFE54-\uFE61\uFE63\uFE68\uFE6A\uFE6B\uFF01-\uFF03\uFF05-\uFF0A\uFF0C-\uFF0F\uFF1A\uFF1B\uFF1F\uFF20\uFF3B-\uFF3D\uFF3F\uFF5B\uFF5D\uFF5F-\uFF65]|\uD800[\uDD00-\uDD02\uDF9F\uDFD0]|\uD801\uDD6F|\uD802[\uDC57\uDD1F\uDD3F\uDE50-\uDE58\uDE7F\uDEF0-\uDEF6\uDF39-\uDF3F\uDF99-\uDF9C]|\uD803[\uDF55-\uDF59]|\uD804[\uDC47-\uDC4D\uDCBB\uDCBC\uDCBE-\uDCC1\uDD40-\uDD43\uDD74\uDD75\uDDC5-\uDDC8\uDDCD\uDDDB\uDDDD-\uDDDF\uDE38-\uDE3D\uDEA9]|\uD805[\uDC4B-\uDC4F\uDC5B\uDC5D\uDCC6\uDDC1-\uDDD7\uDE41-\uDE43\uDE60-\uDE6C\uDF3C-\uDF3E]|\uD806[\uDC3B\uDE3F-\uDE46\uDE9A-\uDE9C\uDE9E-\uDEA2]|\uD807[\uDC41-\uDC45\uDC70\uDC71\uDEF7\uDEF8]|\uD809[\uDC70-\uDC74]|\uD81A[\uDE6E\uDE6F\uDEF5\uDF37-\uDF3B\uDF44]|\uD81B[\uDE97-\uDE9A]|\uD82F\uDC9F|\uD836[\uDE87-\uDE8B]|\uD83A[\uDD5E\uDD5F]"  # noqa: E501 -) - - -# Currently without astral characters support. -def isPunctChar(ch: str) -> bool: -    """Check if character is a punctuation character.""" -    return UNICODE_PUNCT_RE.search(ch) is not None - - -MD_ASCII_PUNCT = { -    0x21,  # /* ! */ -    0x22,  # /* " */ -    0x23,  # /* # */ -    0x24,  # /* $ */ -    0x25,  # /* % */ -    0x26,  # /* & */ -    0x27,  # /* ' */ -    0x28,  # /* ( */ -    0x29,  # /* ) */ -    0x2A,  # /* * */ -    0x2B,  # /* + */ -    0x2C,  # /* , */ -    0x2D,  # /* - */ -    0x2E,  # /* . */ -    0x2F,  # /* / */ -    0x3A,  # /* : */ -    0x3B,  # /* ; */ -    0x3C,  # /* < */ -    0x3D,  # /* = */ -    0x3E,  # /* > */ -    0x3F,  # /* ? */ -    0x40,  # /* @ */ -    0x5B,  # /* [ */ -    0x5C,  # /* \ */ -    0x5D,  # /* ] */ -    0x5E,  # /* ^ */ -    0x5F,  # /* _ */ -    0x60,  # /* ` */ -    0x7B,  # /* { */ -    0x7C,  # /* | */ -    0x7D,  # /* } */ -    0x7E,  # /* ~ */ -} - - -def isMdAsciiPunct(ch: int) -> bool: -    """Markdown ASCII punctuation characters. - -    :: - -        !, ", #, $, %, &, ', (, ), *, +, ,, -, ., /, :, ;, <, =, >, ?, @, [, \\, ], ^, _, `, {, |, }, or ~ - -    See http://spec.commonmark.org/0.15/#ascii-punctuation-character - -    Don't confuse with unicode punctuation !!! It lacks some chars in ascii range. - -    """  # noqa: E501 -    return ch in MD_ASCII_PUNCT - - -def normalizeReference(string: str) -> str: -    """Helper to unify [reference labels].""" -    # Trim and collapse whitespace -    # -    string = re.sub(r"\s+", " ", string.strip()) - -    # In node v10 'ẞ'.toLowerCase() === 'Ṿ', which is presumed to be a bug -    # fixed in v12 (couldn't find any details). -    # -    # So treat this one as a special case -    # (remove this when node v10 is no longer supported). -    # -    # if ('ẞ'.toLowerCase() === 'Ṿ') { -    #   str = str.replace(/ẞ/g, 'ß') -    # } - -    # .toLowerCase().toUpperCase() should get rid of all differences -    # between letter variants. -    # -    # Simple .toLowerCase() doesn't normalize 125 code points correctly, -    # and .toUpperCase doesn't normalize 6 of them (list of exceptions: -    # İ, ϴ, ẞ, Ω, K, Å - those are already uppercased, but have differently -    # uppercased versions). -    # -    # Here's an example showing how it happens. Lets take greek letter omega: -    # uppercase U+0398 (Θ), U+03f4 (ϴ) and lowercase U+03b8 (θ), U+03d1 (ϑ) -    # -    # Unicode entries: -    # 0398;GREEK CAPITAL LETTER THETA;Lu;0;L;;;;;N;;;;03B8 -    # 03B8;GREEK SMALL LETTER THETA;Ll;0;L;;;;;N;;;0398;;0398 -    # 03D1;GREEK THETA SYMBOL;Ll;0;L;<compat> 03B8;;;;N;GREEK SMALL LETTER SCRIPT THETA;;0398;;0398 -    # 03F4;GREEK CAPITAL THETA SYMBOL;Lu;0;L;<compat> 0398;;;;N;;;;03B8 -    # -    # Case-insensitive comparison should treat all of them as equivalent. -    # -    # But .toLowerCase() doesn't change ϑ (it's already lowercase), -    # and .toUpperCase() doesn't change ϴ (already uppercase). -    # -    # Applying first lower then upper case normalizes any character: -    # '\u0398\u03f4\u03b8\u03d1'.toLowerCase().toUpperCase() === '\u0398\u0398\u0398\u0398' -    # -    # Note: this is equivalent to unicode case folding; unicode normalization -    # is a different step that is not required here. -    # -    # Final result should be uppercased, because it's later stored in an object -    # (this avoid a conflict with Object.prototype members, -    # most notably, `__proto__`) -    # -    return string.lower().upper() - - -LINK_OPEN_RE = re.compile(r"^<a[>\s]", flags=re.IGNORECASE) -LINK_CLOSE_RE = re.compile(r"^</a\s*>", flags=re.IGNORECASE) - - -def isLinkOpen(string: str) -> bool: -    return bool(LINK_OPEN_RE.search(string)) - - -def isLinkClose(string: str) -> bool: -    return bool(LINK_CLOSE_RE.search(string)) | 
