From 12cf076118570eebbff08c6b3090e0d4798447a1 Mon Sep 17 00:00:00 2001 From: cyfraeviolae Date: Wed, 3 Apr 2024 03:17:55 -0400 Subject: no venv --- .../site-packages/pygments/lexers/textedit.py | 202 --------------------- 1 file changed, 202 deletions(-) delete mode 100644 venv/lib/python3.11/site-packages/pygments/lexers/textedit.py (limited to 'venv/lib/python3.11/site-packages/pygments/lexers/textedit.py') diff --git a/venv/lib/python3.11/site-packages/pygments/lexers/textedit.py b/venv/lib/python3.11/site-packages/pygments/lexers/textedit.py deleted file mode 100644 index 09defdb..0000000 --- a/venv/lib/python3.11/site-packages/pygments/lexers/textedit.py +++ /dev/null @@ -1,202 +0,0 @@ -""" - pygments.lexers.textedit - ~~~~~~~~~~~~~~~~~~~~~~~~ - - Lexers for languages related to text processing. - - :copyright: Copyright 2006-2023 by the Pygments team, see AUTHORS. - :license: BSD, see LICENSE for details. -""" - -import re -from bisect import bisect - -from pygments.lexer import RegexLexer, bygroups, default, include, this, using -from pygments.lexers.python import PythonLexer -from pygments.token import Comment, Keyword, Name, Number, Operator, \ - Punctuation, String, Text, Whitespace - -__all__ = ['AwkLexer', 'SedLexer', 'VimLexer'] - - -class AwkLexer(RegexLexer): - """ - For Awk scripts. - - .. versionadded:: 1.5 - """ - - name = 'Awk' - aliases = ['awk', 'gawk', 'mawk', 'nawk'] - filenames = ['*.awk'] - mimetypes = ['application/x-awk'] - - tokens = { - 'commentsandwhitespace': [ - (r'\s+', Text), - (r'#.*$', Comment.Single) - ], - 'slashstartsregex': [ - include('commentsandwhitespace'), - (r'/(\\.|[^[/\\\n]|\[(\\.|[^\]\\\n])*])+/' - r'\B', String.Regex, '#pop'), - (r'(?=/)', Text, ('#pop', 'badregex')), - default('#pop') - ], - 'badregex': [ - (r'\n', Text, '#pop') - ], - 'root': [ - (r'^(?=\s|/)', Text, 'slashstartsregex'), - include('commentsandwhitespace'), - (r'\+\+|--|\|\||&&|in\b|\$|!?~|' - r'(\*\*|[-<>+*%\^/!=|])=?', Operator, 'slashstartsregex'), - (r'[{(\[;,]', Punctuation, 'slashstartsregex'), - (r'[})\].]', Punctuation), - (r'(break|continue|do|while|exit|for|if|else|' - r'return)\b', Keyword, 'slashstartsregex'), - (r'function\b', Keyword.Declaration, 'slashstartsregex'), - (r'(atan2|cos|exp|int|log|rand|sin|sqrt|srand|gensub|gsub|index|' - r'length|match|split|sprintf|sub|substr|tolower|toupper|close|' - r'fflush|getline|next|nextfile|print|printf|strftime|systime|' - r'delete|system)\b', Keyword.Reserved), - (r'(ARGC|ARGIND|ARGV|BEGIN|CONVFMT|ENVIRON|END|ERRNO|FIELDWIDTHS|' - r'FILENAME|FNR|FS|IGNORECASE|NF|NR|OFMT|OFS|ORFS|RLENGTH|RS|' - r'RSTART|RT|SUBSEP)\b', Name.Builtin), - (r'[$a-zA-Z_]\w*', Name.Other), - (r'[0-9][0-9]*\.[0-9]+([eE][0-9]+)?[fd]?', Number.Float), - (r'0x[0-9a-fA-F]+', Number.Hex), - (r'[0-9]+', Number.Integer), - (r'"(\\\\|\\[^\\]|[^"\\])*"', String.Double), - (r"'(\\\\|\\[^\\]|[^'\\])*'", String.Single), - ] - } - - -class SedLexer(RegexLexer): - """ - Lexer for Sed script files. - """ - name = 'Sed' - aliases = ['sed', 'gsed', 'ssed'] - filenames = ['*.sed', '*.[gs]sed'] - mimetypes = ['text/x-sed'] - flags = re.MULTILINE - - # Match the contents within delimiters such as // - _inside_delims = r'((?:(?:\\[^\n]|[^\\])*?\\\n)*?(?:\\.|[^\\])*?)' - - tokens = { - 'root': [ - (r'\s+', Whitespace), - (r'#.*$', Comment.Single), - (r'[0-9]+', Number.Integer), - (r'\$', Operator), - (r'[{};,!]', Punctuation), - (r'[dDFgGhHlnNpPqQxz=]', Keyword), - (r'([berRtTvwW:])([^;\n]*)', bygroups(Keyword, String.Single)), - (r'([aci])((?:.*?\\\n)*(?:.*?[^\\]$))', bygroups(Keyword, String.Double)), - (r'([qQ])([0-9]*)', bygroups(Keyword, Number.Integer)), - (r'(/)' + _inside_delims + r'(/)', bygroups(Punctuation, String.Regex, Punctuation)), - (r'(\\(.))' + _inside_delims + r'(\2)', - bygroups(Punctuation, None, String.Regex, Punctuation)), - (r'(y)(.)' + _inside_delims + r'(\2)' + _inside_delims + r'(\2)', - bygroups(Keyword, Punctuation, String.Single, Punctuation, String.Single, Punctuation)), - (r'(s)(.)' + _inside_delims + r'(\2)' + _inside_delims + r'(\2)((?:[gpeIiMm]|[0-9])*)', - bygroups(Keyword, Punctuation, String.Regex, Punctuation, String.Single, Punctuation, - Keyword)) - ] - } - -class VimLexer(RegexLexer): - """ - Lexer for VimL script files. - - .. versionadded:: 0.8 - """ - name = 'VimL' - aliases = ['vim'] - filenames = ['*.vim', '.vimrc', '.exrc', '.gvimrc', - '_vimrc', '_exrc', '_gvimrc', 'vimrc', 'gvimrc'] - mimetypes = ['text/x-vim'] - flags = re.MULTILINE - - _python = r'py(?:t(?:h(?:o(?:n)?)?)?)?' - - tokens = { - 'root': [ - (r'^([ \t:]*)(' + _python + r')([ \t]*)(<<)([ \t]*)(.*)((?:\n|.)*)(\6)', - bygroups(using(this), Keyword, Text, Operator, Text, Text, - using(PythonLexer), Text)), - (r'^([ \t:]*)(' + _python + r')([ \t])(.*)', - bygroups(using(this), Keyword, Text, using(PythonLexer))), - - (r'^\s*".*', Comment), - - (r'[ \t]+', Text), - # TODO: regexes can have other delims - (r'/[^/\\\n]*(?:\\[\s\S][^/\\\n]*)*/', String.Regex), - (r'"[^"\\\n]*(?:\\[\s\S][^"\\\n]*)*"', String.Double), - (r"'[^\n']*(?:''[^\n']*)*'", String.Single), - - # Who decided that doublequote was a good comment character?? - (r'(?<=\s)"[^\-:.%#=*].*', Comment), - (r'-?\d+', Number), - (r'#[0-9a-f]{6}', Number.Hex), - (r'^:', Punctuation), - (r'[()<>+=!|,~-]', Punctuation), # Inexact list. Looks decent. - (r'\b(let|if|else|endif|elseif|fun|function|endfunction)\b', - Keyword), - (r'\b(NONE|bold|italic|underline|dark|light)\b', Name.Builtin), - (r'\b\w+\b', Name.Other), # These are postprocessed below - (r'.', Text), - ], - } - - def __init__(self, **options): - from pygments.lexers._vim_builtins import auto, command, option - self._cmd = command - self._opt = option - self._aut = auto - - RegexLexer.__init__(self, **options) - - def is_in(self, w, mapping): - r""" - It's kind of difficult to decide if something might be a keyword - in VimL because it allows you to abbreviate them. In fact, - 'ab[breviate]' is a good example. :ab, :abbre, or :abbreviate are - valid ways to call it so rather than making really awful regexps - like:: - - \bab(?:b(?:r(?:e(?:v(?:i(?:a(?:t(?:e)?)?)?)?)?)?)?)?\b - - we match `\b\w+\b` and then call is_in() on those tokens. See - `scripts/get_vimkw.py` for how the lists are extracted. - """ - p = bisect(mapping, (w,)) - if p > 0: - if mapping[p-1][0] == w[:len(mapping[p-1][0])] and \ - mapping[p-1][1][:len(w)] == w: - return True - if p < len(mapping): - return mapping[p][0] == w[:len(mapping[p][0])] and \ - mapping[p][1][:len(w)] == w - return False - - def get_tokens_unprocessed(self, text): - # TODO: builtins are only subsequent tokens on lines - # and 'keywords' only happen at the beginning except - # for :au ones - for index, token, value in \ - RegexLexer.get_tokens_unprocessed(self, text): - if token is Name.Other: - if self.is_in(value, self._cmd): - yield index, Keyword, value - elif self.is_in(value, self._opt) or \ - self.is_in(value, self._aut): - yield index, Name.Builtin, value - else: - yield index, Text, value - else: - yield index, token, value -- cgit v1.2.3