diff options
author | monster <monster@ydb.tech> | 2022-07-07 14:41:37 +0300 |
---|---|---|
committer | monster <monster@ydb.tech> | 2022-07-07 14:41:37 +0300 |
commit | 06e5c21a835c0e923506c4ff27929f34e00761c2 (patch) | |
tree | 75efcbc6854ef9bd476eb8bf00cc5c900da436a2 /contrib/python/Pygments/py3/pygments/lexers/sql.py | |
parent | 03f024c4412e3aa613bb543cf1660176320ba8f4 (diff) | |
download | ydb-06e5c21a835c0e923506c4ff27929f34e00761c2.tar.gz |
fix ya.make
Diffstat (limited to 'contrib/python/Pygments/py3/pygments/lexers/sql.py')
-rw-r--r-- | contrib/python/Pygments/py3/pygments/lexers/sql.py | 838 |
1 files changed, 0 insertions, 838 deletions
diff --git a/contrib/python/Pygments/py3/pygments/lexers/sql.py b/contrib/python/Pygments/py3/pygments/lexers/sql.py deleted file mode 100644 index 043a3d17cf6..00000000000 --- a/contrib/python/Pygments/py3/pygments/lexers/sql.py +++ /dev/null @@ -1,838 +0,0 @@ -""" - pygments.lexers.sql - ~~~~~~~~~~~~~~~~~~~ - - Lexers for various SQL dialects and related interactive sessions. - - Postgres specific lexers: - - `PostgresLexer` - A SQL lexer for the PostgreSQL dialect. Differences w.r.t. the SQL - lexer are: - - - keywords and data types list parsed from the PG docs (run the - `_postgres_builtins` module to update them); - - Content of $-strings parsed using a specific lexer, e.g. the content - of a PL/Python function is parsed using the Python lexer; - - parse PG specific constructs: E-strings, $-strings, U&-strings, - different operators and punctuation. - - `PlPgsqlLexer` - A lexer for the PL/pgSQL language. Adds a few specific construct on - top of the PG SQL lexer (such as <<label>>). - - `PostgresConsoleLexer` - A lexer to highlight an interactive psql session: - - - identifies the prompt and does its best to detect the end of command - in multiline statement where not all the lines are prefixed by a - prompt, telling them apart from the output; - - highlights errors in the output and notification levels; - - handles psql backslash commands. - - The ``tests/examplefiles`` contains a few test files with data to be - parsed by these lexers. - - :copyright: Copyright 2006-2022 by the Pygments team, see AUTHORS. - :license: BSD, see LICENSE for details. -""" - -import re - -from pygments.lexer import Lexer, RegexLexer, do_insertions, bygroups, words -from pygments.token import Punctuation, Whitespace, Text, Comment, Operator, \ - Keyword, Name, String, Number, Generic, Literal -from pygments.lexers import get_lexer_by_name, ClassNotFound - -from pygments.lexers._postgres_builtins import KEYWORDS, DATATYPES, \ - PSEUDO_TYPES, PLPGSQL_KEYWORDS -from pygments.lexers._mysql_builtins import \ - MYSQL_CONSTANTS, \ - MYSQL_DATATYPES, \ - MYSQL_FUNCTIONS, \ - MYSQL_KEYWORDS, \ - MYSQL_OPTIMIZER_HINTS - -from pygments.lexers import _tsql_builtins - - -__all__ = ['PostgresLexer', 'PlPgsqlLexer', 'PostgresConsoleLexer', - 'SqlLexer', 'TransactSqlLexer', 'MySqlLexer', - 'SqliteConsoleLexer', 'RqlLexer'] - -line_re = re.compile('.*?\n') -sqlite_prompt_re = re.compile(r'^(?:sqlite| ...)>(?= )') - -language_re = re.compile(r"\s+LANGUAGE\s+'?(\w+)'?", re.IGNORECASE) - -do_re = re.compile(r'\bDO\b', re.IGNORECASE) - -# Regular expressions for analyse_text() -name_between_bracket_re = re.compile(r'\[[a-zA-Z_]\w*\]') -name_between_backtick_re = re.compile(r'`[a-zA-Z_]\w*`') -tsql_go_re = re.compile(r'\bgo\b', re.IGNORECASE) -tsql_declare_re = re.compile(r'\bdeclare\s+@', re.IGNORECASE) -tsql_variable_re = re.compile(r'@[a-zA-Z_]\w*\b') - - -def language_callback(lexer, match): - """Parse the content of a $-string using a lexer - - The lexer is chosen looking for a nearby LANGUAGE or assumed as - plpgsql if inside a DO statement and no LANGUAGE has been found. - """ - lx = None - m = language_re.match(lexer.text[match.end():match.end()+100]) - if m is not None: - lx = lexer._get_lexer(m.group(1)) - else: - m = list(language_re.finditer( - lexer.text[max(0, match.start()-100):match.start()])) - if m: - lx = lexer._get_lexer(m[-1].group(1)) - else: - m = list(do_re.finditer( - lexer.text[max(0, match.start()-25):match.start()])) - if m: - lx = lexer._get_lexer('plpgsql') - - # 1 = $, 2 = delimiter, 3 = $ - yield (match.start(1), String, match.group(1)) - yield (match.start(2), String.Delimiter, match.group(2)) - yield (match.start(3), String, match.group(3)) - # 4 = string contents - if lx: - yield from lx.get_tokens_unprocessed(match.group(4)) - else: - yield (match.start(4), String, match.group(4)) - # 5 = $, 6 = delimiter, 7 = $ - yield (match.start(5), String, match.group(5)) - yield (match.start(6), String.Delimiter, match.group(6)) - yield (match.start(7), String, match.group(7)) - - -class PostgresBase: - """Base class for Postgres-related lexers. - - This is implemented as a mixin to avoid the Lexer metaclass kicking in. - this way the different lexer don't have a common Lexer ancestor. If they - had, _tokens could be created on this ancestor and not updated for the - other classes, resulting e.g. in PL/pgSQL parsed as SQL. This shortcoming - seem to suggest that regexp lexers are not really subclassable. - """ - def get_tokens_unprocessed(self, text, *args): - # Have a copy of the entire text to be used by `language_callback`. - self.text = text - yield from super().get_tokens_unprocessed(text, *args) - - def _get_lexer(self, lang): - if lang.lower() == 'sql': - return get_lexer_by_name('postgresql', **self.options) - - tries = [lang] - if lang.startswith('pl'): - tries.append(lang[2:]) - if lang.endswith('u'): - tries.append(lang[:-1]) - if lang.startswith('pl') and lang.endswith('u'): - tries.append(lang[2:-1]) - - for lx in tries: - try: - return get_lexer_by_name(lx, **self.options) - except ClassNotFound: - pass - else: - # TODO: better logging - # print >>sys.stderr, "language not found:", lang - return None - - -class PostgresLexer(PostgresBase, RegexLexer): - """ - Lexer for the PostgreSQL dialect of SQL. - - .. versionadded:: 1.5 - """ - - name = 'PostgreSQL SQL dialect' - aliases = ['postgresql', 'postgres'] - mimetypes = ['text/x-postgresql'] - - flags = re.IGNORECASE - tokens = { - 'root': [ - (r'\s+', Whitespace), - (r'--.*\n?', Comment.Single), - (r'/\*', Comment.Multiline, 'multiline-comments'), - (r'(' + '|'.join(s.replace(" ", r"\s+") - for s in DATATYPES + PSEUDO_TYPES) + r')\b', - Name.Builtin), - (words(KEYWORDS, suffix=r'\b'), Keyword), - (r'[+*/<>=~!@#%^&|`?-]+', Operator), - (r'::', Operator), # cast - (r'\$\d+', Name.Variable), - (r'([0-9]*\.[0-9]*|[0-9]+)(e[+-]?[0-9]+)?', Number.Float), - (r'[0-9]+', Number.Integer), - (r"((?:E|U&)?)(')", bygroups(String.Affix, String.Single), 'string'), - # quoted identifier - (r'((?:U&)?)(")', bygroups(String.Affix, String.Name), 'quoted-ident'), - (r'(?s)(\$)([^$]*)(\$)(.*?)(\$)(\2)(\$)', language_callback), - (r'[a-z_]\w*', Name), - - # psql variable in SQL - (r""":(['"]?)[a-z]\w*\b\1""", Name.Variable), - - (r'[;:()\[\]{},.]', Punctuation), - ], - 'multiline-comments': [ - (r'/\*', Comment.Multiline, 'multiline-comments'), - (r'\*/', Comment.Multiline, '#pop'), - (r'[^/*]+', Comment.Multiline), - (r'[/*]', Comment.Multiline) - ], - 'string': [ - (r"[^']+", String.Single), - (r"''", String.Single), - (r"'", String.Single, '#pop'), - ], - 'quoted-ident': [ - (r'[^"]+', String.Name), - (r'""', String.Name), - (r'"', String.Name, '#pop'), - ], - } - - -class PlPgsqlLexer(PostgresBase, RegexLexer): - """ - Handle the extra syntax in Pl/pgSQL language. - - .. versionadded:: 1.5 - """ - name = 'PL/pgSQL' - aliases = ['plpgsql'] - mimetypes = ['text/x-plpgsql'] - - flags = re.IGNORECASE - tokens = {k: l[:] for (k, l) in PostgresLexer.tokens.items()} - - # extend the keywords list - for i, pattern in enumerate(tokens['root']): - if pattern[1] == Keyword: - tokens['root'][i] = ( - words(KEYWORDS + PLPGSQL_KEYWORDS, suffix=r'\b'), - Keyword) - del i - break - else: - assert 0, "SQL keywords not found" - - # Add specific PL/pgSQL rules (before the SQL ones) - tokens['root'][:0] = [ - (r'\%[a-z]\w*\b', Name.Builtin), # actually, a datatype - (r':=', Operator), - (r'\<\<[a-z]\w*\>\>', Name.Label), - (r'\#[a-z]\w*\b', Keyword.Pseudo), # #variable_conflict - ] - - -class PsqlRegexLexer(PostgresBase, RegexLexer): - """ - Extend the PostgresLexer adding support specific for psql commands. - - This is not a complete psql lexer yet as it lacks prompt support - and output rendering. - """ - - name = 'PostgreSQL console - regexp based lexer' - aliases = [] # not public - - flags = re.IGNORECASE - tokens = {k: l[:] for (k, l) in PostgresLexer.tokens.items()} - - tokens['root'].append( - (r'\\[^\s]+', Keyword.Pseudo, 'psql-command')) - tokens['psql-command'] = [ - (r'\n', Text, 'root'), - (r'\s+', Whitespace), - (r'\\[^\s]+', Keyword.Pseudo), - (r""":(['"]?)[a-z]\w*\b\1""", Name.Variable), - (r"'(''|[^'])*'", String.Single), - (r"`([^`])*`", String.Backtick), - (r"[^\s]+", String.Symbol), - ] - - -re_prompt = re.compile(r'^(\S.*?)??[=\-\(\$\'\"][#>]') -re_psql_command = re.compile(r'\s*\\') -re_end_command = re.compile(r';\s*(--.*?)?$') -re_psql_command = re.compile(r'(\s*)(\\.+?)(\s+)$') -re_error = re.compile(r'(ERROR|FATAL):') -re_message = re.compile( - r'((?:DEBUG|INFO|NOTICE|WARNING|ERROR|' - r'FATAL|HINT|DETAIL|CONTEXT|LINE [0-9]+):)(.*?\n)') - - -class lookahead: - """Wrap an iterator and allow pushing back an item.""" - def __init__(self, x): - self.iter = iter(x) - self._nextitem = None - - def __iter__(self): - return self - - def send(self, i): - self._nextitem = i - return i - - def __next__(self): - if self._nextitem is not None: - ni = self._nextitem - self._nextitem = None - return ni - return next(self.iter) - next = __next__ - - -class PostgresConsoleLexer(Lexer): - """ - Lexer for psql sessions. - - .. versionadded:: 1.5 - """ - - name = 'PostgreSQL console (psql)' - aliases = ['psql', 'postgresql-console', 'postgres-console'] - mimetypes = ['text/x-postgresql-psql'] - - def get_tokens_unprocessed(self, data): - sql = PsqlRegexLexer(**self.options) - - lines = lookahead(line_re.findall(data)) - - # prompt-output cycle - while 1: - - # consume the lines of the command: start with an optional prompt - # and continue until the end of command is detected - curcode = '' - insertions = [] - for line in lines: - # Identify a shell prompt in case of psql commandline example - if line.startswith('$') and not curcode: - lexer = get_lexer_by_name('console', **self.options) - yield from lexer.get_tokens_unprocessed(line) - break - - # Identify a psql prompt - mprompt = re_prompt.match(line) - if mprompt is not None: - insertions.append((len(curcode), - [(0, Generic.Prompt, mprompt.group())])) - curcode += line[len(mprompt.group()):] - else: - curcode += line - - # Check if this is the end of the command - # TODO: better handle multiline comments at the end with - # a lexer with an external state? - if re_psql_command.match(curcode) \ - or re_end_command.search(curcode): - break - - # Emit the combined stream of command and prompt(s) - yield from do_insertions(insertions, - sql.get_tokens_unprocessed(curcode)) - - # Emit the output lines - out_token = Generic.Output - for line in lines: - mprompt = re_prompt.match(line) - if mprompt is not None: - # push the line back to have it processed by the prompt - lines.send(line) - break - - mmsg = re_message.match(line) - if mmsg is not None: - if mmsg.group(1).startswith("ERROR") \ - or mmsg.group(1).startswith("FATAL"): - out_token = Generic.Error - yield (mmsg.start(1), Generic.Strong, mmsg.group(1)) - yield (mmsg.start(2), out_token, mmsg.group(2)) - else: - yield (0, out_token, line) - else: - return - - -class SqlLexer(RegexLexer): - """ - Lexer for Structured Query Language. Currently, this lexer does - not recognize any special syntax except ANSI SQL. - """ - - name = 'SQL' - aliases = ['sql'] - filenames = ['*.sql'] - mimetypes = ['text/x-sql'] - - flags = re.IGNORECASE - tokens = { - 'root': [ - (r'\s+', Whitespace), - (r'--.*\n?', Comment.Single), - (r'/\*', Comment.Multiline, 'multiline-comments'), - (words(( - 'ABORT', 'ABS', 'ABSOLUTE', 'ACCESS', 'ADA', 'ADD', 'ADMIN', 'AFTER', - 'AGGREGATE', 'ALIAS', 'ALL', 'ALLOCATE', 'ALTER', 'ANALYSE', 'ANALYZE', - 'AND', 'ANY', 'ARE', 'AS', 'ASC', 'ASENSITIVE', 'ASSERTION', 'ASSIGNMENT', - 'ASYMMETRIC', 'AT', 'ATOMIC', 'AUTHORIZATION', 'AVG', 'BACKWARD', - 'BEFORE', 'BEGIN', 'BETWEEN', 'BITVAR', 'BIT_LENGTH', 'BOTH', 'BREADTH', - 'BY', 'C', 'CACHE', 'CALL', 'CALLED', 'CARDINALITY', 'CASCADE', - 'CASCADED', 'CASE', 'CAST', 'CATALOG', 'CATALOG_NAME', 'CHAIN', - 'CHARACTERISTICS', 'CHARACTER_LENGTH', 'CHARACTER_SET_CATALOG', - 'CHARACTER_SET_NAME', 'CHARACTER_SET_SCHEMA', 'CHAR_LENGTH', 'CHECK', - 'CHECKED', 'CHECKPOINT', 'CLASS', 'CLASS_ORIGIN', 'CLOB', 'CLOSE', - 'CLUSTER', 'COALESCE', 'COBOL', 'COLLATE', 'COLLATION', - 'COLLATION_CATALOG', 'COLLATION_NAME', 'COLLATION_SCHEMA', 'COLUMN', - 'COLUMN_NAME', 'COMMAND_FUNCTION', 'COMMAND_FUNCTION_CODE', 'COMMENT', - 'COMMIT', 'COMMITTED', 'COMPLETION', 'CONDITION_NUMBER', 'CONNECT', - 'CONNECTION', 'CONNECTION_NAME', 'CONSTRAINT', 'CONSTRAINTS', - 'CONSTRAINT_CATALOG', 'CONSTRAINT_NAME', 'CONSTRAINT_SCHEMA', - 'CONSTRUCTOR', 'CONTAINS', 'CONTINUE', 'CONVERSION', 'CONVERT', - 'COPY', 'CORRESPONDING', 'COUNT', 'CREATE', 'CREATEDB', 'CREATEUSER', - 'CROSS', 'CUBE', 'CURRENT', 'CURRENT_DATE', 'CURRENT_PATH', - 'CURRENT_ROLE', 'CURRENT_TIME', 'CURRENT_TIMESTAMP', 'CURRENT_USER', - 'CURSOR', 'CURSOR_NAME', 'CYCLE', 'DATA', 'DATABASE', - 'DATETIME_INTERVAL_CODE', 'DATETIME_INTERVAL_PRECISION', 'DAY', - 'DEALLOCATE', 'DECLARE', 'DEFAULT', 'DEFAULTS', 'DEFERRABLE', - 'DEFERRED', 'DEFINED', 'DEFINER', 'DELETE', 'DELIMITER', 'DELIMITERS', - 'DEREF', 'DESC', 'DESCRIBE', 'DESCRIPTOR', 'DESTROY', 'DESTRUCTOR', - 'DETERMINISTIC', 'DIAGNOSTICS', 'DICTIONARY', 'DISCONNECT', 'DISPATCH', - 'DISTINCT', 'DO', 'DOMAIN', 'DROP', 'DYNAMIC', 'DYNAMIC_FUNCTION', - 'DYNAMIC_FUNCTION_CODE', 'EACH', 'ELSE', 'ELSIF', 'ENCODING', - 'ENCRYPTED', 'END', 'END-EXEC', 'EQUALS', 'ESCAPE', 'EVERY', 'EXCEPTION', - 'EXCEPT', 'EXCLUDING', 'EXCLUSIVE', 'EXEC', 'EXECUTE', 'EXISTING', - 'EXISTS', 'EXPLAIN', 'EXTERNAL', 'EXTRACT', 'FALSE', 'FETCH', 'FINAL', - 'FIRST', 'FOR', 'FORCE', 'FOREIGN', 'FORTRAN', 'FORWARD', 'FOUND', 'FREE', - 'FREEZE', 'FROM', 'FULL', 'FUNCTION', 'G', 'GENERAL', 'GENERATED', 'GET', - 'GLOBAL', 'GO', 'GOTO', 'GRANT', 'GRANTED', 'GROUP', 'GROUPING', - 'HANDLER', 'HAVING', 'HIERARCHY', 'HOLD', 'HOST', 'IDENTITY', 'IF', - 'IGNORE', 'ILIKE', 'IMMEDIATE', 'IMMEDIATELY', 'IMMUTABLE', 'IMPLEMENTATION', 'IMPLICIT', - 'IN', 'INCLUDING', 'INCREMENT', 'INDEX', 'INDITCATOR', 'INFIX', - 'INHERITS', 'INITIALIZE', 'INITIALLY', 'INNER', 'INOUT', 'INPUT', - 'INSENSITIVE', 'INSERT', 'INSTANTIABLE', 'INSTEAD', 'INTERSECT', 'INTO', - 'INVOKER', 'IS', 'ISNULL', 'ISOLATION', 'ITERATE', 'JOIN', 'KEY', - 'KEY_MEMBER', 'KEY_TYPE', 'LANCOMPILER', 'LANGUAGE', 'LARGE', 'LAST', - 'LATERAL', 'LEADING', 'LEFT', 'LENGTH', 'LESS', 'LEVEL', 'LIKE', 'LIMIT', - 'LISTEN', 'LOAD', 'LOCAL', 'LOCALTIME', 'LOCALTIMESTAMP', 'LOCATION', - 'LOCATOR', 'LOCK', 'LOWER', 'MAP', 'MATCH', 'MAX', 'MAXVALUE', - 'MESSAGE_LENGTH', 'MESSAGE_OCTET_LENGTH', 'MESSAGE_TEXT', 'METHOD', 'MIN', - 'MINUTE', 'MINVALUE', 'MOD', 'MODE', 'MODIFIES', 'MODIFY', 'MONTH', - 'MORE', 'MOVE', 'MUMPS', 'NAMES', 'NATIONAL', 'NATURAL', 'NCHAR', 'NCLOB', - 'NEW', 'NEXT', 'NO', 'NOCREATEDB', 'NOCREATEUSER', 'NONE', 'NOT', - 'NOTHING', 'NOTIFY', 'NOTNULL', 'NULL', 'NULLABLE', 'NULLIF', 'OBJECT', - 'OCTET_LENGTH', 'OF', 'OFF', 'OFFSET', 'OIDS', 'OLD', 'ON', 'ONLY', - 'OPEN', 'OPERATION', 'OPERATOR', 'OPTION', 'OPTIONS', 'OR', 'ORDER', - 'ORDINALITY', 'OUT', 'OUTER', 'OUTPUT', 'OVERLAPS', 'OVERLAY', - 'OVERRIDING', 'OWNER', 'PAD', 'PARAMETER', 'PARAMETERS', 'PARAMETER_MODE', - 'PARAMETER_NAME', 'PARAMETER_ORDINAL_POSITION', - 'PARAMETER_SPECIFIC_CATALOG', 'PARAMETER_SPECIFIC_NAME', - 'PARAMETER_SPECIFIC_SCHEMA', 'PARTIAL', 'PASCAL', 'PENDANT', 'PERIOD', 'PLACING', - 'PLI', 'POSITION', 'POSTFIX', 'PRECEEDS', 'PRECISION', 'PREFIX', 'PREORDER', - 'PREPARE', 'PRESERVE', 'PRIMARY', 'PRIOR', 'PRIVILEGES', 'PROCEDURAL', - 'PROCEDURE', 'PUBLIC', 'READ', 'READS', 'RECHECK', 'RECURSIVE', 'REF', - 'REFERENCES', 'REFERENCING', 'REINDEX', 'RELATIVE', 'RENAME', - 'REPEATABLE', 'REPLACE', 'RESET', 'RESTART', 'RESTRICT', 'RESULT', - 'RETURN', 'RETURNED_LENGTH', 'RETURNED_OCTET_LENGTH', 'RETURNED_SQLSTATE', - 'RETURNS', 'REVOKE', 'RIGHT', 'ROLE', 'ROLLBACK', 'ROLLUP', 'ROUTINE', - 'ROUTINE_CATALOG', 'ROUTINE_NAME', 'ROUTINE_SCHEMA', 'ROW', 'ROWS', - 'ROW_COUNT', 'RULE', 'SAVE_POINT', 'SCALE', 'SCHEMA', 'SCHEMA_NAME', - 'SCOPE', 'SCROLL', 'SEARCH', 'SECOND', 'SECURITY', 'SELECT', 'SELF', - 'SENSITIVE', 'SERIALIZABLE', 'SERVER_NAME', 'SESSION', 'SESSION_USER', - 'SET', 'SETOF', 'SETS', 'SHARE', 'SHOW', 'SIMILAR', 'SIMPLE', 'SIZE', - 'SOME', 'SOURCE', 'SPACE', 'SPECIFIC', 'SPECIFICTYPE', 'SPECIFIC_NAME', - 'SQL', 'SQLCODE', 'SQLERROR', 'SQLEXCEPTION', 'SQLSTATE', 'SQLWARNINIG', - 'STABLE', 'START', 'STATE', 'STATEMENT', 'STATIC', 'STATISTICS', 'STDIN', - 'STDOUT', 'STORAGE', 'STRICT', 'STRUCTURE', 'STYPE', 'SUBCLASS_ORIGIN', - 'SUBLIST', 'SUBSTRING', 'SUCCEEDS', 'SUM', 'SYMMETRIC', 'SYSID', 'SYSTEM', - 'SYSTEM_USER', 'TABLE', 'TABLE_NAME', ' TEMP', 'TEMPLATE', 'TEMPORARY', - 'TERMINATE', 'THAN', 'THEN', 'TIME', 'TIMESTAMP', 'TIMEZONE_HOUR', - 'TIMEZONE_MINUTE', 'TO', 'TOAST', 'TRAILING', 'TRANSACTION', - 'TRANSACTIONS_COMMITTED', 'TRANSACTIONS_ROLLED_BACK', 'TRANSACTION_ACTIVE', - 'TRANSFORM', 'TRANSFORMS', 'TRANSLATE', 'TRANSLATION', 'TREAT', 'TRIGGER', - 'TRIGGER_CATALOG', 'TRIGGER_NAME', 'TRIGGER_SCHEMA', 'TRIM', 'TRUE', - 'TRUNCATE', 'TRUSTED', 'TYPE', 'UNCOMMITTED', 'UNDER', 'UNENCRYPTED', - 'UNION', 'UNIQUE', 'UNKNOWN', 'UNLISTEN', 'UNNAMED', 'UNNEST', 'UNTIL', - 'UPDATE', 'UPPER', 'USAGE', 'USER', 'USER_DEFINED_TYPE_CATALOG', - 'USER_DEFINED_TYPE_NAME', 'USER_DEFINED_TYPE_SCHEMA', 'USING', 'VACUUM', - 'VALID', 'VALIDATOR', 'VALUES', 'VARIABLE', 'VERBOSE', - 'VERSION', 'VERSIONS', 'VERSIONING', 'VIEW', - 'VOLATILE', 'WHEN', 'WHENEVER', 'WHERE', 'WITH', 'WITHOUT', 'WORK', - 'WRITE', 'YEAR', 'ZONE'), suffix=r'\b'), - Keyword), - (words(( - 'ARRAY', 'BIGINT', 'BINARY', 'BIT', 'BLOB', 'BOOLEAN', 'CHAR', - 'CHARACTER', 'DATE', 'DEC', 'DECIMAL', 'FLOAT', 'INT', 'INTEGER', - 'INTERVAL', 'NUMBER', 'NUMERIC', 'REAL', 'SERIAL', 'SMALLINT', - 'VARCHAR', 'VARYING', 'INT8', 'SERIAL8', 'TEXT'), suffix=r'\b'), - Name.Builtin), - (r'[+*/<>=~!@#%^&|`?-]', Operator), - (r'[0-9]+', Number.Integer), - # TODO: Backslash escapes? - (r"'(''|[^'])*'", String.Single), - (r'"(""|[^"])*"', String.Symbol), # not a real string literal in ANSI SQL - (r'[a-z_][\w$]*', Name), # allow $s in strings for Oracle - (r'[;:()\[\],.]', Punctuation) - ], - 'multiline-comments': [ - (r'/\*', Comment.Multiline, 'multiline-comments'), - (r'\*/', Comment.Multiline, '#pop'), - (r'[^/*]+', Comment.Multiline), - (r'[/*]', Comment.Multiline) - ] - } - - def analyse_text(self, text): - return - - -class TransactSqlLexer(RegexLexer): - """ - Transact-SQL (T-SQL) is Microsoft's and Sybase's proprietary extension to - SQL. - - The list of keywords includes ODBC and keywords reserved for future use.. - """ - - name = 'Transact-SQL' - aliases = ['tsql', 't-sql'] - filenames = ['*.sql'] - mimetypes = ['text/x-tsql'] - - flags = re.IGNORECASE - - tokens = { - 'root': [ - (r'\s+', Whitespace), - (r'--.*?$\n?', Comment.Single), - (r'/\*', Comment.Multiline, 'multiline-comments'), - (words(_tsql_builtins.OPERATORS), Operator), - (words(_tsql_builtins.OPERATOR_WORDS, suffix=r'\b'), Operator.Word), - (words(_tsql_builtins.TYPES, suffix=r'\b'), Name.Class), - (words(_tsql_builtins.FUNCTIONS, suffix=r'\b'), Name.Function), - (r'(goto)(\s+)(\w+\b)', bygroups(Keyword, Whitespace, Name.Label)), - (words(_tsql_builtins.KEYWORDS, suffix=r'\b'), Keyword), - (r'(\[)([^]]+)(\])', bygroups(Operator, Name, Operator)), - (r'0x[0-9a-f]+', Number.Hex), - # Float variant 1, for example: 1., 1.e2, 1.2e3 - (r'[0-9]+\.[0-9]*(e[+-]?[0-9]+)?', Number.Float), - # Float variant 2, for example: .1, .1e2 - (r'\.[0-9]+(e[+-]?[0-9]+)?', Number.Float), - # Float variant 3, for example: 123e45 - (r'[0-9]+e[+-]?[0-9]+', Number.Float), - (r'[0-9]+', Number.Integer), - (r"'(''|[^'])*'", String.Single), - (r'"(""|[^"])*"', String.Symbol), - (r'[;(),.]', Punctuation), - # Below we use \w even for the first "real" character because - # tokens starting with a digit have already been recognized - # as Number above. - (r'@@\w+', Name.Builtin), - (r'@\w+', Name.Variable), - (r'(\w+)(:)', bygroups(Name.Label, Punctuation)), - (r'#?#?\w+', Name), # names for temp tables and anything else - (r'\?', Name.Variable.Magic), # parameter for prepared statements - ], - 'multiline-comments': [ - (r'/\*', Comment.Multiline, 'multiline-comments'), - (r'\*/', Comment.Multiline, '#pop'), - (r'[^/*]+', Comment.Multiline), - (r'[/*]', Comment.Multiline) - ] - } - - def analyse_text(text): - rating = 0 - if tsql_declare_re.search(text): - # Found T-SQL variable declaration. - rating = 1.0 - else: - name_between_backtick_count = len( - name_between_backtick_re.findall(text)) - name_between_bracket_count = len( - name_between_bracket_re.findall(text)) - # We need to check if there are any names using - # backticks or brackets, as otherwise both are 0 - # and 0 >= 2 * 0, so we would always assume it's true - dialect_name_count = name_between_backtick_count + name_between_bracket_count - if dialect_name_count >= 1 and \ - name_between_bracket_count >= 2 * name_between_backtick_count: - # Found at least twice as many [name] as `name`. - rating += 0.5 - elif name_between_bracket_count > name_between_backtick_count: - rating += 0.2 - elif name_between_bracket_count > 0: - rating += 0.1 - if tsql_variable_re.search(text) is not None: - rating += 0.1 - if tsql_go_re.search(text) is not None: - rating += 0.1 - return rating - - -class MySqlLexer(RegexLexer): - """The Oracle MySQL lexer. - - This lexer does not attempt to maintain strict compatibility with - MariaDB syntax or keywords. Although MySQL and MariaDB's common code - history suggests there may be significant overlap between the two, - compatibility between the two is not a target for this lexer. - """ - - name = 'MySQL' - aliases = ['mysql'] - mimetypes = ['text/x-mysql'] - - flags = re.IGNORECASE - tokens = { - 'root': [ - (r'\s+', Whitespace), - - # Comments - (r'(?:#|--\s+).*', Comment.Single), - (r'/\*\+', Comment.Special, 'optimizer-hints'), - (r'/\*', Comment.Multiline, 'multiline-comment'), - - # Hexadecimal literals - (r"x'([0-9a-f]{2})+'", Number.Hex), # MySQL requires paired hex characters in this form. - (r'0x[0-9a-f]+', Number.Hex), - - # Binary literals - (r"b'[01]+'", Number.Bin), - (r'0b[01]+', Number.Bin), - - # Numeric literals - (r'[0-9]+\.[0-9]*(e[+-]?[0-9]+)?', Number.Float), # Mandatory integer, optional fraction and exponent - (r'[0-9]*\.[0-9]+(e[+-]?[0-9]+)?', Number.Float), # Mandatory fraction, optional integer and exponent - (r'[0-9]+e[+-]?[0-9]+', Number.Float), # Exponents with integer significands are still floats - (r'[0-9]+(?=[^0-9a-z$_\u0080-\uffff])', Number.Integer), # Integers that are not in a schema object name - - # Date literals - (r"\{\s*d\s*(?P<quote>['\"])\s*\d{2}(\d{2})?.?\d{2}.?\d{2}\s*(?P=quote)\s*\}", - Literal.Date), - - # Time literals - (r"\{\s*t\s*(?P<quote>['\"])\s*(?:\d+\s+)?\d{1,2}.?\d{1,2}.?\d{1,2}(\.\d*)?\s*(?P=quote)\s*\}", - Literal.Date), - - # Timestamp literals - ( - r"\{\s*ts\s*(?P<quote>['\"])\s*" - r"\d{2}(?:\d{2})?.?\d{2}.?\d{2}" # Date part - r"\s+" # Whitespace between date and time - r"\d{1,2}.?\d{1,2}.?\d{1,2}(\.\d*)?" # Time part - r"\s*(?P=quote)\s*\}", - Literal.Date - ), - - # String literals - (r"'", String.Single, 'single-quoted-string'), - (r'"', String.Double, 'double-quoted-string'), - - # Variables - (r'@@(?:global\.|persist\.|persist_only\.|session\.)?[a-z_]+', Name.Variable), - (r'@[a-z0-9_$.]+', Name.Variable), - (r"@'", Name.Variable, 'single-quoted-variable'), - (r'@"', Name.Variable, 'double-quoted-variable'), - (r"@`", Name.Variable, 'backtick-quoted-variable'), - (r'\?', Name.Variable), # For demonstrating prepared statements - - # Operators - (r'[!%&*+/:<=>^|~-]+', Operator), - - # Exceptions; these words tokenize differently in different contexts. - (r'\b(set)(?!\s*\()', Keyword), - (r'\b(character)(\s+)(set)\b', bygroups(Keyword, Whitespace, Keyword)), - # In all other known cases, "SET" is tokenized by MYSQL_DATATYPES. - - (words(MYSQL_CONSTANTS, prefix=r'\b', suffix=r'\b'), Name.Constant), - (words(MYSQL_DATATYPES, prefix=r'\b', suffix=r'\b'), Keyword.Type), - (words(MYSQL_KEYWORDS, prefix=r'\b', suffix=r'\b'), Keyword), - (words(MYSQL_FUNCTIONS, prefix=r'\b', suffix=r'\b(\s*)(\()'), - bygroups(Name.Function, Whitespace, Punctuation)), - - # Schema object names - # - # Note: Although the first regex supports unquoted all-numeric - # identifiers, this will not be a problem in practice because - # numeric literals have already been handled above. - # - ('[0-9a-z$_\u0080-\uffff]+', Name), - (r'`', Name.Quoted, 'schema-object-name'), - - # Punctuation - (r'[(),.;]', Punctuation), - ], - - # Multiline comment substates - # --------------------------- - - 'optimizer-hints': [ - (r'[^*a-z]+', Comment.Special), - (r'\*/', Comment.Special, '#pop'), - (words(MYSQL_OPTIMIZER_HINTS, suffix=r'\b'), Comment.Preproc), - ('[a-z]+', Comment.Special), - (r'\*', Comment.Special), - ], - - 'multiline-comment': [ - (r'[^*]+', Comment.Multiline), - (r'\*/', Comment.Multiline, '#pop'), - (r'\*', Comment.Multiline), - ], - - # String substates - # ---------------- - - 'single-quoted-string': [ - (r"[^'\\]+", String.Single), - (r"''", String.Escape), - (r"""\\[0'"bnrtZ\\%_]""", String.Escape), - (r"'", String.Single, '#pop'), - ], - - 'double-quoted-string': [ - (r'[^"\\]+', String.Double), - (r'""', String.Escape), - (r"""\\[0'"bnrtZ\\%_]""", String.Escape), - (r'"', String.Double, '#pop'), - ], - - # Variable substates - # ------------------ - - 'single-quoted-variable': [ - (r"[^']+", Name.Variable), - (r"''", Name.Variable), - (r"'", Name.Variable, '#pop'), - ], - - 'double-quoted-variable': [ - (r'[^"]+', Name.Variable), - (r'""', Name.Variable), - (r'"', Name.Variable, '#pop'), - ], - - 'backtick-quoted-variable': [ - (r'[^`]+', Name.Variable), - (r'``', Name.Variable), - (r'`', Name.Variable, '#pop'), - ], - - # Schema object name substates - # ---------------------------- - # - # "Name.Quoted" and "Name.Quoted.Escape" are non-standard but - # formatters will style them as "Name" by default but add - # additional styles based on the token name. This gives users - # flexibility to add custom styles as desired. - # - 'schema-object-name': [ - (r'[^`]+', Name.Quoted), - (r'``', Name.Quoted.Escape), - (r'`', Name.Quoted, '#pop'), - ], - } - - def analyse_text(text): - rating = 0 - name_between_backtick_count = len( - name_between_backtick_re.findall(text)) - name_between_bracket_count = len( - name_between_bracket_re.findall(text)) - # Same logic as above in the TSQL analysis - dialect_name_count = name_between_backtick_count + name_between_bracket_count - if dialect_name_count >= 1 and \ - name_between_backtick_count >= 2 * name_between_bracket_count: - # Found at least twice as many `name` as [name]. - rating += 0.5 - elif name_between_backtick_count > name_between_bracket_count: - rating += 0.2 - elif name_between_backtick_count > 0: - rating += 0.1 - return rating - - -class SqliteConsoleLexer(Lexer): - """ - Lexer for example sessions using sqlite3. - - .. versionadded:: 0.11 - """ - - name = 'sqlite3con' - aliases = ['sqlite3'] - filenames = ['*.sqlite3-console'] - mimetypes = ['text/x-sqlite3-console'] - - def get_tokens_unprocessed(self, data): - sql = SqlLexer(**self.options) - - curcode = '' - insertions = [] - for match in line_re.finditer(data): - line = match.group() - prompt_match = sqlite_prompt_re.match(line) - if prompt_match is not None: - insertions.append((len(curcode), - [(0, Generic.Prompt, line[:7])])) - insertions.append((len(curcode), - [(7, Whitespace, ' ')])) - curcode += line[8:] - else: - if curcode: - yield from do_insertions(insertions, - sql.get_tokens_unprocessed(curcode)) - curcode = '' - insertions = [] - if line.startswith('SQL error: '): - yield (match.start(), Generic.Traceback, line) - else: - yield (match.start(), Generic.Output, line) - if curcode: - yield from do_insertions(insertions, - sql.get_tokens_unprocessed(curcode)) - - -class RqlLexer(RegexLexer): - """ - Lexer for Relation Query Language. - - .. versionadded:: 2.0 - """ - name = 'RQL' - url = 'http://www.logilab.org/project/rql' - aliases = ['rql'] - filenames = ['*.rql'] - mimetypes = ['text/x-rql'] - - flags = re.IGNORECASE - tokens = { - 'root': [ - (r'\s+', Whitespace), - (r'(DELETE|SET|INSERT|UNION|DISTINCT|WITH|WHERE|BEING|OR' - r'|AND|NOT|GROUPBY|HAVING|ORDERBY|ASC|DESC|LIMIT|OFFSET' - r'|TODAY|NOW|TRUE|FALSE|NULL|EXISTS)\b', Keyword), - (r'[+*/<>=%-]', Operator), - (r'(Any|is|instance_of|CWEType|CWRelation)\b', Name.Builtin), - (r'[0-9]+', Number.Integer), - (r'[A-Z_]\w*\??', Name), - (r"'(''|[^'])*'", String.Single), - (r'"(""|[^"])*"', String.Single), - (r'[;:()\[\],.]', Punctuation) - ], - } |