1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
|
"""
Base classes for prompt_toolkit lexers.
"""
from __future__ import annotations
from abc import ABCMeta, abstractmethod
from typing import Callable, Hashable
from prompt_toolkit.document import Document
from prompt_toolkit.formatted_text.base import StyleAndTextTuples
__all__ = [
"Lexer",
"SimpleLexer",
"DynamicLexer",
]
class Lexer(metaclass=ABCMeta):
"""
Base class for all lexers.
"""
@abstractmethod
def lex_document(self, document: Document) -> Callable[[int], StyleAndTextTuples]:
"""
Takes a :class:`~prompt_toolkit.document.Document` and returns a
callable that takes a line number and returns a list of
``(style_str, text)`` tuples for that line.
XXX: Note that in the past, this was supposed to return a list
of ``(Token, text)`` tuples, just like a Pygments lexer.
"""
def invalidation_hash(self) -> Hashable:
"""
When this changes, `lex_document` could give a different output.
(Only used for `DynamicLexer`.)
"""
return id(self)
class SimpleLexer(Lexer):
"""
Lexer that doesn't do any tokenizing and returns the whole input as one
token.
:param style: The style string for this lexer.
"""
def __init__(self, style: str = "") -> None:
self.style = style
def lex_document(self, document: Document) -> Callable[[int], StyleAndTextTuples]:
lines = document.lines
def get_line(lineno: int) -> StyleAndTextTuples:
"Return the tokens for the given line."
try:
return [(self.style, lines[lineno])]
except IndexError:
return []
return get_line
class DynamicLexer(Lexer):
"""
Lexer class that can dynamically returns any Lexer.
:param get_lexer: Callable that returns a :class:`.Lexer` instance.
"""
def __init__(self, get_lexer: Callable[[], Lexer | None]) -> None:
self.get_lexer = get_lexer
self._dummy = SimpleLexer()
def lex_document(self, document: Document) -> Callable[[int], StyleAndTextTuples]:
lexer = self.get_lexer() or self._dummy
return lexer.lex_document(document)
def invalidation_hash(self) -> Hashable:
lexer = self.get_lexer() or self._dummy
return id(lexer)
|