aboutsummaryrefslogtreecommitdiffstats
path: root/contrib/tools/python3/Parser/tokenizer.h
diff options
context:
space:
mode:
authorshadchin <shadchin@yandex-team.com>2024-08-17 21:51:59 +0300
committershadchin <shadchin@yandex-team.com>2024-08-17 22:04:51 +0300
commitee9edbd8878888bafcd0eeb3b528f3ec4311560b (patch)
treed54d8138e50a446904f10a2092719be86af011b7 /contrib/tools/python3/Parser/tokenizer.h
parent72cbe4bad58add0912623ba51351ff1db8587249 (diff)
downloadydb-ee9edbd8878888bafcd0eeb3b528f3ec4311560b.tar.gz
Update Python 3 to 3.12.5
https://docs.python.org/release/3.12.5/whatsnew/changelog.html#python-3-12-5-final de86cdeacd3a8653b9ec36e87975886fafcf6dc2
Diffstat (limited to 'contrib/tools/python3/Parser/tokenizer.h')
-rw-r--r--contrib/tools/python3/Parser/tokenizer.h211
1 files changed, 107 insertions, 104 deletions
diff --git a/contrib/tools/python3/Parser/tokenizer.h b/contrib/tools/python3/Parser/tokenizer.h
index 1e1daa3648..4eeeb1f234 100644
--- a/contrib/tools/python3/Parser/tokenizer.h
+++ b/contrib/tools/python3/Parser/tokenizer.h
@@ -14,133 +14,136 @@ extern "C" {
#define MAXLEVEL 200 /* Max parentheses level */
#define MAXFSTRINGLEVEL 150 /* Max f-string nesting level */
-enum decoding_state {
- STATE_INIT,
- STATE_SEEK_CODING,
- STATE_NORMAL
-};
+enum decoding_state { STATE_INIT, STATE_SEEK_CODING, STATE_NORMAL };
enum interactive_underflow_t {
- /* Normal mode of operation: return a new token when asked in interactive mode */
- IUNDERFLOW_NORMAL,
- /* Forcefully return ENDMARKER when asked for a new token in interactive mode. This
- * can be used to prevent the tokenizer to prompt the user for new tokens */
- IUNDERFLOW_STOP,
+ /* Normal mode of operation: return a new token when asked in interactive mode
+ */
+ IUNDERFLOW_NORMAL,
+ /* Forcefully return ENDMARKER when asked for a new token in interactive mode.
+ * This can be used to prevent the tokenizer to prompt the user for new tokens
+ */
+ IUNDERFLOW_STOP,
};
struct token {
- int level;
- int lineno, col_offset, end_lineno, end_col_offset;
- const char *start, *end;
- PyObject *metadata;
+ int level;
+ int lineno, col_offset, end_lineno, end_col_offset;
+ const char *start, *end;
+ PyObject *metadata;
};
enum tokenizer_mode_kind_t {
- TOK_REGULAR_MODE,
- TOK_FSTRING_MODE,
+ TOK_REGULAR_MODE,
+ TOK_FSTRING_MODE,
};
#define MAX_EXPR_NESTING 3
typedef struct _tokenizer_mode {
- enum tokenizer_mode_kind_t kind;
-
- int curly_bracket_depth;
- int curly_bracket_expr_start_depth;
-
- char f_string_quote;
- int f_string_quote_size;
- int f_string_raw;
- const char* f_string_start;
- const char* f_string_multi_line_start;
- int f_string_line_start;
-
- Py_ssize_t f_string_start_offset;
- Py_ssize_t f_string_multi_line_start_offset;
-
- Py_ssize_t last_expr_size;
- Py_ssize_t last_expr_end;
- char* last_expr_buffer;
- int f_string_debug;
+ enum tokenizer_mode_kind_t kind;
+
+ int curly_bracket_depth;
+ int curly_bracket_expr_start_depth;
+
+ char f_string_quote;
+ int f_string_quote_size;
+ int f_string_raw;
+ const char *f_string_start;
+ const char *f_string_multi_line_start;
+ int f_string_line_start;
+
+ Py_ssize_t f_string_start_offset;
+ Py_ssize_t f_string_multi_line_start_offset;
+
+ Py_ssize_t last_expr_size;
+ Py_ssize_t last_expr_end;
+ char *last_expr_buffer;
+ int f_string_debug;
+ int in_format_spec;
} tokenizer_mode;
/* Tokenizer state */
struct tok_state {
- /* Input state; buf <= cur <= inp <= end */
- /* NB an entire line is held in the buffer */
- char *buf; /* Input buffer, or NULL; malloc'ed if fp != NULL or readline != NULL */
- char *cur; /* Next character in buffer */
- char *inp; /* End of data in buffer */
- int fp_interactive; /* If the file descriptor is interactive */
- char *interactive_src_start; /* The start of the source parsed so far in interactive mode */
- char *interactive_src_end; /* The end of the source parsed so far in interactive mode */
- const char *end; /* End of input buffer if buf != NULL */
- const char *start; /* Start of current token if not NULL */
- int done; /* E_OK normally, E_EOF at EOF, otherwise error code */
- /* NB If done != E_OK, cur must be == inp!!! */
- FILE *fp; /* Rest of input; NULL if tokenizing a string */
- int tabsize; /* Tab spacing */
- int indent; /* Current indentation index */
- int indstack[MAXINDENT]; /* Stack of indents */
- int atbol; /* Nonzero if at begin of new line */
- int pendin; /* Pending indents (if > 0) or dedents (if < 0) */
- const char *prompt, *nextprompt; /* For interactive prompting */
- int lineno; /* Current line number */
- int first_lineno; /* First line of a single line or multi line string
- expression (cf. issue 16806) */
- int starting_col_offset; /* The column offset at the beginning of a token */
- int col_offset; /* Current col offset */
- int level; /* () [] {} Parentheses nesting level */
- /* Used to allow free continuations inside them */
- char parenstack[MAXLEVEL];
- int parenlinenostack[MAXLEVEL];
- int parencolstack[MAXLEVEL];
- PyObject *filename;
- /* Stuff for checking on different tab sizes */
- int altindstack[MAXINDENT]; /* Stack of alternate indents */
- /* Stuff for PEP 0263 */
- enum decoding_state decoding_state;
- int decoding_erred; /* whether erred in decoding */
- char *encoding; /* Source encoding. */
- int cont_line; /* whether we are in a continuation line. */
- const char* line_start; /* pointer to start of current line */
- const char* multi_line_start; /* pointer to start of first line of
- a single line or multi line string
- expression (cf. issue 16806) */
- PyObject *decoding_readline; /* open(...).readline */
- PyObject *decoding_buffer;
- PyObject *readline; /* readline() function */
- const char* enc; /* Encoding for the current str. */
- char* str; /* Source string being tokenized (if tokenizing from a string)*/
- char* input; /* Tokenizer's newline translated copy of the string. */
-
- int type_comments; /* Whether to look for type comments */
-
- /* async/await related fields (still needed depending on feature_version) */
- int async_hacks; /* =1 if async/await aren't always keywords */
- int async_def; /* =1 if tokens are inside an 'async def' body. */
- int async_def_indent; /* Indentation level of the outermost 'async def'. */
- int async_def_nl; /* =1 if the outermost 'async def' had at least one
- NEWLINE token after it. */
- /* How to proceed when asked for a new token in interactive mode */
- enum interactive_underflow_t interactive_underflow;
- int report_warnings;
- // TODO: Factor this into its own thing
- tokenizer_mode tok_mode_stack[MAXFSTRINGLEVEL];
- int tok_mode_stack_index;
- int tok_extra_tokens;
- int comment_newline;
- int implicit_newline;
+ /* Input state; buf <= cur <= inp <= end */
+ /* NB an entire line is held in the buffer */
+ char *buf; /* Input buffer, or NULL; malloc'ed if fp != NULL or readline !=
+ NULL */
+ char *cur; /* Next character in buffer */
+ char *inp; /* End of data in buffer */
+ int fp_interactive; /* If the file descriptor is interactive */
+ char *interactive_src_start; /* The start of the source parsed so far in
+ interactive mode */
+ char *interactive_src_end; /* The end of the source parsed so far in
+ interactive mode */
+ const char *end; /* End of input buffer if buf != NULL */
+ const char *start; /* Start of current token if not NULL */
+ int done; /* E_OK normally, E_EOF at EOF, otherwise error code */
+ /* NB If done != E_OK, cur must be == inp!!! */
+ FILE *fp; /* Rest of input; NULL if tokenizing a string */
+ int tabsize; /* Tab spacing */
+ int indent; /* Current indentation index */
+ int indstack[MAXINDENT]; /* Stack of indents */
+ int atbol; /* Nonzero if at begin of new line */
+ int pendin; /* Pending indents (if > 0) or dedents (if < 0) */
+ const char *prompt, *nextprompt; /* For interactive prompting */
+ int lineno; /* Current line number */
+ int first_lineno; /* First line of a single line or multi line string
+ expression (cf. issue 16806) */
+ int starting_col_offset; /* The column offset at the beginning of a token */
+ int col_offset; /* Current col offset */
+ int level; /* () [] {} Parentheses nesting level */
+ /* Used to allow free continuations inside them */
+ char parenstack[MAXLEVEL];
+ int parenlinenostack[MAXLEVEL];
+ int parencolstack[MAXLEVEL];
+ PyObject *filename;
+ /* Stuff for checking on different tab sizes */
+ int altindstack[MAXINDENT]; /* Stack of alternate indents */
+ /* Stuff for PEP 0263 */
+ enum decoding_state decoding_state;
+ int decoding_erred; /* whether erred in decoding */
+ char *encoding; /* Source encoding. */
+ int cont_line; /* whether we are in a continuation line. */
+ const char *line_start; /* pointer to start of current line */
+ const char *multi_line_start; /* pointer to start of first line of
+ a single line or multi line string
+ expression (cf. issue 16806) */
+ PyObject *decoding_readline; /* open(...).readline */
+ PyObject *decoding_buffer;
+ PyObject *readline; /* readline() function */
+ const char *enc; /* Encoding for the current str. */
+ char *str; /* Source string being tokenized (if tokenizing from a string)*/
+ char *input; /* Tokenizer's newline translated copy of the string. */
+
+ int type_comments; /* Whether to look for type comments */
+
+ /* async/await related fields (still needed depending on feature_version) */
+ int async_hacks; /* =1 if async/await aren't always keywords */
+ int async_def; /* =1 if tokens are inside an 'async def' body. */
+ int async_def_indent; /* Indentation level of the outermost 'async def'. */
+ int async_def_nl; /* =1 if the outermost 'async def' had at least one
+ NEWLINE token after it. */
+ /* How to proceed when asked for a new token in interactive mode */
+ enum interactive_underflow_t interactive_underflow;
+ int report_warnings;
+ // TODO: Factor this into its own thing
+ tokenizer_mode tok_mode_stack[MAXFSTRINGLEVEL];
+ int tok_mode_stack_index;
+ int tok_extra_tokens;
+ int comment_newline;
+ int implicit_newline;
#ifdef Py_DEBUG
- int debug;
+ int debug;
#endif
};
extern struct tok_state *_PyTokenizer_FromString(const char *, int, int);
extern struct tok_state *_PyTokenizer_FromUTF8(const char *, int, int);
-extern struct tok_state *_PyTokenizer_FromReadline(PyObject*, const char*, int, int);
-extern struct tok_state *_PyTokenizer_FromFile(FILE *, const char*,
- const char *, const char *);
+extern struct tok_state *_PyTokenizer_FromReadline(PyObject *, const char *,
+ int, int);
+extern struct tok_state *_PyTokenizer_FromFile(FILE *, const char *,
+ const char *, const char *);
extern void _PyTokenizer_Free(struct tok_state *);
extern void _PyToken_Free(struct token *);
extern void _PyToken_Init(struct token *);