/third_party/python/Lib/test/ |
H A D | test_tokenize.py | 6 NEWLINE, _generate_tokens_from_c_tokenizer, DEDENT) 85 DEDENT '' (5, 0) (5, 0) 647 DEDENT '' (4, 0) (4, 0) 648 DEDENT '' (4, 0) (4, 0) 826 DEDENT '' (4, 2) (4, 2) 834 DEDENT '' (6, 0) (6, 0) 835 DEDENT '' (6, 0) (6, 0) 860 DEDENT '' (3, 0) (3, 0) 908 DEDENT '' (7, 0) (7, 0) 946 DEDENT '' ( [all...] |
/third_party/python/Include/ |
H A D | token.h | 19 #define DEDENT 6 macro 85 (x) == DEDENT)
|
/third_party/python/Tools/scripts/ |
H A D | reindent.py | 287 DEDENT=tokenize.DEDENT, 295 # (NL | COMMENT)* (INDENT | DEDENT+)? 302 elif type == DEDENT:
|
/third_party/python/Lib/ |
H A D | tabnanny.py | 277 DEDENT = tokenize.DEDENT 287 # (NL | COMMENT)* (INDENT | DEDENT+)? 301 elif type == DEDENT:
|
H A D | tokenize.py | 204 elif tok_type == DEDENT: 250 elif toknum == DEDENT: 521 yield TokenInfo(DEDENT, '', (lnum, pos), (lnum, pos), line) 612 yield TokenInfo(DEDENT, '', (lnum, 0), (lnum, 0), '')
|
H A D | token.py | 12 DEDENT = 6 variable
|
H A D | inspect.py | 1208 elif type == tokenize.DEDENT:
|
/third_party/python/Lib/lib2to3/fixes/ |
H A D | fix_metaclass.py | 137 if isinstance(node, Leaf) and node.type != token.DEDENT: 224 suite.children[-1].type == token.DEDENT):
|
/third_party/python/Lib/lib2to3/pgen2/ |
H A D | token.py | 15 DEDENT = 6 variable
|
H A D | tokenize.py | 217 elif toknum == DEDENT: 434 yield (DEDENT, '', (lnum, pos), (lnum, pos), line) 558 yield (DEDENT, '', (lnum, 0), (lnum, 0), '')
|
/third_party/python/Tools/peg_generator/pegen/ |
H A D | tokenizer.py | 70 tok.type < tokenize.NEWLINE or tok.type > tokenize.DEDENT
|
/third_party/python/Lib/lib2to3/ |
H A D | patcomp.py | 30 skip = {token.NEWLINE, token.INDENT, token.DEDENT}
|
/third_party/python/Tools/i18n/ |
H A D | pygettext.py | 446 elif ttype not in [tokenize.COMMENT, token.INDENT, token.DEDENT,
|
/third_party/python/Parser/ |
H A D | pegen_errors.c | 432 if (last_token->type == INDENT || last_token->type == DEDENT) { in _Pypegen_set_syntax_error()
|
H A D | pegen.c | 446 if (token->type != ENDMARKER && (token->type < NEWLINE || token->type > DEDENT)) { in _PyPegen_get_last_nonnwhitespace_token()
|
H A D | parser.c | 4009 // block: NEWLINE INDENT statements DEDENT | simple_stmts | invalid_block 4027 { // NEWLINE INDENT statements DEDENT in block_rule() 4032 D(fprintf(stderr, "%*c> block[%d-%d]: %s\n", p->level, ' ', _mark, p->mark, "NEWLINE INDENT statements DEDENT")); in block_rule() 4044 (dedent_var = _PyPegen_expect_token(p, DEDENT)) // token='DEDENT' in block_rule() 4047 D(fprintf(stderr, "%*c+ block[%d-%d]: %s succeeded!\n", p->level, ' ', _mark, p->mark, "NEWLINE INDENT statements DEDENT")); in block_rule() 4058 p->error_indicator ? "ERROR!" : "-", _mark, p->mark, "NEWLINE INDENT statements DEDENT")); in block_rule() 7334 // | "match" subject_expr ':' NEWLINE INDENT case_block+ DEDENT 7358 { // "match" subject_expr ':' NEWLINE INDENT case_block+ DEDENT in match_stmt_rule() 7363 D(fprintf(stderr, "%*c> match_stmt[%d-%d]: %s\n", p->level, ' ', _mark, p->mark, "\"match\" subject_expr ':' NEWLINE INDENT case_block+ DEDENT")); in match_stmt_rule() [all...] |