Lines Matching refs:tok
183 tok = self._tokenizer.peek()
184 return f"{tok.start[0]}.{tok.start[1]}: {token.tok_name[tok.type]}:{tok.string!r}"
188 tok = self._tokenizer.peek()
189 if tok.type == token.NAME and tok.string not in self.KEYWORDS:
195 tok = self._tokenizer.peek()
196 if tok.type == token.NUMBER:
202 tok = self._tokenizer.peek()
203 if tok.type == token.STRING:
209 tok = self._tokenizer.peek()
210 if tok.type == token.OP:
216 tok = self._tokenizer.peek()
217 if tok.type == token.TYPE_COMMENT:
223 tok = self._tokenizer.peek()
224 if tok.type == token.NAME and tok.string in self.SOFT_KEYWORDS:
230 tok = self._tokenizer.peek()
231 if tok.string == type:
234 if tok.type == exact_token_types[type]:
237 if tok.type == token.__dict__[type]:
239 if tok.type == token.OP and tok.string == type:
261 tok = self._tokenizer.diagnose()
262 return SyntaxError(message, (filename, tok.start[0], 1 + tok.start[1], tok.line))