Lines Matching defs:readline
49 result = stringify_tokens_from_source(tokenize(f.readline), s)
58 tokens = list(tokenize(f.readline))
92 readline = BytesIO(indent_error_file).readline
96 for tok in tokenize(readline):
223 for toktype, token, start, end, line in tokenize(f.readline):
954 result = stringify_tokens_from_source(generate_tokens(f.readline), s)
960 g = tokenize(BytesIO(s.encode('utf-8')).readline) # tokenize the string
1039 def readline():
1048 tokens = list(_tokenize(readline, encoding='utf-8'))[1:-2]
1056 def readline():
1065 tokens = list(_tokenize(readline, encoding=None))[:-2]
1075 def readline():
1082 return readline
1131 readline = self.get_readline(lines)
1132 self.assertRaises(SyntaxError, detect_encoding, readline)
1165 readline = self.get_readline(lines)
1166 self.assertRaises(SyntaxError, detect_encoding, readline)
1222 readline = self.get_readline(lines)
1223 self.assertRaises(SyntaxError, detect_encoding, readline)
1240 readline = self.get_readline((b'print(something)\n',))
1241 encoding, consumed_lines = detect_encoding(readline)
1249 readline = self.get_readline((b'\xef\xbb\xbfprint(something)\n',))
1250 encoding, consumed_lines = detect_encoding(readline)
1254 readline = self.get_readline((b'\xef\xbb\xbf',))
1255 encoding, consumed_lines = detect_encoding(readline)
1259 readline = self.get_readline((b'# coding: bad\n',))
1260 self.assertRaises(SyntaxError, detect_encoding, readline)
1264 readline = self.get_readline((b'print("#coding=fake")',))
1265 encoding, consumed_lines = detect_encoding(readline)
1301 def readline(self):
1312 detect_encoding(ins.readline)
1315 detect_encoding(ins.readline)
1331 def mock_detect_encoding(readline):
1334 def mock__tokenize(readline, encoding):
1339 next_line = readline()
1375 toks = list(tokenize(BytesIO(buf.encode('utf-8')).readline))
1380 tokens = list(tokenize(BytesIO(opstr.encode('utf-8')).readline))
1467 tokens = list(tokenize(BytesIO(source.encode('utf-8')).readline))
1534 readline = iter(code.splitlines(keepends=True)).__next__
1535 tokens5 = list(tokenize(readline))
1645 return untokenize(tokenize(BytesIO(code).readline)).decode('utf-8')