@@ -465,7 +465,6 @@ def cursor_offset(self, value: int) -> None:
465465 self ._set_cursor_offset (value )
466466
467467 if TYPE_CHECKING :
468-
469468 # not actually defined, subclasses must define
470469 cpos : int
471470
@@ -567,7 +566,7 @@ def current_string(self, concatenate=False):
567566 return ""
568567 opening = string_tokens .pop ()[1 ]
569568 string = list ()
570- for ( token , value ) in reversed (string_tokens ):
569+ for token , value in reversed (string_tokens ):
571570 if token is Token .Text :
572571 continue
573572 elif opening is None :
@@ -602,7 +601,7 @@ def _funcname_and_argnum(
602601 # if keyword is not None, we've encountered a keyword and so we're done counting
603602 stack = [_FuncExpr ("" , "" , 0 , "" )]
604603 try :
605- for ( token , value ) in Python3Lexer ().get_tokens (line ):
604+ for token , value in Python3Lexer ().get_tokens (line ):
606605 if token is Token .Punctuation :
607606 if value in "([{" :
608607 stack .append (_FuncExpr ("" , "" , 0 , value ))
@@ -692,7 +691,6 @@ def get_args(self):
692691 # py3
693692 f .__new__ .__class__ is not object .__new__ .__class__
694693 ):
695-
696694 class_f = f .__new__
697695
698696 if class_f :
@@ -1117,7 +1115,7 @@ def tokenize(self, s, newline=False) -> List[Tuple[_TokenType, str]]:
11171115 line_tokens : List [Tuple [_TokenType , str ]] = list ()
11181116 saved_tokens : List [Tuple [_TokenType , str ]] = list ()
11191117 search_for_paren = True
1120- for ( token , value ) in split_lines (all_tokens ):
1118+ for token , value in split_lines (all_tokens ):
11211119 pos += len (value )
11221120 if token is Token .Text and value == "\n " :
11231121 line += 1
@@ -1263,7 +1261,7 @@ def next_indentation(line, tab_length) -> int:
12631261
12641262
12651263def split_lines (tokens ):
1266- for ( token , value ) in tokens :
1264+ for token , value in tokens :
12671265 if not value :
12681266 continue
12691267 while value :
0 commit comments