From acd0a8f879e9cb2477f636a6e5c027988473ef97 Mon Sep 17 00:00:00 2001 From: Fangchen Li Date: Mon, 12 Oct 2020 19:24:24 -0500 Subject: [PATCH 1/2] CLN/TYP: clean unreachable code --- pandas/core/computation/parsing.py | 4 +--- pandas/tests/frame/test_constructors.py | 3 +-- 2 files changed, 2 insertions(+), 5 deletions(-) diff --git a/pandas/core/computation/parsing.py b/pandas/core/computation/parsing.py index 86e125b6b909b..8d8da3dd00310 100644 --- a/pandas/core/computation/parsing.py +++ b/pandas/core/computation/parsing.py @@ -37,9 +37,7 @@ def create_valid_python_identifier(name: str) -> str: special_characters_replacements = { char: f"_{token.tok_name[tokval]}_" # The ignore here is because of a bug in mypy that is resolved in 0.740 - for char, tokval in ( - tokenize.EXACT_TOKEN_TYPES.items() # type: ignore[attr-defined] - ) + for char, tokval in (tokenize.EXACT_TOKEN_TYPES.items()) } special_characters_replacements.update( { diff --git a/pandas/tests/frame/test_constructors.py b/pandas/tests/frame/test_constructors.py index 8ec11d14cd606..00f1a75719443 100644 --- a/pandas/tests/frame/test_constructors.py +++ b/pandas/tests/frame/test_constructors.py @@ -1079,8 +1079,7 @@ def test_constructor_empty_list(self): # Empty generator: list(empty_gen()) == [] def empty_gen(): - return - yield + yield from () df = DataFrame(empty_gen(), columns=["A", "B"]) tm.assert_frame_equal(df, expected) From 2203ef6f0e05ad1360570c5581ed67732c4bb48e Mon Sep 17 00:00:00 2001 From: Fangchen Li Date: Mon, 12 Oct 2020 19:25:37 -0500 Subject: [PATCH 2/2] CLN/TYP: clean unreachable code --- pandas/core/computation/parsing.py | 1 - 1 file changed, 1 deletion(-) diff --git a/pandas/core/computation/parsing.py b/pandas/core/computation/parsing.py index 8d8da3dd00310..ab7c1a9f0cf23 100644 --- a/pandas/core/computation/parsing.py +++ b/pandas/core/computation/parsing.py @@ -36,7 +36,6 @@ def create_valid_python_identifier(name: str) -> str: # toke.tok_name contains a readable description of the replacement string. special_characters_replacements = { char: f"_{token.tok_name[tokval]}_" - # The ignore here is because of a bug in mypy that is resolved in 0.740 for char, tokval in (tokenize.EXACT_TOKEN_TYPES.items()) } special_characters_replacements.update(