-
-
Notifications
You must be signed in to change notification settings - Fork 18.6k
BUG-19214 int categoricals are formatted as ints #24494
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Changes from 9 commits
a2593f0
36e16e1
ee22299
e6ef56e
97472b3
6938616
ee5ab83
bbd2ff0
2771401
2bb35dc
f8e67f9
4057932
7b46fec
7cc02c1
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -1113,6 +1113,36 @@ cast from integer dtype to floating dtype (:issue:`22019`) | |
...: 'c': [1, 1, np.nan, 1, 1]}) | ||
In [4]: pd.crosstab(df.a, df.b, normalize='columns') | ||
|
||
Formatting Categorical Integer Data With ``NaN`` Values | ||
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ | ||
|
||
Categorical integer data with ``NaN`` values will be formatted as integers | ||
instead of floats. :meth:`Series.to_numpy` is not affected (:issue:`19214`) | ||
|
||
*Previous Behavior* | ||
|
||
.. code-block:: ipython | ||
|
||
In [3]: pd.Series([1, 2, np.nan], dtype='object').astype('category') | ||
Out[3]: | ||
0 1.0 | ||
1 2.0 | ||
2 NaN | ||
dtype: category | ||
Categories (2, int64): [1, 2] | ||
|
||
In [4]: pd.Categorical([1, 2, np.nan]) | ||
Out[4]: | ||
[1.0, 2.0, NaN] | ||
Categories (2, int64): [1, 2] | ||
|
||
*New Behavior* | ||
|
||
.. ipython:: python | ||
|
||
pd.Series([1, 2, np.nan], dtype='object').astype('category') | ||
pd.Categorical([1, 2, np.nan]) | ||
|
||
Datetimelike API Changes | ||
^^^^^^^^^^^^^^^^^^^^^^^^ | ||
|
||
|
@@ -1653,6 +1683,7 @@ Reshaping | |
- :meth:`DataFrame.nlargest` and :meth:`DataFrame.nsmallest` now returns the correct n values when keep != 'all' also when tied on the first columns (:issue:`22752`) | ||
- Constructing a DataFrame with an index argument that wasn't already an instance of :class:`~pandas.core.Index` was broken (:issue:`22227`). | ||
- Bug in :class:`DataFrame` prevented list subclasses to be used to construction (:issue:`21226`) | ||
- Calling :func:`pandas.concat` on a ``Categorical`` of ints with NA values now causes them to be processed as objects (formerly coerced to floats) (:issue:`19214`) | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. This only applies when concating a categorical with a different dtype, right? If I concat two integer cats with the same dtype, it’s still categorical right? |
||
- Bug in :func:`DataFrame.unstack` and :func:`DataFrame.pivot_table` returning a missleading error message when the resulting DataFrame has more elements than int32 can handle. Now, the error message is improved, pointing towards the actual problem (:issue:`20601`) | ||
|
||
.. _whatsnew_0240.bug_fixes.sparse: | ||
|
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -240,6 +240,15 @@ def test_categorical_repr_datetime_ordered(self): | |
|
||
assert repr(c) == exp | ||
|
||
def test_categorical_repr_int_with_nan(self): | ||
c = Categorical([1, 2, np.nan]) | ||
c_exp = """[1, 2, NaN]\nCategories (2, int64): [1, 2]""" | ||
assert repr(c) == c_exp | ||
|
||
s = Series([1, 2, np.nan], dtype="object").astype("category") | ||
s_exp = """0 1\n1 2\n2 NaN\ndtype: category\nCategories (2, int64): [1, 2]""" # noqa | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. You can use parenthesis to wrap this line. |
||
assert repr(s) == s_exp | ||
|
||
def test_categorical_repr_period(self): | ||
idx = period_range('2011-01-01 09:00', freq='H', periods=5) | ||
c = Categorical(idx) | ||
|
Uh oh!
There was an error while loading. Please reload this page.