-
-
Notifications
You must be signed in to change notification settings - Fork 18.4k
Read csv headers #37966
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Read csv headers #37966
Conversation
cc @twoertwein |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Great work! The only part missing is the documentation update in pandas/core/shared_docs.py and making sure that all functions that take storage_options
work as expected with it. imho I would only add a new test for parquet, the other functions forward everything directly to get_handle
.
we have tons of fixtures eg s3_resources pls try to not reinvent the wheel if u must mock then use monkeypatch prefer not to mock at all |
The default compression on to_parquet is not installed on one of the test environments. It should work just the same with no compression anyway.
I'm sure folks are busy; I just wanted to make sure no one is waiting on me for anything. I think I've addressed the comments so far, but if I'm missing something or I need to do something else please let me know! |
thanks @cdknox very nice |
Glad to help as a long time pandas user! Just double checking, we're good without the change then from |
def read_parquet( | ||
path, | ||
engine: str = "auto", | ||
columns=None, | ||
storage_options: StorageOptions = None, |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This new parameter should maybe be added after use_nullable_dtypes
black pandas
git diff upstream/master -u -- "*.py" | flake8 --diff