-
-
Notifications
You must be signed in to change notification settings - Fork 18.4k
unable to write JSON to S3 use to_json #28375
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Thanks for the report. If you'd like to investigate and push a PR to fix would certainly be welcome |
I think you can try use StringIO if you are on python3, python2 should be BytesIO
|
Why was this issue closed? The problem still persists |
@rohitkg98 issues are closed when the corresponding PR is merged |
@jreback This is still broken using the latest pandas 1.0.5 on Anaconda Python 3.6.10 on the Amazon Deep Learning Ubuntu image. it is clearly trying to do a simple
|
@rjurney this was merged for 1.1 which has not been released |
@jreback thanks, any idea when it will be released? |
likely in july |
@jreback I think with 1.1.0 this functionality is back in fact! however there is an issue when specifying compression. For example, writing a dataframe to s3 like this: Should I create a new issue to tackle this? |
you can create a new issue |
@manugarri, HAve your already created the issue? Can you link it? Thanks |
@imanebosch i did not, i ended up using dask directly. |
Still facing this issue |
Code Sample, a copy-pastable example if possible
S3 paths work for reading and writing CSV. However, I am not able to write json files using the to_json method. Reading json from an S3 path seems to work just fine.
Problem description
Expected Output
None. Expected output is no error and the file is written to the s3 bucket.
Output of
pd.show_versions()
[paste the output of
pd.show_versions()
here below this line]INSTALLED VERSIONS
commit: None
pandas: 0.24.2
pytest: None
pip: 19.2.2
setuptools: 41.0.1
Cython: None
numpy: 1.16.4
scipy: 1.3.1
pyarrow: None
xarray: None
IPython: 7.7.0
sphinx: None
patsy: 0.5.1
dateutil: 2.8.0
pytz: 2019.2
blosc: None
bottleneck: None
tables: None
numexpr: None
feather: None
matplotlib: 3.1.0
openpyxl: 2.6.2
xlrd: 1.2.0
xlwt: None
xlsxwriter: None
lxml.etree: None
bs4: None
html5lib: None
sqlalchemy: None
pymysql: None
psycopg2: None
jinja2: None
s3fs: 0.2.1
fastparquet: None
pandas_gbq: None
pandas_datareader: None
gcsfs: None
The text was updated successfully, but these errors were encountered: