-
Notifications
You must be signed in to change notification settings - Fork 105
I can't export a very large table #181
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Your issue description doesn't include the text of any Python exception. Can you post the actual traceback? |
Good morning. the traceback is: AttributeError: 'NoneType' object has no attribute 'resultFormat' i have over 20000 rows. |
Thanks. What version of |
the version is 2.8.0 |
I am also having the same problem with version 2.8.0:
I rolled back to 2.7.0 and the problem went away. |
Thankyou william |
@maragondestiny Could you reopen this issue this there is a problem on the latest 2.8.0? I also see this when trying to create a table with |
I've reopened. Downgrading is a decent workaround but if this is a regression then we need to solve it going forward. |
|
What's the priority of this? My team would like to use some functionality that does not work in 2.7.0, but this issue makes the newer versions unusable. Let me know if there is any other information that would be helpful. |
is this still an issue for 2.9.3? |
I am still seeing this issue in 3.1.1 |
Issue is indeed still present in v3.1.1. It seems to be caused by the Cloud Fetch download manager (#146) which was added in v2.8.0. As far as I can see, it's simply due to the fact that the timeout for downloads is set to 60 seconds and it's not possible to configure a different value from outside the library. This means that larger datasets will always result in the is_file_download_successful method returning False. The A retry would not fix the issue either, since some downloads might always take longer than 60 seconds. instead, it should be possible for callers to determine what the accepted timeout is for their queries. In case that the timeout is hit, an exception should be raised instead of simply logging to Temporary workaroundDowngrading to v2.7.0 is not necessary! Instead, it's possible to simply disable Cloud Fetch when creating the client:
|
Bingo. I'd strongly advise putting this into a PR and tagging @benc-db or @kravets-levko for review. It's strange that a cloudfetch download could exceed sixty seconds. I wonder if there is a low-bandwidth connection to the cloud provider where the cloudfetch chunk is stored? These file sizes are quite small. But either way, the connector should give more actionable feedback rather than failing silently. Currently |
I can see that there's already an open PR since February which should fix Cloud Fetch, and also allows for the timeout to be configured using an environment variable. @andrefurlan-db, are you still working on it? |
Hello here 👋 We just released v3.3.0 which includes a refactoring of CloudFetch-related code. Please give it a try and let me know if it helped with your issues or not (remember to enable CloudFetch via use_cloud_fetch=True). If you still see any issues - please enable debug logging (see #383 (comment)) and share log output. Thank you! |
I have a very large table, with more than 100 thousand rows and when I try to download it using this method, it gives an error.
I replace the table name because it doesn't matter.
apparently the code is not treated to receive such a large size.
however when I use select * from table LIMIT 5000 it works
The text was updated successfully, but these errors were encountered: