-
Notifications
You must be signed in to change notification settings - Fork 7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Unable to load CelebA dataset. File is not zip file error. #2262
Comments
This has nothing to do with the loader. We can get the same result with from torchvision import datasets
dataset = datasets.CelebA(".", split="train", download=True,) The underlying problem was reported in #1920: Google Drive has a daily maximum quota for any file, which seems to be exceeded for the CelebA files. You can see this in the response which is mindlessly written to every
@ajayrfhp The only "solution" we can offer is to tell you to wait and try again, since we have no control about your issue. You can ask the author of the dataset to host it on a platform that does not have daily quotas. If you do and he goes through with your proposal please inform us so that we can adapt our code. @fmassa We should check the contents of the response first before we write them to the files and raise a descriptive error message. |
vision/torchvision/datasets/utils.py Lines 160 to 167 in a85f21d
Maybe it is as easy as checking the Problem I see is that we need wait until we have a day where the quota is exceeded and fix it instantly. Furthermore, I have no idea how to test this. |
I see. Thanks, I will download at a later point then. |
@pmeier your fix sounds good to me, but indeed, this might be difficult to test. |
@fmassa I suggest we wait for another issue raising this problem. At least I won't check daily if this quota is exceeded. If there is another issue for this and I miss it or you somehow find a day when we can fix this feel free to tag me in. I'll see what I can do. |
Sounds good, thanks a lot @pmeier ! |
Seems this is a known issue, but wanted to raise this again as per @pmeier 's comment. I didn't want to open another ticket on this though. |
@jotterbach This was fixed in e757d52 but didn't make it in the latest release. |
This is still an issue FYI |
I would just like to add that the authors also include a Baidu drive you can download the data from on their website. The dataset is also available on Kaggle. |
Can a Dataset class (like this) that downloads the data from Kaggle (using the Kaggle API) be a possible solution? |
same |
Run into the same problem. In the original Google Drive shared folder the dataset files are placed in different directories (Anno, Eval, Img), whereas there is no download path indication in celeba.py. This probably causes the error. |
This was fixed in #4109, but the commit is not yet included in a stable release. It will be in the upcoming one. |
This issue is still persisting, is there a way to get the dataset and load it just like we would through |
Problem still exists. (Jun 14) |
the kaggle alternative worked for me |
Hello everyone! Based on this discussion, this steps can help you (for me they perfectly worked):
import torchvision.datasets as dset
img_path = './celeba'
data = dset.celeba.CelebA(root=img_path, split="train", target_type='attr', transform=None, download=False) This tutorial worked for me! |
I had the same issue. I had problems and the installation. |
the celeba loader cannot read celeba in HQ, right? |
In my case it seems that there was a problem when using the wifi. After I connected using the LAN I had no timeouts anymore
|
Problem still exists |
The problem still exists. |
Hello i have uploaded celeba into 🤗 Datasets. |
🐛 Bug
Unable to download and load celeba dataset into a loader.
To Reproduce
Returns
Environment
PyTorch version: 1.5.0+cu101
Is debug build: No
CUDA used to build PyTorch: 10.1
OS: Ubuntu 18.04.3 LTS
GCC version: (Ubuntu 7.5.0-3ubuntu1~18.04) 7.5.0
CMake version: version 3.12.0
Python version: 3.6
Versions of relevant libraries:
The text was updated successfully, but these errors were encountered: