Wendy,
I’m not aware of any specific code changes since 5.14 that would affect this. (There has been a lot of change to add storage quotas, etc. so it’s a bit hard to tell, but I don’t think the basic code to unzip once has changed. Perhaps others will remember something.)
One possibility would just be that you’re running out of temporary space – normal upload would copy the file, and then unzip to get the inner zip, so you might need ~2x the file size to be successful. (There could be enough space in your persistent file store and not in the directories/on the volume assigned as temporary space).I’m not sure that’s consistent with you seeing that a zip with a gzip inside works unless that results in smaller files (or they have just been trying when there’s other activity using up temp space.). Hopefully there is something in the server.log file that would indicate what’s causing things to hang, or you might be able to see disk usage going to 100% for some volume.
In terms of recommendations, it probably is useful to have the inner file be a .zip as there is a Zip Previewer available that would let you see the 38K files inside. (I don’t think .gz is currently supported.) That may be less relevant if you’re not running that previewer.
The other general recommendation that’s more at the installation config level would be to consider S3 storage, which allows direct upload. That by default doesn’t unzip at all and it avoids any temporary copies that could cause issues on the Dataverse server (not to mention faster/more robust upload). I would say beyond a GB is where that starts to be worth the effort, but it is obviously a significant technical change.
Hopefully that helps. If you get more clues from the log, or have a sharable file that consistently fails on 5.14, let us know and we might be able to identify some other issue.
-- Jim
--
You received this message because you are subscribed to the Google Groups "Dataverse Users Community" group.
To unsubscribe from this group and stop receiving emails from it, send an email to
dataverse-commu...@googlegroups.com.
To view this discussion on the web visit
https://groups.google.com/d/msgid/dataverse-community/05553d16-53a7-451b-825c-6c276d6905abn%40googlegroups.com.
On a Mac - a couple of ways:
After using the "right-click" to Compress a folder..... (which creates a zip'd file of the folder contents),
I use the command line and gzip (which zips a zip'd file):
- gzip FileName.zip => creates FileName.zip.gz
- Then upload the file Filename.zip.gz to Dataverse
You can also use the "zip" command (twice), where "originalFolder" is the unzipped folder:
- zip new.zip originalFolder
- Creates a new file "new.zip"
- Then zip that file: zip newdouble.zip new.zip
- Upload "newdouble.zip" to Dataverse - it will do one "unzip" and leave "new.zip in its place
Here are instructions from the Australian Data Archive (which is also a Dataverse installation) on how to double zip via Windows. Note - I have not tried this personally:
--