I would like to upload big files to my self hosted Nextcloud AIO.
Iām looking for both general tips on how to do it as well as what to do with the default 10GB PHP max upload limit.
Nextcloud is already deployed and running on my VPS so Iād need to adjust the limitation ideally without reinstalling the Docker Image. I have limited experience with Docker, so step-by-step explanations would be appreciated!
what is this default limit relevant for? I have a 55GB (archived) file Iād like to back up and ideally, not divide it into smaller limited files.
Iām just a little insecure about the next step because Iām not entirely sure how docker works.
the comment in the forum says:
As with any other docker container you need to stop it first, remove it and then adjust the docker run command by adding this environmental varilable to the docker run command. (but before the last line nextcloud/all-in-one:latest).
so as I have NX AIO already deployed and running, what does āremoveā mean in this case?
I have several containers running at the moment, which I can stop in the AIO interface. but this is still not the main container referred to in the comment right?
so would you first stop the containers and then which container would you remove and how (before running the docker run script again with the relevant env vars)?
and also, what are the consequences the aforementioned instructions? by rerunning the new docker script, is it like reinstalling NX altogether and then recreating my user or is it info being kept?
first Iāve tried uploading this file from the built-in Gnome-NX integration, through Gnomeās Files. it stopped imedietly, claiming that the file is too large (I think this is a standard WebDAV?).
then I tried again via in Firefox using NX web client. there the upload started and was running for some time (Iāve tried 3-4 times every time itād run anywhere from 15 mins to an hour) but then itād eventually stop, providing no reason. so I started reading the uploading large files NX article and found the max upload limit.
so what you are saying, since Iām the admin user, I should have no problem uploading files larger than 10GB?
the only option I havenāt tried yet is via the official NX desktop client. I suppose I could do that by placing the file directly in the synced directory on my drive?
Please check your browser console (Inspect) and network tabs. Also check the Nextcloud Server log. There should be some hint about whatās going wrong somewhere in there.
so what you are saying, since Iām the admin user, I should have no problem uploading files larger than 10GB?
Correct, at least from the NC Web Client UI or any any other official NC client app, it should work. Though there are other variables. For example, if youāre using S3 underneath/etc.
The GNOME Files built-in WebDAV client though doesnāt know anything about the chunking approach AFAIK. That one failing makes sense (I use GNOME regularly for most day-to-day stuff, but not for super large files).
I donāt know what youāre use case is. But if you canāt get the chunking capable clients to work, than non-chunking WebDAV clients are even less likely to work.
So Iād focus on figuring out why the Web UI client isnāt working for you first.
ok. let me try to upload it once more and this time Iāll keep an eye on the console. the browser network console is simple but how do I access to the NC console?
what did you mean by āchunking capable clientsā?
Nextcloud Server log is in the Web UI under Administration settings->Logging or on-disk in your configured datadirectory in nextcloud.log (at least by default).
what did you mean by āchunking capable clientsā?
All of the Nextcloud maintained client apps (Web, Desktop, Android, iOS, etc.) have essentially an extension to WebDAV implemented in them to make uploading large files more reliable and faster. The client breaks the file up into chunks of >= 10MB then the uploads the chunks in parallel. Nextcloud Server re-assemblies the chunks automatically into a single file.
I see! so maybe this is where it failed. since the upload always started and seemed to be running nominally for a while. Iād leave my PC and come back to check on it occasionally and itād be running just until the progress bar would no longer be present. so it seems that the breaking down process was taking place up until a certain point.
generally: is there a logistical difference between uploading one large compressed file and the original unarchived folder?
Iām having the consoles open right now and itās seems to be running in order. Iāll check on it in a bit. Iām not quite sure what sort of signs I should look for in the console/inspect>network
@jtr so it keeps failing. Iāve tried it with the desktop client and the behaviour seems to be similar.
it takes a moment to star syncing and then the syncing progresses seemingly in order until it eventually fails and shows an error.
Iām suspecting itās something to do with my VPS provider. if the PHP max_upload_limit isnāt the issue, there might be some kind of limitation applying by them. I just have no idea how to verify it.
Iāve stopped the whole container. these logs that I shared yesterday are from when the container was stopped.
Iām afraid I was asking too much from my VPS actually. I think the bandwidth they provide wonāt allow me to push such a heavy file and itās nothing to do with NC itselfā¦
If you disable the ClamAV container, you still need to disable antivirus in Nextcloud itself. Thatās why I provided that doc link. Itāll, at a minimum, get rid of that error.