Hi, I wonder if anyone can help me figure out a resolution for this question. I’ve dug in forums and google search and found various discussions but I can’t actually see how to do what I want / was hoping.
I setup a nextcloud instance for a local non profit I help with IT stuff about 6 years ago. It got modest use for about 3-4 years but due to staff turnover, change of ‘business flow’ and ‘preferred work tools’ they no longer use nextcloud at all. One user had been going into it maybe 3-4x per year to refer to an old file here and there from a prior member of staff / for ‘legacy archival access’ sort of thing.
So, I don’t wish to leave this nextcloud instance alive and on the internet as a security liability / and as a maintenance burden / since it is effectively now retired from service.
My goal is to make a “Good clean complete copy” of the full extent of all data folders for all users on the nextcloud instance. Ideally will be something as simple as possible, ie, so that a non-technical user may browse the file structure and then easily find anything-everything which was stored in the old nextcloud.
I know based on the nextcloud data dir size - we have about 50gig as total footprint of this nextcloud instance. This includes all users, all files, all versions, all trash cans.
I know for example one particular old staff user account has ~15gigs of data. I was able to setup and use the ‘rclone’ tool in a ‘nearby current debian VM’ and harvest all the data for that particular user account
but
there are about 20 user accounts
and this is kind of klunky and gross
Originally when I did the rclone attempt, I foolishly assumed this one user could see ~everything but now after doing more digging I realize
we have a few users with 10-15gig of files
many users with ~trivial use (0-100meg)
misc things in trash and version directories
and this all together adds up to the ‘total footprint’
So, I am wondering if someone can hint to me cleanest way to do this
Is there a way I may create a new “mega-access-migrate-data-user” who has inherently simply got all-data-access. Then I can do my rclone sync / harvest ‘latest copy of all files, all folders’ and that will give me what I want. (ie, I am not fussed about versions and trash)
OR
Is it really just a matter of me manually pulling the data tree out of my nextcloud instance, and then manually doing some data tidy up
we can see in the root of the nextcloud data tree - basically a list of (all users - each has their own folder)
and then inside every user I cache / files / files_trashbin
and I’ll just manually copy-mulch-move-simplify
so that my end product is an external USB HDD
with a structure of
USERNAME_1
USERNAME_2
…
USERNAME_10
and inside each username dir is – all the directories of stuff they worked on
Anyhow. End of the day I think I can just do this. But I am hoping to confirm there is no better cleaner nicer way to do this than a
WinSCP > File transfer pull from root user nextcloud data dir
Harvest to local dir on windows machine with USB storage attached
dump the data in there
once it is all tidied up > copy whole mess to a second drive, so we have no less than 2 copies of the archival data for ‘safe keeping’
Many thanks for help / if you have read this far.
Tim