
Honestly, unless you can spend more $, one or two USB disks for the mini pc is probably your only choice.
Honestly, unless you can spend more $, one or two USB disks for the mini pc is probably your only choice.
I’m surprised no one mentioned ansible yet. It’s meant for this (and more).
By ssh keys I assume you’re talking about authorized_keys, not private keys. I agree with other posters that private keys should not be synced, just generate new ones and add them to the relevant servers authorized_keys with ansible.
It would not be for me, but they just sent me this chat message which is concerning:
We are currently seeing unexpected growth across Dropbox Advanced, and as a result are currently only able to grant 1 TB per month per team. We understand this may be frustrating and are working to resolve this for our customers.
thank you, just subsribed
750gb upload, 5tb download per day. However this seems to be another limit, maybe file based max sharing or something.
guys are not downloading enough 😅
User statistics
All-time upload: 143.678 TiB
All-time download: 112.403 TiB
All-time share ratio: 1.27
you need to improve your search skills:
It matters only if “the docker hosts external IP” your dns resolves is a public IP. In that case packets travel to the router which needs to map/send them back to the docker hosts LAN IP (NAT-Reflection). With cgnat this would need to be enabled on the carrier side, where you set up the port forwarding. If that’s not possible, split-DNS may be an alternative.
If “the docker hosts external IP” is actually your docker hosts LAN IP, all of that is irrelevant. Split-DNS would accomplish that.
Are you hosting behind NAT / at home? If so, you may need to enable NAT reflection on your router.
It depends on your usage. If you are downloading hundreds of GB per month or more, a block account does not make sense.
Personally I get almost everything off torrents, so I also have some Block accounts which last me many years for the occassional use.
If you only want online file storage and sync, you may want to try Seafile. It’s a lot faster and has been rock solid since 10+ years for me. Not viable if you need some of the many nextcloud exentions though
I ran into the same problem some months ago when my cloud backups stopped being financially viable and I decided to recycle my old drives. For offline backups mergerfs will not work as far as I understand. Creating tar archives of 130TB+ also doesnt sound like a good option. Some of the tape backup solutions looked to be possible options, but are often complex and use special archive formats…
I ended up writing my own solution in python using json state files. It’s complete enough to run the backup, but otherwise very work-in-progress with no restore at all. So I do not want to publish it.
If you find a suitable solution I am also very interested 😅