• 0 Posts
  • 124 Comments
Joined 1 year ago
cake
Cake day: June 1st, 2023

help-circle










  • Atomic automatic updates with snapshot creation? Maybe consider opensuse microOS if you are going headless…didn’t quite understand from your description. I have a VPS running microOS that has been doing its automatic updates/reboot thing for a year+ now without a single issue. Opensuse’s rolling stuff works very well, and you get native btrfs and snapper integration out of the box.

    Easy to use reverse proxy - I really like Caddy. Reading/writing the config for that clicks better for me than others.

    I like the novelty of using filesystem tools for backups, but can’t shake the feeling that tools like restic and borg are more widely deployed and battle tested.




  • I’m far from an expert, but I don’t know of rclone doing versioning, or a continuous sync like syncthing. Also haven’t used proton, so take my thoughts with a grain of salt.

    Stage 1 Run rclone config to set up the proton remote. rclone config should take you through a wizard and will eventually ask you to authenticate somehow with the remote. Once that is done and saved, you’ll exit the rclone config wizard and be back at the command line.

    Then you would run a test command like: rclone ls :

    If it worked, you should see a list of files/folders on Proton. If not, you’ll have to go back to rclone config and edit the remote to fix whatever went wrong.

    Stage 2

    Test out copying the folders with a command something like: rclone copy localfile/folder remotename:remotepath

    Do some testing to get the hang of the command, but it is pretty straightforward.

    Stage 3

    I don’t know how many files or how big the files are, but I assume not too many and not too big. I also don’t know which version of Linux you have, but I assume you have access to systemd, cron, or both.

    You’ll make a basic shell script that runs the command you practiced in stage 2. Easy peasy, put it in a text file with a shebang at the beginning, make it executable, and give it a go. It should run exactly how it did when you typed the command out manually.

    Finally, you will write a systemd timer or a cron/crontab entry to execute that script at some frequency.

    So just to summarize:

    1. Setup the proton remote in rclone using rclone config
    2. Test out copying files to proton through rclone
    3. Write a basic shell script that runs the command to copy files from the desired local folders to the desired proton folders.
    4. Use one of the tools on Linux that lets you schedule the execution of scripts to automate running your copy to proton script as frequently as makes sense to you.






  • I don’t think it was the point of your post necessarily, but I did want to mention a couple of things that might make the Linux switch a little easier - if not for you, anybody else reading and agreeing.

    First, distrobox (https://github.com/89luca89/distrobox) is a nifty frontend for podman/docker that I think makes it a little more usable/accessible without having a PhD in devops. Basically helper scripts that create a series of simple CLI commands that let you launch a new environment, enter the new (or old) environment, do stuff in the environment, exit the environment. Keeps your core system’s python separated from your development environment(s). Sacrifices the isolation qualities of containers in favor of usability, so probably ok for dev work on a desktop and not so much for production on a server.

    Also, there are GUI applications for point and click management of distrobox - I use BoxBuddy, which is available as a Flatpak on Flathub, so again no interference with the core system provided it can run podman and flatpak.

    Second, I know the php dev world figured out ways to abstract some docker complexity away with stuff like ddev (https://github.com/ddev/ddev) and lando (https://github.com/lando/lando). I wonder whether other languages have or will build that dev environment in a box abstraction on Docker/Podman/whatever.