Hello! I was wondering if running periodically a script to automatically pull new images for all my containers is a good or a bad idea. I’d run it everyday at 5.00AM to avoid interruptions. Any tips?

EDIT: Thanks to everyone for the help! I’ll install Watchtower to manage the updates

  • atzanteol@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    6
    ·
    2 days ago

    Depends on how you like to roll. If you enjoy waking up to a service not working then go for it.

    But it very much depends on what containers you’re using and what tags you’re pulling.

  • SteadyGoLucky@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    25
    ·
    3 days ago

    Some apps have breaking changes. If you can restore a complete backup when that occurs, you can recover. Immich is famous for its breaking changes

    • peregus@lemmy.world
      link
      fedilink
      English
      arrow-up
      9
      ·
      3 days ago

      But from the moment that the script updates and breaks something and the moment he realizes it may be too late for some applications.

      For example I host Traccar to track car/vans and in this case some tracks would be lost. Or maybe SyncThing, he may realize days/weeks later that a sync is not working and if he was synching his smartphone pictures with his server and the smartphone is lost/broke/stolen, he may lose days/weeks or even months of pictures.

      I wouldn’t trust a script. Use Watchtower or What’s up Docker

      @tubbadu@lemmy.kde.social

      • tritonium@midwest.social
        link
        fedilink
        English
        arrow-up
        1
        ·
        edit-2
        21 hours ago

        That’s why you you bind mount all the important data and back it up with a proper backup solution like borg. And why you also have a monitoring and notification system that alerts you if a service goes down. I will get a telegram message within 15 minutes of a service going down.

    • tubbadu@lemmy.kde.socialOP
      link
      fedilink
      English
      arrow-up
      15
      arrow-down
      1
      ·
      3 days ago

      Because I was today years old when I found out this beautiful piece of software exist :D

      thank you very much!

    • haui@lemmy.giftedmc.com
      link
      fedilink
      English
      arrow-up
      3
      ·
      3 days ago

      Pretty solid advice.

      One could argue though that a backup script could pull the particular container image right after doing the backup to have maximum coverage.

      If someone is already that adept at scripting to rely on a script to do automatic backups, they can very well pull the new images and clean old ones.

      I‘m one of those who have a backup script and still use watchtower.

  • redxef@feddit.de
    link
    fedilink
    English
    arrow-up
    6
    ·
    3 days ago

    I get a summary once a week of all the updates. I then check the release notes and if nothing needs any changes just run the ansible playbook that updates to those releases. I don’t want to get up and first thing in the morning read alert emails because an update failed over night, so i sit down for 10 minutes once a week.

  • BrightCandle@lemmy.world
    link
    fedilink
    English
    arrow-up
    4
    ·
    3 days ago

    It really depends on the project. Some of them take breaking changes seriously and don’t do them and auto migrate and others will throw them out on “minor” number releases and there might be a lot of breaking changes but you only run into one that impacts you occasionally. I typically don’t want containers that are going to be a lot of work to keep up to date so I jettison projects that have unreliable releases for whatever reason and if they put out a breaking change its a good time to re evaluate whether I want that container at all and look at alternatives.

    So no its not safe, but depending on the project it actually can be.

  • wjs018@lemmy.world
    link
    fedilink
    English
    arrow-up
    8
    ·
    3 days ago

    I used to have my docker updates done automatically. However, as the services I used to run just for myself have started to be used by other people (family, friends), I am less tolerant of having things break. So, instead of something like watchtower, I run diun these days. I have it set up to ping me in a discord channel when a docker update is available. Then, I can actually perform the update when I have time and attention to troubleshoot any issues that may come up.

    • Earth Walker@lemmy.world
      link
      fedilink
      English
      arrow-up
      5
      ·
      3 days ago

      Agree, if you are running containers on a casual or “just for fun” basis then automatic updates are fine. But the more you or others depend on the service running, the more it makes sense to perform an update manually, when you have time to troubleshoot any problems that may arise. Or, even update on a test setup first to identify issues and then update on your production setup.

  • JASN_DE@lemmy.world
    link
    fedilink
    English
    arrow-up
    7
    ·
    3 days ago

    I run a mixed setup, many of the “less important” containers are on watchtower auto-update, the rest on notification (reverse proxy, Nextcloud, etc).

    But I also have many of them on specific branches instead of “latest”.

  • ShortN0te@lemmy.ml
    link
    fedilink
    English
    arrow-up
    4
    ·
    3 days ago

    I recommend, reading the release changelogs actively. For most services you can just put the github release page in an RSS reader to get a notification when a new release hits, so you can quickly look for any breaking changes, also this will give you info about new features.

    I have been using watchtower for a few years. No problems with auto updates so far, but keep your backup handy.

  • adr1an@programming.dev
    link
    fedilink
    English
    arrow-up
    1
    ·
    2 days ago

    I’m using github.com/mag37/dockcheck for this, with its “-d N” argument. There’s a tradeoff between stability and security, you need to decide for yourself. It will also depend on what services you’re hosting. For example, nextcloud and immich would be disastrous under such a regime.

  • GravitySpoiled@lemmy.ml
    link
    fedilink
    English
    arrow-up
    4
    ·
    3 days ago

    That’s what I do as well. Even with immich. It may break but it’s usually just a simple change in the env file

  • schizo@forum.uncomfortable.business
    link
    fedilink
    English
    arrow-up
    4
    ·
    3 days ago

    I have watchtower configured to update most, but not all containers.

    It runs after the nightly backup of everything runs, so if something explodes, I’ve got a backup that’s recent and revertible. I also don’t update certain types of containers (databases, critical infrastructure, etc.) automatically so that the blast radius of a bad update when I’m not there doing it is limited.

    In the last ~3 years I’ve had exactly zero instances of ‘oops shit’s fucked!’, but I also don’t run anything that’s in a massive state of flux and constantly having breaking changes (see: immich).

  • Konraddo@lemmy.world
    link
    fedilink
    English
    arrow-up
    3
    ·
    3 days ago

    Depends on the application really. For example, I don’t need to update Jellyfin and the arrs as soon as the new updates drop. They work just fine and I’m not waiting for any particular fixes.

  • Oisteink@feddit.nl
    link
    fedilink
    English
    arrow-up
    3
    ·
    edit-2
    3 days ago

    Basically why i feel more comfortable with LXC than docker for my home lab services. It feels more like a VM in management.

    We run a good mix of docker, vm’s and bare metal at work; no containers are auto-updated

  • catloaf@lemm.ee
    link
    fedilink
    English
    arrow-up
    2
    ·
    3 days ago

    I’ve been doing it for a few years and haven’t had any issues. The risk/reward decision is yours.