Curious to know how many people do zero-downtime deployment of backend code and how many people regularly take their service down, even if very briefly, to roll out new code.

Zero-downtime deployment is valuable in some applications and a complete waste of effort in others, of course, but that doesn’t mean people do it when they should and skip it when it’s not useful.

  • koreth@lemm.eeOP
    link
    fedilink
    English
    arrow-up
    10
    ·
    edit-2
    1 year ago

    Answering my own question: My systems do zero-downtime deployment. Some of my services are managed using ECS and some using custom deployment scripts.

    It’s interesting that people mostly focus on the mechanics of launching the new code. To me, the interesting thing about zero-downtime deployment is what happens while the release is in progress, when there will be a mix of the old and new code versions accessing the same resources (databases, microservices, etc.) at the same time.

    For example, you don’t want to just drop a previously-mandatory column from a SQL database: even if your new release no longer references the column, the new code will break if you deploy code before updating the database, and the old code will break if you update the database before deploying code. Obviously there are ways to do this kind of thing (roll out the change in small backward-compatible steps) but they’re extra work and can be easy to get wrong even if you’re using ECS to launch the code. Whereas, if you’re allowed to take downtime, you can do it all in one step without worrying about mixed-version environments.

    • pohart@lemmyrs.org
      link
      fedilink
      English
      arrow-up
      5
      ·
      1 year ago

      if you’re allowed to take downtime, you can do it all in one step without worrying about mixed-version environments.

      You don’t need to wiry about mixed version environments but you need to worry about whether you can roll back your changes without loss of data. It’s not as hard but it seems to get overlooked if there haven’t been any bad deployments lately.

    • nous
      link
      fedilink
      English
      arrow-up
      3
      ·
      1 year ago

      On the flip side, if something goes wrong and your service is backwards compatible you can roll back without any more issues. If you allow downtime and backwards incompatible changes rollback can cause even more problems and result in far longer outages and lots of very stressed programmers.

      You should always be able to roll back code changes. And zero downtime deployment are not that hard to do if you are already enforcing that.

  • nickel
    link
    fedilink
    English
    arrow-up
    7
    ·
    1 year ago

    Zero-downtime for us using Kubernetes. It’s built-in. Deployment gets updated, new pod comes online, once it’s healthy, the old pod goes offline.

    We do have a little code to handle graceful shutdowns to properly finish any active requests before going offline, but that was a trivial addition.

  • Jim
    link
    fedilink
    English
    arrow-up
    6
    ·
    1 year ago

    For our batch workflows, we do have downtime on deploys. It’s by design because 0 downtime doesn’t add any value. Downtime is usually 5 to 10 minutes. For our services, we rely on lambdas or kubernetes rolling deployments so no downtime.

  • funbike
    link
    fedilink
    English
    arrow-up
    6
    ·
    1 year ago

    Zero downtime deployments can get very complex for heavy usage apps, such as blue-green deployment.

    We decided to avoid the complexity with some practical workarounds.

    • Most deployments happen at 4am. “develop” branch merges deploy at 4am, and “master” branch merges deploy immediately.
    • We force browser refresh if the front end detects the back end has had breaking changes. We attempt to re-populate form field values.
    • During database migrations, we send 503 with Retry-After header in response to POSTs. Our client code knows to wait for that time and try again. If the time is too long, the user gets a friendly message that it will try again in X seconds. GETs are handled by an available read-replica, if possible.
    • hascat
      link
      fedilink
      English
      arrow-up
      3
      ·
      1 year ago

      We force browser refresh if the front end detects the back end has had breaking changes. We attempt to re-populate form field values.

      Do users not find this disruptive?

      • funbike
        link
        fedilink
        English
        arrow-up
        3
        ·
        edit-2
        1 year ago

        Yes, but it’s a very rare event. Maintaining state (form fields) makes it less of an issue. As I said, most deploys are at 4am at extremely low usage (usu zero), and even then a refresh is only needed if the backend has had breaking changes. A severe bug requires a mid-day deploy, but in my experience most severe bug fixes are only a few lines and therefore aren’t a breaking change so don’t require a refresh.

        Our way wouldn’t work well if you had 24 hours of heavy load, but most apps I’ve written have been US-only with low nightly usage (HR, K-12 admin, power grid, medical).

  • rickerbh@lemmy.nz
    link
    fedilink
    English
    arrow-up
    4
    ·
    1 year ago

    We have a clustered/load balanced application and do zero downtime deployments with elastic beanstalk on aws. I’m uncertain as to how popular elastic beanstalk is, but it makes managing this sort of stuff really easy.

  • nibblebitA
    link
    fedilink
    English
    arrow-up
    4
    ·
    1 year ago

    Whenever possible, I’ve run projects to have zero downtime deployments. Multiple stateless instances behind a load balancer. Deploy one instance at a time, run a health check and move traffic to the fresh instances. Most cloud providers often have these out of the box. Database migrations are run well in advance. New functionality is hidden behind feature flags.

    Zero downtime is nice, but the real benefit is that you force the teams to really think about deployments as migrations to accomplish this policy.

    Your instrumentation and alerting need to be top-shelf you need to automate deployments fully, which means you can fully automate rollbacks.

    The downside is that you have to build everything twice, deployments are slower and there is a significant descaffolding.

    But that’s a small price to pay not to be on call outside of business hours to deploy.

  • Hexarei
    link
    fedilink
    English
    arrow-up
    2
    ·
    1 year ago

    I just use platforms that have it built in.

  • pohart@lemmyrs.org
    link
    fedilink
    English
    arrow-up
    2
    ·
    edit-2
    1 year ago

    I do not. I’ve got apps that go unused 5pm-9am or 1am-9am depending on the night, apps that have lower usage 5pm-8am and one app that’s basically unused 8am-5pm. Each one gets redeployed in its down time, with hours extra to rollback any problems.

    Actual downtime is usually around twenty seconds which is fine except for emergency midday builds.

    I’ve been trying to get zero downtime deployments, but it’s hard to justify the extra complexity when we’ve got such open service windows. Also, we’re likely to have more downtime from an ISP service outage than our midday builds.

    Other teams have much shorter service windows but deployments that take the whole window.

  • Rev@ihax0r.com
    link
    fedilink
    English
    arrow-up
    2
    ·
    1 year ago

    Yeah zero downtime. You ship out the new features but gate them using some system you can control. When all the new features are shipped you turn up the new features until it gets to 100%. This lets you observe the real world behavior of the new features if they don’t cache well or cause 500s or what have you you can turn it off without having to ship new code.

    Also if you keep all these feature flags, if you have a situation where you have capacity problems you can turn down features for the survival of the service as a whole.

  • Ducks@ducks.dev
    link
    fedilink
    English
    arrow-up
    2
    ·
    1 year ago

    I write data pipeline code and there is zero downtime. We use kafka to buffer messages from dozens of producers to dozens of consumers on kubernetes.

  • DinosaurSr
    link
    fedilink
    English
    arrow-up
    2
    ·
    edit-2
    1 year ago

    Yep. All of them. Each release for a about ~40 legacy asp .net webforms apps

  • Kuiche
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    Very rarely. Most of our services are ECS which will manage rolling deployments. Older/legacy systems are manually taken out of load balancer, upgraded then added back.

    Only times I can think of have been backwards incompatible database changes/database engine upgrades. This is rare.