We have a fairly large application (15-16GB image) and we’ve noticed recently that when we deploy, our application has about 5 minutes of downtime where it seems the new application image is “building” before it can handle requests and serve traffic.
- The application deployment succeeds
- The router deployment succeeds
- Pods are Service status: Running
But we’ve noticed immediately post-deploy that if we look in the pods console, each pod has to incrementally build up (as in, go from 1% to 100% of the full image memory commitment), and pod CPU is fully maxed while this is happening. This takes a few minutes, then once it completes and we reach the full memory size of the image, CPU normalizes and the application is good to go.
Did anything change in terms of how Qovery handles image caching, build rate or anything like that in the past month or two? We used to deploy the same size image and application and the new application build was online almost instantly (maybe the router took a few seconds to resolve). Not aware of anything with our config, cluster, DNS or anything else that would be impacting this. As a result we’ve had to be more judicious about timing deploys for late nights or weekends when we can tolerate being down a few minutes.
Wondering if there’s an underlying reason and solution for this? Thanks