[08:46] cjwatson, guruprasad: It's just risk mitigation, yeah. It's only come up a handful of times, due to fairly low schema velocity, but we've violated it for simple cases. [08:47] If you're just adding a table here and a column there I've no problem combining them. [08:51] wgrant, thanks for the context. I ended up doing 3 back-to-back schema deployments one after the other with the preparations done once, and 3 separate outage windows within a few minutes of each other. [08:51] It was late on a Friday night and I didn't want to find out what might go wrong and then be forced to fix it. [08:51] :) [08:51] That is also fine now that the prep process is so much faster. [08:51] The latency to prep the new code etc. used to be 40 minutes when that process was written. [08:53] Oh okay, that is useful to know because I hear a lot of suggestions/complaints from new people who want to throw away all the "legacy cruft" and fix it with a "modern" replacement :) [08:54] wgrant, if you have a few minutes to spare, can you check my question in https://irclogs.ubuntu.com/2024/09/23/%23launchpad-dev.txt? [08:54] This is not urgent. [08:54] Apologies, I've been out and about the last couple of weeks, let me see. [08:55] guruprasad: I almost certainly removed the walblock pipe and added -j8 to the pg_restore. [08:55] After your departure, we managed to restore the qastaging database from the production copy in ~2+ days time without any issues. The staging one (with all the known iops issues) is still pending. So your tweaks might be helpful to have. [08:56] he says, before reading the third line where he clearly removed walblock [08:56] Thanks, this is very helpful [08:56] The full set of changes had fancier stuff to do it in three phases, but that didn't end up being a big win. [08:56] But removing walblock and adding -j`nproc` is always a good win. [08:57] Even without any of these fixes the qastaging restore went on rails and succeeded. But these are good to have. [08:59] Oh nice. [08:59] How long did it take? [08:59] Like 6 days? [09:00] No. 2+ days. I started it on a Friday evening and it finished in the middle of Monday. [09:00] The qastaging database is on a more performant, much less-lightly-loaded aggregate. [09:00] Ah right yes [09:00] We checked with IS about moving the staging database to the same agg too, but they mentioned that there is no capacity to do so. [09:00] staging primary was on staging AZ1 iirc [09:01] So we are stuck with the status quo [09:01] so was absolutely dreadful %st [09:01] Is qastaging on the prod agg, or just on a less bad AZ? [09:01] I think it is just in a less bad AZ. [09:02] I also asked IS about moving the staging database to a prod agg, and they pushed back strongly and recommended against doing so. [09:03] Particularly since they were concerned about such a change affecting other production/"critical" services. [09:03] LP staging is much more critical than most of them :P [09:03] legit [09:04] God, I miss that snark :) [09:09] :'( [09:10] And it was right on time for the weekly infrastructure meeting, which is happening now :)