Tuesday, 2025-09-16

*** dmellado2 is now known as dmellad07:14
*** dmellad is now known as dmellado07:14
*** janders1 is now known as janders13:12
*** diablo_rojo_phone is now known as Guest2664313:12
opendevreviewsean mooney proposed zuul/zuul-jobs master: make disto and pypi mirror configuration condtional  https://review.opendev.org/c/zuul/zuul-jobs/+/96136913:28
*** janders0 is now known as janders13:49
*** open10k8s_ is now known as open10k8s13:49
*** clarkb is now known as Guest2667313:57
*** acoles_ is now known as acoles14:00
fungithe mirror.ubuntu volume move has been going for 26 hours at this point14:41
fungiif it only takes as long as mirror.ubuntu-ports then it should finish by the time for our meeting, though i have a feeling it will take longer since it still has bionic packages14:42
*** Guest26673 is now known as clarkb14:46
mnasiadkaAh, that's why some Kolla Ubuntu builds are failing ;-)15:13
fungimnasiadka: which ones? arm64 on bionic?15:14
mnasiadkanope, some stale mirror it seems for Noble15:15
mnasiadkahttps://zuul.opendev.org/t/openstack/build/30691c49d8cb485093a2958a59ea2b25/log/kolla/build/000_FAILED_manila-share.log15:15
fungithe ubuntu mirror should only be at most 28 hours stale right now15:16
mnasiadkaI think the discrepancy is Kolla using ubuntu cloud archive upstream and the standard Noble repos from mirror15:18
mnasiadkaIs there a UCA mirror in OpenDev - or should we think of having one after all these moves?15:18
fungiaha, yeah uca may have newer dependencies than are in our mirror if they updated libsqlite3-0 in the past day15:19
fungipackages.ubuntu.com is being really obstinate in returning content for the updated package15:22
fungilooks like it may even be disjoint on their own mirror network15:22
clarkbmnasiadka: we have a uca mirror but it is updated separately of the main repos15:22
clarkbso the same disconnect can occur I think15:23
mnasiadkaoh well15:23
mnasiadkaeven if these would run in the same cron script - I assume disconnects can occur15:24
fungiwell, hopefully the mirror will be up to date by tomorrow15:24
fungihttps://changelogs.ubuntu.com/changelogs/pool/main/s/sqlite3/sqlite3_3.45.1-1ubuntu2.5/changelog is persistently returning a 404 response15:27
fungiseems like this may be an update in progress15:27
clarkbhttps://opendev.org/openstack/kolla-ansible/commit/3be5a3852def1b4580c439809067bda7a2c49730 exists now so something must've triggered broader replication for kolla-ansible15:50
fungieventually consistent replicas15:56
clarkbfungi: I rechecked https://review.opendev.org/c/opendev/system-config/+/958666 so that we can get that ready for merging when you're ready on the afs side of things15:59
fungithanks!15:59
opendevreviewMerged opendev/system-config master: Expand old chrome UA rules  https://review.opendev.org/c/opendev/system-config/+/96039916:04
clarkbI'm starting to look very briefly at summit scheduling and I think Friday night might be the only night I personally have free to try and do a get together16:08
clarkbso if there is interest in having an opendev (and probably zuul) dinner type thing try to keep friday night free!16:08
clarkbfungi and corvus are going to be there not sure who else is ^ but pass that along I guess. Note I don't think I'll be organizing anything super formal more of just a lets aim for that day and then find something that works16:09
*** dan_with__ is now known as dan_with16:10
corvus"try to coalesce into a social blob and imbibe sustenance" sounds like a plan!  :)16:12
fungiinfra-root: 960399 just finished deploying (to gitea, mailman, static, and zuul), so keep an eye out for any reports of rejected web requests on those sites/services16:21
fungii guess we don't update gerrit with those16:21
clarkbwe've been selective in deploying that to things that have had struggles with the crawlers I think16:22
fungii couldn't remember which services used it and whether we remembered to trigger deployment jobs for all of them when the file changes16:26
fungiinteresting... there's a lists01.openstack.org/main01 cinder volume in rax-dfw which isn't in use, created_at is the same day as the lists01.opendev.org server instance17:49
fungiaside from having the wrong name, it's also sata not ssd17:50
clarkbits possible the sata volumes are also speedy? we can probably profiel that somewhere. Or just use ssd because we know we're sensitive to iowait and want to avoid the problem as much as possible17:50
fungii'm checking the current volume of data in /var/lib/mailman to confirm whether the minimum 100gb volume size will still be sufficient or if we need something larger17:51
clarkbdon't forget to check the database size too17:51
fungiit's also in there17:51
clarkbah ok17:52
*** Guest26643 is now known as diablo_rojo_phone17:52
fungi/var/lib/mailman/database on the host is mounted as /var/lib/mysql in the mariadb container17:52
fungi41G     /var/lib/mailman18:13
fungiso 100gb should be plenty18:13
clarkbmight want to give some headroom?18:15
clarkbI think that the indexes in particular maybe a concern if xapian is anything like lucene18:15
clarkbI guess with lvm we can add a second volume and expand the fs without too much fuss if that becomes necessary18:16
fungithat seems like ample headroom anyway, but yes we can always use pvmove onto a larger volume if we want with no downtime18:16
fungii've created a 100gb ssd volume named lists01.opendev.org/main01 and attached it to the server as /dev/xvdb, put a logical volume on it using all available extents and formatted it ext418:18
fungii've temporaily mounted /dev/main/mailman at /mnt and will get an initial rsync going to it from /var/lib/mailman18:19
fungi`rsync -Sax --delete --info=progress2 /var/lib/mailman/ /mnt/` is running in a root screen session on lists0118:22
fungiin other news, the mirror.ubuntu volume move is approaching the 30 hour mark, still going from what i can tell18:26
fricklerlooks like the mirror updates are still running, just the vos release is failing for them? might be better to stop those until the move is finished?18:41
fungiwe could temporarily comment out the ubuntu reprepro cronjob and put the mirror-update server in the emergency disable list, though it's probably not got much longer to go at this point18:52
fungiomw to an early dinner, bbl19:45
opendevreviewClark Boylan proposed opendev/system-config master: Switch generic container role image builds back to docker  https://review.opendev.org/c/opendev/system-config/+/96141019:47
clarkbthis is the chagne to flip our default image builder back to docker (everything should already be using docker due to explicit overrides, this is just ensuring our default matches our intention going forward)19:47
clarkbfungi: I left some notes on your etherpad. Overall looks good just some minor things to consider19:51
fungii suppose i could rename to something other than /var/lib/mailman.old that wouldn't risk getting backed up in the few minutes between those steps21:25
fungiit's on the rootfs so a rename to basically anywhere would still be atomic21:26
clarkbya that would also work21:26
clarkbI just don't want us to accidentally make backups carry data we don't want21:26
fungii'm not super familiar with what we include/exclude. what path would you recommend?21:28
clarkbfungi: https://opendev.org/opendev/system-config/src/commit/03ba936444f4b2e42b08981b549e53b90b267814/playbooks/roles/borg-backup/defaults/main.yaml this is the deafult set of rules21:29
clarkb/var/cache might be a good choice21:29
clarkbI think that isn't mounted as tmpfs or anything21:29
clarkband it is in the exclude list21:30
fungirunning fio on /mnt now following the same options you last used on the rootfs for comparison21:32
clarkbnote it created the four fungi-test.* files. Dont' delete them until you're done as it will reuse them if you run multiple passes21:33
fungiREAD: bw=111MiB/s (116MB/s), 111MiB/s-111MiB/s (116MB/s-116MB/s), io=3337MiB (3499MB), run=30036-30036msec21:33
clarkbfungi: if you look near the top of the output it gies what I think is a better summary as it includes iops21:33
fungibw (  KiB/s): min=109654, max=115386, per=100.00%, avg=113857.77, stdev=330.621:34
fungi2, samples=23921:34
fungiiops        : min=27412, max=28846, avg=28464.11, stdev=82.68, samples=23921:34
fungithat?21:34
clarkbno it looks like read: IOPS=814, BW=3260KiB/s (3338kB/s)(95.8MiB/30085msec)21:35
clarkbso it includes uops and bw on the same line which I feel is a better summary21:35
fungiread: IOPS=28.4k, BW=111MiB/s (116MB/s)(3337MiB/30036msec)21:35
clarkbfungi: and was that for read or randread?21:36
clarkb(I think they both say read: in the output unfortunately)21:36
fungii don't see a randread in the output21:37
clarkbfungi: its in the command you run --rw=read or --rw=randread iirc21:37
clarkbit selects which type of test to perform.21:38
fungioh, i ran it with --rw=read21:38
fungisince that was the last way you ran it21:38
clarkbya I was collecting both sets of data21:38
clarkbread is sequential and randread is random. I think the randread value is more improtant here (at least it was the test that did not great on the existing disk)21:39
clarkbso just rerun the command with the different test selected and it should spit out a similar report21:39
clarkbbut both pieces of info are helpful21:39
fungihttps://paste.opendev.org/show/bwMpEzmUCJPT7TCZTNXR/21:41
fungithat's for randread21:41
clarkbthat looks very similar to mirror randread `read: IOPS=20.0k, BW=78.2MiB/s (82.0MB/s)(2346MiB/30004msec)`21:41
clarkbso I don't think switching to the performance flavor is going to be any better for us21:42
fungiso maybe ssd and sata performance are similar21:42
clarkbfungi: I ran that on the mirrors root disk not its cache volume fwiw21:42
clarkbin any case that is much much better than the current randread ~1k IOPS number so yes this should be an improvement21:42
fungiah, so could also be that the rootfs on performance is ssd or has similar performance characteristics21:42
clarkbyup21:42
fungiplan in the etherpad is updated based on your feedback21:46
clarkbwhat does rsync -S get us here? Just smaller transfers?21:50
clarkbmy main concern is that some things may not like that. I know we can't sparse out swap files anymore for example. Wonder if mariadb in particular might have issues with that21:50
fungioh, old habit, i can drop it and rerun the pre-sync21:50
fungiedited the plan21:51
clarkbya just thinking if the database is preallocating space for $reasons that might not produce a happy result (I have no evidence this is the case but do know databases do wizardly things with disk stuff)21:51
clarkbI think the plan lgtm now21:51
fungii want to say ~ancient rsync did not preserve sparseness and the meaning of the -S option changed over time21:52
clarkbmy local manpage says `turn sequences of nulls into sparse blocks`21:54
clarkboh one other thought: Maybe we shutdown everything, then do a mysqldump backup, then shutdown mariadb21:54
clarkb*shutdown everything but mariadb21:54
clarkbthat way we have a database backup that doesn't rely on the underlying backing files (in theory since we're moving around on the same fielsystem with no applications running that is safe anyway but belts and suspenders if rsync does something we don't expect)21:55
clarkbI don't know how extra careful we want to be21:56
fungior maybe we just don't do the final rm of the original directory we moved until we're sure all is well?21:59
clarkbya that is another good option21:59
clarkbsince a mv should be largely transparent if we need to shift things back22:00
fungiyes, the files are being left entirely untouched, as long as we assume rsync doesn't modify the source side in any way, which has always been my understanding22:01
clarkbya I think we can probably trust rsync there22:03
fungii made a note22:05
funginapkin math, taking the rsync slow reads from the rootfs into account, is ~30 minutes for the outage22:06
fungiinbound deliveries should get temporarily queued by exim while mailman is offline, so posts will only be delayed in theory22:07
clarkband in theory they should retry for like a day or two right?22:07
clarkbbut I think if we announce it even if some deliveries fail thats ok22:08
fungimaybe end of this week, friday afternoon my time? say... 20:00 utc?22:13
fungimail volume tends to be really low then22:14
clarkbI should be around then22:14
fungii'll go ahead and send something to service announce for that time22:18
fungisent22:30
clarkbfungi: any idea if I need to moderate it through? I don't see it in my inbox but I also don't see the moderation request either22:33
clarkbmaybe I just need to be patient22:33
fungii received it right away22:33
clarkbI just got it22:34
clarkbso yes patience was required22:34
fungii blocked out 19:30-21:30 utc on my calendar for it, just in case22:36

Generated by irclog2html.py 4.0.0 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!