Thursday, 2025-08-21

opendevreviewOpenStack Proposal Bot proposed openstack/project-config master: Normalize projects.yaml  https://review.opendev.org/c/openstack/project-config/+/95799502:20
*** clarkb is now known as Guest2459307:31
*** dmellado62 is now known as dmellado13:32
fungiclarkb: yes, i'm keeping them in the emergency list until they're all done14:07
zigoHi there!14:38
zigoI'd like to backport https://review.opendev.org/c/openstack/watcher/+/958207 to unmaintained/zed, though there's no such branch anymore. How can I do, as there's only the EOL tag?14:38
zigo(that's for Watcher's OSSN-0094)14:38
fungizigo: you would carry a local backport14:42
zigoYou mean in Debian? Yeah, I have that already done. Though I would have prefer to share it.14:42
fungiit's really more of an openstack question, not an opendev question, but basically if you need unmaintained branches to stay open for longer then the people serving as caretakers for those branches would probably appreciate the help14:43
fungiif there are no volunteers to keep up minimal testing for them at least, then they get closed down (tagged and deleted) so that people will stop relying on them14:43
zigoWell, then IMO we should stop destroying branches, and keep them open just in case there's a security problem and people start to care again.14:45
zigoIt's very much ok to keep them as unmaintained/X14:45
zigoI just heard Red Hat people are even backporting to Train. Why not sharing these patches then?14:46
fungizigo: that's an excellent question to ask them14:46
fungibut i suspect it's because their work would violate upstream policies since they aren't backporting to newer branches when doing so14:47
fungiwhich leaves users on those versions without a clear upgrade path to newer versions14:47
zigo*I* do the work, and would be happy to share it for watcher unmaintained/zed.14:48
zigoI guess I should open a new thread in the list about this, since that's not the first time this happens.14:48
fungii will say that from an opendev hosting perspective, we don't want projects leaving an ever-growing pile of branches around because every branch is additional configuration in the ci/cd system and it makes pruning old unused configuration impossible14:50
zigoI think it's very much ok to delete all the CI stuff around it, and just let downstream share patches without the CI.14:51
fungiworkflows, processes and policies are built around many years of an assumption that unused branches will be deleted14:51
zigoCan I quote this ? :)14:51
fungifeel free! i know we've had this discussion ad nauseum, and every time openstack tries to appease you on this by coming up with new ways to leave branches open for longer you still complain14:52
fungithe latest attempt is the unmaintained branch policy, which has put a lot of additional strain on project maintainers and our systems14:53
zigoI do each time there's a new security fix that needs backporting, and we have no space to share work. :)14:53
fungiwell, basically the technical committee came up with a way for interested downstream stakeholders to volunteer to take care of those branches, but when nobody volunteers to do that they get closed14:56
fungihttps://governance.openstack.org/tc/resolutions/20230724-unmaintained-branches.html which was further amended by https://governance.openstack.org/tc/resolutions/20231114-amend-unmaintained-status.html14:57
fungiif that's not adequately solving the problem, then #openstack-tc would be a good place to discuss it (or on the openstack-discuss mailing list, but i'd recommend adding at least [tc] in the subject line)14:57
Guest24593re the system strain: it wouldn't be so bad if peopl were actively caring for the branches as presumably the ci jobs would get pruned or updaetd as necesasry to keep things mostly working. The problem is when we create the branch under the assumption it will be cared for, then it is ignored which orphans the system configurations leaving others to clean them up when say opendev wants15:01
Guest24593to drop a test platform. Or if zuul changes some syntax15:01
Guest24593arg I'm a guest again15:01
frickleryeah, we could consider to more actively clean up broken zuul configs, like delete them completely after a while15:03
Guest24593the thing I was trying to argue for was not to open the branches in the first place. Wait until someone volunteers. But I'm not sure what effect that may have on say tempest testing and branch defaults15:04
*** Guest24593 is now known as clarkb15:05
clarkbok sorry about that I am me again15:05
clarkblooks like rax flex iad3 did use a couple of instances overnight. Not a lot of load there but non zero15:08
clarkbfwiw I'm not sure if we can easily tell zuul to ignore branches either15:08
fungii suppose they could merge a change to replace all their pipeline configs with noop jobs, but that's still configuration on every branch. and even without any configuration on a branch at all zuul still needs to evaluate the branch contents to determine there is no configuration15:11
clarkbexplicitly using a noop config is probably a nice way to represent it for humans if zuul won't complain15:13
clarkbinfra-root https://review.opendev.org/c/opendev/system-config/+/957950 is a relatively straightforward change to drop testing of bionic servers with system-config15:18
clarkbI checked our fact cache and as far as I can tell we don't have any bionic servers any longer15:18
clarkb(then once zuul drops ansible 9 we can drop bionic from zuul-launcher completely and clean up our mirrors etc)15:18
mattcrees[m]Hi all. In the Blazar project, we still have a stable/pike branch available. I understand this is because the branch was made before Blazar was managed by opendev. Does anyone know how we'd go about removing this branch?15:19
clarkbmattcrees[m]: in general the openstack release team has permissions to manage branches within openstack projects. If they want you to clean it up then you'd need extra gerrit acl permissions on the project or have a gerrit admin do it for you15:21
clarkbI would check with them first. I believe they already have tools that script branch cleanups which does the eol tag first then drops the branch15:21
mattcrees[m]I see, thanks clarkb. I'll reach out to them15:22
fricklermattcrees[m]: clarkb: likely the branch was created before release-management was in place, so we'd need to delete it manually anyway. I'll add that to my todo list15:22
fungiyeah, i think this one may fall into a grey area where they've avoided managing existing branches that pre-date a project's inclusion in openstack, but it would be good to confirm with them first15:22
fungiah sounds like we just did ;)15:22
mattcrees[m]Nice, thanks frickler 15:23
fricklerconfirmed, deliverables/pike/blazar.yaml doesn't exist15:23
fricklerfor the zuul config issue, I created https://review.opendev.org/c/openstack/tacker/+/958219 as an example, seems to work fine. waiting for feedback from elodilles but maybe that can be a simple workaround for the pile of zuul config errors we still have15:24
clarkbfrickler: that looks promising15:34
elodillesfrickler: well, tacker has zuul config errors (and broken gates?) even on stable branches. so probably tacker team should start with those, as i guess we don't want to set noop for the whole project...15:35
clarkbnot for the whole project but unmaintained branches seems like a good idea if tyhey are broken since they are well unmaintained15:36
elodillesclarkb: but as i said, stable branches are broken too15:36
clarkbyes and those should be fixed15:37
elodillesi agree15:38
clarkbthe extra old branches present extra problems because they tend to be even less cared after and also rely on old ci constructs that need to go away. Replacing them with noop jobs nicely addresses both problems15:38
elodillesanyway, i'm not against dropping the complete CI on unmaintained in this case, but i feel it a bit drastic when there are broken stable branches, too15:40
fricklerthe thing is we do not notice the broken stable branches when the huge amount of issues for unmaintained branches overwhelm that list. plus, it is an explicit requirement to keep unmaintained branches open, even though we are slacking at enforcing that requirement15:46
fungiyeah, stable branches with broken testing need to be fixed, unmaintained branches with broken testing are supposed to just get deleted15:49
fungibut it's an acceptable compromise to remove the testing on the unmaintained branches instead of deleting them immediately15:50
elodillesfrickler: well, i could name a couple of places where we could add the noop and after that not the unmaintained branches will be the majority of the zuul config errors (like monasca-* and openstack-ansible-tests)15:50
elodillesfungi: yepp, now that frickler is proposed the monasca 2023.1-eol patches we will be one step closer to that.15:51
fungihopefully monasca ceases to be a problem soon (either because the person offering to adopt it fixes the jobs, or because the tc decides to go ahead with retiring it)15:51
elodillesyepp15:53
fricklerelodilles: yes, but monasca will hopefully be retired, so I simply chose the next best other example I found16:00
frickleralso someone seems to be actively working on fixing tacker at least for master https://review.opendev.org/c/openstack/tacker/+/95645816:00
fricklerwhich I haven't seen happen for many of the unmaintained branches (though maybe I didn't look closely enough?)16:01
clarkbok I'm popping out now for the eyeball inspection. I'll be back in a bit16:31
fungihope you come back with as many as you left with!16:31
fungi(or at least as many)16:32
fungilooks like backups failed today on kdc0317:05
fungiaha, i think we install borg into a venv and the python version has changed, so we'll need ansible to recreate that venv. sound right?17:14
fungithe log has a traceback for importlib.metadata.PackageNotFoundError: No package metadata was found for borgbackup17:14
fungifollowing https://docs.opendev.org/opendev/system-config/latest/afs.html#no-outage-server-maintenance for afs01.ord.openstack.org it looks like there are no rw volumes on it, so not moving any before upgrading17:29
fungihaving to rm -rf /var/lib/docker/aufs on these too17:32
clarkbfungi: yes that sounds right17:40
clarkbre having ansible recreate the venv17:40
clarkband then you'll have to do that again with the jump to noble.17:41
fungii have a feeling with the additional work required for the afs01.dfw server, it will make the most sense to upgrade it from focal to jammy and then from jammy to noble, and start working my way back through the noble upgrades on the others17:42
fungithat way we don't have to move rw volumes off and back onto it more than once17:42
clarkbmakes sense17:42
clarkbone thing I wondered about is if the dkms stuff is reinstalling the pacakges on these upgrades in order to rebuild against the new kernels17:43
fungiso basically having it be the last focal->jammy upgrade and then be the first jammy->noble upgrade17:43
clarkbit must as I'm pretty sure this is hwo we upgraded them in the past17:43
fungiand yes it is, i'm wayching it right now17:43
clarkbcool17:43
fungipart of what takes so long with the upgrades17:43
fungias for holding writes to the rw volume on afs01.dfw, i wonder whether we need to put zuul executors on hold too somehow17:47
fungier, rw volumes17:47
fungithe mirror-update server can just be shut down temporarily, but it's not the only system we have doing writes into afs17:48
clarkbisn't that why we change the rw volume to the other srver?17:48
clarkbor do we have to hold writes to do that?17:48
fungioh! it's an either/or in the doc i guess17:49
fungimake sure i'm not misreading that17:49
fungiso if i move rw volumes from afs01.dfw to afs02.dfw then that in theory happens transparently and i don't have to block anything from writing17:50
clarkbthat was my understanding though I have't reread the docs17:51
clarkbbut yes I thought the idea was to always keep the rw volumes up so that we didn't have to stop writers. Do the work only on the ro side and then let it resync17:51
fungiand https://grafana.opendev.org/d/9871b26303/afs seems to indicate that they both have the same amount of available space and sizes17:52
fungiafs01.ord.openstack.org is on jammy now, working on afs02.dfw.openstack.org next and saving afs01.dfw.openstack.org for last17:53
fungiafs seems to be at least functional on afs01.ord17:54
fungistill reporting all its same ro volume sites17:54
fungiall the volumes on afs02.dfw are confirmed to be ro too so no need to move any yet17:55
clarkb"Basically what we need to do is make sure that either no one needs the RW volumes hosted by a fileserver before taking it down or move the RW volume to another fileserver."18:04
clarkbyes I read that as two options are available to us and ensuring all of the RW volumes are on one fileserver and taking down the other avoids needing to stop all the writers18:04
clarkbthe other options requires stopping all writers18:05
fungicool, so i think we're fine here. the main unknown is how long the rw volume moves will take, but hopefully not long since there are synced ro equivalents of all of them18:06
clarkbI wonder if disabling cron jobs on mirror-update will make that go more smoothly/quickly18:11
clarkbit may still be worth doing if not strictly necessary if it speeds the process up18:11
fungicould just... stop crond too18:17
fungior whatever systemd replaced it with18:17
fungiafs02.dfw.openstack.org is up on jammy and afs seems to work there still. now on to afs01.dfw.openstack.org, going to start moving its rw volumes to afs02.dfw.openstack.org18:21
fungii'll move a small one initially and double-check its still functional18:22
clarkb++18:22
fungiafs01.dfw.openstack.org had 57 rw volumes and 55 ro volumes18:25
clarkbthat sounds right if all the rw volumes are there since I think one or two don't have ro volumes18:25
clarkbiirc its ok for those volumes to go down. But its worth double checking18:25
fungidocs-old and mirror.logs18:25
clarkbhrm is mirror.logs the volume that hosts: https://mirror.dfw.rax.opendev.org/logs/ ?18:26
fungiyeah18:26
clarkbif so then we probably do end up having writers to that volume and we need to do something about that18:26
clarkbafsmon and afs-release run super often. Then the others are the mirror cron jobs18:26
fungioh, also the "service" volume only exists rw on afs01.dfw and there are no ro sites18:27
clarkbshould we maybe add an ro volume for mirror.logs on afs02.dfw and then it can become the rw site?18:27
clarkbI think docs-old is unlikely to be an issue. And I'm not sure about server18:27
clarkb*service18:27
fungiand then there's a test.corvus volume which is ro on afs02.dfw but has no rw site at all18:27
clarkbI suspect that is partial cleanup that orphaned the RO volume18:28
fungilooks like there are a few volumes which don't have an ro replica on afs02.dfw18:28
clarkbfungi: some may be on afs01.ord18:28
fungiwell, i mean there are some where the only ro volume is also on afs01.dfw, no other servers have a replica18:29
fungiafs02.dfw.openstack.org has 47 ro volumes18:29
clarkbgot it. I was just calling out that afs02 may be the RO site but also afs01.ord could be18:29
fungiand afs01.ord.openstack.org has 1718:30
corvusi don't need test.corvus18:30
fungii figured, looked like it was just a missed deletion of a replica18:30
fungibut yeah, i'll need to audit the volumes on afs01.dfw to see which ones only have local replicas and no remote ones18:31
fungithis'll take a few minutes18:31
fungiokay, so these are the volumes with no remote replica, residing only on afs01.dfw: docs-old, mirror.logs, service, user18:38
fungithe have remote replicas on afs01.ord but not afs02.dfw: docs, docs.dev, mirror, project, project.airship, root.afs, root.cell, user.corvus18:38
fungiall other volumes with rw on afs01.dfw have a ro replica on afs02.dfw18:39
clarkbI think its ok to have RO in ord and not dfw we just have to make that site the new RW site temporarily18:39
clarkbwhich means the main thing to consider is whether docs-old, mirror.logs, service, and user need secondary sites. I think mirror.logs having a secondary site is a good idea so that we don't have to turn off all the logging and afs monitoring18:40
clarkbI suspect that not having access to user.corvus (because user is down) is something that won't be a big deal. Any idea what "service" is for?18:41
fungii can probably still move its rw volume, it has no ro replica not even locally18:41
clarkboh ya maybe that is the case. I guess in my head you had to have an RO first that gets promoted to RW but that is probably an inaccurate mental model18:42
fungii think it just ends up having to move all the data18:42
clarkbgot it that makes sense. Wheraes if you have an up to date RO copy its a matter of flipping some attributes18:43
fungivos move -id mirror.logs -toserver afs02.dfw.openstack.org -topartition vicepa -fromserver afs01.dfw.openstack.org -frompartition vicepa -localauth18:43
fungi55M     /afs/.openstack.org/mirror/logs18:44
fungiprobably wouldn't take too long18:44
clarkbya and they are in the same region too18:45
fungi5 directories and 707 files18:45
clarkbI'm guessing service and user are even smaller18:46
clarkbdocs-old is maybe old enough to not worry about?18:46
fungiright, i don't think we hook any readers up to it18:46
fungiand we definitely don't write to it18:46
fungiwe kept it just in case18:47
fungimaybe it's time to tar it up and stick the file somewhere for posterity18:47
clarkbbut also if size isnt massive could just move it too18:47
fungialso just realized that on afs01.dfw i have to use 104.130.138.161 instead of afs01.dfw.openstack.org in commands18:49
fungimoving mirror.logs rw volume to afs02.dfw.openstack.org took 16 seconds18:49
clarkbany idea why we need the ip?18:49
fungihas to do with the host lookup18:50
fungivos: server 'afs01.dfw.openstack.org' not found in host table18:50
clarkbah it gets back 127.0.0.118:50
fungiyeah18:50
clarkbI smell lunch so will be afk for a bit18:51
fungi/afs/.openstack.org/mirror/logs/ still has contents so i think we're good with that18:51
clarkbwill you do the same for service and user? seems like a good idea given I'm still not quite sure what is under service18:52
fungii've moved service and user to afs02.dfw as well just now, yes. they took only a few seconds each18:53
clarkback18:53
fungii think it's just an anchor for the various service.foo volumes18:53
fungii'm not going to bother moving docs-old18:53
fungii'll work through the 8 that have ro replicas on afs01.ord next, moving the corresponding rw volumes there18:54
Clark[m]Oh if it is the anchor for those volumes we probably want ro copies on a second server?18:54
fungithough i also wonder why root.afs and root.cell have replicas on afs01.ord but not afs02.dfw18:55
fungii assume those are special in some way18:55
Clark[m]Ord was the original alternative. But the small window size made it unworkable for large mirrors18:56
Clark[m]So we added the second dfw server after and prefer it for new stuff18:56
fungiright, basically i was wondering if those were forgotten18:56
fungisome volumes have remote replicas on both18:56
Clark[m]Or just unnecessary to move them since they are small18:57
fungii'm switching docs rw to afs01.ord now, seeing how long it takes18:57
fungianswer so far: it's definitely not instantaneous (still in progress after ~10 minutes), i've got it running in a root screen session on afs01.dfw in case anyone needs that19:06
fungistill going after half an hour19:30
fungiand still going... i'll grab a bite to eat and check back in, hopefully that'll give it enough time. bbiab19:48
clarkbI've been doing laptop surgery. New wifi card seems to work a lot more reliably than the old one. Now to reinstall everything to clear out the mess of debugging steps I ahd previously applied20:37
*** mtreinish_ is now known as mtreinish20:37
*** keekz_ is now known as keekz20:39
fungiand the docs volume move is still in progress20:46
*** ShadowJonathan_ is now known as ShadowJonathan20:55
*** keekz_ is now known as keekz20:55
*** dviroel_ is now known as dviroel20:55
*** clayg_ is now known as clayg20:55
*** naskio_ is now known as naskio20:55
*** dan_with__ is now known as dan_with_20:55
*** thystips_ is now known as thystips20:55
*** rpittau_ is now known as rpittau20:55
*** tonyb_ is now known as tonyb20:55
*** gmaan_ is now known as gmaan20:55
*** clarkb is now known as Guest2467920:58
Guest24679is anyone else having conenctivity issues to oftc or is it more likely on my client side?21:39
Guest24679the other irc networks I'm connceted to don't seem to be bothered. But maybe its an ipv4 vs ipv6 problem21:39
tonybmy connection dropped about 40 mins ago but had been stable since then.21:40
fungii think everyone fell off the wagon21:41
fungiseems the wheels are back on now21:41
Guest24679ack this is just the third time in 2 days so I've started to wonder if it was me or the servers21:41
fungiit's the first time it's gotten me, fwiw21:42
fungior at least that i've noticed21:42
fungihttps://meetings.opendev.org/irclogs/%23opendev/latest.log.html also hasn't been updating that i can see21:42
Guest24679taking longer this time because I can't just identify I have hto igure out how to ghost myself21:43
*** Guest24679 is now known as clarkb21:44
fungiah, though the raw version at https://meetings.opendev.org/irclogs/%23opendev/%23opendev.2025-08-21.log has content at least21:44
clarkblooks like reclaim and regain are aliases of one another21:45
clarkbfungi: I assume afs is still working on that docs volume?21:47
corvusi'm going to restart the launchers, schedulers, and web in order to pick up some zuul provider configuration syntax changes21:48
clarkback21:49
fungiclarkb: yeah, it almost feels like having a ro replica at that location doesn't speed up moving the rw there at all and it's just transferring all the data anyway21:49
clarkbfungi: I wonder if that is because we're publishing docs updates via zuul jobs regularly21:50
fungino idea, if it doesn't finish before i knock off in a bit, i'll pick it back up in the morning. at least i ran it under time so i'll know how long it took to finish21:51
clarkboh good idea. I wouldn't have done that :)21:52
corvusit'd be cool if a shell would let you retroactively "time" something21:53
corvuslike "time %1"21:53
fungiyeah, seems like it wouldn't take much for a shell to track that internally21:53
corvusapparently fish has a $CMD_DURATION variable set after each interactive command21:54
fungioh cool!21:54
tonybhistory -w and look at the timestamps21:54
fungithat's a tempting selling point for it21:55
fungitonyb: that gets you start time but not end time, i think?21:55
fungiso if you catch the process right after it terminates that's mostly usable21:55
fungiif you leave and come back later and want to know when the process ended, i don't think history gets you that21:55
tonybtrue.   it'd be an approximation21:56
corvusi've been trying out fish on one of my machines... i'm growing addicted to its command completion21:56
clarkbtonyb: corvus anyone want to weigh in on https://review.opendev.org/c/opendev/system-config/+/957950 which gets system-config out of needing ansible 9 overrides?21:57
tonybyou could add something to a prompt_command to record timestamps21:57
tonybbut that's getting kinda clunky.   fish's solution is much neater21:58
fungii just noticed mirror.debian last updated 12 days ago. looking at the reprepro logs, it's complaining about a missing pubkey for signatures by B8E5F13176D2A7A75220028078DBA3BC47EF2265, so probably we need to add another key to our config but i don't think i have time for that tonight22:01
corvusclarkb: lgtm22:01
fungisee the end of /var/log/reprepro/debian.log on mirror-update.o.o for details22:02
clarkbfungi: I'm guessing the new key came in via trixie and they started signing the old stuff with it too? I can take a look22:02
fungilikely, that's what i'm assuming anyway22:02
fungitiming about lines up22:02
tonybclarkb: lgtm.  I guess approve at will 22:03
opendevreviewJames E. Blair proposed opendev/zuul-providers master: Update bindep-fallback path  https://review.opendev.org/c/opendev/zuul-providers/+/95824722:04
clarkbthanks I've approved it22:04
corvusthe change to remove the nodepool elements from openstack/project-config merged, but that exposed a place where zuul-providers was referencing that repo.  https://review.opendev.org/958247  is needed to fix our image builds, and we need to fix our image builds in order to get images for rax-flex iad3  (i think we have probably exceeded our 3 day timeout for keeping the objects around)22:05
corvussince clarkb +2d that i went ahead and approved it.  that's got several hours of work in the gate ahead of it still if anyone else has thoughts....22:07
corvusactually it's failing in gate because22:07
corvus2025-08-21 22:07:11,093 ERROR zuul.Launcher:   zuul.exceptions.LaunchStatusException: Server in error state22:07
corvuswe're getting that from osuosl22:07
clarkbcorvus: if we can get the error message we can pass that along to Ramereth[m]22:08
corvusclarkb: zuul logs it if it gets one22:08
corvusand i don't see one22:08
corvusworth double checking on bridge though in case there's a bug in that22:09
clarkbcorvus: via server show you mean?22:09
corvus(or maybe there are errors that only show up in detailed server gets, not server listings)22:09
corvusyeah22:09
corvusi'm on that22:09
corvusthis command right? /usr/launcher-venv/bin/openstack --os-cloud opendevzuul-osuosl --os-region RegionOne server list22:10
corvusthat's empty, i think zuul is deleting the error servers very quickly22:11
clarkbya that command looks right22:11
corvusi think it may be time to consider this change: https://review.opendev.org/95579722:11
clarkbwe may need to boot something out of band and check the error or see if Ramereth[m] sees anything on the cloud side22:12
clarkbcorvus: for 955797 what causes zuul to ignore the failed image builds? They won't have the archive info listed because the failed jobs don't get to that point?22:12
corvusyep22:13
corvusthey either produced and artifact or didn't, that's all we really care about22:13
clarkb+2 from me22:14
corvusanother thing we could consider is doing another set of pipelines for the arm images.  would let us group them together in buildsets.  that's an alternative to 797, but we can also do both.22:16
corvus797 gives us the swiss-cheese model of image builds: whatever builds, we use.  making another set of pipelines lets us do all-or-none for x86, and all-or-none for arm.  both are different from our current global-all-or-none.22:17
opendevreviewClark Boylan proposed opendev/system-config master: Add Trixie keys to reprepro config  https://review.opendev.org/c/opendev/system-config/+/95824822:17
corvus(and, if we do both things, then we get swiss-cheese buildsets for each)22:17
clarkbcorvus: considering each image build is largely its own thing (even each release within a distro is pretty independent) I think taking what we can get is probably best22:17
clarkbrather than splitting it up by arch or distro or whatever22:18
corvuswfm22:18
clarkbfungi: something like 958248 maybe22:18
opendevreviewJames E. Blair proposed opendev/zuul-providers master: Remove build_diskimage_image_name variable  https://review.opendev.org/c/opendev/zuul-providers/+/95637322:19
clarkbfungi: the key hash string you provided above doesn't seem to match the key hashes I found but I'm guessing thats just because its hashing something different22:20
opendevreviewMerged opendev/zuul-providers master: Use label-defaults  https://review.opendev.org/c/opendev/zuul-providers/+/95694622:20
corvusi have not seen this error building images before: https://zuul.opendev.org/t/opendev/build/6cdc17c3b50c41d2a9daa5149813c1c722:21
fungiclarkb: looks like it's a related subkey: https://lists.debian.org/debian-devel-announce/2025/04/msg00001.html22:21
corvusother builds got past that point... so it seems like maybe a fluke, but i don't know the origin22:22
clarkbcorvus: let me try cloning that repo locally from each backend I guess22:22
corvusclarkb: oh good idea22:22
clarkbcould be gitea backend specific22:22
fungiclarkb: 958248 is adding 225629DF75B188BD which is the corresponding master key for that subkey, so that looks right to me22:23
clarkbfirst thing I notice is that repo appears to be somewhat large....22:23
clarkblike bigger than nova22:23
corvusoO22:24
clarkbhttps://opendev.org/cfn/computing-offload/src/branch/master/LingYaoSNIC/BCLinux they are just shoving rpms in there22:24
clarkb456MB22:25
clarkbso maybe not bigger than nova but in the same range22:25
corvusgive it time22:26
clarkbfungi: do you know who to talk to about this? I know you interacted with them a couple times then I tried to respond to them on the list. But this is really not what gerrit or git is good at...22:27
clarkbanyway gitea09 clones cleanly22:27
fungii do not, sorry22:27
fungihorace may have some contacts22:27
fungiif memory serves, he follows some of their development effort22:28
clarkbthanks I'll followup there once I have a bit more info22:28
fungihe might also be able to get their attention more easily through wechat/weixin22:31
opendevreviewMerged opendev/system-config master: Drop Bionic testing for OpenDev services, playbooks and roles  https://review.opendev.org/c/opendev/system-config/+/95795022:35
clarkbcorvus: all 6 gitea backends clone that repo for me successfully right now22:35
clarkbcorvus: the only other thought I've got is maybe a zuul merger is copying data over that is unhappy and that is propogating into the image builds?22:36
clarkbexcept I think we're just copying the old cache over not using the mergers?22:36
clarkbcould be the repo is large enough to have hit a bit flip?22:36
clarkbI've asked horace if we can see about helping them use our tools more effectively (code review to prevent bad patches in the first place, artifact build and storage, git/gerrit/gitea, etc)22:47
clarkbcorvus: tonyb: re arm64 there is a bookworm arm64 node booted right now that is active and not in an error state and I was able to ssh into it23:37
clarkbI also manually booted an ubuntu noble node and that worked (using our image)23:38
clarkbso maybe whatever the issue is has been resolved or is image specific?23:38
clarkbI'm going to delete my test node now23:39
corvusmaybe23:39
corvusi can recheck something in a sec23:39
corvusbut i just got distracted by the fact that all the x86 image builds failed somehow23:39
corvusoh23:40
opendevreviewJames E. Blair proposed opendev/zuul-providers master: Update bindep-fallback path  https://review.opendev.org/c/opendev/zuul-providers/+/95824723:40
corvusclarkb: ^ i got the path wrong23:41
clarkb+223:41
corvusthat will produce a number of arm requests23:41
clarkbthe last server listing I did shows 4 servers. 2 noble, 1 jammy, 1 bookwrom all active on the cloud side23:42
corvuswell, they're 2 minutes in the building stat, that's promising23:43
corvusand some in use now23:43
corvusso yes, i guess the anomaly was temporally bounded :)23:44

Generated by irclog2html.py 4.0.0 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!