Tuesday, 2025-06-24

clarkbjust about meeting time. We'll get started in a minute or two18:59
clarkb#startmeeting infra19:00
opendevmeetMeeting started Tue Jun 24 19:00:19 2025 UTC and is due to finish in 60 minutes.  The chair is clarkb. Information about MeetBot at http://wiki.debian.org/MeetBot.19:00
opendevmeetUseful Commands: #action #agreed #help #info #idea #link #topic #startvote.19:00
opendevmeetThe meeting name has been set to 'infra'19:00
clarkb#link https://lists.opendev.org/archives/list/service-discuss@lists.opendev.org/thread/YJ4OIEZ4ARUKB5DRBWTRKT3U6BNBCFME/ Our Agenda19:00
clarkb#topic Announcements19:00
clarkbThank you fungi for running the last couple of meetings. I had some family stuff pop up and wasn't able to consistently be around19:00
clarkbbut I'm slowly catching back up now and finding my routine again19:01
fungiwelcome back!19:01
clarkbthe open infra summit's forum session proposals are open right now19:01
clarkbI don't know that we need any opendev forums sessions but figured I'd mention that19:02
clarkbanything else to announce?19:02
clarkbalright lets jump into the agenda then19:04
clarkb#topic Zuul-launcher transition19:04
clarkbI think that auto hold stuff has rolled out or is rolling out19:04
clarkbwhich means we're largely ready to start migrating more jobs over to zuul launcher from nodepool. Early changes to make that happened merged just before the meeting19:04
corvusyep...19:05
corvusand i think we want to revert it19:05
corvusrax-flex-sjc3 is very sad19:05
corvusand in a new way that z-l isn't handling well19:05
fungideploy was at 18:50 utc, for the record19:05
corvusso i'm going to propose a revert, manually merge it, then add some exception handling.19:05
fungithanks for spotting that quickly19:06
clarkback19:06
corvusbut yeah, we're really close :)19:06
corvusand after we merge this change: https://review.opendev.org/95272119:06
clarkbfwiw looking at the nodepool grafana graphs (not the zl grafana graphs) it looks similar to what we saw yseterday which was the cloud having some issues that needed to be resolved on the cloud side to get successful boots19:06
clarkbbut having zuul launcher handle that gracefully is a good idea (nodepool mostly does)19:06
corvuswe will only have ubuntu-focal and ubuntu-xenial images left in nodepool19:07
corvusi forget -- where did we land on whether those were used, and what do we want to do before turning off nodepool?19:07
clarkbfor xenial one of the biggest users is/was opendev. But I did some cleanup of that on the opendev side and I think we can rip the bandaid off for opendev19:08
clarkbfocal is more difficult19:08
clarkbor did you mean bionic?19:08
clarkbin either case I Think we have openstack stable stuff and translation stuff relying on either bionic or focal (maybe both)19:08
corvusoh sorry i meant bionic19:09
corvushttps://review.opendev.org/952726 is the change i meant to link19:09
clarkbya I think the translation sync stuff runs on bionic due to java versions or something along those lines19:09
corvusno wait19:09
corvusit's bionic and focal that are left19:09
corvusand focal?19:10
fungipossible we could do something with a tarred-up bionic chroot or container image, i suppose19:10
clarkbgot it. I see my comment on 952726 now too. To summarize I think we can rip the bandaid off for xenial and live with the fallout. We may need to check bionic/focal things more carefully. fungi not sure if you know how many if any stable branches for openstack rely on either of those.19:10
clarkbfungi: yes I think one option for the small number of things that need the old tooling would be to use a container to run the tools19:10
corvusfungi: that would almost certainly be *harder* than porting over the build, no?19:11
fungiporting it over to what?19:11
corvusdon't we have working bionic and focal dib builds with nodepool-launcher?19:11
corvuser nodepool-builder19:11
clarkbcorvus: yes19:11
clarkbso yes another option available to us is to add those images to zuul launcher19:11
fungii mean using a bionic chroot with zanata tools installed on a newer (jammy or noble) test node19:12
corvusyeah, which is probably like 5m of work19:12
fungiif the goal is to drop the bionic nodes19:12
clarkbI think that is a good goal (and why I think we should proceed wtih xenial that way). But I recognize there is some improtant functionality relying on it so maybe comprimise is better there19:13
clarkbif its only 5m of work that seems like a reasonable thing to do19:13
clarkbanything else on this topic?19:14
corvusyeah, so if opendev is happy to continue hosting bionic and focal images, then we just need someone to copy over the config to zuul-providers19:14
fungiright now some openstack projects are still testing on their unmaintained/victoria branches as the oldest not yet eol. victoria's official test runtime included ubuntu 20.04 (focal), so bionic should not be needed if it weren't for the translation jobs requiring older java, i think19:14
clarkbfungi: ack that is good info19:15
clarkbthat tells me we should probably go ahead and add focal images. Then bionic is a matter of deciding about translation jobs but probably easy enough to do when we do focal19:15
fungiin theory the bionic requirement for those jobs goes away once openstack's i18n team finishes migrating to weblate19:15
tonybI'm planning on helping the i18n team with that in July.19:16
corvusi think we need an action item about the images19:16
corvusfrom my pov, we sent out an email a few months ago and got some help, and now all but those 2 are done19:16
clarkbdo we have a volunteer to add focal and xenial images to zuul launcher?19:16
corvusfocal and bionic, right? :)19:17
clarkbyes right sorry19:17
corvusi don't think anyone has volunteered19:17
clarkbmnasiadka isn't in here but did do a big part of that move for openstack. I can ask mnasiadka if they would be willing to do focal and bionic too19:17
corvusok, then if not, maybe send a "last call" email19:17
clarkb++19:17
clarkb#action clarkb to ask mnasiadka about adding focal and bionic image builds to zuul launcher19:18
corvusthanks... i'm happy to send the last-call email if needed19:18
corvusi think that's all the decision making i need right now19:18
clarkblets continue on with our agenda since we're already 1/3 of the way through our time19:19
clarkb#topic Gerrit Shutdown Process and 3.11 Upgrade19:19
fungifor further data points, openstack unmaintained/zed is the most recent branch officially relying on focal nodes (it's been unmaintained for over a year now), and i think the openstack position on it's going to be that if opendev wants to pull the rug out from under those jobs because nobody who claims to care about those branches volunteered to make this work, then that's how19:19
fungiit goes19:19
clarkbI'm going to combine the next two topics into one as I'm largely driving them and don't have any real updates19:19
clarkbI am hoping that I will finally have the ability to update the gerrit command line string to remove the h2 cache cleanup timeout change and restart gerrit this week19:20
clarkband use that process to test manual sighup signalling and possibly a followup to do a docker compose driven sigint as well19:20
clarkbthen assuming I get that done I'm hoping Gerrit upgrade planning and testing can start up again early july19:21
clarkbno real updates other than tyhis is still on my radar and I plan to look at it again as I catch up19:21
clarkb#topic Upgrading old servers19:21
clarkbcorvus replaced zuul mergers and zuul launchers with new noble nodes19:21
corvusplanning to do schedulers and executors later this week/weekend19:22
clarkbthat seems to have largely gone well. The one hiccup was reducing the size of zuul launchers which reduced their epehermal disk sizes which a fair chunk is needed for to shuffle image data around19:22
clarkbthis has since been corrected with some cinder volumes19:22
corvusalso, the restart script needed a change because "docker-compose ps -q" behavior changed19:22
corvusi think that command is only used for the zuul servers; does that sound right?19:23
clarkbcorvus: yes I think that is the only situation where we use that19:23
clarkbI'm also like 95% of the way through replacing mirror-update02 with a noble mirror-update0319:23
clarkbI thoguht I was going to delete 02 before this meeting then discovered a problem with debian reprepro config that I wanted to fix first. The fix is in place and one of the two mirror locations is happy again. The other is waiting for the next run in about 45 minutes19:24
clarkbonce that is successful I'll proceed with cleaning up the old mirror-update02 node19:24
clarkbcorvus: worht noting that mirror-update03 is using openafs to read and write data on noble which is a good indicator for executors19:24
clarkbpreviously we were doing read only operations on mirrors I think19:25
tonybI don't know if the centos team have done it yet but they have a ticket to fix the IP situation19:25
clarkbtonyb: yup that seems to be working now too19:25
fungitonyb: yeah, that's been resolved since around 13:00z19:25
tonyb\o/ 19:26
clarkbmy hope is to start on the zookeeper cluster tomorrow aswell19:27
clarkbcurrently we have zk04, zk05, and zk06. Do you want zk07-09 or zk01-03 as the new servers?19:27
clarkbchanging things in place is maybe more complicated with zk since it uses the digits suffix as the cluster member id or whatever it is called19:27
clarkbI think new servers should have new ids as a result19:28
clarkbI'm guessing 01-03 is preferred. But let me know if not and I can do something else19:28
corvusi don't feel strongly; slight pref for 1-319:28
fungisounds good to me as well19:29
clarkbother things to note: don't forget to use the --config-drive flag to launch node when booting our noble image in rax classic (it is required there but not in other clouds)19:29
clarkband fungi was there any refstack update?19:29
funginope!19:29
clarkbthat was all I had on this topic. Anything else before I go to the next one?19:29
clarkb#topic OFTC Matrix bridge no longer supporting new users19:31
clarkbthe last time I was around to discuss this I volunteered to write a spec to explicitly call out what we're trying to achieve with a move to matrix for opendev comms and a plan for testing that through doing it19:31
clarkbI have not done that. But doing so is one of the things I intend to do as I catch up19:31
clarkbis there anything new to consider from the last couple of weeks before I do that?19:32
corvusno news i'm aware of19:33
funginot really. there were discussions in the kubernetes community about moving off slack, but they ruled out matrix basically straight away and most people who were involved in the conversation seemed to want to use discord, predictably19:33
clarkback I'll proceed with the plan from a couple of weeks ago as soon as I dig through enough of my backlog19:34
clarkbfungi: ya they seem to have immediately dismissed matrix as being incapable of handling their load/demand/users19:34
clarkbmattermost and zulip people got invovled in the discussion as well but they seem to have been dismissed too19:34
fungiit could be useful to dig into the size/performance concerns they raised about matrix, though for opendev's purposes we're unlikely to get anywhere near where it would start to be impacted19:35
clarkbya I don't think that is a concern for us19:35
clarkbif they can quantify that somehow (rather than just asserting it as fact) that could be useful info generally19:35
clarkb#topic Adding CentOS 10 Stream Support to Glean, DIB, and Nodepool19:36
clarkbglean is done19:36
clarkb#link https://review.opendev.org/c/openstack/diskimage-builder/+/949942 DIB functional testing without Nodepool19:36
clarkbthis change and its children to have dib stop relying on nodepool for testing image builds is the next step. This will allow dib greater control over devstack and its nova cpu configuration so that centos 10 can be booted with all of the cpu flags it requires19:37
clarkbonce those are in we should be able to land support for centos 10 and rocky 1019:37
clarkb#link https://review.opendev.org/c/openstack/diskimage-builder/+/934045 DIB support for CentOS 10 Stream19:37
clarkband at this point all the related changes are ready for review. I intend on reviewing those today19:37
clarkbtonyb: any other concerns or things to call out?19:38
tonybThe series adds devstack builds along side of the nodepool ones, but the removal is more complex due to job users in other repos so I'll rework that today so we can drop the nodepool testing19:38
tonybassuming that sounds fair to others19:38
clarkbdo we define those jobs in dib? I thought we were just consuming them19:39
clarkbbut yes we need to update glean as well to stay in sync19:39
tonybwe define and use jobs like dib-nodepool-funcational-openstack-$distro ; which are uses in openstacksdk19:40
clarkbgotcha19:40
tonybI'm less worried about glean19:40
tonybopenstacksdk is a little harder as it has several open branches ;P19:41
fungiah, yeah i guess when we pulled shade out of nodepool we wanted to test that it didn't break, and then when shade was merged into openstacksdk we kept that testing19:41
clarkbI wonder if the motivation from the openstacksdk side is ensuring they don't break nodepool or if they want to test the sorts of operations nodepool does19:41
fungithe former, i'm almost certain19:41
clarkbin that case cleanup is probably most appropriate at this point?19:42
clarkbsince nodepool is going away and I don't know that zuul wants to be that tightly coupled to openstacksdk19:42
fungii believe so19:42
clarkbbut maybe there is some middle ground I'm not considering19:42
tonybI'll double check with them and proceed as appropriate19:42
clarkbsounds good thanks19:42
clarkb#topic projects.yaml normalization19:42
tonybI have the patches out there to switch from nodepool to devstack but cleanup would be better19:43
corvusyeah, i don't think that's necessary anymore19:43
clarkb#link https://sourceforge.net/p/ruamel-yaml/tickets/546/19:43
corvus(from a zuul perspective)19:43
clarkb#link https://review.opendev.org/95200619:43
clarkbcorvus: ack thanks for confirming19:43
clarkbsorry I'm moving ahead since we're now 3/4 of the way through our time19:43
clarkband I'm not caught up on this one so want to make sure we talk about it19:43
corvus++19:44
clarkbbasically it seems there is some bug in ruamel and maybe we worked around it?19:44
fungii believe we excluded the recent releases19:44
clarkb952006 says that there is some other thing that may fix things but then unfortunately links to itself rather than the actual chagne that I think fixed things19:44
corvusconfused by that last comment, suspect link is wrong19:44
clarkbcorvus: yes that. This is where I got lost trying to followup on this19:44
corvuspossibly https://review.opendev.org/c/openstack/project-config/+/95231519:45
clarkbok so basically we're not needing to do a big normalization pass beacuse we're using an old library verison which produces output consistent with what we current have19:45
corvusyes, also, the output of the new version is wrong19:46
clarkbI guess we can rollforward for now as is. Monitor the upstream bug (sourceforge!) and see if upstream is going to fix it19:46
corvus(so the choices we feel acceptable are (1) pin and wait/hope for them to fix; (2) do some more wrapping of ruamel output to work around it)19:47
corvuswe chose 1... then 2 if that fails19:47
corvusthe option of accepting the output was rejected19:47
clarkbgot it. And ya I agree that new output is not how I would format things19:47
corvusthe bot hasn't updated that change.. so maybe that means the pin worked?19:47
clarkbI suspect so. It runs daily iirc19:48
fungii believe that's the reason, yes19:48
corvusi think contents of repo == ideal means no update is needed19:48
corvuswe should probably abandon that change now19:48
clarkb++19:49
corvusdone19:49
clarkbok now i feel caught up19:49
fungithough as an unrelated bug, it became apparent that existing change == generated diff doesn't prevent it from pushing a new patchset19:49
clarkbfungi: we rely on gerrit to reject it?19:49
fungino, commit dates are different each time so it doesn't get rejected19:50
clarkbgotcha so we get a new patchset each day as long as there is a delta to master19:50
corvusyep19:50
fungifixing that is probably a trivial line or two in the script, but nobody's had time to work out the patch19:51
clarkbgood to know. Anything else on this item?19:51
fungii don't think so19:52
clarkb#topic Working through our TODO list19:52
clarkb#link https://etherpad.opendev.org/p/opendev-january-2025-meetup19:52
clarkbBefore my schedule went upside downI said I should formaize this better19:52
clarkbwhich is still my intention but until then here is a friendly reminder that this list exists in that etherpad19:52
clarkb#topic Pre PTG Planning19:52
clarkbwhich is related to the last thing on the agenda: Planning pre ptg dates19:53
clarkbOpenStack's next release is happening October 1, The summit is October 17-19, and the actual ptg is October 27-3119:53
clarkbconsidering all of that I thought that October 6-10 might work for picking 2-3 days to try and have a PTG if we don't feel that October is already full of stuff19:53
fungiwfm19:54
clarkbI think it was helpful to do the january thing at the beginning of the year and doing a quick checkup before the ptg again would be useful19:54
fungii've got a talk at all things open just prior to the summit, but the proposed pre-ptg dates are early enough not to conflict19:55
clarkbI'll probably pencil in those dates on my own calendar but please say something if that doesn't work for some reason and we can look at late september intsead potentially19:55
tonybWorks for me.  There's a solid chance I'll be in the US at that time which possibly simplifes the time-of-day selection19:55
clarkbI suspect we'd do 2 or three days with blocks of a couple of hours19:55
clarkbvery similar to what we did in january19:56
clarkb#topic Open Discussion19:56
clarkbAnything else?19:56
fungii don't think i had anything19:56
clarkbthanks again everyone for helping out when I wasn't able to be around much. Really appreciate it19:57
clarkband thank you for helping keep opendev up and running19:57
corvusgood to have you back :)19:57
fungiyes, i don't mind pitching in, it's my pleasure19:58
clarkbsounds like that may be everything. I expect to be back here same time and location next week19:58
clarkb#endmeeting19:58
opendevmeetMeeting ended Tue Jun 24 19:58:59 2025 UTC.  Information about MeetBot at http://wiki.debian.org/MeetBot . (v 0.1.4)19:58
opendevmeetMinutes:        https://meetings.opendev.org/meetings/infra/2025/infra.2025-06-24-19.00.html19:58
opendevmeetMinutes (text): https://meetings.opendev.org/meetings/infra/2025/infra.2025-06-24-19.00.txt19:58
opendevmeetLog:            https://meetings.opendev.org/meetings/infra/2025/infra.2025-06-24-19.00.log.html19:58
fungithnks clarkb!19:59

Generated by irclog2html.py 4.0.0 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!