15:01:29 #startmeeting ironic 15:01:29 Meeting started Mon Dec 11 15:01:29 2023 UTC and is due to finish in 60 minutes. The chair is JayF. Information about MeetBot at http://wiki.debian.org/MeetBot. 15:01:29 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 15:01:29 The meeting name has been set to 'ironic' 15:01:38 #topic Announcements/Reminder 15:01:45 #info Standing reminder to review patches tagged ironic-week-prio and to hashtag any patches ready for review with ironic-week-prio: https://tinyurl.com/ironic-weekly-prio-dash 15:03:01 o/ 15:03:07 dtantsur: w/r/t ncsi, just an option afaik 15:03:24 No action items from last week, skipping that one 15:03:40 #topic Caracal release schedule 15:03:50 #info Next milestone C-2, Jan 11 15:03:59 #topic Meeting schedule for holiday 15:04:22 I propose we cancel the 12/25 and 1/1 meetings. Well, I propose I won't be here, you all can do what you want with Christmas + New Years Day :D 15:05:20 If I don't hear an objection going to move ahead with that plan 15:05:31 o/ 15:06:33 #action JayF to email mailing list about Christmas + New Year's Day meetings being cancelled 15:06:39 #topic Review Ironic CI Status 15:06:55 So Ironic CI is currently busted, BFV job was broken by some Ironic driver changes that landed 15:06:55 The bfv fix hasn't merged yet.. 15:07:09 the fix is in the nova gate, if it gets clogged up we can push a quick -nv to get things free 15:07:28 I'm a bit confused how this landed broken, I had a patch w/Depends-On on the tip that passed all our jobs 15:07:51 Verification of a change to openstack/ironic-python-agent master failed: Fix referencing to the raid_device var which is not set https://review.opendev.org/c/openstack/ironic-python-agent/+/900324 15:08:07 I don't know what patches in question, but it looked like it was lacking a key we expected to be present 15:08:13 possibly a race condition?! 15:08:47 Well, there was a comment on the original change that it could be racey, but with dtantsur's fix, it doesn't look like it could've ever worked 15:08:56 so I'm thinking maybe a follow up just to ensure our jobs are running nova from git 15:09:04 well, we know it is, because they broken 15:09:29 Either way, we know the path forward, path behind matters less, not like we push major changes to Ironic<>Nova driver often. 15:09:39 Anything else on CI/Nova breakage? 15:10:21 We cannot run nova NOT from git 15:11:03 fair 15:11:08 moving on 15:11:11 #topic Bug Deputy 15:11:21 I was the bug deputy. I triaged some bugs. I did not put together a report. 15:11:35 I'll note for whoever is taking it next (I can go another week if needed), there's something mildly screwy with the dashboard 15:11:55 I didn't have time to look, but either launchpad API returns bad data or something in our bug deduplication breaks on non-Ironic projects 15:12:07 but I managed to triage most new bugs 15:12:32 I can go for it this week 15:12:44 #action rpittau to be bug deputy this week 15:12:53 well, I say I didn't have time to look; I looked; I didn't solve lol 15:12:59 :D 15:13:04 We have lots of RFEs up 15:13:09 including some I proxied as bug deputy 15:13:13 #topic RFE Review 15:13:23 dtantsur: has three, going to link them in and let you talk about them 15:13:29 #link https://bugs.launchpad.net/ironic/+bug/2045548 15:13:34 #link https://bugs.launchpad.net/ironic/+bug/2045551 15:13:46 both carryovers from last week, we didn't have many folks in the meeting and wanted more people to see these 15:13:51 they were provisionally approved 15:13:54 Yeah, thanks! 15:14:07 #link https://bugs.launchpad.net/sushy-tools/+bug/2046153 "Testing the minimal subset of Redfish features" is the new one for dtantsur 15:14:10 no objection to 2045548, only concern about the kernel command line length limit 15:14:23 TheJulia: do you remember the limit from the top of your head? 15:14:31 1024 chars 15:14:34 total 15:14:43 I think we're quite good still 15:15:02 yeah, maybe move it to the end *just in case* 15:15:34 I think the kernel truncates it, if memory serves it is also configurable and at one point was 256 charts 15:15:36 chars 15:15:38 I'll how doable that is (without relying on Python dict ordering) 15:15:59 I was thinking template wise 15:16:00 fwiw 15:16:15 I think all these arguments end up in a dict.. anyway, technical details 15:16:47 they do, but the actual line draws from several different fields in the tempalate, the actual consumption side, yes hits dict and entires may be truncated 15:16:47 how long until ipa-b64-config=dgjskljdfhgksljdhg 15:17:05 lol 15:17:17 I'm only half-joking, we're getting more complex in the preboot configuration we need 15:17:21 with virtual media, we should start embedding it as a file already (we actually have all the code there) 15:17:32 that is worthy of being written up 15:17:52 * JayF notes he is +1 to all the proposed RFEs 15:18:01 I'll try to.. but I cannot do everything at once 15:18:20 The second rfe, I guess I'm trying to understand why we feel the need to create the second script/why we feel it is explicitly needed. I.e. is there a path not to have it, and do we need more, or less complexity to get there 15:18:31 everything at once is impossible 15:18:33 2045551 is a bit more interesting. It's continuation of the whole "self-configuring ironic" thing 15:18:51 to be clear: this script exists now, it's just normally hand-written 15:19:08 we *could* roll this logic into boot.ipxe, but then we won't be able to reuse the ipxe_script_template 15:19:49 like, I don't want https://github.com/metal3-io/ironic-image/blob/main/ironic-config/inspector.ipxe.j2 to be handwritten any more 15:19:51 I'm personally pro all together, if possible 15:20:39 Not sure I know what you mean 15:21:10 from my pov, there is no technical reason why we've kept that pattern, out of "this is the way we always did it" 15:21:24 and we should evolve that. What that looks like exactly, more so depends on the overall requirements 15:21:30 it falls the loop , , ..., fallback 15:21:49 where do you suggest it goes, https://opendev.org/openstack/ironic/src/branch/master/ironic/drivers/modules/boot.ipxe ? 15:21:57 yes, except what did we do in bifrost? 15:22:04 same as in metal3: a separate file 15:22:20 it's also the approach we document for ironic-inspector 15:22:29 we could take boot_failed path 15:22:38 oh, so bifrost got changed way from the single file ? 15:22:44 if we add that to boot.ipxe, then the operators who customize boot.ipxe and ipxe_config.template will need to change their templates 15:22:50 yes 15:22:51 bifrost has always worked like that 15:22:53 that is unavailable 15:23:03 nothing I touched implemented inspection as part of boot.ipxe 15:23:10 (I don't think TripleO either) 15:23:12 maybe bifrost for the past five years, but a very long time ago it was a single file 15:23:33 possibly? if so, it's long gone 15:23:38 yeah 15:23:45 I'm not sure why we'd ask people to duplicate their templates though 15:23:52 it worked quite nicely, from my point of view, but that was a long time ago 15:23:58 nor am I 15:23:58 (not my pain any more since we don't customize ipxe_config.template, but still) 15:24:02 other than hardware discovery 15:24:54 I'm afraid your memory misleads you 15:25:04 https://opendev.org/openstack/bifrost/commit/c1f9beac6358efd70a4197f57f71ef98499fa7d6 is your patch that introduced inspector support, and it uses the separate file 15:25:05 sweet! 15:25:48 The value I see in the separate file is to reuse ipxe_config.template out of box 15:26:13 * TheJulia shrugs 15:26:13 so, people who customized something for cleaning, deployment and managed inspection will get the same thing for the unmanaged inspection 15:26:42 I'm worried about managed and unmanaged inspection ever diverging 15:26:52 yeah, I just worry how much debt we're carrying for that and if we should make it more streamlined, or not 15:27:11 I need to step away for a minute; if you all get done with this next up are those two sushy-tools RFEs which I'm +1 with very low context on (I put them on the agenda from seeing them in triage) 15:27:14 #chair TheJulia dtantsur 15:27:14 Current chairs: JayF TheJulia dtantsur 15:27:14 I guess i always thought we were going to try and drift everything towards managed 15:27:26 anyway, we should carry on 15:27:32 We will try, but we don't succeed right away 15:27:40 naturally :) 15:27:42 People want auto-discovery all the time... 15:27:48 true 15:27:59 and we have to still soft of have that as a path, but that doesn't mean things can't evolve 15:28:09 Onward? 15:28:18 Mmmm, so no consensus? 15:28:27 dtantsur: lazy consensus? 15:28:38 I thought you were against the RFE as it's written? 15:29:01 eh, not a fan, but doesn't mean I'm going to take a hardline stance on it 15:29:09 I'm more asking question for the big picture 15:29:34 #link https://bugs.launchpad.net/sushy-tools/+bug/2045906 15:29:36 #link https://bugs.launchpad.net/sushy-tools/+bug/2045908 15:30:05 I don't have anything against extending sushy-tools as long as we have some understanding why people are doing it 15:30:11 i.e. they're not doing it in production 15:30:11 ++ 15:30:25 The first one, seems like a "non-issue"/"just do it" 15:30:27 I think these are the folks who came by here the other day 15:30:41 someone dropped in and asked if we'd be OK with getting features into sushy-tools 15:30:50 yeah, SKU,Serial are not even RFE-worthy 15:30:56 and I gave them the usual interrogation; sounds like they are building a product in the category of Ironic 15:31:05 and want to enhance the testing suite in sushy-tools more 15:31:06 the latter one, yeah, I'm trying to understand why 15:31:20 if it's for testing their product - no problem 15:31:41 as long as it does not end up putting more burden on us 15:31:43 TheJulia: I could easily see a datacenter management product monitoring power usage, especially if it's oriented differently than OpenStack 15:32:02 dtantsur: if anything, getting another project depending on sushy{,-tools} may reduce our overall burden 15:32:07 JayF: of course, which I think kind of goes to dtantsur's comment 15:32:19 by the way, we should rename sushy-tools 15:32:24 it causes too much confusion with sushy proper 15:32:27 ++++ 15:32:45 * dtantsur has recently learned there is an alternative to sushy-tools called ksushy 15:32:58 ..... 15:33:05 ikr? 15:33:10 too bad we didn't get OIF to (tm) sushy 15:33:12 lol; 15:33:16 :D 15:33:25 I'd be OK with a sushy-tools rename, maybe start a next-PTG etherpad with that? 15:33:42 possibly 15:33:44 or send something to the mailing list 15:33:49 deferring to a PTG is a bad habit 15:33:59 because it forces discussion to halt until high bandwidth discussion 15:34:01 Maybe; but in this case I'd say halfway thru the cycle is a bad time to rename it 15:34:03 I'm +1 to both proposals if they don't end up dumping a lot of obscure code on us 15:34:04 which casues the topic to drop 15:34:14 which is the only reason I punted it; not for the high-bw but for the "after release" :) 15:34:27 JayF: would need to be well communicated and we would already be into the next cycle once we reach the PTG 15:34:50 I've left a comment on the first sushy-tools rfe 15:34:55 good point 15:34:56 We could have renaming by forking fwiw 15:35:00 so we are approving all those RFEs, right? 15:35:04 Also dropping the static emulator that I'm afraid nobody uses 15:35:06 yeah 15:35:18 #info All RFEs evaluated at meeting are approved. 15:35:21 #topic Open Discussion 15:35:24 mmm, also mine? :) 15:35:26 uhh, we did't get to the last one 15:35:29 I don't mind it :D 15:35:31 #undo 15:35:31 Removing item from minutes: #topic Open Discussion 15:35:33 #undo 15:35:33 Removing item from minutes: #info All RFEs evaluated at meeting are approved. 15:35:46 * JayF presses << on the tape player 15:35:49 What last one TheJulia? 15:35:59 TheJulia: we batched dtantsur' 15:36:12 #link https://bugs.launchpad.net/sushy-tools/+bug/2046153 Testing the minimal subset of Redfish features 15:36:12 **we batched dtantsur's three together, so even though we did them out of order, I thought we discussed all 5? 15:36:24 It's not even really an RFE, more of a heads-up 15:36:38 I want us to avoid relying on a redfish implementation having ALL features we support 15:36:38 the last one wasn't linked until when dtantsur just did so 15:36:43 thought we were going in order 15:36:52 for that, I want a CI job that runs with a bare sushy-tools with no extras at all 15:36:59 objections, concerns? 15:37:20 dtantsur: none really 15:37:46 it *should* be a test scenario in the tempest plugin explicitly instead of a separate CI job itself, if at all possible. 15:37:50 TheJulia: some mornings I can hide when I didn't get my coffee made before meeting... others... :) 15:38:04 TheJulia: I was thinking to just changing one non-vmedia job 15:38:06 or just, run the emulator on minimal and make sure none of the advanced features are turned on 15:38:29 all I care about is that we don't e.g. depends on firmware versions being available as we've just accidentally done 15:38:29 ok 15:38:43 (context: https://review.opendev.org/c/openstack/ironic/+/903185) 15:39:27 (in the past, we accidentally made EthernetInterfaces required, breaking mmm Cisco?) 15:39:42 sweet! 15:39:51 sounds like a pretty good diea then 15:40:28 yeah, every redfish feature we do, we should be checking "if the value is not none" 15:40:35 stupidly easy thing to miss though 15:40:47 on that note, I also want to document which resources we require and which we can use if they are present 15:41:00 (quite unfortunately, that the profiles work stagnated..) 15:41:30 so a profile merged, but it is not quite clear 15:41:51 Background: from time to time, our partners ask me for a list of Redfish APIs we require 15:42:04 so I essentially have this document downstream. I can just publish it. 15:42:48 makes sense, the profile is rather broad 15:43:00 And I'm not sure anyone is using it :\ 15:43:01 IDK how to go about this; but if we could document quirks too that'd be cool 15:43:11 e.g. the bug around gigabyte servers wanting a different payload for boot mode than anything else 15:43:22 We're trying to fix such things instead :) 15:43:24 Did we get a reply to that one? 15:44:06 https://bugs.launchpad.net/ironic/+bug/2045191 yes 15:44:17 In fact it looks very good 15:44:22 I left it incomplete because I wanted someone more redfishy to verify it was enough info 15:44:24 but it LGTM 15:44:48 I also reached out a handful of different ways trying to get a contact in Gigabytes' server engineering team 15:45:16 #topic Open Discussion 15:45:23 (we are basically already doing this, just reflecting the reality) 15:45:43 another fishy item, just as a heads-up: https://review.opendev.org/c/openstack/sushy-tools/+/903331 15:45:46 oh joy, yeah, what we have to do with them was semi-incompatiable with some of the other vendors 15:46:04 ..... hmmmmmm 15:46:19 This is part of why I'm trying to work out a contact in their server group :D 15:46:31 Seems like the sort of thing that ... collaboration could improve future versions of 15:46:36 yeah 15:46:39 (I do not think I was/will be successful FWIW) 15:47:39 I don't know if I'll have time to take a look this week, this week is shaping up already to be a very busy week 15:48:01 I'm going to be focusing primarily on nova driver bits, trying to get sharding spec mergable and getting all the code lined back up 15:48:07 so I can work on those tempest tests again 15:48:13 I suspect we could likely just teach ironic to recognize such a constrant and send back the same value or something slightly wicked 15:48:31 but where in that sequence is sort of a question, I'll need to look a little deeper 15:49:49 Makes sense. 15:50:00 Oh, one thing I wanted to mention for open discussion 15:50:19 right now there are patches up in openstack/releases to retire Ironic Train/Ussuri branches due to the unmaintained branches migration 15:50:36 AIUI devstack/various supporting CI cast things are going away for Train/Ussuri 15:50:55 Riccardo Pittau proposed openstack/bifrost master: Fix disable-dhcp option in playbook https://review.opendev.org/c/openstack/bifrost/+/903135 15:51:02 if anyone wants to keep these up, and in unmaintained/* branches, they have [slightly less than, at this point] a month to go -1 that patch and volunteer to be the steward of those branches 15:51:27 https://review.opendev.org/c/openstack/releases/+/903199 15:51:59 #info Train is scheduled to be EOL: https://review.opendev.org/c/openstack/releases/+/903199 -- if you wish it to not EOL, you must -1 that patch and volunteer to steward the branch. 15:52:08 Anything else about that or for open discussion? 15:52:29 I guess the Train is soon departing the last station 15:53:51 4+ years is too good of a run for a single release of software :D 15:53:58 I'm going to close out the meeting if there's nothing else? 15:54:11 ... unless we're required to release for 5 years 15:54:48 https://www.digitalsme.eu/cyber-resilience-act-the-eu-strikes-a-deal-on-security-requirements-for-digital-products/ 15:55:34 I think this is confirmation we're off the Ironic path for the day :D 15:55:36 #endmeeting