Friday, 2025-11-21

JayFcardoe: https://youtu.be/Vtrdr-nxvAA#t=15s00:00
TheJuliaso much troublemaking :)00:07
TheJuliacardoe: two questions: 1) could I get the ironic api log just to be super sure, specifically I just want to see the high level requests/transations for the node. 2) Can you confirm there are no orphaned vif records?00:21
cardoeyes.. by orphaned vif you mean a neutron port right? or something in nova?00:22
opendevreviewMerged openstack/ironic stable/2025.2: Filter null NIC firmware versions from cache  https://review.opendev.org/c/openstack/ironic/+/96677300:52
TheJuliacardoe: yes, recorded on an ironic node01:20
*** mdfr5 is now known as mdfr04:30
rpittaugood morning ironic! happy friday! o/07:30
opendevreviewDavid Nwosu proposed openstack/ironic master: Move configdrive functions to configdrive_utils and configdrive tests to test_configdrive_utils  https://review.opendev.org/c/openstack/ironic/+/96588008:42
opendevreviewDavid Nwosu proposed openstack/ironic master: Enable codespell to use pyproject.toml for spelling exceptions  https://review.opendev.org/c/openstack/ironic/+/96796508:51
rpittauwe really need to merge https://review.opendev.org/c/openstack/ironic/+/96782109:45
rpittauIt's just bringing things back to what they were before the change in oslo.service, this is blocking ironic and metal3 CI :/09:45
*** tkajinam is now known as Guest3183611:05
alegacycid: JayF: dtantsur: TheJulia: ... friendly reminder to those of you that did a 1st pass on my patchset to swing by and take a look at the latest revision! please and thank you:  https://review.opendev.org/q/topic:%22feature/standalone-networking%22+status:open14:01
dtantsuron my plans for today14:01
alegacydtantsur: thank you!14:02
zigoHi. scciclient fails because it's not pysnmp 7 compatible:14:04
zigohttps://bugs.debian.org/111742614:04
zigoIs it still maintained?14:04
TheJuliaalegacy: also my time to spend some time reviewing today as well14:05
alegacyTheJulia: perfect!14:06
cardoeTheJulia: I'll get ya those logs today but someone just asked me about... "The mac address 14:23:f3:f4:c7:e0 is in use." when a build hit one of the servers that didn't clean up the other day.14:08
cardoeSo I'm gonna guess that the answer to your "was a vif left around" is gonna be yes14:08
TheJuliazigo: for your purposes, the answer is likely no, and I believe a patch is up to deprecate the fujitsu driver which would ultimately remove python-scciclient as a dependency14:09
zigoTheJulia: FYI, because of that, Ironic was removed from Debian Testing. :/14:09
TheJuliawe literally just got them to agree in writing this week. removal/deprecation wise, our hands are tied by openstack community process14:11
zigoOk, thanks for the info.14:12
TheJuliacardoe: mac in use in neutron, oh absolutely, I was meaning on the ironic node, which I don't think generates that error, I think that one comes from neutron14:12
zigoThough it's only the SNMP part, which isn't much.14:12
zigoIt's a shame, we should only remove that part, IMO.14:12
cardoethe internal info on the baremetal port still shows a neutron vif as well.14:12
zigoFYI, that's what I'm currently doing.14:12
skrobulcardoe: yup, the vif was left around - https://gist.githubusercontent.com/skrobul/21b9961d2bdb686cf26819ed966108bb/raw/4b1f582ea31ae1988f842203cc7904d4b998abd2/leftover.txt  14:13
cardoeTheJulia: ^ that answers that14:13
zigo(ie: remove snmp tests for scciclient, and fix the pysnmp imports)14:13
TheJuliacardoe: not really14:13
TheJuliacardoe: that is neutron, you crashed neutron14:13
TheJuliaI'm talking about vif attachment records in ironic14:13
TheJuliaopenstack baremetal node vif list <node_uuid>14:13
skrobulTheJulia: in that case no, it's empty. The 'internal info' in 'baremetal port list --long' is empty too.14:17
TheJuliacool cool14:21
TheJuliaThanks14:21
zigoThis one also worries me: https://bugs.debian.org/1117734 is pykmip still used by Ironic ?14:21
zigoFrom the bug report:14:22
zigo"There is a suggested fix as a PR on the upstream github14:22
zigoat https://github.com/OpenKMIP/PyKMIP/pull/707, but it seems like the14:22
zigoproject is abandoned."14:22
* TheJulia raises an eyebrow14:22
JayFIf we use pykmip it's a transitory dependency, not a direct one14:23
zigoMy bad, it looks like barbican uses it, not Ironic.14:26
zigoListed as extras, and in tests.14:27
zigoThe proposed patch seems easy enough to add, I'll try.14:27
cardoethe pysnmp stuff is "solved" by dropping proliantutils14:28
TheJuliacardoe: python-scciclient is also impacted14:30
TheJuliaSpeaking of, who was going to put the deprecation in for ilo stuffs?14:32
cardoeIt's there14:34
cardoehttps://review.opendev.org/c/openstack/ironic/+/96500914:35
cardoeIt's unfortunately gonna cause me to not upgrade to 2026.1 for a while14:36
cardoeCause I've got ilo5 stuff still14:36
cardoeFor my test gear14:36
TheJuliaThats removal, we need to deprecate first14:46
opendevreviewMerged openstack/ironic master: Fix singleprocess launcher compatibility with oslo.service 4.4+  https://review.opendev.org/c/openstack/ironic/+/96782115:23
JayFTheJulia: to deal with that is the same as irmc right: we're only deprecating it until the SNMP driver is dead next cycle15:39
TheJuliaYeah, I think so15:39
dtantsurTheJulia: the VXLAN spec draft will be easier to read if you link it from the index (and thus cause it to actually get rendered)17:00
dtantsurit does seem to render locally even now, I wonder why I get 404 on the preview in the CI17:02
TheJuliaweeeeird17:09
TheJuliaack, I thought I fixed that17:09
TheJuliaThanks for the heads up17:09
* TheJulia finally gets off vxlan call and falls over dead17:09
dtantsurSure. I'm afraid my contribution to this document will be a few readability nits.17:10
TheJuliaThat is prefectly okay17:10
dtantsurSo far, there is no paragraph that I've understood :D17:10
TheJuliaThat is perfectly okay17:10
dtantsurAll I understand is that you people are way too smart for me17:10
TheJuliaIt is a subset of context acquired by drinking delicious beverages17:11
TheJuliaMeanwhile, livers complain17:11
dtantsurYeah, mine won't survive enough alcohol17:11
TheJuliafair enough17:11
TheJuliaDoug started a call close to 2 hours ago and I finally got off a call with the last neutron person a few minutes ago17:12
dtantsurImpressive dedication17:12
TheJuliaMy spec proposes two separate aspects, an broad idea of connectivity as it relates to OVN+neutron, and then the NGS side impact. The broad connectivity idea *may* end up in something like networking-baremetal or elsewhere, but what cardoe has been pushing to bring clarity for first is the port binding and segmentation issues first17:14
TheJuliathen the actual layer 2 logical segment bridging as a secondary aspect17:14
dtantsurThe spec could actually use some sort of "dummy's guide to this spec" :D as in, summary for people, for whom FRR is the sound that a cat can make.17:16
TheJulialol, okay17:18
TheJuliaIf this becomes a thing *in* ironic, we must name modules after colors of cats and their behaviors. I'm just unsure which thing should be the orange cat or the black cat (commonly referred to as a void)17:19
dtantsurOrange cat should regularly send packages where they're not supposed to be17:20
dtantsur"Why are 10% of packages dropped?" "Have you tried looking under the fridge?" "Oh heck, there are gigabytes of them here!!"17:24
cardoeThat's the thing. It's complicated. And every time I make a simple representation people tell me "well why can't it just be X".17:24
dtantsurMaybe calling it "Laymen summary" will be offensive enough to people who know networking to prevent them from commenting?17:25
dtantsurLayhumans/Peasants/Dmitry-who-should-have-known-networking-after-all-these-years-here summary, pick yours17:26
TheJuliacardoe: my favorite is "just use type5, that is what we're standardizing on"17:27
TheJuliadtantsur: would "Friends, Romans, Countrymen, lend me your ears" style work?17:29
dtantsurPerfect!17:29
TheJulia"As if your Marcus Antonius, please give me a simple, easy to understand primer"17:30
* TheJulia expects latin as the result17:30
TheJuliadtantsur: does https://review.opendev.org/c/openstack/neutron/+/965415 help as a primer?17:44
dtantsurI'll have to read it on Monday with a fresh head, thanks17:47
TheJuliaok17:48
cardoedtantsur: https://cardoe.com/neutron/evpn-vxlan-network/admin/data-center-networks.html is a rendered copy of 965415 as well.17:53
dtantsurnice!17:54
dkingdtantsur: I have an unimportant question about BMO. For https://github.com/metal3-io/baremetal-operator/blob/main/config/base/rbac/kustomization.yaml, I see that most of the rbac .yaml aren't being used. Is that intentional? Am I missing something, or do those get picked up on some specific kustomize?17:58
dtantsurdking: I suspect they're auto-generated examples (you mean stuff like *_role.yaml right?)18:00
dtantsurhttps://github.com/metal3-io/baremetal-operator/blob/main/config/base/rbac/role.yaml is what the actual service user uses18:00
dkingdtantsur: Correct. I see that role.yaml and about 8 other files are used, but that's out of about 25 files.18:02
TheJuliaclif: i just -1'ed to request a quick revision on 964895. If you want to do it as a follow-up, thats fine. Hopefully test order won't change and will just work as well, but that might be another reason to use id value over uuid :) Just let me know if you want to do it as a follow-up or not18:06
opendevreviewMerged openstack/ironic master: Add a script to copy inspection data between Swift buckets  https://review.opendev.org/c/openstack/ironic/+/96689918:11
clifTheJulia: So you mean sort by ID in descending order (greatest/highest ID would be first)?18:14
TheJuliaYeah, that was also the way I was interpretting JayF's comment as as well18:15
clifJayF didn't seem to explicitly favor one over the other but newer first may make more sense18:17
cardoeSo https://review.opendev.org/c/openstack/ironic-python-agent-builder/+/966192 can we land that?18:23
cardoeor maybe https://review.opendev.org/c/openstack/ironic/+/964570 with a cherry on top?18:24
TheJuliacardoe: I workflowed it and then saw the other comment and un-workflowed the first one18:28
TheJuliaso.. if people are okay we can workflow it18:28
cardoeSo I think there's value in the less combined thing... but already the current one is combined enough people cannot use it18:28
cardoeSo I thought if there's interest in it... we make a "base"18:29
TheJuliafair enough18:29
cardoeThe only reason we wanted the less combined thing is cause the virtual-media not waiting 30 seconds bug18:30
cardoeWe needed to turn off dhcp-all-interfaces18:30
TheJuliathats super bizzar to me18:31
TheJuliasince when we've tested, and maybe its just that specific hardware, I dunno18:32
opendevreviewClif Houck proposed openstack/ironic master: Trait Based Networking Filter Expression Parsing and Base Models  https://review.opendev.org/c/openstack/ironic/+/96149818:37
opendevreviewClif Houck proposed openstack/ironic master: Configuration file for Trait Based Networking  https://review.opendev.org/c/openstack/ironic/+/96259818:37
opendevreviewClif Houck proposed openstack/ironic master: Generate network plan based on trait based networking config  https://review.opendev.org/c/openstack/ironic/+/96489518:37
opendevreviewClif Houck proposed openstack/ironic master: Trait Based Networking Simulator  https://review.opendev.org/c/openstack/ironic/+/96620218:37
opendevreviewClif Houck proposed openstack/ironic master: WIP: Add configuration options for trait based networking  https://review.opendev.org/c/openstack/ironic/+/96805418:37
cardoeTheJulia: what's the ironic-api log snippet you're looking for?18:47
cardoeCause I included the one with the 500 error18:48
TheJuliacardoe: anything that includes the additional calls made by nova afterwards with their corresponding error codes18:49
TheJuliasince that is just the call itself, the thing which needs to be done is ensure/verify where things went sideways on the unwind18:50
cardoe2025-11-20 12:19:03.160 8 INFO ironic.api [None req-899167ff-7be6-4a7c-b72f-ce5c4ca9becd 6dff409ebb31414299c3d0cd837eea9e 32e02632f4f04415bab5895d1e7247b7 - - a6f7dcd63c9b4940915062f57a48df77 7f46f53fcb3c4625a343eaa35b5e0d04] 10.64.48.142 - DELETE /v1/nodes/86eb7354-cc10-4173-8ff2-d1ac2ea6befd/vifs/ecf586d0-99c2-4867-b516-6fa2c39c30b1 - 204 (1537.92ms)18:52
cardoeThat seems good an validated what skrobul said above18:52
TheJuliaOkay, any others, specifically looking for a PATCH call18:52
cardoe2025-11-20 12:19:03.889 8 INFO ironic.api [None req-7880e28e-bbc9-4c00-8e88-aa946c5500d6 6dff409ebb31414299c3d0cd837eea9e 32e02632f4f04415bab5895d1e7247b7 - - a6f7dcd63c9b4940915062f57a48df77 7f46f53fcb3c4625a343eaa35b5e0d04] 10.64.48.142 - PATCH /v1/nodes/86eb7354-cc10-4173-8ff2-d1ac2ea6befd - 200 (609.26ms)18:53
TheJuliaWUT18:53
TheJuliaoh, that is a different failure18:53
TheJulialook at the times on https://gist.github.com/cardoe/b0aefe21b1fc7b81c38bed8dad8e14b218:53
cardoeLook at my fail-y copy and paste with the different times.... it's that node. grafana copy and paste got fixed18:54
cardoeThat's 2 seconds after the failure18:54
cardoe2025-11-20 12:19:01.544 8 INFO ironic.api [None req-2415f97c-664f-4a61-8241-12161991eb12 6dff409ebb31414299c3d0cd837eea9e 32e02632f4f04415bab5895d1e7247b7 - - a6f7dcd63c9b4940915062f57a48df77 7f46f53fcb3c4625a343eaa35b5e0d04] 10.64.48.142 - POST /v1/nodes/86eb7354-cc10-4173-8ff2-d1ac2ea6befd/vifs - 500 (7047.78ms)18:55
cardoeThat's the one that started it18:55
TheJuliaoh, okay18:56
TheJuliaso... hmm18:56
TheJuliaand that patch is the last call on that node?>18:57
cardoeugh18:57
cardoeyou're right its a race18:57
cardoecause that one looks like it didn't have the instance-uuid left behind18:57
TheJuliabut another attempt did, so I bet it has a different error code on the patch18:57
TheJuliaI *bet* the lock was still unwinding from the vif removal18:58
TheJuliaThat is super rapid succession18:58
TheJuliaand maybe the thing to do is resolve the request and unwind the lock before returning the response18:58
TheJuliaI *bet* that would actually fix it18:58
cardoe2025-11-20 12:37:00.921 8 INFO ironic.api [None req-75fab2b4-b31f-4ff4-9609-b18b4b81fadf 6dff409ebb31414299c3d0cd837eea9e 32e02632f4f04415bab5895d1e7247b7 - - a6f7dcd63c9b4940915062f57a48df77 7f46f53fcb3c4625a343eaa35b5e0d04] 10.64.48.142 - PATCH /v1/nodes/86eb7354-cc10-4173-8ff2-d1ac2ea6befd - 409 (2588.32ms)18:59
cardoeThat's a run with instance-uuid left behind18:59
cardoeEverything else is the same error wise.. its the same 500 error with the Could not seed network configuration for VIF19:00
opendevreviewJulia Kreger proposed openstack/ironic master: doc: trivial: Quick revision of README  https://review.opendev.org/c/openstack/ironic/+/96805619:00
TheJuliaYeah, a held lock will do that19:01
TheJuliacardoe: do you have defaults for [conductor]node_locked_retry_interval and [conductor]node_locked_retry_attempts ?19:13
TheJulialikely not since it should be 1 and 3 respectively19:15
cardoenot set in the file19:18
TheJuliaso, its running long, but it shoudl be taking longer than 2.588 seconds19:19
cardoethis was on a box that's got everything converged on it and hosting virtual nodes so its fairly slammed 19:19
TheJuliacould be one extra retry is getting in there I guess19:19
opendevreviewJulia Kreger proposed openstack/ironic master: WIP: Downgrade the lock on vif detach  https://review.opendev.org/c/openstack/ironic/+/96806119:34
TheJuliaoh, 2.5ms makes sense because of the way tenaicy works, the first one counts immediately and 1 second between each19:39
TheJuliaso yeah, its just unwinding.19:39
TheJuliaSo yeah, I think downgrading the lock is only the real viable solution, the alternative is to play with extending the retries19:46
opendevreviewJulia Kreger proposed openstack/ironic master: WIP: Downgrade the lock on vif detach  https://review.opendev.org/c/openstack/ironic/+/96806119:51
TheJuliacardoe: give ^^^ a spin if you wouldn't mind on your ironic conductor. Otherwise the lock us unwound in __exit__ which I'm wondering might be a bit too late. The alternative is really tuning, or going going a "this is somehow imperative, and we need to keep retrying", update_node and task reservation locking code is only the basic retries on the update_node19:55
JayFTheJulia: +1 to the suggested solution of ensuring the lock is unwound before the response is given (in general we should see if we can enforce this pattern)19:55
JayFlol I'm a little late I see 19:55
opendevreviewMerged openstack/ironic-python-agent stable/2025.2: Test advertised ip reachability before assigning it  https://review.opendev.org/c/openstack/ironic-python-agent/+/96667119:55
TheJuliaThe call is "exception or success through no exception", __exit__ should be unwinding it all sooner, but yeah... 19:56
TheJuliaat least, sooner than the call but that might jsut be racey overall19:56
TheJuliaThe alternative is largely to add some more loggin to understand why the node lock was still held19:57
TheJulialogging19:57
JayFIt's weird to me, given the retry config, that the lock could be held so long19:57
JayFbecause if we 403 to nova it should retry 5 times which takes about 8 seconds19:58
TheJuliawell19:58
TheJuliawe 40919:58
JayF409 is what I meant19:58
JayF409/503 gets retried by default in osdk19:58
TheJuliaWe 500, then 409, I'm not sure nova is retrying on a 409, it has a general exception catch its in19:58
TheJuliaoh, news to me19:58
JayFthis is the road cid and I was going down19:58
JayFwhen we saw the servicing bug19:59
JayFby default Ironic osdk clients retry 503/409 up to 5 times19:59
TheJuliabut functionally what is happening is we get vif detach, we're returning it but the lock is still somehow there, in rapid succession nova is immediately kicking back. As doug noted, if the machine is under load, things might get a little funky19:59
JayFI really suspect that cardoe is running something that's subduing retries19:59
TheJuliaThat... might also do it20:00
JayFlike configuring it to retry less either intentionally or unintentionally20:00
TheJuliaI just don't know why that lock was sitting around for so long20:00
TheJuliamaybe power state sync snuck in?!20:00
JayFnot saying we should require the retry to work properly by any means, but that the story here is missing the "why did retries not happen" and/or "why was the lock held for several seconds so retries didn't matter"20:00
TheJuliaThe *only* alternative I can think of is "recognize this is a complete detach of the instance metadata due to instance_info and instance_uuid both being reset, and hard lock the update or somehow have more aggressive internal retry logic20:01
TheJuliaBut that feels super complex possibly20:01
TheJuliaand leads down a path of wanting to over-engineer20:01
JayFwell where do we die with the lack of lock20:01
opendevreviewMarcus Furlong proposed openstack/sushy master: remove tests for boot and actions missing attributes  https://review.opendev.org/c/openstack/sushy/+/96806920:02
JayFinternal retries on requests that involve removal of instance_uuid is a really, really clever idea depending on where it'd blow up20:02
JayFmaybe with deferred task framework, we could one day do a "accepted" type of response for undeploys in cases of locked node20:02
JayFlike have a class of things that if locked, we will do it later instead20:02
JayFIDK20:02
* JayF spitballing 20:03
TheJuliaeverything points to update_node call handling inside manager.py20:05
TheJuliaI dunno, I'd rather not modify the task_manager with an optional retry override20:05
TheJuliabut if cardoe can give some insight, that will help in the end20:06
opendevreviewVerification of a change to openstack/ironic-python-agent-builder master failed: Add simple-init by default  https://review.opendev.org/c/openstack/ironic-python-agent-builder/+/96619220:39
opendevreviewNahian Pathan proposed openstack/ironic master: Reduce API calls when collecting sensor data with redfish  https://review.opendev.org/c/openstack/ironic/+/95548420:51
opendevreviewJulia Kreger proposed openstack/ironic-python-agent master: ci: disable vnc on IPA jobs  https://review.opendev.org/c/openstack/ironic-python-agent/+/96807820:56
opendevreviewJulia Kreger proposed openstack/ironic-python-agent master: Remove PReP support  https://review.opendev.org/c/openstack/ironic-python-agent/+/96539021:03
JayFthere was a change already up for that?21:10
JayFhttps://review.opendev.org/c/openstack/ironic-python-agent/+/96420021:10
JayFcid proposed it before the ptg even hit21:10
JayFoh, deprecate vs remove21:11
JayFhm21:11
opendevreviewJulia Kreger proposed openstack/ironic master: deprecate PReP partition support  https://review.opendev.org/c/openstack/ironic/+/96808121:21
JayFTheJulia: ^21:22
JayFI'm not sure which direction we wanna go, but the patches in flight right now are duplicating work :/21:22
TheJuliaoh, I had no idea cid was going to do that21:25
TheJuliawell, Odds are we can safely remove from IPA as long as we don't remove the call handling in ironic until the deprecation, but *shrugs*21:25
JayFit was done pre-ptg :)21:26
TheJuliaoh, really?!21:26
TheJuliawow21:26
TheJuliaokay21:26
JayFyeah, it's not labelled -prio21:26
JayFso it's sorta in a weird limbo21:26
TheJuliaI mean, we should review patches outside of -prio, but prio does give us a smaller list :)21:28
JayFI actaully stopped using prio dash for the most part21:30
JayFunless I'm really, really low on time21:30
JayFstarted noticing too many non-core patches getting big review delays, so I just use my "anything in ironic repo I haven't reviewed" search21:31
TheJuliacid, on https://review.opendev.org/c/openstack/python-ironicclient/+/955102, you need to update the client's max version, its a bit hidden but... i think in http.py ;)21:32
TheJuliaYeah, sort of the same21:33
JayFspeaking of review; I need to make 100% sure alain gets a review from me today21:33
* JayF goes and does that21:33
cardoeI've been adding stuff to ironic-week-prio that seems ready.21:34
cardoeBut the queue is backed up21:34
JayFyeah I'm not saying doing that or using that dashboard is bad21:34
JayFbut as someone with a lot of review b/w I try to cast a larger net21:35
TheJuliasome of the items also need to be fixed/rebased/adjusted21:37
JayFIMO as cores we should drop the tag from anything needing code changes that haven't been updated in a couple of days21:38
JayFbut maybe that's a bit aggro :)21:38
TheJuliaYeah21:44
TheJuliaI'm sort of pondering trying to take a day, and that might be something I do this next week, to take a look at a couple of them and just do a "I'm updating it for you"21:44
TheJuliaIts something I tend to do at times but haven't done in recent time21:44
JayFI am like 50/50 on if I'm working Monday, if you wanted to carve some time out for a group review/rebase party, I would make time for it21:45
* JayF is out most of next week going home to NC to visit family21:45
TheJuliaJayF: enjoy!21:45
TheJuliaI'm sort of hoping next week is nice and quiet, I've got some open downstream items which really require people to re-appear who are on PTO right now, so... *shrug*21:46
JayFif we got a third core we could just go land half the world, one person to update, two to review, bang out some patches21:46
TheJulia+++21:46
TheJuliaI might be able to make some time on monday21:46
JayFif so we need to pick a time since I won't just be around all day :) lmk, no rush really21:46
TheJuliaIt looks like, and I'll double check, but I think I only have 1 meeting on Monday21:46
TheJuliamonday morning after the team meeting?21:46
JayFI think I have a vet appt on Monday (for my cats :P) but otherwise I should be clear. Unsure when that is, I hope I put it on the calendar lol21:47
JayFI should be able to make that. Not gonna set an alarm for the meeting though :D 21:47
TheJulialol21:47
TheJuliasounds good21:47
JayFcid: if you are working monday, and can come to the review jam right after team meeting, please do :D 21:48
JayFcardoe: also ^^^ similarly a chance for you to help us shave down that list21:48
TheJuliathat list should shrink, I aggressively approved some backports21:49
TheJuliaBut, yeah.21:49
cid,,,,21:50
opendevreviewMerged openstack/ironic master: Support segmented port ranges  https://review.opendev.org/c/openstack/ironic/+/96772721:50
* cid tests, for some reason I couldn't send messages here for a while now.21:50
cidTheJulia, patch '955102,, ack'ed21:50
cidJayF, sure, I will be working Monday.21:50
opendevreviewcid proposed openstack/python-ironicclient master: A new `instance_name` field to the node object  https://review.opendev.org/c/openstack/python-ironicclient/+/95510221:51
opendevreviewDoug Goldstein proposed openstack/ironic master: fix: glance image member lookup resulted in an empty list always  https://review.opendev.org/c/openstack/ironic/+/96808721:51
cardoeJayF: yep21:51
cardoeI've been slowly approving backports21:51
JayFcardoe: I literally spent my first like, 6 months at GR-OSS just backporting every bugfix since the beginning of time. Satisfying but time consuming work for something so simple.21:52
cardoe^ that's not even what I'm trying to fix but its spamming my logs... so that's fixed now... it needs to be backported to 2025.221:52
cardoebut in all seriousness... do we want to hold off on https://review.opendev.org/c/openstack/ironic/+/964570 ?21:53
cardoeI'm waiting on that to decide how I approach some of the other segment issues21:53
cardoeAs well as refreshing my patch on deleting the check that JayF wanted clif to delete.21:53
JayFI don't have enough context to say one way or the other re 96457021:54
cardoeprobably a TheJulia poke21:55
TheJuliamerge it21:56
TheJuliaMMMEEERGE!21:56
TheJuliacardoe: unless you don't want us to?!21:56
cardoewell I didn't think it right to +W my own patch21:57
TheJuliafair enough21:57
* TheJulia does it21:57
TheJuliadone21:57
JayFcardoe: I don't hesitate to +W my own patch if it has 2x+2 from other cores, as long as it's not something I JUST posted up 21:58
TheJuliasame21:59
TheJuliaIf its been up for a while, I'm okay with it, otherwise I only do stuff like that if its a CI fix or something semi-critical so other stuff can flow21:59
opendevreviewMerged openstack/ironic-python-agent stable/2025.1: Test advertised ip reachability before assigning it  https://review.opendev.org/c/openstack/ironic-python-agent/+/96677622:08
opendevreviewMerged openstack/ironic-python-agent bugfix/11.1: Test advertised ip reachability before assigning it  https://review.opendev.org/c/openstack/ironic-python-agent/+/96677422:08
opendevreviewMerged openstack/ironic-python-agent bugfix/11.0: Test advertised ip reachability before assigning it  https://review.opendev.org/c/openstack/ironic-python-agent/+/96677522:08
opendevreviewMerged openstack/ironic-python-agent bugfix/11.1: Fix RuntimeError when stopping heartbeater in rescue mode  https://review.opendev.org/c/openstack/ironic-python-agent/+/96735622:08
opendevreviewMerged openstack/ironic-python-agent bugfix/11.0: Fix RuntimeError when stopping heartbeater in rescue mode  https://review.opendev.org/c/openstack/ironic-python-agent/+/96735722:08
cardoeMy favorite messages in this channel.22:09
JayFbugfix/11.* made me do a double take22:15
JayFuntil I saw IPA lol22:15
cardoeokay now that this member list thing is quiet... here's the other weird one22:18
cardoe2025-11-21 16:24:09.848 13 ERROR ironic.drivers.modules.image_cache [None req-1642551b-ad53-4692-927d-eca2d6cdf90f 6dff409ebb31414299c3d0cd837eea9e 32e02632f4f04415bab5895d1e7247b7 - - a6f7dcd63c9b4940915062f57a48df77 7f46f53fcb3c4625a343eaa35b5e0d04] Could not link image 5048de36-4c01-4047-b638-195e639cb1e3 from /var/lib/openstack-helm/ironic/master_images/5048de36-4c01-4047-b638-195e639cb1e3.converted to 22:18
cardoe/var/lib/openstack-helm/ironic/images/5180e19d-c3c6-4afb-b626-08d70ec1f456/disk, error: [Errno 2] No such file or directory: '/var/lib/openstack-helm/ironic/master_images/tmpgmxnx_5h/5048de36-4c01-4047-b638-195e639cb1e3.converted': FileNotFoundError: [Errno 2] No such file or directory: '/var/lib/openstack-helm/ironic/master_images/tmpgmxnx_5h/5048de36-4c01-4047-b638-195e639cb1e3.converted'22:18
JayFI have a present for you cardoe 22:20
cardoeThe weekend?22:20
JayFcardoe: https://review.opendev.org/c/openstack/ironic/+/964502 I suspect you're hitting the bug this fixes22:20
cardoehttps://opendev.org/openstack/ironic/src/commit/27c805b97ffd7a4a159860537311815f579d9cf3/ironic/drivers/modules/image_cache.py#L243 I'm trying to figure out if its the first os.link or the 2nd22:20
JayFcardoe: I spent a hilariously long time trying to get a big enough image into glance to test that before ragequitting it for a while, and I haven't gotten back around22:20
cardoeokay so it doesn't happen on 2025.1 and it doesn't happen when I turn caching off.22:22
cardoeso what you're saying is I gotta stop slamming this over loaded box?22:24
cardoeIt's actually not overloaded. Someone played a funny on me and set the k8s resource limit on CPU per pod to like nada.22:25
JayFI think there's a case that can happen (emphasis on the /think/, this is 100% untested hypothesis territory) where your conductor gets timed out while doing the hash22:26
JayFour code that deletes an image after a failed deployment fires22:26
JayFand you sorta end up in a loop of sadness22:26
JayFI know johnthetubaguy has reported having deployments fail the first time and succeed later, but I think that's with a version of ironic old enough to not clear cache on failures22:26
cardoeI'm being an absolute stick in the mud and wanting 2025.2 to work happy on this box that's clearly struggling and not just on regular hardware.22:27
opendevreviewVerification of a change to openstack/ironic-python-agent-builder master failed: Add simple-init by default  https://review.opendev.org/c/openstack/ironic-python-agent-builder/+/96619222:53
cardoeJayF: yes jsonschema works with yaml23:02
opendevreviewMerged openstack/ironic master: Include inspector conf groups in sample conf/docs  https://review.opendev.org/c/openstack/ironic/+/95233823:09
opendevreviewMerged openstack/ironic master: pass along physical_network to neutron from the baremetal port  https://review.opendev.org/c/openstack/ironic/+/96457023:28
JayFcardoe: clif: Then we should probably use jsonschema to do the validation unless there's a specific reason we cannot.23:48

Generated by irclog2html.py 4.0.0 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!