Wednesday, 2021-09-29

opendevreviewVerification of a change to openstack/ironic master failed: Use an ImageCache for provided boot/deploy ISO images  https://review.opendev.org/c/openstack/ironic/+/81024101:20
*** pmannidi is now known as pmannidi|brb04:00
*** pmannidi|brb is now known as pmannidi04:19
opendevreviewMerged openstack/ironic master: require_exclusive_lock: log traceback that lead to an error  https://review.opendev.org/c/openstack/ironic/+/81061505:31
opendevreviewMerged openstack/ironic master: Refactor: move base_iso handling from create_boot_iso  https://review.opendev.org/c/openstack/ironic/+/81007605:33
opendevreviewVerification of a change to openstack/ironic master failed: Use an ImageCache for provided boot/deploy ISO images  https://review.opendev.org/c/openstack/ironic/+/81024105:47
iurygregorygood morning janders and Ironic o/06:10
opendevreviewVerification of a change to openstack/ironic master failed: Use an ImageCache for provided boot/deploy ISO images  https://review.opendev.org/c/openstack/ironic/+/81024106:22
rpittaugood morning ironic! o/06:59
iurygregorymorning rpittau o/07:15
rpittauhey iurygregory :)07:15
opendevreviewAija Jauntēva proposed openstack/ironic stable/xena: Fix iDRAC configuration mold docs  https://review.opendev.org/c/openstack/ironic/+/81149507:29
opendevreviewAija Jauntēva proposed openstack/ironic stable/wallaby: Add iDRAC configuration mold docs  https://review.opendev.org/c/openstack/ironic/+/81149607:30
*** pmannidi is now known as pmannidi|AFK07:38
jandersgood morning iurygregory rpittau and Ironic o/07:44
rpittauhey janders :)07:44
iurygregoryjanders, o/07:44
dtantsurTheJulia: ack re molteniron07:48
dtantsurgood morning ironic07:48
rpittauhey dtantsur :)07:48
iurygregorymorning dtantsur 07:48
dtantsurayoung: welcome!07:48
iurygregoryanything very urgent we need to take a look?07:49
* dtantsur no idea07:49
jandershey dtantsur o/07:50
rpittauthe issue wit the CI should be fixed, so maybe some rechecks are in order ?07:52
iurygregoryrpittau, ack07:52
dekeHi07:59
dekeWhat is the current status of Xilinx drivers for Ironic?07:59
iurygregoryhumm I don't recall any Xilinx driver in Ironic08:04
dekesorry I actually meant cyborg08:05
dekewrong channel08:05
iurygregoryoh no worries deke =)08:05
rpittaustill something to do with metal :P08:06
dekeclose enough haha08:06
dekewhile I'm here though, quick question08:07
iurygregorygo ahead =)08:07
dekehas there been any discussion about integration between ironic and MAAS?08:07
dtantsurdeke: I don't think anyone has ever expressed any interest in that08:08
iurygregorynot that I'm aware08:08
dtantsurI given the past rhetoric of Canonical.. I doubt it08:08
dekeIf we deploy Openstack with juju on MAAS, it seems silly to also then deploy ironic on top of that when MAAS is already there doing baremetal provisioning. 08:09
dekejust a thought08:09
dekewhat rhetoric do you mean?08:09
dtantsurdeke: there have been discouraging public statements, let me leave that at this08:09
dtantsurif you already have bare metal provisioning, why you need ironic?08:10
dekewell, for end users of the cloud who don't have access to MAAS08:10
dekeIs the relationship between canonical and openstack rocky? I was unaware of this08:11
dtantsurI wouldn't put it like that. But there is no cooperation between the MAAS team and Ironic, as far as I'm aware.08:11
dekeSay I have an end user who needs a dedicated baremetal server orchestrated via the horizon dashboard. Right now, my only option is to deploy ironic and enable that for them08:11
dtantsurdeke: what you need in theory, is a Nova driver for MAAS, not Ironic integration08:12
dekeyea that sounds more accurate08:12
rpittauor install ironic with juju08:12
* dtantsur -> doctor appt, bbl08:12
rpittauand keep maas and ironic separated08:12
dekedo they play nice?08:12
dekeif I have a baremetal machine setup with MAAS and then I use ironic to deploy something else, will MAAS pick it back up afterward no problem? I have not tested this08:13
rpittaunever tried it :)08:13
rpittaubut I'm pretty sure you will find some guide on that08:13
dekeinteresting08:14
dekethanks guys!08:14
opendevreviewAija Jauntēva proposed x/sushy-oem-idrac stable/xena: Update .gitreview for stable/xena  https://review.opendev.org/c/x/sushy-oem-idrac/+/81161108:27
opendevreviewAija Jauntēva proposed x/sushy-oem-idrac stable/xena: Update TOX_CONSTRAINTS_FILE for stable/xena  https://review.opendev.org/c/x/sushy-oem-idrac/+/81161208:27
opendevreviewRiccardo Pittau proposed x/sushy-oem-idrac master: Add Python3 yoga unit tests  https://review.opendev.org/c/x/sushy-oem-idrac/+/81161308:33
opendevreviewAija Jauntēva proposed x/sushy-oem-idrac master: Add Python3 Yoga unit tests  https://review.opendev.org/c/x/sushy-oem-idrac/+/81161408:34
ajyauff08:34
rpittaud'oh08:34
rpittauI will abandon mine, it's ok08:35
ajyaok, thanks, was a bit too slow :) now just don't abandon mine 08:35
ajya(I'll do the same for python-dracclient just in case)08:36
rpittauok! :)08:36
mgoddarddtantsur: hi. Regarding https://storyboard.openstack.org/#!/story/1737947, we generally use the dnsmasq filter these days. Unsure if the iptables filter is still affected08:47
opendevreviewMerged x/sushy-oem-idrac stable/xena: Update .gitreview for stable/xena  https://review.opendev.org/c/x/sushy-oem-idrac/+/81161108:58
opendevreviewMerged x/sushy-oem-idrac stable/xena: Update TOX_CONSTRAINTS_FILE for stable/xena  https://review.opendev.org/c/x/sushy-oem-idrac/+/81161209:00
opendevreviewVerification of a change to openstack/ironic master failed: Use an ImageCache for provided boot/deploy ISO images  https://review.opendev.org/c/openstack/ironic/+/81024109:19
dtantsurmgoddard: so, does it work okay with the dnsmasq filter?10:00
mgoddarddtantsur: yes, seems to AFAIK10:01
dtantsurthanks!10:01
dtantsurdeke: you have to be aware of potential DHCP conflicts10:01
opendevreviewMerged openstack/ironic master: PXE: avoid trailing dots in exception messages  https://review.opendev.org/c/openstack/ironic/+/81082810:30
opendevreviewDerek Higgins proposed openstack/ironic master: Exit ipxe script if enable_netboot_fallback failed  https://review.opendev.org/c/openstack/ironic/+/81171010:42
opendevreviewDerek Higgins proposed openstack/ironic master: Exit ipxe script if enable_netboot_fallback failed  https://review.opendev.org/c/openstack/ironic/+/81171010:48
opendevreviewVerification of a change to openstack/ironic master failed: Use an ImageCache for provided boot/deploy ISO images  https://review.opendev.org/c/openstack/ironic/+/81024110:52
janderssee you tomorrow Ironic o/12:01
dtantsurwow, what a thunderstorm we have here13:02
ayoungCan the CLI output be made to show a nested column value?  LIke, to get the ipmi address of a node, I need driver_info.ipmi_address.  -c  driver_info  works, but it gives the whole dictionary.  I know I can pull a value out of there using jq, but was wondering if there is some format that allows the CLI to do that in a shorter format?13:29
ayoungI mean this is not that bad13:30
ayoungfor UUID in `openstack baremetal node list --provision-state "clean failed"  -f value -c UUID` ; do openstack baremetal node show  $UUID -f json | jq -r  '.driver_info | .ipmi_address' ;done13:30
dtantsurthis looks about right13:35
TheJuliadtantsur: Well, fall storms are a thing there right?13:35
dtantsurTheJulia: pretty much, although usually without thunder13:35
dtantsuralso good morning TheJulia 13:35
TheJuliaayoung: pathces welcome?13:35
iurygregorygood morning TheJulia =)13:37
TheJuliadtantsur: well handled w/r/t the maas questions13:39
dtantsur:)13:39
TheJuliaPity there were no keynote video links13:39
TheJuliabut I think the worst actually didn't get posted by OIF staff13:40
dtantsurheh, yeah, I didn't want to start the whole archaeology 13:41
TheJuliayeah13:41
TheJuliaa waste of energy really13:41
ayoungTheJulia, I'd actually prefer there not be a way to do it, so I can feel justified relying on my preloaded jq know-how.13:41
ayoungis there a write up on what happens during the cleaning process?  How to debug etc?13:44
TheJuliayeah, quick searching did not yield a video13:44
TheJuliaayoung: https://docs.openstack.org/ironic/latest/admin/cleaning.html by default erase_devices or erase_devices_metadata based upon what your settings are13:45
TheJuliaat least, a stock deploy13:45
ayoungto be clear, we get an automated if we do:   openstack baremetal node maintenance unset ${i}13:46
ayoung  openstack baremetal node manage ${i}13:46
ayoung  openstack baremetal node provide ${i}13:46
ayoungright?13:46
TheJuliaare you in maintenance mode right now?13:47
TheJuliathe latter to commands, yes. Between provide and the node becoming available, automated cleaning is run13:48
TheJuliadtantsur: https://storyboard.openstack.org/#!/story/list?status=active&project_group_id=75&tags=rfe <-- down to 19513:48
dtantsur\o/13:48
dtantsurvery nice13:48
ayoungso if I do a provide (from manage) I ge an automated clean.  Is there a way to list the steps that are (or will be) performed?13:48
TheJuliaayoung: it is generated dynamically from the hardware managers loaded in IPA. The in-built default includes erase_devices and erase_devices_metadata and global configuration knobs in your conductor(s) ironic.conf files 13:50
TheJuliawhich allows you to tune if they run at all or which one to run13:50
TheJuliaobviously scrubbing disks can take a while13:50
TheJuliametadata is easy13:50
ayoungSo there would be no way to know until it is actually run, as it needs to PXE boot the "clean me" Kernel image, which loads IPA?13:51
dtantsuryep13:53
TheJuliayup13:54
TheJuliathere have been various proposals to make an api to figure it out13:55
opendevreviewMerged openstack/ironic master: Use an ImageCache for provided boot/deploy ISO images  https://review.opendev.org/c/openstack/ironic/+/81024113:55
TheJuliabut... given the hardware manager nature, and the demands of "i want it absolutes right now", we reached a conflict over operational consistency and continuity of workloads because the proposals were basically ending up in kill the running workload if queried and boot the ramdisk to find out because caching was unacceptable.13:57
ayoungAPI is a step beyond where I need.  I just need to get smart on what would be done right now, if I have full access to everything13:59
ayounglike, can I restart a clean from a 'clean wait' state?13:59
ayoungThe requested action "manage" can not be performed on node "5411e7e8-8113-42d6-a966-8cacd1554039" while it is in state "clean wait". (HTTP 400)13:59
dtantsurayoung: you need to "abort" it first13:59
TheJuliaand you may be doing a full erase_deviecs14:00
ayoungyeah but how would I know?14:00
TheJuliaif you look at the entire node detail you should see clean_steps14:00
ayoungI think I can attach to the machine via ipmi SOL14:00
dtantsurayoung: the current step is in the 'clean_step' field14:01
ayoungdriver_internal_info14:02
ayoungopenstack baremetal node show  5411e7e8-8113-42d6-a966-8cacd1554039 -f json  | jq '.driver_internal_info | .agent_cached_clean_steps'14:03
ayoungthat is hte "plan" I take it?14:03
dtantsurayoung: I think during the process you have just "clean_steps" which are the steps that will be run14:04
ayoung"clean_step": {},14:05
dtantsurthis means no step is running now14:05
ayoungso I assume that this means it is waiting for PXE to finish14:06
ayoungAnd a call back from IPA on the node14:06
dtantsurcorrect14:06
ayoungthis is good stuff14:07
ayoungOK  let me see if I can get something off the node14:07
ayoungSOL is just hanging.  How do I abort the process to kick it off again?14:07
dtantsurayoung: baremetal node abort <node>14:08
ayoungk14:08
dtantsuryou may have to configure where the serial console goes via the kernel params14:08
dtantsurI don't remember if we have docs for that...14:08
ayoung openstack baremetal node show  5411e7e8-8113-42d6-a966-8cacd1554039 -c provision_state14:08
ayoung+-----------------+--------------+14:08
ayoung| Field           | Value        |14:08
ayoung+-----------------+--------------+14:08
ayoung| provision_state | clean failed |14:08
ayoung+-----------------+--------------+14:08
dtantsuryou may also try runtime SSH configuration via something like https://docs.openstack.org/diskimage-builder/latest/elements/dynamic-login/README.html14:08
ayoungkoolio14:08
ayoungso I can do the three above operations to kick it off again...14:09
TheJuliaayoung: yes, just it woudl likely help to do what dmitry notes with your ipa image and/or look at the logs uploaded to the conductor14:16
TheJuliathe logs *should* yield why14:16
dtantsurif it's a PXE timeout, there will be no logs uploaded14:17
TheJulia++14:17
TheJuliaand things like SOL would be useful then14:17
ayoungOK...logs.  I have the server side logs, but should there be an ipa log on the node?  I can get in via SOL14:32
ayoungdebian instance is running, ipa is running14:32
TheJuliaayoung: so if uploaded, /var/log/ironic/deploy_logs I think14:32
TheJuliaif it never heartbeated and just timed out14:32
ayoungon the server?14:32
TheJuliaLike...after 10 minutes, then that would mean network connectivity issues *or* the ramdisk didn't boot14:33
ayoungNO ip address on the node14:33
TheJuliaayoung: on the conductor, on the ipa image, it gets written out to the system logger so journald or whatever file for the service 14:33
TheJuliaayoung: \o/ does your OS support your nics?14:33
TheJuliahow was the ramdisk built is another question14:34
dtantsurayoung: how did you build your IPA image?14:34
dtantsurI think debian does not DHCP by default, I needed to add the dhcp-all-interfaces element (or simple-init as an alternative)14:34
ayoungrunning dhclient gets me an ip address14:35
dtantsuryep, it sounds like nothing DHCPs. on RH systems NetworkManager tries to do something (sometimes wrongly), on Debian there is nothing by default.14:35
ayoungand in conductor  2021-09-29 07:35:09.402 6 INFO ironic.conductor.task_manager [req-1af9106c-9850-4ab2-93a3-811f758b3b94 - - - - -] Node 5411e7e8-8113-42d6-a966-8cacd1554039 moved to provision state "available" from state "cleaning"; target provision state is "None"14:35
ayoungI did not build the IPA image.  I can ask the person who did14:36
dtantsur++14:36
TheJuliaayoung: that seems like successful cleaning14:36
TheJuliaconductor log wise14:36
dtantsuryeah, probably DHCP was the issue14:36
ayoungAnd now SOL is locked,  guessing it is powered down14:37
dtantsurshould be, unless you disabled it explicitly14:37
ayoung| 5411e7e8-8113-42d6-a966-8cacd1554039 | jade05-r097     | None                                 | power off   | available  14:38
ayoungOK, these are the kind of problems we like to have14:38
dtantsurokay, I need to leave now.. and I don't feel like going outside at this moment :(14:56
TheJuliadtantsur: :(14:59
dtantsurokay, the thunderstorm seems to relax a bit, see you o/15:12
rpittaubye dtantsur 15:12
rpittaugood night! o/16:12
TheJuliao/16:13
stevebakermorning19:49
TheJuliagood morning!19:51
TheJuliastevebaker: are you available tomorrow around this time. Well, starting in like 23 hours 38 minutes ?19:52
stevebakerTheJulia: I certainly am19:56
opendevreviewJulia Kreger proposed openstack/ironic master: Yoga: Change default boot mode to uefi  https://review.opendev.org/c/openstack/ironic/+/81024920:03
TheJuliastevebaker: excellent, I'll send a meeting invite!20:03
TheJuliaand ^^^ was ouchy20:03
TheJuliathe patch that was20:03
TheJuliasent!20:05
TheJuliastevebaker: fwiw, your not on downstream irc :)20:15
opendevreviewJulia Kreger proposed openstack/ironic master: Remove debian packages file for devstack  https://review.opendev.org/c/openstack/ironic/+/81031420:52
TheJuliajust a rebase since the main bindep file should work20:52
opendevreviewJulia Kreger proposed openstack/ironic master: Yoga: Change default boot mode to uefi  https://review.opendev.org/c/openstack/ironic/+/81024921:20
opendevreviewJulia Kreger proposed openstack/ironic master: DNM/WIP - Add FIPS jobs  https://review.opendev.org/c/openstack/ironic/+/79773921:43
* TheJulia goes and exercises21:47
*** pmannidi|AFK is now known as pmannidi23:58

Generated by irclog2html.py 2.17.2 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!