15:00:18 #startmeeting openstack_ansible_meeting 15:00:19 Meeting started Tue Jun 8 15:00:18 2021 UTC and is due to finish in 60 minutes. The chair is noonedeadpunk. Information about MeetBot at http://wiki.debian.org/MeetBot. 15:00:20 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 15:00:23 The meeting name has been set to 'openstack_ansible_meeting' 15:00:26 #topic rollcall 15:00:38 \o/ 15:00:56 hm, btw topic is not actually changed with meetingbot :( 15:04:51 hello 15:05:11 #topic office hours 15:05:59 CentOS 8.4 <- this has been solved right now I think. CI is passing again, which is awesome. But timing as usual was awful 15:06:27 but we also need to backport to V and I think also to U? 15:06:50 oh yes i expect so 15:06:56 thankfully they changes are small 15:07:06 yeah, just one symlink:) 15:07:17 Dmitriy Rabotyagov proposed openstack/openstack-ansible-lxc_hosts stable/victoria: Add CentOS 8.4 support https://review.opendev.org/c/openstack/openstack-ansible-lxc_hosts/+/795282 15:07:48 Dmitriy Rabotyagov proposed openstack/openstack-ansible-lxc_hosts stable/ussuri: Add CentOS 8.4 support https://review.opendev.org/c/openstack/openstack-ansible-lxc_hosts/+/795283 15:07:53 and thankfully 8.4 does not fail on libvirt firmware like stream still does 15:08:23 as https://review.opendev.org/c/openstack/openstack-ansible-os_nova/+/793007 still failing.... 15:09:30 So next we have 2 topics, that I expect to be merged by the end of the day 15:09:37 #link https://review.opendev.org/q/topic:%22osa%252Fpki%22+(status:open%20OR%20status:merged) 15:09:52 #link https://review.opendev.org/q/topic:%22osa%252Fbullseye%22+(status:open%20OR%20status:merged) 15:11:07 Once this is done, we do RC1. And probably we can branch as well, but I'd wait till next week with branching roles at least 15:11:49 yes - feels like all the major things are done now 15:11:50 we also should do bunch of rechecks... 15:13:22 as our gates were not really stable lately 15:14:32 ok, then regarding https://bugs.launchpad.net/openstack-ansible/+bug/1930276 15:14:34 Launchpad bug 1930276 in openstack-ansible "Nova API not restarted when nova policy is updated" [Undecided,Triaged] - Assigned to Dmitriy Rabotyagov (noonedeadpunk) 15:15:03 I'd say that it's really oslo bug. But I'm not sure if we should workaround it with triggering handlers on older branches 15:15:12 as it can be reproduced only with json 15:15:24 but it's also pretty much work I'd say.. 15:16:05 I wasn't able to look into oslo.policy code yet :( But will try to this week 15:16:40 I hope that just catching exception in correct place will do job 15:17:03 as otherwise we should also trigger service restart when we're removing policy.yaml 15:18:05 noonedeadpunk jrosser i have installed aio_ovn_metal senario and seeing this - http://paste.openstack.org/show/806463/ 15:18:09 But I a bit doubt regarding possibility of fixing json file, since they're deprecated for a while 15:18:16 inventory is totally empty is that ok ? 15:18:48 sorry we are in middle of meeting. 15:20:11 even though the json is deprecated it's still supported in a lot of stable branchs 15:20:40 i'd expect there to be an interest in fixing it if we can get a good bug report 15:20:44 I missed oslo meeting yestarday to raise this :( 15:22:03 fair ppint 15:22:06 *point 15:22:15 lets ask them about our bug 15:23:50 I was asking but haven't pinged hberaud so get no answer:) 15:24:16 ah well i just did, lets see if anything happens 15:25:26 I guess they mostly following on channel during meetings 15:26:30 regarding https://specs.openstack.org/openstack/openstack-ansible-specs/specs/xena/protecting-plaintext-configs.html - I think it's worth creating new repo for vault only after branching 15:27:25 I saw folks have been asking here about backporting this work to previous branches, which is kind of important point for them. But I'd say that's not possible considering amount of work and requirement of vault role 15:29:17 yes it looks awkward 15:29:34 Oh! And regarding DDoS! What I think we can do - in python_venv_build, check for amount of hosts in the play and fail, in case there're more then let's say 5 hosts, and wheels won't be built for $reasons 15:29:56 yes like safety net of some sort 15:31:09 and maybe link to some doc about troubleshooting this? And suggest workaround of setting osa_serial to lower then 5 if it can't be fixed for any reason 15:31:14 we can probably print out useful stuff in a fail: message about number of hosts, number of repo servers and stuff 15:31:26 yeah, totally 15:31:32 OS releases / architecture etc, to show why it's not worked 15:32:09 yeah, really good idea. it would be easier to understand for deployer what he does wrong 15:32:45 Also I was thinking of moving https://docs.openstack.org/openstack-ansible/rocky/admin/upgrades/distribution-upgrades.html to master as well in regards to focal upgrade? 15:33:06 and do some critical note about that you can't just upgrade computes without having repo container 15:33:38 but yeah, having repo container might be really tricky. Smth we probably should figure out in terms of the best way to sort this out 15:33:52 ah - does that page not appear in all branches 15:33:56 nope 15:34:00 it's only for rocky 15:34:11 it has been merged directly there 15:34:28 right - rocky was xenial->bionic, but i guess that W is a really huge transition release for all OS 15:34:36 yep 15:35:04 well, for centos it would be relatively easy migration, but debian and ubuntu... 15:35:42 theres a similar thing to bring arm nodes into an existing x86_64 cloud 15:35:51 nothing prevents that other than missing a repo server 15:36:00 eventually we have planned bionic->focal upgrade for autumn, so it will be adjusted anyway, but would be good to look into it in advance 15:36:26 and the biggest thing here is actually lsync 15:36:36 i feel like we maybe miss something here to bring in temporary / alternative repo hosts for specific cases 15:37:30 you mean for being standalone (ie under proxy)? 15:38:16 right, it's easy to point the build host at something else for an ansible group 15:38:29 but not so easy to get the build results to appear at the LB 15:38:52 ah 15:39:08 I think we can leverage ACLs or smth like that 15:39:17 an be a bit smarter here... 15:39:53 actually nothing stopps us from having several group of backends 15:40:06 oh well actually i wonder if overriding openstack_repo_url for a host group would work 15:41:15 it probably should... but it's not that neat.. As you need to set it manually and define extra group for new OS 15:41:50 and well, if you have more then 1 repo container per specific case... 15:42:06 that would be not handy really 15:44:17 oh, btw. I saw really weird failures during facts gathering after switching to min 15:45:25 oh dear - anything specific? 15:45:26 https://700a7e3ec90813bffe3e-a140c73e010d1e58296165dc52d0a2c3.ssl.cf2.rackcdn.com/788031/11/check/openstack-ansible-deploy-aio_lxc-debian-buster/c9ca1d5/logs/ara-report/results/3366.html 15:45:48 And I saw it periodicly on different patches 15:46:10 no details in stdout either 15:47:19 so have literally no idea wtf is going on... 15:48:46 those are the worst 15:49:06 sometimes you can get the actual traceback by adding -vvv 15:49:09 dmsimard: have you seen smth like that in other places? 15:49:20 ansible eats tracebacks unless it's -vvv 15:49:43 yeah... but you need to catch it first... and until it's random and not really reproducable... 15:50:04 at least I couldn't reproduce that relatively easily... 15:50:43 it's delegating the setup task that fails ? maybe one of the item from the list doesn't work for some reason 15:51:54 well, it's a container on the same host... and in case of CI item consist of single element 15:52:10 but that's actually what I also thought about 15:54:55 I could imagine this happening in case of OOM but haven't found any proof in the log 15:57:43 not found anything suspicious at all here 15:59:36 but I also don't see it happening today... 16:01:01 #endmeeting