15:00:40 #startmeeting openstack_ansible_meeting 15:00:40 Meeting started Tue Feb 21 15:00:40 2023 UTC and is due to finish in 60 minutes. The chair is noonedeadpunk. Information about MeetBot at http://wiki.debian.org/MeetBot. 15:00:40 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 15:00:40 The meeting name has been set to 'openstack_ansible_meeting' 15:00:44 #topic rollcall 15:00:49 \o/ 15:00:58 hi! 15:06:34 #topic bug triage 15:06:42 We have a couple of new bugs here 15:07:39 #link https://bugs.launchpad.net/openstack-ansible/+bug/2007296 15:08:57 Basically idea/proposal here was to create folder under inventory/group_vars for each group we have basically, and move playbooks/defaults/repo_packages there 15:09:19 but some naming convention for files should be present, so that bump script could find them and update 15:09:50 This will also affect haproxy thing I beleive, as instead group_vars file a directory worth to be used 15:09:57 any thoughts on that? 15:10:50 IMO it's ok, we should leverage group_vars more often. That's also what i did for separated haproxy service config 15:12:43 I'd say it would be a bit more tough to find version that's being used, as file location will depend on group 15:13:00 But not sure it matters much to be frank 15:15:11 Ok, next one 15:15:34 #link https://bugs.launchpad.net/openstack-ansible/+bug/2007849 15:15:59 I don't have anything to say here... I wasn't really digging deep into code of our linear implementation 15:16:20 But it looks like it's not even required after all? 15:18:03 i also didn't dig deeper into this, but https://review.opendev.org/c/openstack/openstack-ansible/+/874482 looks good without it 15:18:18 It's hard to say also if there's any benefit in execution speed... At the moment it looks like load on nodepool workers is still high, so we have long executions overall 15:18:50 there was a timeout for ceph scenario but it happens very often nowadays so i believe it's not relevant 15:18:58 nah, it's not. 15:19:29 o/ sorry am late :) 15:19:38 I was trying to roughly compare time spent on LXC jobs of this patch and others 15:19:51 no worries Neil! 15:20:29 hi Neil! 15:21:17 yeah, i'm not sure how to compare performance looking at zuul becuse i believe it may strongly depend on a servers' provider 15:21:27 I think worth trying to calculate execution time on some more predictable AIO deployment 15:21:28 maybe i should do some tests locally and compare results 15:21:39 and see if there's any benefit from custom strategy 15:21:48 yeah, would be great 15:21:53 ok, i'll do that 15:22:00 #topic office hours 15:22:38 So haproxy role was updated after last review. I still haven't reviewed it as last 2 days were quite tough internally 15:23:41 Dmitriy Rabotyagov proposed openstack/openstack-ansible-haproxy_server master: Accept both HTTP and HTTPS also for external VIP during upgrade https://review.opendev.org/c/openstack/openstack-ansible-haproxy_server/+/864785 15:23:49 no worries, there is also neutron and glance PKI/TLS support waiting for reviews 15:23:50 https://review.opendev.org/c/openstack/openstack-ansible-os_neutron/+/873654 15:23:52 https://review.opendev.org/c/openstack/openstack-ansible-os_glance/+/821011 15:24:14 currently I'm working on TLS support for nova but it's a bit complicated due to already existing TLS support for consoles 15:25:34 That's the topic for review 15:25:36 #link https://review.opendev.org/q/topic:separated-haproxy-service-config+status:open 15:26:05 #link https://review.opendev.org/q/topic:tls-backend+status:open 15:26:25 damiandabrowski: it's not only consoles but also libvirt 15:26:55 as we do encrypt live migrations and libvirt makes cert auth 15:28:09 yeah..theoretically speaking we can share the same certs for API, libvirt and console if all of them reside on the same host, right? 15:29:18 well. I think consoles do reside on APIs, but they can use different interface iirc. 15:30:45 i believe in most cases the do reside on the same host, that's why I'm thinking of sharing the same cert 15:30:52 they* 15:31:15 I've made some progress on cloud-init v22.2+ for RHEL 9 and friends.. hoping in the next week or so 15:31:27 cc jrosser 15:31:49 And I think we still haven't backported curl hassle to stable branches 15:33:08 Also zuul result is quite confusing here: https://review.opendev.org/c/openstack/openstack-ansible/+/873289 15:33:34 But we still need reviews on dependant patch - maybe it will make zuul happier... 15:34:21 Eventually - we need plenty of reviews. Since Andrew is not around, damiandabrowski can you take a round of reviews on current patches? 15:34:42 yeah, ofc 15:35:13 Another thing I was going to discuss. I started looking at quorum queues for rabbit as a replacement of our HA queues that are going to be removed from rabbit 4 15:35:58 And the thing is, that exchange must be removed in order to create quorum queues, since as of today exchange is not durable while it should be for quorum 15:36:42 And removing exchange is quite a hussle, as then you need to stop all services at the same time using this exchange and have a user with broad permissions 15:37:56 So what I was thinking - maybe we can create a new "clean" vhost, for example without leading `/` (it's sooooo confusing to be frank to have that `/`) and make vhost name conditional depending on usage of quorum queues or not 15:38:24 This way it should be possible to switch back and forth as well without stopping service for a really long time 15:39:53 But yes, service will be desynced until role is finished anyway, as members will be configured with different vhosts 15:40:22 The thing is that easiest way I found to drop exchange is along with vhost.... 15:40:45 As I failed to drop exchange using rabbitmqadmin with administrator user... 15:41:31 i'm not a rabbitmq expert but looks good at first glance. I believe you know what to do :D 15:41:49 I hope I do lol 15:42:03 Will know soon :D 15:42:05 you're not a rabbitmq expert if you think you're a rabbitmq expert 15:42:15 ^ soooo true 15:42:15 so you're on the right track damiandabrowski :) 15:42:45 haha :D 15:45:03 So that's kind of it from my side 15:46:06 btw. don't you think we have quite many intermittent gating failures/timeouts these days? 15:46:18 for ex. I had to trigger recheck 5 times for https://review.opendev.org/c/openstack/openstack-ansible/+/871189 15:47:16 damiandabrowski: regarding time outs - it's known issue that affects literally every project as of today 15:48:46 My thinking is that it's related to high load on providers we're using for CI, or our CI is a noisy neighbour for itself 15:49:25 and afaik some quite big provider stopped donating infra for our CI, so load on others has increased 15:50:34 ahhh okok, makes sense 16:00:46 #endmeeting