15:00:40 #startmeeting openstack_ansible_meeting 15:00:40 Meeting started Tue May 17 15:00:40 2022 UTC and is due to finish in 60 minutes. The chair is noonedeadpunk. Information about MeetBot at http://wiki.debian.org/MeetBot. 15:00:40 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 15:00:40 The meeting name has been set to 'openstack_ansible_meeting' 15:00:47 #topic rollcall 15:00:55 o/ 15:01:46 hi! 15:04:17 o/ hello 15:04:38 #topic hours 15:05:09 so, evenctually, as you might all guess, I haven't done anything from what I promised to do previous week which upsets me a lot.... 15:06:02 Like 10 mins before meeting I started looking into PKI stuff for octavia as that seemed pretty close 15:07:06 Didn't have chance to play with gluster stuff though 15:07:19 but from codebase I don't see any big issues right now 15:08:21 i nearly have the whole stack passing CI now https://review.opendev.org/q/topic:osa-gluster 15:08:40 just problem with uninstalling rsync on metal jobs breaks the CI log upload :) 15:09:25 but only for ubuntu /o\ 15:10:44 I'm not sure I actually also realized whole mess with phased packages :( https://review.opendev.org/c/openstack/openstack-ansible-lxc_hosts/+/840518 15:11:05 Do we need to enable phased updates also for Focal? 15:11:37 oh yes there was a big mess with that 15:11:48 phasing is only after Focal 15:12:32 then we should add some conditional inside script for os release? 15:12:49 as now it's a global prep script 15:13:00 (which also targets Debian) 15:13:32 we could do, though it does seem to be OK 15:13:46 ok then 15:14:02 I guess it's last patch to land for yammy? 15:14:03 APT appears to disregard that config on the other OS 15:14:30 yes, expect for this https://review.opendev.org/c/openstack/openstack-ansible/+/839483 15:14:45 and i don't know how many jobs we want to introduce right now 15:15:25 that only puts jammy in the check pipeline and NV as well 15:15:59 which might be reasonable for 'experimental' support then it gets switched to voting/gate for Z? 15:16:38 yup, I guess it's fair atm 15:16:52 i expect that the patches generated during branching will throw up a bunch of role failures for jammy that we've not tried yet 15:16:52 And indeed we should switch to voting in Z 15:17:13 I won't be surprised :) 15:17:38 And indeed NV sounds quite suitable right now for it 15:17:53 considering there's no maria/rabbit at very least 15:18:13 then we have centos-9 i guess 15:18:43 https://review.opendev.org/c/openstack/openstack-ansible/+/823417 \o/ 15:19:07 and a few small things left in here https://review.opendev.org/q/topic:osa%252Fcentos-9 15:19:17 And eventually we should make it voting kind of.... 15:19:28 as centos-8 is being dropped in Z 15:19:39 (or well, py3.6 is) 15:19:40 right - though we have to decide what to do about LXC 15:20:00 let me guess - it's absent for it? 15:20:04 i just don't know, other than have some experimental code path that uses a snap, and just say "don't use this" 15:20:36 i could add code to lxc_hosts to install it via snap 15:20:42 use at own risk :) 15:21:47 so it's provided only by lxd? 15:22:55 well yeah, lxc for centos 8 that we're using also not maintained actually 15:23:06 oh hmm 15:23:18 maybe i misunderstand that lxc can't be installed on it's own with a snap 15:26:31 oh, so you mean there's only lxd option now? 15:28:00 snapcraft.io only seems to show a LXD snao 15:28:02 snap 15:28:24 thats going to bring in a whole load of stuff we don't want to manage, i think 15:28:39 annoyingly there is lxc4.x in fedora 15:32:06 i wonder how hard it is to use copr to build what we want 15:32:18 as really thats what we already use for c8 i think 15:33:14 i wonder what long-term solutions do we have to make our lives easier...switch to LXD? maybe systemd-nspawn? 15:33:22 NeilHanlon: would you have any tips on how we could get lxc packages for centos9? 15:33:29 given that we already do this https://github.com/openstack/openstack-ansible-lxc_hosts/blob/9a4004169450a0bb68ffe0fec31e6887a9c075f3/vars/redhat-host.yml#L19-L20 15:33:45 damiandabrowski[m]: i made some initial start on LXD support maybe a year ago 15:34:00 but it would suffer from the same problem that the only viable installation on all platforms is a snap 15:34:06 and thats got it's own problems 15:35:17 damiandabrowski[m]: we jsut got rid of nspawn as it's support is absent on centos because of dropped btrfs 15:35:41 everything leads to containerd lol 15:35:51 jrosser: ahh :/ have we noted these problems somewhere? 15:35:53 fwiw we use LXD here for a bunch of things outside OSA, and before we "deliberately broke" the mandatory auto-update thing we had * break simultaneously because of new buggy lxd versions then get pushed onto you 15:36:02 noonedeadpunk: fair point, thanks 15:36:53 yeah, auto-update of snap is nightmare 15:37:03 I can imagine that for galera as example.... 15:37:25 jrosser: you;re right, fedora has even lxc4.... 15:37:59 i understand zero about copr but i am guessing we can use that to compile the .spec file for fedora into a centos9 package 15:38:02 Basically, if sombody would volunteer to maintain lxc in epel for centos/rocky that would be awesome 15:38:05 or something similar 15:39:18 I also just found that https://koschei.fedoraproject.org/package/lxc?collection=epel8 - is it same thing? 15:40:10 that looks like where it comes from i think 15:40:55 that looks like we should even find lxc3 for centos8 in epel 15:40:56 Oh, ok I even submitted a bug there.... https://bugzilla.redhat.com/show_bug.cgi?id=2034709 15:44:08 it's all here https://src.fedoraproject.org/rpms/lxc/blob/f36/f/lxc.spec 15:45:01 anyway maybe we should look at bugs, we're going to be out of time 15:46:17 #topic bug triage 15:46:31 I'm looking at https://bugs.launchpad.net/openstack-ansible/+bug/1967270 at the moment 15:47:28 I think it's quite fair 15:48:01 indeed 15:48:50 * jrosser fixes 15:49:41 Another one that bothers me was https://bugs.launchpad.net/openstack-ansible/+bug/1971179 15:49:51 but I don't find it _that_ trivial 15:50:07 andrewbonney wrote that related fix is not really covering issue 15:50:23 Jonathan Rosser proposed openstack/openstack-ansible-os_octavia master: Fix condition for deleting old amp images https://review.opendev.org/c/openstack/openstack-ansible-os_octavia/+/842145 15:50:55 but I'm didn't have time to have a deep dive into CSP. And X default worked for us nicealy, but we don't use horizon in fact 15:51:30 i think andrewbonney may have addresses the CSP stuff 15:52:43 i have seen other reports of that RBD size mismatch too and never understood what was actually happening there 15:53:52 damiandabrowski[m]: was playing with that for a while 15:54:07 But basically it's indeed all about chunking 15:54:25 disabling uwsgi helps here if anybody facing it 15:54:47 basically glanceclient doesn't support chunking, while openstackclient does. 15:55:08 So there can be great confusion/mess between services who does use what 15:55:11 yeah, but i haven't found out yet why uwsgi is causing these issues :/ 15:55:28 it is a shame there is no activity from the glance people on that bug for > 1 year 15:55:51 there was some stuff from the debian packaging which was about enabling chunking for uwsgi? 15:55:56 well... they were not very active in fixing uwsgi at all 15:56:16 well yes, but client should also support that 15:56:20 (I guess) 15:57:36 "If you plan to put uWSGI behind a proxy/router be sure it supports chunked input requests (or generally raw HTTP requests)." 15:57:42 At least in openstacksdk I see that https://opendev.org/openstack/openstacksdk/src/branch/master/openstack/image/iterable_chunked_file.py 15:58:02 And there's nothing close to that in glanceclient. And it has own implementation of client object as well 15:58:48 but on the other hand, if i recall correctly, i noticed these issues only when using uwsgi+ceph backend 15:58:57 uwsgi+file backend worked fine 15:59:07 so i don't really understand what's going on 15:59:52 so damiandabrowski[m] was able to reproduce issue nicely using glance image-create, but not with openstack image create 16:00:41 well yeah, maybe ceph adds some extra thing here as well.... 16:01:27 As then you upload client -> haproxy -> uwsgi -> ceph and each should chunk correctly 16:01:38 maybe we're somewhere out of sync in that chain.... 16:01:52 that bug is talking about snapshots written to ceph backend isnt it? 16:02:16 if you take snapshot from instance - it's kind of same? 16:02:40 as you have a file you need to upload to glance as an image 16:03:01 but potentially a different client code there? 16:03:07 yeah 16:03:26 +1 16:03:51 right now i'm not totally clear what is / is not working 16:03:52 and I guess horizon still uses glance client as well 16:04:02 might be worth updating the bug with what we know 16:04:15 damiandabrowski[m]: can you do that? 16:04:45 like if `openstack` cli is fine then thats at least one data point 16:04:50 yeah, i plan to come back to this issue when I'll have some time 16:04:58 our internal ticket about that is still open 16:05:26 yeah, I know :) But it's good point we should put that we're working and what's the progress (even if it's close to none) 16:06:07 and we're out of time... 16:06:10 #endmeeting