15:00:18 #startmeeting kolla 15:00:18 Meeting started Wed Mar 4 15:00:18 2020 UTC and is due to finish in 60 minutes. The chair is mgoddard. Information about MeetBot at http://wiki.debian.org/MeetBot. 15:00:19 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 15:00:21 The meeting name has been set to 'kolla' 15:00:24 #topic rollcall 15:00:26 \o 15:00:28 o/ 15:00:35 o/ 15:00:40 o/ 15:00:40 /o 15:00:58 o/ 15:01:05 crowdy 15:03:13 #topic agenda 15:03:24 * Roll-call 15:03:26 * Announcements 15:03:28 * Review action items from last meeting 15:03:30 * CI status 15:03:32 * OpenDev PTG in Vancouver - will anyone attend? 15:03:34 * Kolla SIG (aka Kolla Klub?) https://etherpad.openstack.org/p/kolla-sig 15:03:36 * Discuss: Ansible may run more handlers than originally planned https://bugs.launchpad.net/kolla-ansible/+bug/1863510 15:03:37 Launchpad bug 1863510 in kolla-ansible "Ansible may run more handlers than originally planned" [Medium,Triaged] 15:03:38 * Time to drop CentOS 7 jobs? 15:03:40 * Ussuri release planning (kayobe) 15:03:42 * Ussuri release planning (kolla & kolla ansible) 15:03:44 #topic announcements 15:03:46 #info CI broken in myriad ways 15:03:49 We'll cover that next 15:03:59 Any others? 15:04:25 CirrOS 0.5.0 released 15:04:33 patch for k-a in gerrit 15:04:53 #info CirrOS 0.5.0 released 15:04:56 nice work hrw 15:05:00 any new features? 15:05:46 can neutron finally stick to using cirros for their advanced networking testing instead of relying on ubuntu 14.04? :D 15:05:48 5.3 kernel, graphical console working on aarch64, lot of virtio included (rng as well) 15:06:19 yoctozepto: you know that cirros is simple image to test does your cloud works? 15:06:30 yoctozepto: convince 14.04 -> 18.04 move? 15:06:45 hrw: I just asked 15:07:06 ubuntu is bloated 15:07:13 no idea why cirros was not sufficient 15:07:28 #topic Review action items from last meeting 15:07:30 let's move on 15:07:39 dougsz to write bug report about nova SSH nproc issue 15:07:44 dougsz: did you get to it? 15:07:58 I didn't see it 15:08:54 #action dougsz to write bug report about nova SSH nproc issue 15:08:59 #topic CI status 15:09:06 ok now for some fun 15:09:15 we have a few issues at the moment 15:09:25 and to top it off some mirrors seem to be down 15:09:48 yeah, today is fun all around 15:10:12 let's start with kolla 15:10:28 kolla.common.utils.tempest: ERROR: Package 'patrole' requires a different Python: 2.7.5 not in '>=3.6' (C7@master/Ussuri) 15:11:15 we can drop this from kolla after oslo gets fixed 15:11:23 because we need k-a working 15:11:35 example: https://e3487461d10f0d2cc5e1-c866b68a342a1b98ad74b0accf39cca4.ssl.cf1.rackcdn.com/706042/2/check/kolla-build-centos-source/ae638c7/kolla/build/000_FAILED_tempest.log 15:12:12 sanity check: any idea why it hit master and not train? 15:12:45 we theoretically build train in c7 15:12:53 even on master branch 15:12:56 have we done it today? 15:13:17 nope, not touched yet 15:13:25 in any dimension :-) 15:15:04 do we understand the underlying issue here? 15:15:43 tempest dropped py2 support? 15:16:51 seems like it happened a while ago though 15:17:14 mhm, patrole dropped it 15:17:23 and we just take that branch 15:17:29 it's not from upper-constraints 15:17:37 hence it killed master c7 15:17:47 ok 15:17:54 it's centos7, we don't care too much 15:18:04 but will it break stable branches? 15:18:11 or will UC save us there? 15:18:34 let's see tomorrows periodic 15:18:53 I guess we either drop c7 or drop c7 tempest 15:19:21 drop c7 altogether 15:19:25 drop c7 15:19:26 it's a mix-and-match 15:19:37 it's coming up next 15:19:42 ok, onto k-a 15:19:47 stables not affected because they pin to released version 15:19:50 if someone want c7 then train supports it 15:20:13 (2020-03-04) Ussuri 100% fail at keystone 15:20:20 this is due to oslo.cache 2.1.0 15:20:32 hberaud proposed a patch, we're testing it now 15:20:52 https://review.opendev.org/#/c/711227/ 15:20:52 patch 711227 - kolla - DNM: Test oslo.cache fix - 3 patch sets 15:21:09 affects py3 jobs on master 15:21:16 finally (I hope) 15:21:26 ARA 0.x broken on py2 15:21:35 I raised https://bugs.launchpad.net/kolla-ansible/+bug/1866079 15:21:37 Launchpad bug 1866079 in kolla-ansible train "Stable branch CI fails when gathering facts - invalid syntax (fake_filesystem.py, line 1195)" [Critical,In progress] - Assigned to Mark Goddard (mgoddard) 15:21:47 Dincer Celik proposed openstack/kolla-ansible stable/stein: Fixes gnocchi-api script name for Ubuntu/Debian https://review.opendev.org/711249 15:21:58 proposed a workaround for us https://review.opendev.org/711239 15:21:58 patch 711239 - kolla-ansible (stable/train) - CI: Pin pyfakefs to <4 for Python 2 - 1 patch set 15:22:08 and a fix for ARA: https://review.opendev.org/711240 15:22:08 patch 711240 - recordsansible/ara (stable/0.x) - Fix ansible execution on Python 2 - 1 patch set 15:22:46 oh, we missed one 15:23:15 we're deploying master containers on triain 15:23:17 https://launchpad.net/bugs/1866054 15:23:18 Launchpad bug 1866054 in kolla-ansible train "Kolla Ansible train deploys master containers by default" [High,In progress] - Assigned to Mark Goddard (mgoddard) 15:23:55 this actually means we're subject to the oslo.cache issue on train too (which is how yoctozepto caught it) 15:24:13 indeed 15:24:28 I say we merge your ara and fix to release 15:24:35 and approve 15:24:37 as one 15:24:47 it's not to backport 15:24:52 and reno is quite clear 15:25:07 will need to backport ara 15:25:10 can do that first 15:25:33 well, patches are never removed 15:25:41 so does not matter too much for order 15:25:56 ok 15:26:04 I think we are done with CI 15:26:07 kayobe green \o. 15:26:21 kolla cli red 15:26:25 functional down due to ceph gone (yoctozepto) 15:26:43 I saw you were pushing kolla-cli patches yoctozepto, was it for that? 15:27:04 mgoddard: yeah, sorry, forgot to remove it 15:27:18 nice and green 15:27:20 ok, onwards 15:27:30 mgoddard: gmann proposed py2.7 drop in k-c 15:27:37 mgoddard: and it got stuck 15:27:52 hmm 15:28:00 mgoddard: hence needed unblock 15:28:11 mgoddard: k-c sees no continued work 15:28:17 indeed 15:28:41 feel free to ping interested parties rather than fix yourself 15:28:56 #topic OpenDev PTG in Vancouver - will anyone attend? 15:29:07 nope 15:29:17 mgoddard: I have no idea who is interested in here :D 15:29:23 not going myself 15:29:31 #1 budget, #2 coronavirus corpo rules 15:29:32 too expensive 15:29:34 Once again I was omitted from the mailing list asking which projects will attend Vancouver. Thankfully priteau noticed 15:29:50 hmph, no love for kolla :/ 15:30:00 Vancouver is nice city 15:30:07 to be fair, kolla has shown no love for ptg recently :) 15:30:45 we haven't really had productive development discussions in person for $years 15:30:49 was kolla at any PTG after Dublin? 15:30:56 it's good for ops feedback etc 15:31:08 but that's summit not PTG 15:31:19 mgoddard: we need european ptg to make it cheaper 15:31:35 maybe, but that just makes it expensive for everyone else :) 15:31:37 next is Berlin I think? 15:31:41 well 15:31:56 I may attend Vancouver (love my company) 15:32:30 osmanlicilegi: great. Just be prepared that there will probably be no official kolla discussions 15:32:40 I'm sure it will still be interesting and useful though 15:32:52 I think a virtual PTG works well for us 15:32:57 and works well for our planet too 15:33:02 +1 for vptg 15:33:03 +1 15:33:28 +1 15:33:42 and we can couch as much as we want ;D 15:34:36 seems we have a rough consensus - no official presence from kolla, but there may be some koalas present for informal discussions 15:34:45 I'll pass it on 15:35:13 Gonna mix up the agenda order a little, keep you on your toes 15:35:15 #topic Time to drop CentOS 7 jobs? 15:35:47 To turn it on its head - are there any reasons why we shouldn't drop centos 7 support in master? 15:35:58 1 - kayobe still uses it :D 15:36:17 but we are three patches away from not using it 15:37:05 drop, drop, drop, drop 15:37:08 +2 for drop 15:37:09 oh, there is also bifrost 15:37:11 :) 15:37:18 +2 for dropping bifrost too 15:37:21 but I think we can live without it for a time 15:37:21 ;D 15:37:29 I'm working on it 15:38:43 I'm not hearing many reasons to keep it 15:38:50 so how do we drop? 15:39:00 k-a first? 15:39:16 zuul jobs first? 15:39:24 zuul, k-a+k 15:39:25 does anyone want to do it? 15:39:40 I'm sure it will be therapeutic 15:40:36 first zuul, project order not relevant 15:41:39 looking for a volunteer... 15:42:32 I thought you needed a therapy after today 15:42:54 but could be me if you insist ;p 15:42:54 I'll get the zuul ball rolling 15:43:03 that at least will be easy 15:43:23 I have little incentive to remove support given kayobe needs it for now :) 15:44:10 #agreed Drop CentOS 7 jobs, starting with kayobe 15:44:20 #topic Discuss: Ansible may run more handlers than originally planned https://bugs.launchpad.net/kolla-ansible/+bug/1863510 15:44:22 Launchpad bug 1863510 in kolla-ansible "Ansible may run more handlers than originally planned" [Medium,Triaged] 15:44:36 what are our options? 15:44:40 Thanks for spotting this yoctozepto 15:44:48 It's not great tbh 15:44:51 I don't think we should revert the dry patch 15:45:11 I'd really rather not 15:45:20 when did it land? Stein? 15:45:31 or Train, not earlier 15:45:35 let's check then 15:45:55 train 15:45:57 phew 15:46:05 https://review.opendev.org/647699 15:46:06 patch 647699 - kolla-ansible - Simplify handler conditionals (MERGED) - 13 patch sets 15:46:07 mhm 15:47:03 added to bug report 15:47:08 ditto :) 15:47:36 Radosław Piliszek (yoctozepto) wrote 36 seconds ago: 15:47:39 Mark Goddard (mgoddard) wrote 34 seconds ago: 15:47:44 late by 2 seconds! 15:47:59 it's quite nasty, as it affects all our config file generation and container checking tasks 15:48:08 yoctozepto suggested include_tasks in a loop 15:48:30 this would be a significant rewrite of every role's config.yml and check-containers.yml 15:48:49 as was the one that introduced the bug 15:49:12 not really a rewrite, it mostly removed lines 15:49:24 a lot easier to reason about 15:49:36 (although apparently still flawed) 15:49:59 there are also performance concerns with include_tasks at scale 15:50:07 which we haven't fully investigated 15:50:25 but committing to it in a big way seems scary 15:50:58 hmm 15:52:10 there is another sort of related issue, I wonder if we could fix both 15:52:23 if I run kolla-ansible genconfig, then kolla-ansible deploy 15:52:37 if my config changed but container did not, currently the container is not restarted 15:52:44 I was thinking about how to address this 15:53:04 true 15:53:05 what if we touched a file in /etc/kolla/ 15:53:14 we should address both at once 15:53:26 otherwise it gets extra work for nothing 15:53:37 then we have tasks/handlers that remove said file, and trigger a restart if it existed 15:54:46 just a thought 15:55:49 another option is kolla_set_configs --check 15:56:07 checks for changes in config files 15:56:45 yet another consideration is podman 15:57:08 if we were to be forced to switch at some point, would we need to change our model? 15:57:15 sorry guys, I got lost. can patches fixing the current CI issues get RP+2? 15:57:17 that would prevent backporting 15:57:18 good thinking 15:57:40 hrw: will do when we have working patches 15:57:40 hrw: mgoddard needs to merge them 15:57:55 ok 15:58:08 mgoddard: just to remember that loops with notify will not save us 15:58:14 I had 2-3 other discussion during meeting ;( 15:58:21 as they are b0rken 15:59:26 yoctozepto: true, but if there is something simple we can do once for each container it would help 15:59:57 that's the one thing we should not rely on, the rest is a mystery :D 16:00:08 so to sum up 16:00:15 let's add to bug report 16:00:20 what we should remember when fixing 16:00:33 anyway, I don't think it's critical, but it is quite an important problem we should solve 16:00:50 ok. who is updating bug report? 16:00:54 wouldn't want to dup 16:02:00 I will 16:02:03 thanks 16:02:07 let's wrap it up now 16:02:10 and that brings us to the end 16:02:14 thanks all 16:02:19 #endmeeting