15:00:04 #startmeeting neutron_ci 15:00:04 Meeting started Tue Aug 3 15:00:04 2021 UTC and is due to finish in 60 minutes. The chair is slaweq. Information about MeetBot at http://wiki.debian.org/MeetBot. 15:00:04 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 15:00:04 The meeting name has been set to 'neutron_ci' 15:00:25 hey again 15:00:56 hi 15:02:01 let's wait for lajoskatona, obondarev and maybe others to join 15:03:12 Hi, sorry 15:03:24 np lajoskatona :) 15:04:02 Grafana dashboard: http://grafana.openstack.org/dashboard/db/neutron-failure-rate 15:04:09 please open and we can start 15:04:16 #topic Actions from previous meetings 15:04:23 slaweq to check dns nameserver issues in cirros 15:04:35 I was checking that and I proposed patch https://review.opendev.org/c/openstack/tempest/+/802005 15:04:38 it's now merged 15:04:50 I didn't saw any such issue in last few days 15:04:58 so hopefully this patch solves that problem 15:05:02 *solved 15:05:20 but if You would see same issue again, please let me know 15:05:25 crossing fingers 15:05:42 next one 15:05:44 slaweq to blacklist failing test_established_tcp_session_after_re_attachinging_sg in LB job 15:05:49 done https://review.opendev.org/c/openstack/neutron-tempest-plugin/+/801602 15:06:01 and last one 15:06:03 slaweq to check FIP set to DOWN issue 15:06:18 there was old bug reported already https://bugs.launchpad.net/neutron/+bug/1815585 15:06:30 but I didn't had time to look at it deeper 15:06:57 hi, here, sorry 15:07:09 obondarev: no worries :) hi 15:07:15 #topic Stadium projects 15:07:24 lajoskatona: any new issues with ci in stadium? 15:07:38 or anything related stadium's ci to discuss? 15:08:02 just the payload changes, but at least happens something during the summer :-) 15:08:13 LOL 15:08:26 :) 15:08:40 elod also reported some periodic job failing in stadium, 1 sec 15:09:03 I pasted mail in https://etherpad.opendev.org/p/neutron-periodic-stable-202107 15:09:15 +1 15:09:30 He just mentioned it on our internal chat :-) 15:10:11 thx bcafarel 15:10:27 it seems we have issues with old branches on bgpvpn 15:10:33 and vpnaas 15:10:59 in case of bgpvpn maybe we can EOL pike and queens branches if there will be no volunteers to check those issues 15:11:01 wdyt? 15:11:48 yeah, can eol them 15:11:59 that was my initial thinking too, vpnaas most can probably be fixed (and failures are in newer branches) 15:12:40 ok, I will check failures in bgpvpn and will start process of EOLing it if needed 15:13:01 #action slaweq to check networking-bgpvpn stable branches failures 15:13:07 any volunteer for vpnaas? 15:13:47 o/ (if I do not have time until PTO I will tell you) 15:13:48 I can check 15:14:01 ok bcafarel wins :P 15:14:06 thank You both :) 15:14:09 yes \o/ 15:14:19 #action bcafarel to check vpnaas failures in stable branches 15:15:26 ok, so I think stadium projects are covered already 15:15:29 now next topic 15:15:34 #topic Stable branches 15:15:44 bcafarel: any new issues in stable branches of neutron? 15:16:00 mostly quiet with the summer, and the active backports went OK 15:16:04 except ussuri with https://bugs.launchpad.net/neutron/+bug/1938262 15:16:34 I posted more details in bug, but overall we seem to hit a kernel regression with IPv6+GRE 15:16:45 (though I still do not understand why train is fine) 15:16:50 do we need to wait for new kernel in Ubuntu to unblock it? 15:17:06 I have https://review.opendev.org/c/openstack/neutron/+/802552 as workaround in the meantime (2 tests to skip) 15:17:36 bcafarel: isn't Train running on older Ubuntu? 15:18:33 Train and ussuri are both 18.04 (Bionic), we switched with Victoria 15:18:38 ok 15:18:46 I thought that maybe train was still on 16.04 15:19:19 so we need Your skip patch only on the ussuri, right? 15:19:31 for now yes, maybe it will appear later in train? 15:19:56 but until then yes only ussuri has issue 15:20:11 ok, I think we will need mlavalle's +2 power to approve https://review.opendev.org/c/openstack/neutron/+/802552 :) 15:20:23 as ralonsoh is off this week 15:22:26 ok, let's move on 15:22:27 #topic Grafana 15:23:30 http://grafana.openstack.org/dashboard/db/neutron-failure-rate 15:23:43 I see there that dvr grenade job is failing pretty often recently 15:24:17 and also scenario dvr ha job is failing almost 100% of runs (again) 15:24:39 obondarev: do You know about any issues which could cause that many failures recently? 15:25:23 I do 15:25:26 I think we had a few bugs reported recently on DVR 15:25:49 please check https://bugs.launchpad.net/neutron/+bug/1938685 15:26:01 and fix https://review.opendev.org/c/openstack/neutron/+/803211 15:26:16 so far 3 successful ddvr-ha jobs in a row 15:27:27 obondarev: good catch 15:27:36 thanks bcafarel for adding Sean and Rodolfo 15:27:45 qq is this about all ports? 15:27:54 they are best candidates for review this one 15:28:01 :) np they will know better about it indeed 15:28:07 I think that recently Sean and Rodolfo added binding for such ports on purpose in ovs agent 15:28:31 slaweq: ah, I missed that probably 15:28:45 obondarev: I'm not sure now if they did that in dvr too 15:28:55 but I remember that there was some work on that recently 15:29:14 so, I agree with bcafarel that they should review it before we will move on with that patch :) 15:29:21 but for DVR it's clearly an issue: OFPActionOutput with ofport == -1 will fail 15:29:45 Done 15:29:57 obondarev: I will take a look at that patch more deeply too 15:29:59 :) 15:30:01 mlavalle: thx a lot 15:30:09 slaweq: thanks! 15:30:29 obondarev: added to my todo list for tomorrow morning :) 15:30:37 cool! 15:30:53 obondarev: thx for working on that 15:31:00 sure 15:31:06 anything else regarding grafana? or can we move on? 15:32:34 ok, so let's move on 15:32:37 next topic 15:32:45 #topic fullstack/functional 15:32:58 I found out that functional tests job is failing pretty often recently 15:33:08 and in most cases it was due to same error 15:33:15 I opened bug https://bugs.launchpad.net/neutron/+bug/1938766 15:33:41 I already pinged jlibosva and lucasagomes to take a look into that one 15:33:57 and I think jlibosva already did some investigation 15:34:56 so it will probably be good by next CI meeting 15:35:00 I think it's a regression in python-ovs 15:35:12 after this commit - https://github.com/openvswitch/ovs/commit/c39751e44539a014e642bcd930cb9e3a33af1805 15:36:23 jlibosva: I just checked that in functional job we are using ovs from commit 0047ca3a0290f1ef954f2c76b31477cf4b9755f5 15:36:37 it's been there for a while tho and also https://review.opendev.org/c/openstack/neutron/+/794892 was supposed to workaround that 15:38:19 so it seems that this patch didn't workaround that issue :/ 15:39:14 yeah, I will have to investigate further, maybe it's a combination of things since we started to see it recently but the OVS patch has been in for over a year 15:39:40 jlibosva++ ok, thx for working on that 15:40:50 that was the only issue which I had to discuss today 15:41:01 maybe I still didn't catch up enough after week off :) 15:41:18 do You have any other issues/things to discuss today? 15:42:46 not that I know of 15:43:43 if there are no other topics for today, I think we can finish few minutes earlier today 15:43:55 thx for attending the meeting and for keeping our ci green :) 15:44:02 o/ 15:44:03 have a great week and see You all online 15:44:05 o/ 15:44:07 o/ 15:44:08 #endmeeting