15:01:22 #startmeeting neutron_ci 15:01:23 Meeting started Wed Jun 17 15:01:22 2020 UTC and is due to finish in 60 minutes. The chair is slaweq. Information about MeetBot at http://wiki.debian.org/MeetBot. 15:01:24 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 15:01:25 hi 15:01:27 The meeting name has been set to 'neutron_ci' 15:01:28 o/ 15:01:34 hi 15:01:54 yeah o/ 15:02:53 ok, let's start 15:03:10 Grafana dashboard: http://grafana.openstack.org/dashboard/db/neutron-failure-rate 15:03:22 please open link and we can move on 15:03:24 #topic Actions from previous meetings 15:03:36 0o/ 15:03:40 \o 15:03:57 first action from last meeting 15:04:01 slaweq to check failure in test_ha_router_failover: https://storage.gra.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_6d0/726168/2/check/neutron-functional/6d0b174/testr_results.html 15:04:09 I didn't saw this issue any time later so I didn't check it. I will if it will start failing more often 15:04:34 or maybe You saw it more often and it's just me who missed that 15:04:40 so please tell me :) 15:04:47 641590 15:04:57 ^^ sorry :) 15:05:34 nice score, I will not try to compete :) 15:06:07 :) 15:07:24 ok, so next one 15:07:26 slaweq to add additional logging for fullstack's firewall tests 15:07:31 For now I just did DNM patch https://review.opendev.org/#/c/736034/1 to see exactly when this test is failing and then I will add some additional logs to know more what's going on there. 15:07:43 so I will assign it to myself for next week too 15:07:46 #action slaweq to add additional logging for fullstack's firewall tests 15:07:57 and the last one 15:07:59 maciejjozefczyk to check failing test_port_security_macspoofing_port test 15:09:55 * maciejjozefczyk looking for the link :) 15:10:35 #link https://bugs.launchpad.net/tempest/+bug/1728886 15:10:35 Launchpad bug 1728886 in tripleo "test_port_security_macspoofing_port doesn't work when l2pop and arpresponding are enabled" [Critical,Fix released] 15:11:30 maciejjozefczyk: so it's fixed and test is passing now, right? 15:12:10 slaweq, the problem was in core-ovn and it was fixed there 15:12:24 in some stable releases I needed to blacklist it, because those are not using the newest ovn 15:12:46 ok 15:12:49 thx maciejjozefczyk 15:13:54 ok, lets move on 15:13:56 #topic Stadium projects 15:14:01 standardize on zuul v3 15:14:06 Etherpad: https://etherpad.openstack.org/p/neutron-train-zuulv3-py27drop 15:14:12 there are only 2 missing things there 15:14:15 networking-midonet 15:14:21 and ovn grenade job 15:14:32 for ovn grenade job I have it in my backlog 15:15:12 and for networking-midonet, I asked project core reviewers this week, and they sadly agreed that we should probably deprecate this project as we did with fwaas 15:15:28 so IMO we can cross it out from this list here 15:15:48 +1 15:16:08 time has come then 15:16:23 +1 15:16:29 ok 15:16:34 for ovn grenade job is it still working? neutron grenade has issues in stable branches (where it is still legacy atm) 15:16:54 with virtualenv/pip removal from base image (and also our friend uwsgi) 15:17:03 bcafarel: currently we are not running this ovn grenade job so I don't know if it's really working or not 15:17:10 :) 15:17:17 I will need to make it working and migrate to zuulv3 15:18:16 and we have new topic which also needs to be done in stadium projects 15:18:22 "Switch to Ubuntu Focal" 15:19:13 is there any volunteer to drive this? 15:19:56 I started it but I think with my other focuses someone else might be a good candidate to push forward 15:20:19 with the zuul v3 job almost done I think our job inheritance will do almost all of this 15:20:30 as the base jobs are modified 15:20:45 but the process of debugging any changes is unknown 15:20:52 s/changes/issues/ 15:21:02 njohnston: yes, I think it shouldn't be a lot of work really 15:21:27 I can help it will be a change from fixing xenial jobs :) 15:21:40 bcafarel: :D 15:21:42 thx a lot 15:22:00 looks like the change I did for fullstack/functional is breaking on the step of compiling OVN 15:22:00 https://storage.gra.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_12b/734304/3/check/neutron-functional/12b22c0/job-output.txt 15:22:14 that change is https://review.opendev.org/#/c/734304/ 15:22:48 njohnston, with moving to Focal we can stop compiling the modules 15:23:04 maciejjozefczyk: Perfect, that was my hope 15:23:06 njohnston, cause the kernel in Focal has OVS metric actions that we need :) 15:23:11 maciejjozefczyk: unless we will have another feature which we will need there :P 15:23:13 meter* 15:23:20 slaweq, unless :D 15:23:31 but for now it should be good way to go probably 15:23:46 one step at a time yep 15:24:05 So probably the first step is to disable the module compilation if focal is detected? 15:24:08 wait, the error in njohnston link is about "ipv6_dst_lookup" 15:24:24 so 15:24:25 #link https://review.opendev.org/#/c/735799/ 15:24:27 should fix it 15:24:45 maciejjozefczyk: I see other error in https://storage.gra.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_12b/734304/3/check/neutron-functional/12b22c0/job-output.txt 15:24:57 RuntimeError: dictionary keys changed during iteration 15:24:59 RuntimeError: dictionary keys changed during iteration 15:25:12 first! \o/ :P 15:25:19 right, that is in the compile_ovs action 15:25:20 http://paste.openstack.org/show/794904/ 15:25:31 slaweq: that's why you are PTL :) 15:25:32 the dictionary keys error, I mean 15:25:43 yhmmm, allright 15:25:48 bcafarel: I hope it's not the only reason ;P 15:26:01 and njohnston's question about disabling module compilation? 15:26:10 is there a patch for this? 15:26:18 not yet I think 15:26:23 no, I think that would be a new action item 15:26:28 ok, I'll propose one 15:26:29 we need to start discussion again if we want to comiple ovs/ovn at all, that seems to be outside module (error is related to ovsdb) 15:26:30 first focus was on https://review.opendev.org/#/c/735799/ 15:26:53 we can also add a knob for this, apart from checking the version 15:27:18 +1 15:28:01 (ok we have OVN_BUILD_MODULES) 15:28:36 njohnston: I think You can simply add new variable to https://github.com/openstack/neutron/blob/master/roles/configure_functional_tests/tasks/main.yaml#L13 15:29:00 and then set it in the job's definition to not compile ovs from source 15:29:28 the variable in script is here https://github.com/openstack/neutron/blob/master/tools/configure_for_func_testing.sh#L65 15:29:42 +1 if possible get rid of depenency to out of tree error sources 15:29:49 slaweq, I can work on that part... cause I started doing something similar in: https://review.opendev.org/#/c/727193/ 15:29:52 slaweq: sounds good, I will iterate on that after the meeting 15:30:01 * njohnston yields to maciejjozefczyk 15:30:02 thx maciejjozefczyk and njohnston 15:30:57 actually we need to make a cleanup around compile_ovs functions, because we have this one both in devstack ovn lib and devstack ovs lib ... 15:31:23 maciejjozefczyk: yes, so we need cleanup 15:31:32 will You take care of this? 15:31:34 yes 15:32:00 thx a lot 15:32:33 ok, and bcafarel will You prepare some list of the other jobs which we need to switch to Focal? 15:33:07 slaweq: that's the plan yep! 15:33:12 thx a lot 15:33:21 so we should be good with that topic :) 15:33:31 so next one 15:33:33 #topic Stable branches 15:33:39 Ussuri dashboard: http://grafana.openstack.org/d/pM54U-Kiz/neutron-failure-rate-previous-stable-release?orgId=1 15:33:41 Train dashboard: http://grafana.openstack.org/d/dCFVU-Kik/neutron-failure-rate-older-stable-release?orgId=1 15:33:53 we are still broken with this uwsgi issue in stable branches, right? 15:34:29 * bcafarel looks for links 15:34:42 uwsgi should be good in recent branches now though there are other issues 15:34:53 ouch 15:35:31 #link https://review.opendev.org/#/c/735523/ 15:35:46 uwsgi should be fixed up to stein 15:36:09 good :) 15:36:25 ++ 15:36:57 then we need https://review.opendev.org/#/c/735948/1 (grenade job) + ovs_version fix to backport 15:39:09 by ovs_version You mean the patch from jlibosva https://review.opendev.org/#/c/735799/ right? 15:39:38 sorry, laptop issue 15:39:43 slaweq: that's the one yes 15:40:16 hmm, but will we be able to merge them? as one requires another 15:40:27 so we are in deadlock in ussuri, aren't we? 15:40:36 I think so 15:40:38 we need both 15:40:57 good point, yes they will most probably need to be squashed 15:41:16 so we can combine them too in ussuri, or (better IMO) propose patch to set all those jobs non-voting in ussuri 15:41:25 then merge both patches and revert first one 15:42:43 what do You think? 15:43:29 +1 if we only merge the ovs patch with those jobs stopped 15:44:00 (well, and the zuul migration one) 15:44:28 sounds good, each fix patch can (and should) re-enable relevant jobs 15:44:53 ok, bcafarel will You propose patch to make those jobs non-voting temporary or do You want me to do that? 15:46:04 slaweq: I have the relevant tab already open I can send it (neutron-grenade-multinode* and ovn jobs) 15:46:12 thx bcafarel 15:46:43 I think that should be all related to the stable branches, right? 15:46:51 or do You have anything else? 15:47:40 older branches will need some fixes too, but one branch at a time! 15:47:47 bcafarel: ok :) 15:48:10 lets move on to the next topic as we are late today :) 15:48:12 #topic Grafana 15:48:49 as we already discussed, there are 2 main issues which are causing many jobs failing: uwsgi problem and ovs compilation problem 15:49:13 uwsgi should be fixed in master and fix for ovs issue is in the gate already 15:49:51 there was also problem with designate job but it is also going down with failure rate todayu 15:50:02 so seems like it's fixed already 15:50:27 yes frickler was commenting about it earlier in #openstack-neutron it should be good 15:51:03 do You have anything else regarding grafana and our dashboard? 15:52:22 ok, so lets move on 15:52:25 next topic 15:52:32 #topic fullstack/functional 15:52:44 I found only one new issue in fullstack job for today 15:52:49 Ovs compilation error: https://storage.bhs.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_410/735799/1/check/neutron-fullstack/410977b/job-output.txt 15:53:04 but this one seems different than those which we discussed earlier already 15:53:11 maciejjozefczyk: can You take a look at it? 15:53:26 it's not happening all the time for sure 15:53:58 slaweq, yes 15:54:22 ahh, wait 15:54:29 it was on https://review.opendev.org/#/c/735799/ 15:54:29 bcafarel, found the same in https://review.opendev.org/#/c/735799/ 15:54:50 yeah... I'll take a look tomorrow morning on it 15:54:54 thx a lot 15:55:08 sigh so this is reproducible then 15:55:15 ;/ 15:55:18 bcafarel: no, it's on the same patch 15:55:25 :) 15:55:36 oh true 15:55:37 I found it also in same job run 15:55:57 ok, lets move on quickly to the last topic for today 15:56:02 #topic Tempest/Scenario 15:56:11 first of all, I sent today 2 patches: 15:56:16 https://review.opendev.org/736177 15:56:18 https://review.opendev.org/736186 15:56:26 both are follow-ups from the PTG discussion 15:56:46 first one is move of not stable dvr multinode job to experimental queue 15:56:58 the problem should be solved by https://github.com/openvswitch/ovs/commit/0c4d144a989a444d038d58272d8571e97e00e86f 15:57:08 and the second one is to remove tempest-integrated-networking job with multinode job 15:57:08 so maybe we can use this tag instead the one proposed by Kuba 15:57:16 maciejjozefczyk++ 15:57:16 (oh sorry, too late) :D 15:57:38 slaweq, yeah, sorry for disturbing, I'll update the patch 15:57:46 maciejjozefczyk: no problem at all 15:57:52 thx for looking into that 15:58:22 +1 to those two patches 15:58:27 ok, and that's all what I have for this week 15:58:29 thx ralonsoh 15:58:47 anything else You want to raise here quickly in last minute? 15:59:36 ok, if not then thx for attending the meeting 15:59:41 have a great evening 15:59:43 o/ 15:59:44 bye 15:59:46 #endmeeting