15:01:37 #startmeeting neutron_ci 15:01:37 Meeting started Tue Apr 13 15:01:37 2021 UTC and is due to finish in 60 minutes. The chair is slaweq. Information about MeetBot at http://wiki.debian.org/MeetBot. 15:01:38 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 15:01:40 The meeting name has been set to 'neutron_ci' 15:01:40 hi 15:01:41 hi 15:02:18 o/ 15:02:59 ping lajoskatona 15:03:48 ok, lets start 15:03:55 Grafana dashboard: http://grafana.openstack.org/dashboard/db/neutron-failure-rate 15:04:06 #topic Actions from previous meetings 15:04:13 ralonsoh to check failed qos scenario test 15:04:18 Hi, I am here 15:04:30 yes, I didn't replicate the error 15:04:36 and I didn't see it again 15:04:54 and reviewing the logs, I didn't see why the QoS was not applied 15:05:05 sorry, no solution for this one 15:05:07 ralonsoh: ok, so let's not bother with this for now 15:05:27 we will get back to it when it will be happening more often 15:05:32 ok 15:05:37 ralonsoh: thx for checking it 15:05:39 next one 15:05:41 bcafarel to update grafana dashboards with stable/wallaby 15:05:58 sent and merged (I think), looking for the link 15:06:42 yep https://review.opendev.org/c/openstack/project-config/+/786008 merged, also good time to ping other for release-related changes 15:07:06 like https://review.opendev.org/c/openstack/neutron-tempest-plugin/+/785092 (switch to -wallaby variant jobs) 15:07:08 thx bcafarel 15:08:02 ok, next one 15:08:04 ralonsoh to check failed test_get_egress_min_bw_for_port functional test 15:08:18 yes, one sec 15:08:33 #link https://review.opendev.org/c/openstack/neutron/+/785158 15:08:47 this actually a pre-check that must be always true 15:09:00 this is not solving the error 15:10:03 If I understand well we will have clean feedback if there are policies on port 15:10:31 sorry, someone else pinged me 15:10:52 this patch will provide info if any other qos is set before the test 15:11:38 so this may help to understand what the issue is there 15:11:42 sure 15:11:49 which is good step forward IMO :) 15:11:51 thx ralonsoh 15:12:07 ok, and the last one 15:12:08 ralonsoh to update LB installation guide with info about legacy ebtables 15:12:12 done 15:12:15 one sec 15:12:49 I don't find it 15:13:02 https://review.opendev.org/c/openstack/neutron/+/785137 15:13:17 thx 15:13:21 it's merged already 15:13:51 and that were all action items from last week 15:13:57 let's move on 15:13:59 #topic Stadium projects 15:14:18 lajoskatona: any updates? 15:14:23 or issues 15:14:24 Nothing what I have seen, so seems to be quiet everywhere 15:14:38 ++ 15:14:40 thx 15:15:58 #topic Stable branches 15:16:17 as bcafarel already mentioned there are some patches https://review.opendev.org/c/openstack/neutron-tempest-plugin/+/785092 and https://review.opendev.org/c/openstack/neutron/+/785093 15:16:23 which are for wallaby's ci 15:16:34 and I will need to do similar patches for stadium projects 15:16:45 I will do them this week 15:17:02 #action slaweq to update wallaby's scenario jobs in neutron-tempest-plugin 15:17:21 bcafarel: anything else regarding stable branches ci? 15:18:03 recent branches look OK overall 15:18:14 stein has https://bugs.launchpad.net/neutron/+bug/1923412 small fix needed 15:18:15 Launchpad bug 1923412 in neutron "[stable/stein] Tempest fails with unrecognized arguments: --exclude-regex" [Critical,Triaged] 15:18:40 rocky I wonder, it seems iptables_hybrid job is consistently failing (like on https://review.opendev.org/c/openstack/neutron/+/770794 ) 15:18:54 but I did not have time to check this week sorry :( 15:20:00 I'll try to check it 15:20:13 there are some failing tests there 15:20:24 but also it fails on collecting logs after job's failure 15:21:18 there is some error in subunit2html.py 15:21:37 I wonder when we will be able to EOL those branches which runs on python 2.7 still 15:21:50 and on Ubuntu Xenial 15:22:34 hmm 15:22:37 interesting 15:22:51 Ubuntu Xenial support ends in April 2021 according to https://lists.ubuntu.com/archives/ubuntu-announce/2016-April/000207.html 15:23:05 and we are still using it for testing Queens and Rocky 15:23:19 I think we can try to bump this library to 1.1.0 15:23:31 we are still using 1.0.0 15:23:37 sorry, wrong link, good one is https://wiki.ubuntu.com/Releases 15:24:04 ok, bcafarel can You report an issue regarding that rocky jobs? 15:24:28 slaweq: sure will do later tonight (I will check on other backports too to confirm) 15:24:30 and I will send an email to the community to ask what we should do with that 15:24:37 thx bcafarel 15:24:50 #action bcafarel to report stable/rocky ci failures on LP 15:25:05 #action slaweq to send email about testing older branches on Ubuntu Xenial 15:25:23 ahh, wait 15:25:37 Xenial's standard support ends this month 15:25:42 but EOL will be in 2024 15:25:46 so we still can use it :) 15:25:52 I will not send any email about it 15:25:55 sorry for the noise 15:26:29 ok, I think we can move on 15:26:32 #topic Grafana 15:26:36 http://grafana.openstack.org/dashboard/db/neutron-failure-rate 15:27:12 it seems we are better with functional job finally 15:27:23 but we have some issue with neutron-tempest-plugin-scenario jobs 15:27:43 as most of them failing at around 20-30 % of times 15:29:12 except that I think that things are pretty ok 15:30:09 let's talk about some specific issues 15:30:11 #topic fullstack/functional 15:30:24 I found (again) this week failure of test_keepalived_spawns_conflicting_pid_vrrp_subprocess 15:30:28 https://7c4af2962e649c78a0b8-8dbbf393fa4fd2548b6eb8c45bf2a7e6.ssl.cf2.rackcdn.com/785011/1/check/neutron-functional-with-uwsgi/dc727d9/testr_results.html 15:30:38 but I think we marked those tests as unstable, didn't we? 15:30:45 yes, same problem as before 15:30:59 the sigterm signal do not reply 15:31:08 and the thread never ends 15:31:27 yes, but is the patch which markes that as unstable merged already? 15:31:32 do You remember that patch? 15:31:39 nope, I think we removed it 15:31:44 we'll need to propose it again 15:31:51 ahh, ok 15:32:08 maybe we should then mark them as unstable until we will fix that issue completly 15:32:31 ralonsoh: will You propose it? 15:32:32 perfect 15:32:34 yes 15:32:37 thx 15:32:54 #action ralonsoh to mark test_keepalived_spawns_conflicting_pid_vrrp_subprocess functional test as unstable 15:33:07 ok, regarding fullstack 15:33:19 I just want to ask You about review of the patch https://review.opendev.org/c/openstack/neutron/+/783748 15:33:24 it's new fullstack test to add 15:33:29 for segmentation_id updates 15:33:41 no rush but if You will have few minutes, please take a look 15:33:53 next topic 15:33:55 #topic Tempest/Scenario 15:34:07 here there is most "hot" issue for this week 15:34:34 I noticed that in all our neutron-tempest-plugin scenario jobs random tests are failing due to ssh failure 15:34:48 every time there is problem with accessing metadata service 15:35:00 https://f237087178b608be7cbc-a3abe88e6ab2911093ccea19e77537d0.ssl.cf5.rackcdn.com/785749/1/check/neutron-tempest-plugin-scenario-linuxbridge/f746672/testr_results.html 15:35:02 https://5e09181bcc1a50499619-17764b56a5c622705c872e3c7dca2597.ssl.cf2.rackcdn.com/769788/6/check/neutron-tempest-plugin-scenario-openvswitch/32b4182/testr_results.html 15:35:04 https://f3719601e38156e4ec26-626e176c8fffa963d6c301ab5a223935.ssl.cf1.rackcdn.com/785633/2/check/neutron-tempest-plugin-scenario-linuxbridge/4fcd101/testr_results.html 15:35:06 https://storage.gra.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_da7/785895/1/check/neutron-tempest-plugin-scenario-linuxbridge/da7ea5e/testr_results.html 15:35:09 https://5b963027deeaf11430aa-de753611fe18d17817cd3cc57c427456.ssl.cf1.rackcdn.com/785895/1/check/neutron-tempest-plugin-scenario-openvswitch/ce74a5b/testr_results.html 15:35:10 https://storage.gra.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_785/785889/3/check/neutron-tempest-plugin-scenario-openvswitch/785f8bb/testr_results.html 15:35:12 https://3acd25be4b1698636631-3e445c1a5b8a8c84f71b5a61cff5a09b.ssl.cf5.rackcdn.com/785950/1/check/neutron-tempest-plugin-scenario-linuxbridge/5c9ea3c/testr_results.html 15:35:14 those are examples of the failures 15:35:27 all those jobs runs with "ha=True" so keepalived is involved in the router's namespace 15:35:41 any volunteer to investigate it? 15:35:51 I could do it 15:36:24 I think that maybe https://review.opendev.org/c/openstack/neutron/+/782679 may be useful for debugging that issue 15:36:44 ahh, no 15:36:45 sorry 15:36:49 it's only for grenade jobs 15:37:01 we should do similar change for neutron-tempest-plugin jobs also 15:37:06 ok 15:37:15 ralonsoh: I will open LP for that issue 15:37:22 so you think this is a problem in the iptables rule? 15:37:23 I forgot to do it before the meeting 15:37:32 idk but maybe can be 15:37:39 ping me with the LP link 15:37:59 sure, I will 15:39:02 #action slaweq to report LP with metadata issue in scenario jobs 15:39:17 #ralonsoh to check metadata issue in scenario jobs 15:41:29 ok, that's all regarding scenario jobs for today 15:41:35 #topic Periodic 15:41:46 periodic jobs looks pretty ok recently 15:42:01 one thing I want to ask You is to review https://review.opendev.org/c/openstack/neutron/+/785660 15:42:07 sure 15:42:21 as we recently discovered that periodic jobs based on tripleo are testing neutron Victoria every day 15:42:28 which basically doesn't makes any sense :) 15:42:30 ups... 15:42:47 so this patch ensures that neutron is from master branch in that jobs always 15:43:14 ok, that's all from my side 15:43:15 he, nice catch 15:43:28 do You have anything else to discuss today? 15:43:49 no, thanks 15:44:18 ahh, one last thing - next week's meeting will be cancelled as we have ptg in the same time 15:44:38 ok, so if there are no other topics, we can finish earlier today 15:44:43 thx for attending the meeting 15:44:49 #endmeeting