15:00:34 #startmeeting neutron_ci 15:00:34 Meeting started Wed Jul 8 15:00:34 2020 UTC and is due to finish in 60 minutes. The chair is slaweq. Information about MeetBot at http://wiki.debian.org/MeetBot. 15:00:35 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 15:00:37 hi 15:00:37 The meeting name has been set to 'neutron_ci' 15:00:52 hi 15:01:02 o/ 15:01:46 o/ sorry somehow my irc client now had #openstack-meeting3 (without "-") 15:02:01 \o 15:02:49 Grafana dashboard: http://grafana.openstack.org/dashboard/db/neutron-failure-rate 15:02:59 please open link and we can start :) 15:03:06 #topic Actions from previous meetings 15:03:13 ralonsoh will check get_datapath_id issues in https://storage.bhs.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_63a/711425/11/check/neutron-functional/63ac4ca/testr_results.html 15:03:31 I didn't find the root cause of the timeout 15:03:44 but I've pushed a patch for ovsdbapp 15:03:58 to add a bit more information when the txn fails 15:04:08 if that depends on the RXN or the TXN queue 15:04:33 sorry for that... it's not easy to track those errors 15:05:09 ok, so it's WIP 15:05:12 thx ralonsoh 15:05:18 yes 15:05:29 please keep us updated if You will find anything 15:05:32 sure 15:05:40 btw. do we have LP for that? 15:05:45 no 15:05:53 just a patch in gerrit 15:06:10 If I find another error like this one, I'll open a LP bug 15:06:16 ok, thx 15:06:23 ok, next one 15:06:25 slaweq will check errors with non existing interface in https://storage.bhs.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_63a/711425/11/check/neutron-functional/63ac4ca/testr_results.html 15:06:32 I did not have time for that one 15:06:44 and logs from the link above are already gone 15:07:05 as I didn't saw it later, I will not bother with it anymore for now 15:07:18 if it will happen again, I will try to check it 15:07:29 next one 15:07:31 slaweq to move non-voting jobs to experimental queue in EM branches 15:07:37 Rocky: https://review.opendev.org/#/c/739668/ 15:07:39 Queens: https://review.opendev.org/#/c/739669/ 15:07:41 Pike: https://review.opendev.org/#/c/739910/ (thx bcafarell) 15:07:55 and one more question - what about Ocata? 15:08:04 do we still support this branch? 15:08:15 I think so but I couldn't find any open patch for it 15:08:22 that was set as EOL 15:08:26 if I'm not wrong 15:08:30 I am looking for link I think it is EOL now 15:09:19 EOL IIRC 15:09:24 ah "Moving the stable/ocata to 'Unmaintained' phase and then EOL" 15:09:42 I don't see ocata-eol tag in the repo 15:09:57 maybe the paperwork is not fully completed for https://releases.openstack.org/ (and tags) 15:10:21 but at least on neutron side we did not see recent backport requests 15:10:44 as I recall there was something about how individual projects can EOL it at their individual discretion 15:10:52 ok, I will check its state after I will come back from the PTO 15:11:36 ok, next one 15:11:37 slaweq to investigate failures in test_create_router_set_gateway_with_fixed_ip test 15:11:43 I didn't have time for that one too 15:11:58 it's not very urgent as it happens in non-voting job 15:12:11 but if someone wants to check, feel free to take it ;) 15:12:27 I'll ping you tomorrow about this one 15:12:33 ralonsoh: sure, thx 15:12:40 ok, next one 15:12:42 maciejjozefczyk to check neutron_tempest_plugin.scenario.test_connectivity.NetworkConnectivityTest.test_connectivity_through_2_routers failure in ovn jobs 15:13:07 I havent time to check that one ;/ I'm going to take a look tomorrow 15:14:28 ok 15:14:36 I will assign it to You for next week 15:14:40 #action maciejjozefczyk to check neutron_tempest_plugin.scenario.test_connectivity.NetworkConnectivityTest.test_connectivity_through_2_routers failure in ovn jobs 15:14:58 (busy week for everyone) 15:15:02 yeah 15:15:07 as usual 15:15:21 lets move to the next topic 15:15:23 #topic Stadium projects 15:15:29 zuul v3 migration 15:15:51 I finally pushed patch for last missing neutron job: https://review.opendev.org/729591 - lets see how it will work 15:16:09 cool! 15:16:16 it still didn't even started 15:16:27 but I will keep an eye on it 15:16:31 it will before your PTO :) 15:16:35 hehehe 15:16:56 bcafarel: I hope so 15:16:58 :) 15:17:09 ok, anything else regardig stadium for today? 15:19:52 ok, I guess it means "no" 15:19:56 so lets move on 15:19:58 #topic Switch to Ubuntu Focal 15:20:01 any updates on that? 15:20:21 devstack patch was merged 15:20:22 https://review.opendev.org/#/c/704831/ 15:20:41 but https://review.opendev.org/#/c/734304 is failing 15:20:59 I need to review the logs for FT/fullstack 15:21:33 we need this one first to land https://review.opendev.org/#/c/737984/ 15:22:05 maciejjozefczyk: will it fix errors like those seen in https://review.opendev.org/#/c/738163/ ? 15:22:07 Exception: Invalid directories: /usr/local/share/ovn, /usr/share/ovn, /usr/local/share/openvswitch, /usr/share/openvswitch, None 15:22:52 (some other jobs failing too, but I did not have time to dig further) 15:23:16 maciejjozefczyk, but apart from your patch, FTs/fullstack are failing because in Focal python-openvswitch does not exist 15:23:17 E: Unable to locate package python-openvswitch 15:23:31 :/ 15:23:45 should we ask someone from Ubuntu team for help with that? 15:24:04 we should, yes 15:24:07 ahh 15:24:14 python3-? 15:24:28 maciejjozefczyk, I'll check that today 15:24:37 ralonsoh: I have a quick hack to pull python3- for test in 738163 (hardcoded just for testing) 15:26:12 ok, so package name was renamed to python3-openvswitch in Focal? Is that correct? 15:26:31 I need to check that 15:26:40 yup 15:26:46 ok, thx ralonsoh and bcafarel for working on this 15:27:26 I think we can move on to the next topic 15:27:29 #topic Stable branches 15:27:35 Ussuri dashboard: http://grafana.openstack.org/d/pM54U-Kiz/neutron-failure-rate-previous-stable-release?orgId=1 15:27:37 Train dashboard: http://grafana.openstack.org/d/dCFVU-Kik/neutron-failure-rate-older-stable-release?orgId=1 15:27:51 stable branches maybe, stable ci not these days 15:28:06 LOL 15:28:21 ;/ 15:28:30 bcafarel: so we can say (not so)stable branches 15:28:32 ok? 15:28:34 :D 15:28:41 :) approved! 15:28:47 for the previous issues we are almost good now finally (pike https://review.opendev.org/#/c/739456/ waiting on second +2) 15:29:55 but new pep8 error (isort) needs a requirements backport which failed CI https://review.opendev.org/#/c/739912 15:29:59 bcafarel: +W 15:30:32 and we need same backport all the way to stein, right? 15:30:50 apparently yes 15:30:56 :/ 15:31:37 at least we are good with pep8 issue in master now :) 15:31:49 what are those errors in py36 tests? 15:31:58 https://storage.gra.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_513/739912/1/check/openstack-tox-py36/51376c8/testr_results.html 15:32:39 yes I am not sure where they come from (but appear in both py36/py37 and all other backports) 15:33:53 please, keep an eye on https://github.com/PyCQA/pylint/issues/3722 15:34:14 maybe yet another ci blocker issue (because we still have not enough) ;) 15:35:16 ok, lets move on 15:35:17 #topic Grafana 15:35:24 #link http://grafana.openstack.org/dashboard/db/neutron-failure-rate 15:36:44 I see one thing which worry me a bit 15:36:53 neutron-ovn-tempest-full-multinode-ovs-master is failing 100% of times since some time 15:37:04 I opened LP for that https://bugs.launchpad.net/neutron/+bug/1886807 15:37:04 Launchpad bug 1886807 in neutron "neutron-ovn-tempest-full-multinode-ovs-master job is failing 100% times" [High,Confirmed] 15:37:41 idk if that is job configuration or something like that but seems that in many tests it's failing due to ssh issues 15:38:50 other than that, I think that things are more or less ok in master branch now 15:38:56 yes thats strange because I don't remember anything that could break this... We also didn't changed the ovn branch 15:38:58 yes 15:39:16 I'm gonna take a look tomorrow what was pushed to ovn master recently 15:39:28 maciejjozefczyk: at first glance I suspect some issue with infra, it's multinode job and maybe ssh to vms which lands on the second node don't work 15:39:41 that's something what I would try to check first 15:39:45 yep 15:40:09 btw https://review.opendev.org/737984 just merged :) 15:40:25 #action maciejjozefczyk to check failing neutron-ovn-tempest-full-multinode-ovs-master job 15:40:34 thx maciejjozefczyk for looking into that 15:40:38 ack 15:40:39 ;) 15:41:17 ok, lets move on 15:41:19 #topic fullstack/functional 15:41:29 I found one new (for me at least) issue 15:41:34 neutron.tests.functional.test_server.TestWsgiServer.test_restart_wsgi_on_sighup_multiple_workers 15:41:38 https://f7a63aeb9edd557a2176-4740624f0848c8c3257f704064a4516f.ssl.cf2.rackcdn.com/736026/4/gate/neutron-functional/d7d5c47/testr_results.html 15:41:50 did You saw something like that recently? 15:43:20 not that I remember (and that code was not touched recently) 15:43:39 the worker didn't restart, isn't it? 15:43:42 bcafarel: yes, but I saw it I think twice this week 15:44:32 hmm 15:44:39 I see such error in the test logs: https://f7a63aeb9edd557a2176-4740624f0848c8c3257f704064a4516f.ssl.cf2.rackcdn.com/736026/4/gate/neutron-functional/d7d5c47/controller/logs/dsvm-functional-logs/neutron.tests.functional.test_server.TestWsgiServer.test_restart_wsgi_on_sighup_multiple_workers.txt 15:44:56 do You think it may be related? 15:46:39 actually I think I saw something similar in the past 15:46:51 when You look at the test's code: https://github.com/openstack/neutron/blob/master/neutron/tests/functional/test_server.py#L163 15:47:13 it failed on condition which checkes if some specific file was created and had got expected size 15:47:33 so it could be that size was wrong or there was no this file at all 15:47:46 I will change that condition to add some logging to it 15:47:55 so will be easier to check what happend there 15:49:34 +1 15:49:36 #action slaweq to change condition in the TestNeutronServer to have better logging 15:50:09 ok, we can get back to this issue when we will know more what happen there 15:50:18 +1 15:50:26 lets move on 15:50:28 #topic Tempest/Scenario 15:50:45 I had only this issue related to ovn job but we talked about it already 15:50:54 so just 2 short info 15:51:00 and ask for reviews :) 15:51:40 Please review https://review.opendev.org/#/c/736186/ 15:51:49 I had to rebase it to resolve some conflict there 15:52:01 and also I increased timeouts in singlenode tempest jobs: https://review.opendev.org/739955 15:52:06 please take a look at this one too 15:52:34 and that's all from me about scenario jobs 15:52:44 #topic On demand agenda 15:53:00 do You have anything else You want to talk about today? 15:53:18 no thanks 15:53:40 I have one quick item 15:53:52 as You know next 2 weeks I will be on PTO 15:54:01 slaweq, enjoy :) 15:54:06 do You want me to cancel this ci meeting or ralonsoh can You chair it? 15:54:17 I can (probably) 15:54:32 I'm asking You because You are listed as co-chair of this meeting already :) 15:54:43 ralonsoh: thx a lot 15:55:05 so I will not cancel it, please run it if You will have time or cancel if You want 15:55:15 that's all from my side 15:55:24 thx for attending today 15:55:30 and have a great rest of the week 15:55:33 bye 15:55:34 #endmeeting