15:00:10 #startmeeting neutron_ci 15:00:11 Meeting started Wed Feb 12 15:00:10 2020 UTC and is due to finish in 60 minutes. The chair is slaweq. Information about MeetBot at http://wiki.debian.org/MeetBot. 15:00:12 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 15:00:13 hi 15:00:15 The meeting name has been set to 'neutron_ci' 15:00:16 hi 15:01:22 bcafarel: is on pto but maybe njohnston will join soon 15:01:27 I am here 15:01:31 hi :) 15:01:36 sorry, also on a videoconference with the Octavia folks 15:01:40 so lets start 15:01:44 njohnston: no problem 15:01:51 #topic Actions from previous meetings 15:02:02 slaweq to backport https://review.opendev.org/#/c/695834/ to stable branches in neutron-vpnaas 15:02:11 I was checking that and I had some doubts about it 15:02:31 finally gmann and amotoki fixed stable/rocky branch in other way so this wasn't needed 15:03:02 next one 15:03:04 slaweq to update grafana dashboard with missing jobs 15:03:11 patch https://review.opendev.org/706271 15:03:17 waiting for second +2 now 15:03:46 next one 15:03:48 slaweq to open LP related to fullstack placement issue 15:03:55 Bug reported: https://bugs.launchpad.net/neutron/+bug/1862177 15:03:56 Launchpad bug 1862177 in neutron "Fullstack tests failing due to problem with connection to the fake placement service" [High,In progress] - Assigned to Lajos Katona (lajos-katona) 15:04:09 lajoskatona and rubasov are on it already 15:04:20 so we are in good hands :) 15:04:55 next one 15:04:56 slaweq to open LP related to "hang" neutron-server 15:05:03 Bug reported: https://bugs.launchpad.net/neutron/+bug/1862178 15:05:04 Launchpad bug 1862178 in neutron "Fullstack tests failing due to "hang" neutron-server process" [High,In progress] - Assigned to Rodolfo Alonso (rodolfo-alonso-hernandez) 15:05:18 two patches 15:05:27 #link https://review.opendev.org/#/c/707151/ 15:05:34 #link https://review.opendev.org/#/c/707222/ 15:05:52 (second one is good stuff) 15:06:04 and do You think this will solve this issue? 15:06:19 I've tested the second one with my dev env 15:06:31 and I see a big performance improvement 15:06:53 a) the delete operation (bulk), removing unneded physnets 15:07:04 b) adding new VLAN tags in bulk mode 15:07:16 and of course, reducing the number of tags 15:07:43 that;s all 15:08:07 ok, thx ralonsoh for those patches - I hope this will help with this issue in fullstack tests too 15:08:24 ok, next one 15:08:26 ralonsoh to check missing project_id issue 15:08:36 no luck sorry.... 15:08:42 np 15:08:48 I can't find were/why/how this is happening... 15:08:59 *where 15:09:47 should we simply maybe send some patch to log in such case both expected and actual dict 15:10:01 maybe we will at least will know in which one it's missing 15:10:33 we can see that in the response 15:11:00 the problem is to find where this project_id is filtered and deleted from this returned dict 15:11:15 and, why is this happening 1 out of 500 tests? 15:12:28 ok, I see now 15:12:42 it's filtered out from the "actual" dict than 15:12:48 *then 15:13:10 exactly, but no one (SDK, client or server) should do this 15:13:27 I tried to find something related to the project_id/tenant_id migration 15:13:45 because in some places, some black magic is done to convert one into the other 15:13:55 but project_id should be always there 15:14:05 (not tenant_id, that should be removed) 15:15:24 but afaik tempest has got own implementation of clients 15:15:36 I know 15:15:46 it don't use OpenStack SDK or neutronclient 15:15:52 maybe there is some bug there? 15:16:03 let me check again this 15:16:17 ralonsoh: ok, thx 15:16:19 but this is the nth time I try to find this bug 15:16:40 or maybe we should add some debug log of every response which is going to be send from neutron-server 15:16:59 so we than can confirm if that wasn't send from server or was filtered out on client's side 15:17:21 we could 15:18:06 ok, so You will take another look into it this week, right? 15:18:14 sure! 15:18:19 thx a lot 15:19:05 #ralonsoh to check again mystery of vanishing project_id 15:19:12 hahahaha 15:19:21 :) 15:19:50 ok, that's all actions from last week 15:20:01 do You have anything else to add here or can we move on? 15:20:14 no thanks 15:20:25 ok, so lets move on 15:20:34 #topic Stadium projects 15:20:43 standardize on zuul v3 15:20:45 Etherpad: https://etherpad.openstack.org/p/neutron-train-zuulv3-py27drop 15:20:50 there was slow progress this week 15:20:56 So I added a summary section at the top of that etherpad 15:21:04 so it's easy to find the pending updates 15:21:25 thx njohnston - that's really helpful 15:21:28 for zuulv3 we have 3 projects with pending changes, 2 that have no activity 15:21:37 and we are really close to finish it 15:21:54 yep! 15:21:56 I also noted a few remaining py27 things, including the neutron-tempest-plugin change 15:22:20 yes, neutron-tempest-plugin is something what I'm aware of 15:22:29 patch is ready 15:22:34 yep! 15:22:44 but I would like to first merge few fixes, release last version with support for py27 15:22:51 and than drop this support 15:24:16 njohnston: and I will check this midonet patch today 15:24:44 anything else regarding stadium projects? 15:24:56 nope, things are looking good. 15:26:06 I just commented in https://review.opendev.org/#/c/695094/ 15:26:15 can You check it and tell me what do You think about it? 15:26:43 sure 15:27:00 will do 15:27:34 thx 15:28:01 ok, so lets move on 15:28:03 #topic Grafana 15:28:08 #link http://grafana.openstack.org/dashboard/db/neutron-failure-rate 15:30:05 functional jobs were recently failing pretty often 15:30:24 but as I checked many of those failures were in patches related to ovn 15:32:56 other than that I think that graphs are pretty good 15:33:25 I think so 15:34:01 ok, lets talk about few issues which I found recently 15:34:04 #topic fullstack/functional 15:34:20 here I have only one thing to mention 15:34:27 Unauthorized commands like "ping" or "ncat": 15:34:28 https://2a0154cb9a3e47bde3ed-4a9629bf7847ad9c8b03c9755148c549.ssl.cf1.rackcdn.com/705660/4/check/neutron-functional/2e5030b/testr_results.html 15:34:30 https://656129f4adff35088518-c39e8636195a8a58924c560773952ce4.ssl.cf1.rackcdn.com/705480/3/check/neutron-functional/e49a784/testr_results.html 15:34:33 sorry for that 15:34:39 and also probably same issue https://ad181adc6d8db459c7ce-fbb316944f0ca23c676e132d61555672.ssl.cf1.rackcdn.com/705237/4/check/neutron-functional/ed9cfb5/testr_results.html 15:34:43 bug opened and patch submitted 15:34:50 I wonder why it not happens all the time 15:34:57 #link https://review.opendev.org/#/c/707368/ 15:35:05 yes, that was my question 15:35:17 and I have no answer 15:35:25 but this patch should solve this problem 15:35:59 and also in https://656129f4adff35088518-c39e8636195a8a58924c560773952ce4.ssl.cf1.rackcdn.com/705480/3/check/neutron-functional/e49a784/testr_results.html error was because of unauthorized "ping" command 15:36:06 but ping is in this rootwrap filters file 15:36:24 ?? 15:37:01 ralonsoh: first failed test in the link above 15:37:06 it failed with error: 15:37:11 yes yes 15:37:17 but why?? 15:37:20 Unauthorized command: ip netns exec test-19869f4e-e878-47fe-8cf8-90b60f6269e1 ping 192.178.0.2 -W 1 -c 3 (no filter matched) 15:37:44 ralonsoh: yeah, that's the question 15:37:54 oook ok 15:37:56 maybe I229e926341c5e6c8b06f59950e3ae09864d0f1f6 15:38:01 is the problem 15:38:08 let me review this patch 15:38:21 https://review.opendev.org/#/c/705065/ 15:38:47 ok 15:38:56 so I will assign it as an action to You, ok? 15:39:01 ok 15:40:07 #action ralonsoh to check issues with unauthorized ping and ncat commands in functional tests 15:41:06 ok, so lets talk about scenario jobs 15:41:08 #topic Tempest/Scenario 15:41:25 first, I proposed patch to increase timeout for tempest-ipv6-only job: https://review.opendev.org/707356 15:41:40 I will respond to haleyb's comment there in a minute 15:41:53 :) 15:43:02 OVN slow is taking almost 3 hours and sometimes I've seen timeouts 15:43:11 haleyb: answered :) 15:43:31 ralonsoh: yes, but that's slow tests 15:43:48 tempest-slow-py3 also takes almost 3 hours 15:44:25 ok, lets move on quickly to other issues 15:44:33 I saw couple of times ssh problem due to "socket.timeout: timed out" error in various tests, like: 15:44:43 https://storage.gra.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_9aa/703376/5/check/neutron-tempest-dvr/9aa046a/testr_results.html 15:44:45 https://db85fb7d3af8e26f7154-0d96b608ecbdac6c8248619c1dff0910.ssl.cf5.rackcdn.com/704833/4/check/neutron-tempest-dvr/67fd708/testr_results.html 15:44:47 https://19574e4665a40f62095e-6b9500683e6a67d31c1bad572acf67ba.ssl.cf1.rackcdn.com/705982/6/check/neutron-tempest-dvr/8f3fbd0/testr_results.html 15:44:49 https://storage.gra.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_37a/705452/5/check/neutron-tempest-dvr-ha-multinode-full/37a65a9/testr_results.html 15:44:51 mostly in neutron-tempest-dvr job 15:44:59 but I saw it also in multinode dvr job 15:45:15 I think it's new issue and we should investigate it 15:45:19 any volunteers? 15:46:08 ok, I will report new bug and will try to look into logs 15:46:29 #action slaweq to report issue with ssh timeout on dvr jobs and check logs there 15:47:00 other than that, I saw one issue with nova's revert_resize, like e.g. in 15:47:06 https://3e447a3e4caf9c486a4d-b84d447537aa444ce20bcf5414a5ef0e.ssl.cf5.rackcdn.com/707248/1/check/neutron-tempest-dvr-ha-multinode-full/057233d/testr_results.html 15:47:07 https://3ceba9124358a5c9851b-33ba471340a760763569a038d91ca6b5.ssl.cf2.rackcdn.com/706875/2/check/neutron-tempest-dvr-ha-multinode-full/6d77e6d/testr_results.html 15:47:13 so I will report bug for nova about that 15:47:35 and that's all from my side regarding scenario jobs 15:47:41 anything else You want to add? 15:47:54 no thanks 15:47:59 nope 15:48:10 ok, thx 15:48:24 so (almost) last thing for today 15:48:29 #topic Periodic 15:48:38 neutron-ovn-tempest-ovs-master-fedora is failing everyday 15:48:48 we should check this job 15:49:25 it's failing on deploying devstack 15:49:25 do you have the link for the periodic jobs? 15:49:26 https://9fc08b4308a330f341b7-ee326d1edc43244c4c522686856ef03c.ssl.cf2.rackcdn.com/periodic/opendev.org/openstack/neutron/master/neutron-ovn-tempest-ovs-master-fedora/07d0c6f/job-output.txt 15:49:44 error is like: 15:49:46 2020-02-11 07:05:25.363036 | controller | + lib/infra:install_infra:32 : virtualenv -p python3 /opt/stack/requirements/.venv 15:49:47 2020-02-11 07:05:25.441404 | controller | ERROR:root:ImportError: cannot import name ensure_text 15:50:12 I can check it 15:50:17 I think it's some issue related to fedora 29 15:50:25 ralonsoh: thx 15:50:53 #action ralonsoh to check periodic neutron-ovn-tempest-ovs-master-fedora job's failures 15:51:10 other periodic jobs looks fine 15:51:28 ok, and I have one more topic for tody 15:51:31 *today 15:51:32 #topic Open discussion 15:52:02 some time ago I asked on ML about future of lib/neutron and lib/neutron-legacy in devstack 15:52:31 I think haleyb said that there shouldn't be much additional work needed to finally make lib/neutron usable 15:53:10 so I would like to ask if there is any volunteer to work on that or maybe we should simply remove lib/neutron and rename lib/neutron-legacy to not be "legacy" anymore? 15:53:24 as now things can be confusing for users :) 15:53:30 any thoughts? 15:54:12 I'm ok with this but not a priority now 15:54:42 I can take it but not now 15:54:52 ralonsoh: sure, it's not top priority but IMO we should at least decide what to do with it finally 15:55:09 as now we are "in the middle" of 2 solutions since few years 15:55:21 I'm ok with finishing the migration 15:55:44 great 15:56:07 so lets keep this in mind and maybe move it slowly forward 15:56:19 ok, that's all from my side for today 15:56:27 anything else You want to talk today? 15:56:48 no 15:57:04 * njohnston just finished the other meeting, is catching up 15:57:09 nothing from me 15:57:16 ok, thx for attending 15:57:22 see You online 15:57:24 o/ 15:57:27 #endmeeting