16:00:31 #startmeeting neutron_ci 16:00:32 Meeting started Tue Feb 19 16:00:31 2019 UTC and is due to finish in 60 minutes. The chair is slaweq. Information about MeetBot at http://wiki.debian.org/MeetBot. 16:00:33 hi 16:00:34 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 16:00:36 The meeting name has been set to 'neutron_ci' 16:00:40 o/ 16:00:49 o/ 16:01:26 lets wait few minutes for others 16:01:44 maybe bcafarel and hongbin will join too 16:01:52 I know that haleyb is on PTO today 16:01:58 o/ 16:02:33 thanks for the ping slaweq, I was writing some doc (easy to forget the clock then) 16:02:45 bcafarel: :) 16:03:00 ok, let's start then 16:03:06 #topic Actions from previous meetings 16:03:14 mlavalle to check if adding KillFilter for neutron-keepalived-state-change will solve issues with L3 agent in dvr jobs 16:03:39 I porposed patch https://review.openstack.org/#/c/636710/ 16:03:51 o/ 16:03:59 hi hongbin :) 16:03:59 and actually dougwig made an ineteresting comment 16:04:10 which I intend to explore 16:04:36 yeah, that is interesting 16:05:03 but patch basically looks like is helping to solve some of failing tests in multinode job, right? 16:05:25 I agree in that we should keep the door as narrowly open as possible 16:05:48 +1 16:05:54 ++ 16:06:01 slaweq: I haven't had time to chack on that since yesterday? Did you look? 16:06:55 I looked at test's results right now: http://logs.openstack.org/10/636710/2/check/neutron-tempest-plugin-dvr-multinode-scenario/f44d655/testr_results.html.gz 16:07:10 it looks that "only" 2 tests failed which is much better than it was 16:07:47 oh yeah, in fact the trunk lifecycle test is passing 16:07:56 which I haven't seen in a long time 16:08:08 so it looks we are moving in the right direction 16:08:10 also from quick look at l3-agent's log, it looks much better and IMO agent was working all the time properly 16:08:22 so there really was a rootwrap filter missing all along? 16:08:31 it seems so 16:08:32 bcafarel: for some time at least 16:08:45 it was only for some months 16:08:46 bcafarel: I removed them when I removed old metadata proxy code 16:09:05 slaweq: ok, missing for some months makes more sense :) 16:09:28 :) 16:09:36 bcafarel: when I started suspecting the filters were the cause, I had the exact same question in my mind 16:09:54 how come did this work before? 16:10:10 but we really introduced the bug recently 16:10:43 ok, so I'll play with some more rechecks of the patch 16:10:51 I think that this bug was exposed by switch to python3 16:11:01 on python2.7 it is "working" somehow 16:11:19 and I'll explore dougwig's suggestion, which is very sensible 16:11:21 at least agent is not dying when this issue occurs 16:11:46 yeah, I think that in python2.7 we still miss the filter 16:11:56 but for some reason we don't kill the agent 16:12:12 anbd therefore we don't have a chain reaction in the tests 16:12:21 yep 16:13:16 ok, moving on to the next one then 16:13:18 bcafarel to continue work on grafana jobs switch to python 3 16:13:37 s/grafana/grenade/ 16:13:53 bcafarel: right :) 16:14:39 I commented on #openstack-neutron yesterday on it, I think it will be better to have grenade job definition in grenade repo once it is full zuul v3 (which will allow devstack_localrc vars to work) 16:14:50 so I updated https://review.openstack.org/#/c/636356/ to simply switch to py3 in the meantime 16:15:13 just got zuul +1 :) 16:15:43 and now it looks that it is running on python 3 indeed :) 16:15:54 for example here: http://logs.openstack.org/56/636356/7/check/neutron-grenade-multinode/0ea452a/logs/screen-q-l3.txt.gz#_Feb_19_14_14_16_092790 16:16:24 yes, wrong variable in my first try sorry, now it looks good 16:16:41 +2 it already 16:16:56 IMO it will be good if we just have single node grenade on python 2.7 16:17:01 and others on python3 16:17:31 and the nature of the grenade job is that it upgrades from an old version running py27 to a new version running py3 16:18:28 so regardless of what code is running the grenade harness we get some testing on both versions of neutron code 16:18:29 njohnston_: I think that in those jobs it runs on python3 for both old and new 16:18:43 see e.g. here: http://logs.openstack.org/56/636356/7/check/neutron-grenade-multinode/0ea452a/logs/old/local_conf.txt.gz 16:19:24 if it would be possible to do upgrade from py27 to py3 that would be great IMO 16:19:37 but maybe that should be discussed with QA team? 16:20:00 I checked the grenade output, for example 16:20:09 njohnston_: in grenade-py3 job (defined in grenade repo) it is also like that, all on py3: http://logs.openstack.org/56/636356/7/check/grenade-py3/b075864/logs/old/local_conf.txt.gz 16:20:43 old: http://logs.openstack.org/56/636356/7/check/grenade-py3/b075864/logs/grenade.sh.txt.gz#_2019-02-19_13_45_48_138 "New python executable in /opt/stack/old/requirements/.venv/bin/python2" 16:20:56 new: http://logs.openstack.org/56/636356/7/check/grenade-py3/b075864/logs/grenade.sh.txt.gz#_2019-02-19_14_19_45_871 "New python executable in /opt/stack/new/requirements/.venv/bin/python3.5" 16:21:11 if the upgrade from py3 to py3 success, is there any scenario that upgrade from py2 to py3 break? 16:22:28 njohnston_: but is devstack deployed in venv in grenade job? I really don't think so 16:22:41 but I'm not grenade expert so may be wrong here :) 16:22:57 hongbin: we don't have any other jobs which tests upgrade 16:23:55 slaweq: I can ask the QA team to be sure 16:24:10 njohnston_: ok, thx 16:24:21 ok, from me, i think testing from py3 to py3 upgrade is enough 16:24:39 njohnston_: but still, patch from bcafarel is "consistent" with this single node grenade job so we can IMO go forward with it :) 16:24:56 slaweq: Absolutely, no disagreement there 16:25:36 njohnston_: great :) 16:25:58 ok, can we move on? 16:27:08 yep 16:27:17 ok, next one then 16:27:23 slaweq to propose patch with new decorator skip_if_timeout in functional tests 16:27:33 I did patch https://review.openstack.org/#/c/636892/ 16:27:40 please review it if You can 16:28:40 looks like functional tests failed on the last run 16:29:22 I guess that would be the run still running 16:29:31 njohnston_: where? 16:29:40 http://logs.openstack.org/92/636892/1/gate/neutron-functional/85da30c/logs/testr_results.html.gz ? 16:30:13 ahh, in the gate 16:30:21 ok, so maybe I will need to investigate it more :) 16:30:23 thx 16:30:25 http://logs.openstack.org/92/636892/1/gate/neutron-functional/85da30c/job-output.txt.gz#_2019-02-19_15_32_02_801108 16:31:17 so it looks that it maybe don't work properly, I will investigate that tomorrow morning then 16:31:36 #action slaweq to fix patch with skip_if_timeout decorator 16:31:45 you did get your test failure in the end at least :) 16:31:56 LOL, indeed 16:32:20 ok, that's all for actions from last week 16:32:29 next topic is: 16:32:31 #topic Python 3 16:33:02 as we already talked today, we have (now in the gate even) patch for grenade multinode jobs to switch to py3 16:33:06 thx bcafarel :) 16:33:30 there is also patch https://review.openstack.org/633979 for neutron-tempest-dvr-ha-multinode-full job 16:33:39 but this one is still failing with some tests 16:34:19 I'm not sure if that is issue with tests or maybe similar problem like we have in neutron-tempes-plugin-scenario-dvr-multinode job 16:34:46 so I will probably split it into 2 pieces: migration to zuulv3 and then second patch with switch to py3 16:35:00 do You agree? 16:35:04 yes 16:35:10 sounds good yes 16:35:12 thanks :) 16:35:19 +1 16:35:33 #action slaweq to split patch https://review.openstack.org/633979 into two: zuulv3 and py3 parts 16:36:00 and I think that this will be all for switch to py3 16:36:25 we will still have some experimental jobs to switch but that can be done slowly later I think 16:37:31 yeah it would be a bad sign if important tests we needed for proper python3 support are hidden in experimental jobs 16:37:42 so we can do these "leisurely" 16:37:58 I like leisurely 16:38:04 LOL 16:38:08 me too 16:38:49 ok, any other questions/something to add about python3? 16:38:54 or can we move on? 16:38:57 not from me 16:39:12 I need to send an email to to openstack-discuss to ask the stadium projects about their py3 status 16:39:12 all good here 16:39:14 * mlavalle will have to drop off at 45 minutes after the hour 16:39:26 njohnston_++ 16:39:45 #action njohnston to ask stadium projects about python3 status 16:39:49 I can help with some, but I wouldn't want to but in to projects like midonet that I know little about 16:40:14 njohnston_: yes, same for me 16:40:24 I believe I sent an email a while back and got no requests for help, but let's see if anyone is so motivated now 16:40:36 thx njohnston_ 16:40:48 ok, lets move on quickly 16:40:50 #topic Grafana 16:40:56 #link http://grafana.openstack.org/dashboard/db/neutron-failure-rate 16:42:05 I was looking at grafana today and TBH I don't see anything very bad on it in last few days 16:42:26 but maybe You noticed something what You want to discuss about 16:43:32 even all tempest jobs are in quite good shape this week 16:43:35 looks pretty good to me 16:44:18 the only thing I see if functional py27 in gate 16:44:35 mlavalle: yes, functional tests are not very good currently 16:44:48 and I have cuplrits of it 16:44:55 #topic fullstack/functional 16:45:16 we recently noticed at least 3 bugs in functional tests: 16:45:18 - https://bugs.launchpad.net/neutron/+bug/1816239 - patch proposed https://review.openstack.org/#/c/637544/ 16:45:19 Launchpad bug 1816239 in neutron "Functional test test_router_processing_pool_size failing" [High,In progress] - Assigned to Brian Haley (brian-haley) 16:45:20 - https://bugs.launchpad.net/neutron/+bug/1815585 - if there will be no anyone to look at this, I will try to debug it 16:45:21 Launchpad bug 1815585 in neutron "Floating IP status failed to transition to DOWN in neutron-tempest-plugin-scenario-linuxbridge" [High,Confirmed] 16:45:22 - https://bugs.launchpad.net/neutron/+bug/1816489 - same here, we need volunteer for this one 16:45:23 Launchpad bug 1816489 in neutron "Functional test neutron.tests.functional.agent.l3.test_ha_router.LinuxBridgeL3HATestCase. test_ha_router_lifecycle failing" [High,Confirmed] 16:45:53 first one is hitting as the most often I think and it has already fix proposed 16:46:06 so we should be better after this will be merged 16:46:15 other 2 needs volunteers to debug :) 16:46:27 oh, oh, do we depend on haleyb for this? we are in trouble :-) 16:46:57 I can take a look at https://bugs.launchpad.net/neutron/+bug/1815585 but someone else could look at the last one maybe :) 16:46:58 Launchpad bug 1815585 in neutron "Floating IP status failed to transition to DOWN in neutron-tempest-plugin-scenario-linuxbridge" [High,Confirmed] 16:47:17 ok, I'll try to look at the last one 16:47:20 mlavalle: no, fortunatelly fix was done by liuyulong :) 16:47:24 thx mlavalle 16:47:35 LOL 16:47:35 #action mlavalle to check bug https://bugs.launchpad.net/neutron/+bug/1816489 16:47:36 Launchpad bug 1816489 in neutron "Functional test neutron.tests.functional.agent.l3.test_ha_router.LinuxBridgeL3HATestCase. test_ha_router_lifecycle failing" [High,Confirmed] 16:47:47 #action slaweq to check bug https://bugs.launchpad.net/neutron/+bug/1815585 16:47:48 ok guyas I got to leave 16:47:53 ok, thx mlavalle 16:47:55 see You later 16:47:59 o/ 16:48:06 basically that is all what I have for today 16:48:07 o/ mlavalle 16:48:17 other jobs are in pretty good shape 16:48:28 so do You have something else You want to talk about today? 16:49:13 nothing from me 16:49:26 ok, so lets have 10 minutes back today :) 16:49:36 thx for attending and see You all around 16:49:41 o/ 16:49:43 \o/ 16:49:45 #endmeeting