16:00:11 #startmeeting neutron_ci 16:00:11 Meeting started Tue Oct 30 16:00:11 2018 UTC and is due to finish in 60 minutes. The chair is slaweq. Information about MeetBot at http://wiki.debian.org/MeetBot. 16:00:12 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 16:00:15 The meeting name has been set to 'neutron_ci' 16:00:19 hello everyone 16:00:43 hi mlavalle :) 16:00:50 o/ 16:00:54 am I late? 16:01:04 no, You are first actually 16:01:14 if yes, blame frickler and bcafarel 16:01:15 (except me who started this meeting) 16:01:29 I know, it's always bcafarel's fault :P 16:01:56 haleyb, njohnston: hongbin: manjeets: CI meeting - are You around? 16:01:59 not this time, I even held on a question I have for mlavalle :p 16:02:11 o/ 16:02:11 and o/ btw 16:02:15 o/ 16:02:28 * njohnston was lurking 16:02:42 ok, lets start then 16:02:47 #topic Actions from previous meetings 16:02:56 slaweq to continue checking how jobs will run on Bionic nodes 16:03:10 I was checking it a bit 16:03:34 I created etherpad https://etherpad.openstack.org/p/neutron_ci_on_bionic which I want to use to track the progress 16:04:01 * manjeets is in other meeting 16:04:10 +1 for the etherpad 16:04:11 and also I want to have separate topic about this on today's meeting, so let's talk about it later, ok for You? 16:05:09 I take it as "yes" :) 16:05:14 next action then: 16:05:16 mlavalle to continue debugging issue with not reachable FIP in scenario jobs 16:05:33 you guys are not going to believe this..... 16:05:48 no hist over the past seven days since yesterday 16:05:49 slaweq, so all the jobs mentioned in etherpad should use ubuntu bionic ideally ? 16:06:02 mlavalle: really? wow. 16:06:27 This is the query I'm using http://logstash.openstack.org/#/dashboard/file/logstash.json?query=message:%5C%22line%20143,%20in%20test_trunk_subport_lifecycle%5C%22&from=7d 16:06:49 mlavalle: are You sure? 16:06:58 I found something like http://logs.openstack.org/59/596959/6/check/neutron-tempest-plugin-dvr-multinode-scenario/fbc011b/testr_results.html.gz from yesterday for example 16:07:16 isn't it issue like You were looking for? 16:09:02 yeah, it's the same issue 16:09:14 the query is not catching it 16:09:41 ah, you know why? 16:09:46 it's the line number 16:10:03 I should remove the line number from the query 16:10:47 I took the query that slaweq left originally in the bug (that includes the line number) and just added the 7days 16:10:48 :) 16:11:01 sorry for that then 16:11:16 in that case.... 16:11:21 I'll go back to it 16:11:27 LOL 16:11:32 ok, I will assign it to You again :) 16:11:39 mlavalle to continue debugging issue with not reachable FIP in scenario jobs 16:11:45 #action mlavalle to continue debugging issue with not reachable FIP in scenario jobs 16:12:01 thx mlavalle for taking care of it 16:12:04 next one was: 16:12:07 slaweq to check if failing test_ha_router_namespace_has_ipv6_forwarding_disabled is related to bug https://bugs.launchpad.net/neutron/+bug/1798475 16:12:07 Launchpad bug 1798475 in neutron "Fullstack test test_ha_router_restart_agents_no_packet_lost failing" [High,Confirmed] 16:12:26 and I totaly forgot about this one as I didn't create card on my trello for it :/ 16:12:28 sorry for that 16:12:39 I will assign it to myself for next week then 16:12:46 #action slaweq to check if failing test_ha_router_namespace_has_ipv6_forwarding_disabled is related to bug https://bugs.launchpad.net/neutron/+bug/1798475 16:13:07 ok, next one 16:13:10 slaweq to increase neutron-tempest jobs timeouts 16:13:24 patch is merged already https://review.openstack.org/#/c/612809/ 16:14:17 did You saw any new failures because of timeouts in those jobs in last few days? 16:14:51 I haven't seen anything going into TIMED_OUT state at least 16:15:20 in general things look healthier 16:15:45 njohnston: it wasn't TIMED_OUT, it was usually FAILED 16:15:56 but in job-output.txt.gz there was info about timeout 16:16:21 OK, I haven't seen any of those lately but I haven't tried looking systematically 16:19:12 I don't see anything like that in logstash in last few days so should be better IMO 16:19:23 and lets just check as njohnston said :) 16:19:44 ok, so that's all from last week 16:19:57 #topic Python 3 16:20:12 lets talk about switch CI jobs to python3 16:20:26 njohnston: I think You are most up to date with it 16:20:51 I've been working on the change to the neutron-fu;;stack job we discussed previously: https://review.openstack.org/604749 16:21:52 my goal for this week is to start going through the jobs and adding python3, either by changing their ancestor zuul template or just by adding USE_PYTHON3 depending 16:21:54 * bcafarel actually reading through it, as zuul seems happy now 16:22:15 I noticed that there are differences in the zuul templates - for example if you base things off of "tempest-full" you 16:22:25 ok, but didn't we agree in Denver that we should just switch all jobs to py3 and left only UT and functional with py27 too? 16:23:04 sorry by "adding python3" I meant adding it to existing jobs to convert them to py3 jobs 16:23:14 ok :) 16:23:25 basically what I think is that we should do: 16:23:31 1. etherpad to track progress 16:23:56 So 'tempest-full' will get more tests run than if you base off of 'tempest-full-py3' because the latter has things like object store turned off, which disables object tests - https://git.openstack.org/cgit/openstack/tempest/tree/.zuul.yaml#n114 16:24:01 2. rename existing neutron-functional job to neutron-functional-python27 and switch neutron-functional to be py3 16:24:16 slaweq: Will do, I'll get an etherpad together today 16:24:19 3. Start switching other jobs to py3 as we decided 16:24:29 what do You think about it mlavalle and njohnston ? 16:24:44 sounds like a good plan 16:25:21 and IMO we shouldn't do python35/36/37 jobs as it may depends on OS on which job is running 16:25:47 so we should just have our jobs like e.g. neutron-fullstack to run using python3 16:26:04 yes 16:26:19 mlavalle: ok for You? 16:26:21 we may have unit tests for the subversions like in Zane's proposal 16:26:27 but not for all the flavors of testing 16:26:41 yes 16:26:53 yes, UT are fine and we already have different versions but we shouldn't do it for all other jobs 16:27:00 njohnston: when you say Zane's proposal, you refer to the messsage in the ML? 16:27:50 it's now a governance change: https://review.openstack.org/#/c/613145/ 16:28:15 I encourage everyone to read it and give feedback 16:28:18 "Resolution on keeping up with Python 3 releases" 16:29:22 yeah, about the same thing we stated in the message 16:29:39 so, yes, overall, I'm in agreement 16:30:03 great :) njohnston will You do etherpad to track it? 16:31:46 already started 16:31:49 thx 16:32:14 and do You want to propose this change for functional tests? or do You want me to do it? 16:32:51 I'll do it 16:32:57 thx 16:33:17 #action njohnston rename existing neutron-functional job to neutron-functional-python27 and switch neutron-functional to be py3 16:33:18 #action njohnston to create new neutron-functional-python27 job and switch existing one to python3 16:33:25 #undo 16:33:47 I think that undid mine... we'll have to see :-) 16:33:54 ahh, sorry :) 16:33:56 #action njohnston make py3 etherpad 16:34:07 ok, I will remember only one of them ;) 16:34:47 so next week we will be able to check how it's going and continue this work 16:35:04 do You have anything else related to python3 to talk about? 16:35:23 nope, thanks 16:35:32 ok, so lets move on 16:35:34 next topic 16:35:41 #topic Ubuntu Bionic in CI jobs 16:36:13 as I said, etherpad for it is created https://etherpad.openstack.org/p/neutron_ci_on_bionic 16:36:49 today I sent patch for neutron-tempest-plugin: https://review.openstack.org/#/c/614216/ to switch our jobs to Bionic 16:36:58 it's for now DNM patch as I want to check how it will be 16:37:06 ahh, and one most important thing 16:37:29 there is devstack patch https://review.openstack.org/#/c/610977/ which adds Bionic nodesets which we can use 16:38:02 in etherpad I listed what nodeset should be replaced with what new one (but it may not be perfect yet) 16:38:35 I also did some DNM patch https://review.openstack.org/#/c/610997/ 16:39:02 it looks that for many jobs we are fine and they works good on Bionic 16:39:22 there are some small issues with fullstack jobs for example but it's not something really big 16:40:04 and now the question is: how You want to perform that switch? should we first switch all jobs to python3 and then to Bionic? or do it "in parallel"? 16:40:12 or Bionic first and then python3? 16:40:18 any thoughts? 16:40:51 I like keeping things simple 16:40:58 one change at a time 16:41:14 agreed 16:41:33 if possible, let's convert a job to python3 16:41:37 by the way the problematic part for fullstack - where ovs is compiled - can be discarded when we move to bionic 16:41:45 and then we move that to bionic 16:42:16 that was also my idea :) lets move to python3 first and then start switching to Bionic 16:42:31 do we know if grenade is ready for py3 yet? 16:42:44 no clue 16:42:45 I don't know 16:43:14 #action njohnston check if grenade is ready for py3 16:43:21 thx njohnston :) 16:43:41 I mean we'll discover it if our changes fail for grenade jobs, but it'd be nice to find out the plan 16:44:24 njohnston: I agree 16:44:48 ok, so I guess we can move on to next topic now 16:44:51 #topic Grafana 16:44:59 #link http://grafana.openstack.org/dashboard/db/neutron-failure-rate 16:46:06 fullstack is about 30% of failures now 16:46:20 but it is related to this issue with stopping processes IMHO 16:46:38 https://bugs.launchpad.net/neutron/+bug/1798472 16:46:39 Launchpad bug 1798472 in neutron "Fullstack tests fails because process is not killed properly" [High,Confirmed] 16:46:45 and to https://bugs.launchpad.net/neutron/+bug/1798475 16:46:45 Launchpad bug 1798475 in neutron "Fullstack test test_ha_router_restart_agents_no_packet_lost failing" [High,Confirmed] 16:47:21 and both aren't assignment yet 16:47:41 I will try to check at least one of them if I will have some time this week 16:48:28 in functional tests db_migration test is happening less often I think 16:48:28 I can try to help with the other one 16:48:34 thx mlavalle 16:48:36 contingent on time availaility 16:48:44 which one do you want slaweq? 16:49:02 so I will take https://bugs.launchpad.net/neutron/+bug/1798472 16:49:02 Launchpad bug 1798472 in neutron "Fullstack tests fails because process is not killed properly" [High,Confirmed] 16:49:28 #action slaweq to check Fullstack tests fails because process is not killed properly (bug 1798472) 16:49:44 ok 16:50:22 #action mlavalle to check bug 1798475 16:50:22 bug 1798475 in neutron "Fullstack test test_ha_router_restart_agents_no_packet_lost failing" [High,Confirmed] https://launchpad.net/bugs/1798475 16:51:10 getting back to functional tests, db migration tests are less often after I increased timeout for them to 600 seconds, but unfortunatelly I saw it at lest once that it happend event with such timeout: * http://logs.openstack.org/31/613231/1/gate/neutron-functional/441128f/job-output.txt.gz#_2018-10-25_18_54_53_992690 16:51:39 so I think that it mayb be something different than only slow node 16:52:28 let's observe it for few more days and if You will spot it again, please add it to bug report: https://bugs.launchpad.net/neutron/+bug/1687027 16:52:28 Launchpad bug 1687027 in neutron "test_walk_versions tests fail with "IndexError: tuple index out of range" after timeout" [High,Confirmed] - Assigned to Slawek Kaplonski (slaweq) 16:54:15 ack 16:54:32 and that's all from fullstack/functional tests 16:54:39 anything else You want to add? 16:54:46 no, thanks 16:55:05 ok, so let's now talk about 16:55:07 #topic Periodic 16:55:37 from grafana I see that openstack-tox-py35-with-oslo-master is failing constantly since few days 16:55:55 example from today: http://logs.openstack.org/periodic/git.openstack.org/openstack/neutron/master/openstack-tox-py35-with-oslo-master/7ae4398/testr_results.html.gz 16:57:01 anyone wants to fix that? 16:57:07 or I should take it? 16:57:17 I don't have time this week 16:57:26 ok, I will check it 16:57:43 #action slaweq to check issue with openstack-tox-py35-with-oslo-master periodic job 16:58:11 ok, so that's all from my side for today 16:58:26 anything else You want to add/ask maybe? 16:58:32 not from me 16:59:00 ok 16:59:04 thx for attending 16:59:07 o/ 16:59:09 #endmeeting