21:01:38 #startmeeting Neutron 21:01:38 aloha 21:01:39 Meeting started Mon Nov 25 21:01:38 2013 UTC and is due to finish in 60 minutes. The chair is markmcclain. Information about MeetBot at http://wiki.debian.org/MeetBot. 21:01:40 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 21:01:43 The meeting name has been set to 'neutron' 21:02:29 #link https://wiki.openstack.org/wiki/Network/Meetings 21:02:38 #topic Announcements 21:02:49 #info Icehouse-1 is December 5th 21:03:12 hi all! 21:03:19 It's next week which means code needs to merged by Tuesday Dec 3rd 21:03:26 hi 21:03:37 #link https://launchpad.net/neutron/+milestone/icehouse-1 21:03:45 hi 21:03:49 hi 21:03:59 * salv-orlando regrets the week off taken during I-1 21:04:20 the first milestone always seems to feel rushed 21:04:35 * mestery nods in agreement. 21:04:43 and even more so since many will be taking a few days off at the end of this week 21:05:24 #topic Bugs 21:05:27 anteaya: around? 21:05:53 #link https://bugs.launchpad.net/neutron/+bugs?search=Search&field.importance=Critical&field.status=New&field.status=Confirmed&field.status=Triaged&field.status=In+Progress 21:06:31 I'd add bug #1254236 to the list 21:06:33 Launchpad bug 1254236 in neutron "TestNetworkQuotaBasic tearDownClass fails while deleting port" [Undecided,New] https://launchpad.net/bugs/1254236 21:06:37 not sure if anteaya is active at the moment.. she has gone through and listed the gate blocking bugs in the agenda 21:06:44 I am going to mention really quickly that the way jenkins processes test logs after tests run has been updated to speed up the neutron unittest jobs. py26 is ~40 minutes now and py27 is ~27 minutes 21:06:49 I found out about it last week 21:06:50 it's getting hot pretty fast 21:06:55 clarkb: awesome 21:06:57 shouldn't see anymore job timeouts on those 21:07:02 woooo! 21:07:13 armax: thanks 21:07:15 cool! 21:07:17 clarkb: thanks 21:07:29 clarkb: so do you think that review https://review.openstack.org/#/c/58022/ is no longer worth it? 21:08:30 armax: correct, we should keep an eye on it but I think we are well below variability in test times 21:08:45 *well below the timeouts considering variability in test times 21:09:06 clarkb: ok sounds good 21:09:15 salv-orlando: is working on https://bugs.launchpad.net/swift/+bug/1224001 21:09:17 Launchpad bug 1224001 in neutron "test_network_basic_ops fails waiting for network to become available" [High,In progress] 21:09:33 I think I and otherwise guy merged the patches that should fix it 21:09:41 cool 21:09:43 we hope it won't show up again, but I know it will 21:09:52 (more later when we discuss parallel testing) 21:10:12 k 21:10:35 https://bugs.launchpad.net/neutron/+bug/1251448 21:10:37 Launchpad bug 1251448 in neutron "BadRequest: Multiple possible networks found, use a Network ID to be more specific. " [High,New] 21:10:43 marun: any progress? 21:11:00 mlavalle seems to have narrowed the issue to not having tenant isolation turned on 21:11:19 is there a reason why any of the gates turn it off? 21:11:41 when it was activate earlier it actually caused more gate failures 21:11:52 i thought that was parallelization 21:11:55 :\ 21:12:10 damned if we do, damned it we don't 21:12:43 yeah well as we work on improving the test situation we will uncover a few issue along the way 21:13:12 what are we considering the fundamental error there? 21:13:15 marun, markmcclain, on this note of tenant isolation I might be dumb, but the 'isolated' job we run 21:13:20 that the tearDown fails? 21:13:33 it is a cleanup issue, yeah 21:13:34 which set allow_tenant_isolation = True - aren't they using isolation? 21:13:37 beagles: yes 21:14:06 or does bug 1251448 occur only on non-isolated jobs? 21:14:08 Launchpad bug 1251448 in neutron "BadRequest: Multiple possible networks found, use a Network ID to be more specific. " [High,New] https://launchpad.net/bugs/1251448 21:14:14 sdague or jog0: ^ 21:15:03 salv-orlando: but if any job doesn't use isolation, then we have potential races in cleanup 21:15:13 it looks like it occurs on non-isolated jobs 21:15:40 armax: that was what I was looking for. If we can confirm that it does not happen with isolation, then the issue is nailed 21:15:44 so the question is, are the non-isolated jobs going to continue? 21:15:55 salv-orlando: http://logstash.openstack.org/#eyJzZWFyY2giOiJtZXNzYWdlOlwiIHBvc3NpYmxlIG5ldHdvcmtzIGZvdW5kLCB1c2UgYSBOZXR3b3JrIElEIHRvIGJlIG1vcmUgc3BlY2lmaWMuIChIVFRQIDQwMClcIiAgICAgQU5EIGZpbGVuYW1lOlwiY29uc29sZS5odG1sXCIiLCJmaWVsZHMiOltdLCJvZmZzZXQiOjAsInRpbWVmcmFtZSI6IjYwNDgwMCIsImdyYXBobW9kZSI6ImNvdW50IiwidGltZSI6eyJ1c2VyX2ludGVydmFsIjowfSwic3RhbXAiOjEzODU0MTQxMjYzNjcsIm1vZGUiOiJ0ZXJtcyIsImFuYWx5emVfZmllbGQiOiJidWlsZF9uYW1lIn0= 21:15:57 marun: should be discontinued, this is my understanding 21:16:08 yes, only "non issolated" jobs 21:16:19 so yeah, not sure we want to bother fixing. isolation is the ideal fix 21:16:34 sdague: thanks for showing me how to the query - I was hopelessly trying to do that by myself 21:16:34 marun: so we're not fully isolated on the isolated jobs right now 21:16:56 sdague: ? 21:17:20 last I knew the API tests aren't actually properly issolated, because the load pushed neutron over the edge 21:17:31 mtreinish has the details 21:17:40 sdague: ah, ok 21:17:55 so it will be necessary to ensure name uniqueness 21:18:03 no that we've activated minimize polling we might was to look and see if the load is still too high 21:18:11 s/was/want/ 21:18:15 sdague: yeah the neutron api tests don't have isolation turned on 21:18:36 last time I did that the gate for neutron failed basically 100% of the time 21:18:56 it needs a rebase but: https://review.openstack.org/#/c/53459/ 21:19:43 mtreinish, sdague, marun, markmclain and whoever else: I'm working on that too as a part of the parallel testing effort 21:19:52 cool 21:20:00 the polling switch has made a difference in runtimes at least 21:20:12 Any other critical bugs we need to discuss? 21:20:13 i'm seeing 20-25% less execution time 21:20:22 so neutron gate is failing at 33% right now 21:20:32 .gate-tempest-devstack-vm-neutron 21:21:42 jog0: good to know.. was hoping for a lower number with all of the recent changes 21:22:17 there were some outstanding patches in the nova side that should help 21:22:21 let me hunt those down 21:22:41 me and arosen pinged the nova room earlier with those 21:22:42 https://review.openstack.org/#/c/57711/ , https://review.openstack.org/#/c/57069/6 21:23:00 jog0: ^ 21:23:03 arosen: thanks thats it 21:23:12 arosen: I will review those 21:23:18 jog0: thanks! 21:23:22 jog0: awesome thanks 21:23:38 Any other critical bugs the team needs to track? 21:23:53 #topic Docs 21:24:01 https://bugs.launchpad.net/neutron/+bug/1250168 https://bugs.launchpad.net/neutron/+bug/1251448 https://bugs.launchpad.net/neutron/+bug/1235435 21:24:04 Launchpad bug 1250168 in neutron "gate-tempest-devstack-vm-neutron-large-ops is failing" [Critical,Confirmed] 21:24:07 #undo 21:24:07 Removing item from minutes: 21:24:14 are all gate bugs, none are listed as critical ATM 21:24:28 not sure where the root causes of those bugs are though 21:25:06 Partial fix for bug 1235435 is review 57069 21:25:07 Launchpad bug 1235435 in nova "'SubnetInUse: Unable to complete operation on subnet UUID. One or more ports have an IP allocation from this subnet.'" [Medium,In progress] https://launchpad.net/bugs/1235435 21:25:20 one of the two that aaron pointed you at 21:25:32 s/aaron/arosen 21:26:06 bug 1235435 is because isolations isn't turned on 21:26:08 Launchpad bug 1235435 in nova "'SubnetInUse: Unable to complete operation on subnet UUID. One or more ports have an IP allocation from this subnet.'" [Medium,In progress] https://launchpad.net/bugs/1235435 21:26:15 salv-orlando: wanted to talk about parallel testing 21:26:17 armax: cool thanks 21:26:30 work may help solve that issue 21:26:33 markmcclain: I'm fine either way if you want to move it oto open discussion 21:26:43 and continue with the normal agenda 21:26:49 markmcclain: we just saw bug 435 or the query we have is bad https://review.openstack.org/#/c/58352/ 21:26:50 Launchpad bug 435 in ethereal "ethereal uses gksu instead of gksudo for root menuitem" [Low,Invalid] https://launchpad.net/bugs/435 21:26:57 anyway thanks 21:26:57 markmclain: not in all cases 21:27:11 but we can talk about it offline if you want to learn more 21:27:31 ok.. we can discuss the actually bug in teh room after the meeting, lets keep moving so that we don't run out of time 21:27:48 #topic Docs 21:27:55 hi 21:28:11 Metering docs have merged.. thanks to sfchain 21:28:22 emagana: anything else to add that's not on the agenda? 21:28:28 yes! 21:28:56 just a head ups about a Doc Bug Day 21:29:19 #info https://wiki.openstack.org/wiki/Documentation/BugDay Dec 20th 21:29:43 It could be great to have some neutrons participating 21:29:53 I also added a section about the ml2 scenarios, with to start 2 scenarios with the l2 population mechanism driver 21:30:16 emagana: fine but I really don't feel comfortable around electrons 21:30:39 emagana: thanks for the reminder about the doc bug day 21:30:46 salv-orlando: heh 21:30:47 salv-orlando: keep neutral.. I mean neutron 21:31:02 nothing else markmcclain 21:31:07 thanks for the update 21:31:10 emagana: so we don't have network bugs open against either the admin-guide or netconn-api? 21:31:11 #topic Nova Parity 21:31:16 #undo 21:31:17 Removing item from minutes: 21:31:33 Yes, we do! 21:32:01 I did not believe we want to discuss all of them, at least we have a high priority ones 21:32:14 as we had with metring part 21:32:24 s/metring/metering 21:32:46 just wanted to keep it short, but we can go over more details if needed, advise! 21:33:10 we still have a lot of material to cover, so let's keep moving for now 21:33:15 #topic Nova Parity 21:33:51 somebody is playing my song 21:34:00 yep 21:34:36 I was on PTO last week so I apologize for my silence... however before I left someone proposed creating a draft list of milestones related to parity 21:34:54 yes will be helpful to track progress 21:35:00 I've posted a link to the ethepad in the agenda... it was a complete off the cuff affair,but it fits the schedule (I think) 21:35:03 agreed 21:35:04 and we can fill it in with action items 21:35:16 #link https://etherpad.openstack.org/p/icehouse-neutron-nova-parity-prelim-outline 21:35:17 I'd appreciate suggestions and refinements 21:35:37 thanks a lot beagles 21:36:01 my pleasure: I've mentioned some things in there that warrant conversations offline 21:36:07 do you reckong we are on schedule with your plane? 21:36:10 or out-of-the-meeting 21:36:11 plane/plan 21:36:22 we are a little behind but not dire I think 21:36:48 ok.. we can discuss offline farm out the work if necessary to get us back on track 21:37:26 Anything else for this week? 21:37:46 ipv6? 21:38:23 sc68cal: yeah we'll have to look at gaps there too 21:38:29 #topic Tempest 21:39:02 Lots of updates to the etherpad 21:39:04 #link https://etherpad.openstack.org/p/icehouse-summit-qa-neutron 21:39:22 mlavalle or salv-orlando who wants to update first? 21:39:43 I'd let mlavalle go first - he's the topic lead, I think 21:39:59 * salv-orlando at least that's what the wiki page says :) 21:40:02 markmcclain: as indicated in the ether pad, I have taken responsibility for the api testing 21:40:19 ok 21:40:40 my plan is to have a plan by this week, so we can let other contrabute to this effort 21:41:13 great… I'll follow up with you later in the week 21:41:18 I also worked over the weekend on https://bugs.launchpad.net/neutron/+bug/1251448 21:41:20 Launchpad bug 1251448 in neutron "BadRequest: Multiple possible networks found, use a Network ID to be more specific. " [High,New] 21:41:42 I've updated the but in launchpad with my analysis 21:41:48 great thanks 21:42:01 it seem to me it's a race condition 21:42:18 when deleting ports 21:42:36 yeah the test changes have definitely revealed several races 21:42:53 so by the time the tea tries to delete a subnet, it finds ports still active 21:43:11 so a net and a subnet are left alive in the demo user 21:43:21 that cascades to the other tests after that 21:43:42 yeah.. that will do it 21:44:00 tonight I will attempt to create a fix….. i might ask a lot of questions tomorrow 21:44:07 mlavalle: sounds great 21:44:12 mlavalle: I'll be around 21:44:16 that's all I have today 21:44:22 mlavalle: thanks 21:44:23 in terms of "items" pertaining to testing 21:44:44 on the etherpad we have at least three we need to discuss here: parallel, full, and grenade testing 21:45:08 as I do things without any logic, let's start from the bottom 21:45:13 ok 21:45:28 grenade - the issue is simple, there's no grenade testing, and we're looking for an owner 21:45:41 last week I said "we have no owner" and somebody said "hold on" 21:45:46 does this mean we have an owner now? 21:46:23 I don't think we've had anyone step up 21:46:30 silence == no owner, I guess. markmcclain, can we prioritise finding a volunteer? 21:46:38 nobody wants to dive on the grenade? 21:46:44 cue crickets 21:46:50 #action markmcclain to find grenade lead 21:47:03 I work on that this week 21:47:04 full neutron devstack test 21:47:17 * beagles nods 21:47:20 Does grenade piece also include the OVS -> ML2 migrations? 21:47:21 This work item is a mix of issues with neutron and features missing from neutron. 21:47:22 markmcclain: please coordinate with jlibosva 21:47:32 marun: will do 21:47:34 not sure if that's out of scope for grenade 21:47:38 At the moment we have rossella_s performing an assessment 21:48:00 sc68cal: the ml2 team will need to add migration scripts 21:48:02 she's not online, but as she's in my timezone I will ping her tomorrow 21:48:04 salv-orlando, there is an overlap with the parity related activities outlined for this week 21:48:34 beagles can you coordinate with rossella_s? 21:48:37 sc68cal: Yes, we're aware of the migration requirements from OVS to ML2, but have no specific plans at the moment. 21:48:38 beagles: I am aware of that; I thought I was good to get someone started anyway on the assessment of what needs to be done to get that test passing 21:48:42 absolutely :) 21:48:45 salv-orlando: a couple of months ago I put together (with the help of others) this https://etherpad.openstack.org/p/gate-tempest-devstack-vm-quantum-full 21:48:47 thanks 21:48:53 salv-orlando, no argument there :) 21:49:06 mlavalle: thanks for that I will add it to other etherpad 21:49:13 and finally… parallel testing 21:49:38 salv-oralando: rosella_s can ping me if clarification is needed 21:49:49 salv-orlando: you've been digging into this one right? 21:49:51 mlavalle: I will ell there 21:50:08 Basically there are 3 issues, 1 in devstack/tempest, and 2 in neutron 21:51:05 devstack/tempest give neutron only 16 public IPs. this is fine for nova-network, but neutron uses a public IP for each tenant router. If you add the fact that some tests do not cleanup routers, you can easily run out of IPs 21:51:06 salv-orlando: i could help with grenade, even if i have never contributed yet 21:51:24 armax has a patch for 2 test cases leaving a router behind 21:51:36 I have a patch proposing to start defaulting the public network to /27 21:51:53 any questions before we move to the neutron side? 21:52:10 we have 8 mins, so I think we have to move on 21:52:15 salv-orlando: I think the number of ip's is configurable in the tempest conf file 21:52:41 salv-orlando: the patches you're talking about are: https://review.openstack.org/#/c/57805/ and https://review.openstack.org/#/c/58054/ 21:52:43 mlavalle: this is about the public network created by devstack I think, but we can take this detail offline I think 21:52:52 cool 21:53:06 SpamapS: hey - whats the bug number about that race with new networks? 21:53:12 * salv-orlando is funny to say offline while you'll still be online talking about this 21:53:17 SpamapS: thinking it would be topical to raise it here as something tempest should be testing 21:53:55 I think we're out of time for the tempest discussion - I will update on neutron issues separately, unless you feel it's important 21:54:26 yeah.. I think we discuss on this on the mailing list 21:54:42 * markmcclain fingers won't work today 21:55:08 #topic Subteams 21:55:31 We've got several new subteams in addition to our existing ones. 21:55:43 Please check the agenda for updates 21:56:05 #topic Open Discussion 21:56:28 Any items need to cover in 4 minutes or less? 21:57:48 Is lbbas api still experimental? 21:58:09 no with Havana that api will need to follow the deprecation cycle 21:58:31 markmcclain: Thanks 21:58:45 it looks there is a patch which change beheivior 21:58:49 of LBaaS API 21:59:01 link? 21:59:13 https://review.openstack.org/#/c/53821/ 21:59:47 health monitor was cascate deleted before 21:59:56 This patch changes that beheivior 22:00:06 thanks for the link.. I'll follow up with teh sub-team 22:00:12 markmcclain: Thanks 22:00:19 we're out of time for this week 22:00:24 I want to thank salv-orlando, armax, arosen, jog0, marun, anteaya for digging into the gate problems 22:00:40 markmcclain: thank you 22:00:50 to everyone fixing them 22:00:58 I hope I just don't add more ;) 22:01:05 ditto :) 22:01:10 Have a great week and for everyone traveling for the holidays later this week.. safe travels 22:01:12 #endmeeting