17:00:14 #startmeeting qa 17:00:15 Meeting started Thu Oct 3 17:00:14 2013 UTC and is due to finish in 60 minutes. The chair is sdague. Information about MeetBot at http://wiki.debian.org/MeetBot. 17:00:16 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 17:00:18 The meeting name has been set to 'qa' 17:00:23 ok, who's around for the qa meeting? 17:00:28 Hi 17:00:36 hi 17:00:39 Hi 17:00:50 hi 17:00:55 #link Agenda - https://wiki.openstack.org/wiki/Meetings/QATeamMeeting#Proposed_Agenda_for_October_3_2013 17:01:15 hi 17:01:40 ok, lets get started on blueprint cleanup 17:01:50 #topic Blueprints (sdague) 17:02:01 hi 17:02:11 so, I started going through and closing out blueprints that I think are done this morning 17:02:21 https://blueprints.launchpad.net/tempest 17:02:22 sdague: I have closed some 17:02:36 however, we have a ton in Unknown state 17:02:50 sdague: anything pending more than 6 months and with out any action should be closed 17:02:52 so it would be really good if people could update those 17:02:56 agreed 17:03:23 sdague: I think a lot of them are Started based on patches/reviews 17:03:26 mlavalle: is this still "Not Started" - https://blueprints.launchpad.net/tempest/+spec/quantum-basic-api 17:03:43 sdague: Not started 17:04:10 ok, that I think we still want, what's your target in icehouse for it? 17:04:37 sdague: icehouse 2 17:04:37 hi 17:05:10 sdague, I'm not in the launchpad group to be able to work on the blueprints 17:05:20 giulivo: ok, I can fix that 17:05:23 sdague: I want to tackle first the neutron job thing 17:05:25 will do after 17:05:36 mlavalle: that's fine, I marked it medium / icehouse-2 17:05:45 cool 17:05:57 #action sdague to put giulivo into tempest-drivers 17:06:11 ok, other updates on blueprints there? 17:06:28 do we know about stevebaker's bp? 17:06:30 I think there are some duplicates about the neutron APIs 17:06:50 I reviewed quite a few changes actually implementing those which weren't using the appropriate topic 17:07:15 so the blueprints would need some love around that, I'd be happy to do that 17:07:24 giulivo: ok, great, if you could collapse some of those, that would be great 17:07:45 in future I'd like to not use so many small blueprints, as I think they carry a lot more weight to manage 17:07:54 sdague: +1 17:08:17 honestly, I'll probably see how ttx's storyboard project is progressing at icehouse, and see if it would be good to make tempest an early user of it 17:08:24 as a replacement for blueprints 17:08:34 at icehouse summit that is 17:08:59 the inability for us to target blueprints to multiple projects (like tempest + core project) is really annoying 17:09:26 sdague: normally you can get around that with dependecies though 17:09:26 ok, I'll send out an email today to the list giving people a week to update items in Unknown state, and then plan to script purge them 17:09:39 mtreinish: you can... but it gets a little clunky 17:09:46 yes it is 17:10:00 alright, any other blueprint discussions? 17:10:23 #topic Neutron job status (sdague) 17:10:25 sdague: I was going to do some work on the log errors blueprint Friday but don't want to conflict with any one. 17:10:43 dkranz: honestly, I won't be able to touch if for a bit, so go ahead 17:11:14 sdague: OK, you had some ideas already so perhaps you can tell me out-of-meeting, or I coujld just proceed. 17:11:14 ok, on the neutron jobs, mlavalle you have an update on where we are with the full job? 17:11:27 dkranz: lets chat in irc after the meeting 17:11:31 sdague: k 17:11:41 sdague: I've been running it. We still have 14 modules with failures 17:12:05 sdague: a couple of them are related to quotas 17:12:15 mlavalle: is that more or less than before? 17:12:15 mlavalle: ok, is it in a state where we should turn it back on non voting so it's easier to work the issues? 17:12:44 Yes, that will make it easier for me 17:13:04 mlavalle: ok, lets do that. You want to submit the review? or we need someone else to? 17:13:15 I'll do it 17:13:29 if I need help, i'll yell 17:13:37 #action mlavalle to submit review to turn neutron-full back on non-voting 17:13:46 great 17:14:00 sdague, the bogus fd patch is holding up 17:14:07 we're now running 4 devstack/tempest runs on neutron jobs to help prevent races from sliding through 17:14:35 dims: we're not running any tenant isolation jobs atm (because I screwed up the config) 17:14:37 I guess this bug needs to be fixed in order to pass the quota tests: https://bugs.launchpad.net/neutron/+bug/1189671 17:14:38 Launchpad bug 1189671 in neutron "default quota driver not suitable for production" [High,In progress] 17:14:45 so that might come back when we get that changed again 17:14:51 salv-orlando had patches 17:14:54 sdague, will keep an eye out 17:15:09 afazekas: correct 17:15:31 ok, any other neutron things? 17:15:42 the other issue is the inerfaces extension, can we skip it until it is not fixed ? 17:16:12 if we get down to just a skip or two that we need, we can definitely propose that 17:16:42 ok, next up... 17:16:47 I possible workaround for the quota issue is to configure it by devstack 17:16:47 #topic Elastic Recheck Top issues (mtreinish) 17:17:05 sdague: so this was what the big offenders are now? 17:17:11 mtreinish: yep 17:17:32 well right now bug 1226337 is still the biggest one I think 17:17:33 Launchpad bug 1226337 in cinder "tempest.scenario.test_volume_boot_pattern.TestVolumeBootPattern 'qemu-nbd: Failed to bdrv_open'" [Critical,Fix committed] https://launchpad.net/bugs/1226337 17:18:01 followed by a few neutron ones 17:18:09 mtreinish: so do we need to reopen the cinder bug? 17:18:12 the biggest of which was avoided by stopping tenant isolation 17:18:12 sdague: http://status.openstack.org/elastic-recheck/ 17:18:16 it's in fix released 17:18:25 jog0: yeh, I was about to share that :) 17:18:50 sdague: not sure I haven't really been watching things closely enough the past day or so 17:18:54 sorry for beating you to the punch 17:18:57 I just have the aggregate numbers 17:18:57 though it occurs to me that we actually should scale the metrics based on the % chance that it fails it jobs 17:19:17 because the neutron bugs don't look as bad now that they only trip up in neutron jobs 17:19:27 it looks like it's fixed based on the graphs 17:19:51 yeah the neutron ones have a bunch of attention right now 17:19:53 sdague: yeah this isn't perfect for showing how severe but does show when something is fixed 17:20:06 mtreinish: yeh 1226337 actually does look probably fixed 17:20:21 though... the gate's been really quiet today 17:20:37 with the RCs going out 17:21:04 15 jobs up on zuul ATM which is quiet 17:21:23 yeh, the nova db patch had the gate all to itself this morning 17:21:42 ok, anything else on ER? 17:21:52 no I don't think so 17:22:07 which has already shown great value, can't wait to see how it evolves 17:22:28 I have one thing 17:22:33 ok 17:22:36 https://bugs.launchpad.net/nova/+bug/1233923 17:22:38 Launchpad bug 1233923 in neutron "FAIL: tempest.thirdparty.boto.test_ec2_instance_run.InstanceRunTest.test_run_stop_terminate_instance -- nova VM fails to go to running state when using neutron" [Undecided,Confirmed] 17:23:03 we don't have a query for that bug because don't know why what to look for in the logs to find it 17:23:28 its one of a few reasons why test_run_stop_termiante_instance can fail 17:23:53 right, this is because we don't know why the network setup is wrong? 17:24:25 well I think its a two part issue both nova and neutron 17:24:36 ok 17:24:43 but its very unclear to me whats going on so extra eyes would be useful 17:24:44 https://review.openstack.org/#/c/49439/ it might fix/workaround a lot of neutron issue, I wonder why not monky patch the mysql module 17:25:15 jog0: It is Unassigned 17:25:18 #action extra eyes asked for on https://bugs.launchpad.net/nova/+bug/1233923 17:25:20 Launchpad bug 1233923 in neutron "FAIL: tempest.thirdparty.boto.test_ec2_instance_run.InstanceRunTest.test_run_stop_terminate_instance -- nova VM fails to go to running state when using neutron" [Undecided,Confirmed] 17:25:43 although it hasn't been seen almost 24 hours, but that may be a quiet gate issue 17:25:59 even if people don't manage to figure out the whole thing, we did a pretty good job on the tenant isolation issue by lots of people adding little bits to the bug 17:26:00 jog0: I'll take a look and ping you if i can help 17:26:10 thanks mlavalle ! 17:26:32 thanks 17:26:38 ok, next topic 17:26:55 #topic Stable branch timing (sdague) 17:27:10 so this came up in IRC today, and I figured we could use a plan 17:27:41 sdague: haven't we just done it around release day in the past 17:27:43 last time we basically waited until the week before release and set the stable branch on tempest / devstack / devstack-gate that day 17:28:08 mtreinish: well in grizzly we actually did it once we broken stable/grizzly bitrot :) 17:28:15 but I think 1 week prior 17:28:24 is a good plan 17:28:25 sdague: I think that makes sense to delay as long as we can 17:28:31 dkranz: ok, sure 17:28:32 new tests are pouring in 17:28:43 dkranz: that's actually my concern 17:28:49 well this brings up the backport policy 17:28:51 master is now open for 1/2 the projects for icehouse 17:28:52 sdague: I would like to touch on the post-release issue as well , which is related 17:28:58 do we have a specific one for new tests 17:29:19 mtreinish: I think our decision was new tests could backport if people wanted to do the work 17:29:19 mtreinish: I think some folks will be maintaining havana for a while 17:29:28 so btw that is about cutting a line we don't have strict rules about what we can or can't add in the period preceding the branch, is that correct? 17:29:30 though I'd say only for the last stable 17:29:37 The question is whether they should be encouraged to add tests upstream or down 17:29:54 i.e. we stop landing grizzly tests once havana is out 17:30:00 sdague: Yes 17:30:03 sdague: that's reasonable 17:30:07 sdague, not even cherry pick? 17:30:15 giulivo: not 2 releases back 17:30:24 ah okay, missed that sorry 17:30:57 so cherry pick from icehouse to havana is okay 17:31:01 yep 17:31:03 sdague: Do you have an opinion on my question above? 17:31:17 what's the post-release issue? 17:31:38 dkranz: I think master is priority if that's what you're asking 17:31:40 sdague: Whether folks continuing to add tests to havana should be encouraged to do that upstream or down 17:31:58 right, they need to go master first 17:32:01 mtreinish: Of course they would go in trunk first 17:32:10 but about the backporing 17:32:14 backporting 17:32:28 other than that, havana backporting is fine 17:32:42 Do we want a lot of actgivity on stable/havana, or not is the question 17:33:11 dkranz: I'm not sure I have a huge opinion on that 17:33:45 sdague: ok 17:34:00 sdague: I am concerned about the review burden for backports 17:34:05 I'm completely ok with people doing havana backports. I think if we think it's becoming a distraction for other work, we can sort it out later 17:34:08 sdague: It is a tradeoff 17:34:18 sdague: OK, works for me. 17:34:18 I think this time around the amount of grizzly activity was fine 17:34:35 sdague: Yes, but I predict havana will be more. 17:34:45 FYI, I did close a blueprint this morning that someone put in for backporting quantum tests to folsom 17:34:46 sdague: But we can see. 17:34:57 sdague: :) 17:35:02 which is crazy talk :) 17:35:17 dkranz: yeh, I'd rather deal with problems when they come up 17:35:31 ok so 17:35:42 #agreed wait as long as possible to set stable branch 17:35:58 next topic... 17:36:03 #topic Design Summit Initial Planning (sdague) 17:36:17 #link https://etherpad.openstack.org/icehouse-qa-session-planning 17:36:48 so, this isn't going to be a top focus for 2 weeks, as we don't need a schedule till after the release 17:36:59 however I wanted to get people thinking about design summit 17:37:06 who all is planning on being there? 17:37:10 o/ 17:37:20 o/ 17:37:30 * afazekas me 17:37:30 o/ 17:37:34 me 17:37:52 * giulivo isn't sure 17:38:10 sdague: I think mkoderer said he was going to be there earlier (I don't think he's around now) 17:38:19 ok 17:38:44 so we'll eventually need things in summit.openstack.org 17:38:59 but like last time, I think planning for a coherent track is better in etherpad 17:39:17 so I'm starting to fill out ideas that I think we surely need to discuss 17:39:27 sdague: I think multi-node test runs upstream would be a good topic (again) but not sure if that is us or infra 17:39:50 dkranz: so I think it's actually an infra talk, because what it really is is a nodepool discussion 17:40:00 about whether node pool can allocate us a set of machines 17:40:11 sdague: Yeah, I just want it to happen :) 17:40:27 dkranz: that being said, we've talked about that at every summit I've ever been at 17:40:27 I am waiting to get my final approval for the trip 17:40:33 sdague: But it also raises the issue of "install" 17:41:00 so I'm going to be weary of taking sessions that have shown up before and gotten no progress on them 17:41:02 sdague: Which we have ignored with all-in-one only 17:41:16 dkranz: so devstack will do multi node 17:41:19 dkranz: devstack has multinode support 17:41:29 heat + devstack sounds cool :) 17:41:45 yeh with heat and nodepool I feel like we have more tools this time 17:42:01 sdague: I think it is worth a topic either here or infra 17:42:16 dkranz: sure, can you put something in the etherpad in proposed for it 17:42:22 sdague: The chance of succes if more now 17:42:25 yep 17:42:27 sdague: Yes 17:42:42 we've got 13 sessions this time, and we won't overlap with infra 17:42:46 we'll be in the same room 17:42:51 so that's good 17:43:08 personally, my top priority in HK is a solid neutron plan 17:43:15 sdague: how many max Qa sessions? 17:43:22 sdague: +1 17:43:27 and markmcclain1 and I have been talking about that already 17:43:38 that may actually be in the neutron track 17:43:40 ravikumar_hp: 13 17:44:20 I'd like to be in Neutron if everyone can attend 17:44:42 markmcclain1: yep, I think that's a good call. I was actually going to look at schedule blocks later 17:44:58 I would like to speak about the stress tests.. but it's still not sure if the trip will be approved 17:45:14 I think that's an important enough session we should do it in neutron when there is a QA slot and just drive all the QA folks into the neutron room 17:45:36 mkoderer: you have an idea when you'll know about approval? 17:45:45 for now I'll save you a slot, because I think that's important 17:45:46 sdague: So "dual session"? 17:45:53 dkranz: basically 17:45:57 sdague: hopefully next week 17:46:01 that will ensure we aren't conflicting 17:46:10 sdague: ok sounds great 17:46:10 sdague: I won't go to HK but I want to be part of the Neutron plan 17:46:20 sdague: Good idea 17:46:36 mlavalle: ok, no problem, we'll make sure to get etherpad notes up in advance 17:47:06 reminder to everyone proposing things, we want an etherpad URL to link to with outline of the discussion 17:47:25 sdague: how soon? 17:47:25 I'm not going to give final approval to anything that doesn't have a good etherpad with agenda, as we want to make the most of our time 17:47:32 mtreinish: 2 weeks 17:47:36 ok 17:47:38 sdague: Let's put the urls in a canonical place 17:47:54 I'll start evaluating on release day 17:47:57 Oct 17 17:47:59 sdague: Last time around I created them all when populating the schedule 17:48:24 dkranz: ok, well how about getting folks to link them in https://etherpad.openstack.org/icehouse-qa-session-planning 17:48:34 I'll build a template on the train tomorrow 17:48:49 sdague: sounds good 17:48:49 I'm down in NYC for most of the day, so I have a train ride to do stuff like that 17:49:05 we can revisit next week as well 17:49:39 ok, any other design summit questions / conversations? 17:50:41 ok... 17:50:46 #topic Critical Reviews 17:51:00 any critical reviews that need to get eyes on? 17:51:43 I've one not critical per-se but which doesn't pass jenkins failing because of devstack errors on the neutron VM (this is on stable/grizzly) 17:52:05 giulivo: right, I wonder how those regressed 17:52:14 because we had stable/grizzly working again 17:52:24 do we know when it started failing? 17:52:36 and what could have caused it? 17:52:40 it's the https://bugs.launchpad.net/neutron/+bug/1233264 right? 17:52:41 Launchpad bug 1233264 in python-neutronclient "stable branch patches failing in check queue due to missing 'find_resourceid_by_name_or_id'" [High,Fix released] 17:52:47 sdague: on the stable branch note there is this: https://review.openstack.org/#/c/47337/ with your -2 17:53:06 sdague, mine was pushed on Oct 3, 2013 10:53 and never got the +1 17:53:13 sorry, the https://code.launchpad.net/bugs/1234181 17:53:15 (well, not from neutron) 17:53:16 Launchpad bug 1234181 in neutron "stable/grizzly patches are failing jenkins in check-tempest-devstack-vm-neutron" [Undecided,New] 17:53:23 mtreinish: ok, I'll put that back in the check queue 17:54:02 sdague, people who could work on it I can refer to, to help? 17:54:14 #openstack-infra? 17:54:44 giulivo: honestly the #openstack-neutron channel might be better 17:55:05 got it, thanks :) 17:55:21 Has this particular devstack exit happened more than once? 17:55:33 dkranz: yes 17:55:39 dkranz, yeah it seems to be blocking actually 17:55:49 giulivo: so something I experienced is jenkins kills the script late 17:55:55 so the actual fail might be 100 lines up 17:56:01 dkranz: it's already on rechecks ... Affecting changes: 48300, 49485, 49487, 49490, 49491 17:56:03 you have a link to a fail log? 17:56:06 sdague, I see, thanks 17:56:11 sure 17:56:18 http://logs.openstack.org/91/49491/2/check/check-tempest-devstack-vm-neutron/1fbbb16/ 17:56:23 giulivo: actually, lets take that to -qa after the meeting 17:56:28 we have 4 mins left 17:56:33 anything else from folks? 17:56:56 stress tests sometimes failing in the nightly 17:57:18 I will enhance the logging to get some more information about it 17:57:46 mkoderer: great! 17:58:03 ok, I have to run to another meeting. so thanks everyone for coming 17:58:10 we'll talk more in -qa 17:58:12 ciao 17:58:13 #endmeeting