14:02:22 #startmeeting nova 14:02:23 Meeting started Thu Feb 6 14:02:22 2014 UTC and is due to finish in 60 minutes. The chair is russellb. Information about MeetBot at http://wiki.debian.org/MeetBot. 14:02:24 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 14:02:27 The meeting name has been set to 'nova' 14:02:33 hi 14:02:40 \o 14:02:42 hi everyone! 14:02:47 Hi 14:02:47 o/ 14:02:57 #topic general 14:03:05 first, the schedule 14:03:12 #link https://wiki.openstack.org/wiki/Icehouse_Release_Schedule 14:03:23 rkukura, #openstack-meeting is open now 14:03:32 blueprint approval deadline was this week 14:03:43 code for blueprints must be posted in 2 weeks from today 14:03:54 i3 quickly closing in on us, so be aware :) 14:04:08 other big thing ... nova meetup next week 14:04:15 #link https://wiki.openstack.org/wiki/Nova/IcehouseCycleMeetup 14:04:29 sounds good 14:04:31 if you have any questions, let me know 14:04:42 we've been collecting topics on https://etherpad.openstack.org/p/nova-icehouse-mid-cycle-meetup-items 14:05:02 i'm going to start organizing that into a very rough cut at schedule for some of the big things 14:05:02 anyone know what time we start on Mon? 14:05:06 n0ano: 9am 14:05:14 too civilized :-) 14:05:26 i'm not a great morning person 14:05:50 i want to tag specific big topics to cover in the mornings, at least 14:06:10 like, tasks ... want to see if we can bang through review/approval of design, at least 14:06:17 and finalizing plans for taking v3 to final 14:06:32 neutron interaction is another big one 14:06:35 +1 for task discussion 14:06:42 process discussion would be another 14:06:50 so if you have any topic requests, please record them 14:07:09 hi 14:07:12 russellb, in the etherpad? 14:07:17 ndipanov: yeah etherpad 14:07:33 we can discuss more in open discussion if we have time 14:07:36 #topic sub-teams 14:07:40 \0 14:07:45 * n0ano gantt 14:07:53 hartsocks: you're up 14:07:56 * johnthetubaguy raises xenapi hand 14:08:04 #link https://wiki.openstack.org/wiki/NovaVMware/Minesweeper 14:08:32 Just a quick note there. The CI guys are putting notes on that Wiki. Some status update info is linked on that page as well. 14:08:50 cool 14:08:56 do you have your tempest config published? 14:09:00 what tests you exclude, if any? 14:09:23 We had some infra trouble earlier in the week. And, we'll link that info and info about configs, excludes etc. on that wiki. 14:09:27 features that are not supported are not tested - for example security groups 14:09:41 * russellb nods 14:09:42 hartsocks: gary: i'd like to see more vmware people scoring on this before the rest of us are approving it: https://review.openstack.org/#/c/70137/ 14:09:54 just would be good to have that reference, i think all the driver CI systems should 14:10:05 mriedem: sure 14:10:14 it kind of blew up 14:10:18 and it's blocking your other fixees 14:10:31 mriedem: what do you mean it blew up? 14:10:38 we are going to go through a set of reviews as a team. 14:10:43 just got bigger than the oriinal couple of patch sets 14:10:50 lots of back and forth between vmware people it seemed 14:11:01 so i'm waiting for that to settle before getting back on it 14:11:03 mriedem: correct - it was a result of comments on the patch sets 14:11:10 yeah, i know 14:11:10 mriedem: understood 14:11:43 if possible i'd lile to talk about the deferral of the image - caching 14:12:07 i sent a mail to the list. not sure if you guys have seen it. not sure whu it was pushed out of I? the code has been reasy since december 14:12:29 garyk: blueprint wasn't approved I guess? 14:12:37 johnthetubaguy: bp was apporved 14:12:49 if it was approved, then i shouldn't have deferred it 14:12:50 link? 14:12:57 i didn't think it was 14:13:01 it was then pending a BP that was partially implemented and the proposer of the BP went missing 14:13:09 sec 14:13:26 https://blueprints.launchpad.net/openstack/?searchtext=vmware-image-cache-management 14:13:40 not approved 14:13:42 "Review" 14:14:04 it was approved then move due to https://blueprints.launchpad.net/nova/+spec/multiple-image-cache-handlers which is completely unrelated 14:14:44 so what do we need to move from review to approved? this is a parity feature. as part i implemned the generic aging code 14:15:05 I think I stated it need more info for the docs team, or something like that? 14:15:17 the reason i put it in "Review" originally was i wanted the plan to be around making existing code generic where possible 14:15:23 and i never got back to review it again 14:15:35 looks like johnthetubaguy also wanted docs added 14:15:51 but perhaps more as a note before code merges 14:15:57 russellb: thanks for the clarification. basically it is exactly the same as libvirts aging. no new config var. 14:16:10 sharing some code now? 14:16:19 where it makes sense? 14:16:22 can you please clarify what i need to add? 14:16:46 or did it just end up being sharing config? 14:16:47 the code that is possible to share is shared. can you please look at the mail on the list. i guess we can take it from there 14:17:06 heh ok if you wish, i mean you have my attention now 14:17:18 anyway it's probably fine 14:17:25 the config is the same as libvirts. the aging is just implemented in the driver - based on the generic code 14:17:31 ok. 14:17:36 anything else on vmware? 14:17:49 We could go on about our BP? :-) 14:17:50 i do not want to take too much of the meeting time and i think i interupted theminesweepd update 14:18:07 hartsocks: open discussion if you'd like 14:18:12 johnthetubaguy: let's talk xenapi 14:18:18 johnthetubaguy: i'm very interested to check in on CI status ... 14:18:30 I'll talk to that. CI is annoyingly close. 14:18:39 BobBall: go for it 14:18:55 Our initial efforts were on getting something integrated with the -infra stuff 14:19:07 the key thing for me, is that we are putting for an external system now, since -infra stuff has stalled I guess? 14:19:16 and we've got a bunch of patches up to enable that - but getting review time on that has been tricky 14:19:44 For the last two? weeks we've been focused on setting up our own 3rd party system rather than using -infra 14:19:46 yeah they've been pretty slammed 14:19:48 which is annoying close 14:20:03 Got nodepool running and deploying xenserver nodes and another script that runs the tests on them 14:20:06 any timeframe for turning it on? 14:20:08 joining the dots ATM 14:20:23 I'f you'd asked me at the beginning of this week I would have said "Friday" ;) 14:20:46 So it really could be any day now. 14:20:46 how about confidence level that you'll have it running by icehouse-3? 14:20:54 High 14:20:58 ok, great 14:21:12 we have tempest running outside of the automation right? 14:21:23 and seen it run in VMs that nodepool should be creating 14:21:35 so it really is just joining the dots now, we hope 14:21:36 We have full tempest running + passing in VMs in RAX using the -infra scripts 14:21:52 yes, indeed. 14:21:52 yeah, so almost, but no cigar 14:22:03 thats all from xenapi 14:22:06 ok, well cutting it close 14:22:09 but sounds like good progress 14:22:19 Too close, yes :/ 14:22:37 while we're still on driver CI ... hyper-v CI caught a bug yesterday, so that was cool to see 14:22:37 yeah, aimed for gate integrated, and that gamble didn't quite work this time, but at least thats much closer for Juno 14:22:47 johnthetubaguy: great goal though 14:23:04 johnthetubaguy: bet you can get more attention early cycle 14:23:05 yeah, should be awesome soon :) 14:23:16 So, yes, we distracted ourselves from integration to get the 3rd party stuff working so we don't miss the Icehouse deadline. 14:23:26 sounds like a good plan 14:23:42 n0ano: alright, gantt / scheduler stuff 14:23:51 tnx, couple of things... 14:23:53 n0ano: i saw your mail about devstack support, i assumed that means devstack support merged? 14:24:18 russellb, yep, it's there and I sent an email out on how to configure it (with a typo that just got corrected) 14:24:25 great 14:24:31 couple of things from the meeting... 14:25:11 no_db scheduler - still working on it, turn out removing the compute_node table from the DB is harder than originally thought but they'll get it eventually 14:25:32 it's starting to feel a bit risky for icehouse anyway 14:25:41 not sure how others feel about it ... 14:25:59 but massive architectural change to critical component better early in the cycle than toward the end 14:25:59 for icehouse - as an option maybe, not as a default I hope 14:26:06 ok, perhaps 14:26:14 guess i was assuming we'd do it as *the* way it works 14:26:24 what about a little caching, that gets no db during a user call to the scheduler? 14:26:36 johnthetubaguy: your bp? 14:26:40 * johnthetubaguy feels dirty for plugging his patch 14:26:43 yeah... 14:26:51 johnthetubaguy: i was going to say, i think that got deferred, but based on your description, i'd be happy to sponsor it 14:26:55 https://review.openstack.org/#/c/67855/ 14:27:03 johnthetubaguy: because your approach sounds simple / low risk 14:27:15 I think the current plan is a a complete cut to the no_db way, but if you think there's a way to stepwise develop that would be good 14:27:17 yeah, that one is approved, for a single host solution 14:27:24 it just races with multiple hosts 14:27:32 i have a few issues with that one when the scheduling deals with data that is not homgenous, but we can take it offline 14:27:34 n0ano: i'm not sure there's a sane way to make it optional 14:27:56 we need to find one, I had a rough idea, but now is not the time 14:27:58 russellb, note I said `if', I not sure there's a way either. 14:27:58 johnthetubaguy: ah, you approved it yourself :-p 14:28:08 basically status update from compute, and stats fetch driver 14:28:27 russellb: yeah, probably... 14:28:39 that's fine, in this case anyway, heh 14:29:29 i guess we can leave the full no_db thing targeted for now 14:29:39 but the closer we get, the less comfortable i will be with it merging 14:29:44 we'll see 14:29:47 anyway, we also had a bit of a debate about scheduler request rate vs. compute node update rate, I don't expect that to be resolved until we have a working no_db scheduler with performance numbers. 14:30:02 yes, performance numbers are key 14:30:29 code forklift - devstack changes merged in, still working on getting the pass the unit test changes merged 14:30:30 the debate was about how those things scale 14:30:32 yeah, performance testing show the complex cache didn't work, so I stripped it down to what is up for review now 14:30:48 * n0ano fighting gerrit gets interesting 14:30:56 so on the forklift, couple of things 14:31:06 1) i responded to your devstack message with some next steps on getting gate testing running 14:31:18 i'd love to see an experimental job against nova that runs tempest with gantt instead of nova-scheduler 14:31:28 it's not actually that much work, just the right tweaks in the right places 14:31:35 saw that, after the unit test patches are in I intend to work on that 14:31:42 OK great 14:31:48 ok, the unit test patches .... 14:31:57 my only issue was whether or not we should be including scheduler/rpcapi.py 14:32:04 longer term, that needs to be in its own repo 14:32:24 we also need to have the scheduler to use objects 14:32:26 we could merge it now, and do that later, but that makes the change history against the current gantt more complicated to replay on regenerating the repo later 14:32:35 russellb, yeah, I think I dealt with that this morning but, due to gate testing, I had to add that file in the first patch and then delete it in the second 14:32:57 if the scheduler uses objects then the foklift will be a lot easier in the long run and it will be backward compatiable with nova 14:33:06 it will help with the database parts 14:33:26 anyone working on that? 14:33:34 garyk we're just doing what the current scheduler does 14:33:39 objects? 14:33:48 i started. initial patches git blocked 14:33:59 i need to go back and speak with dan smith about those 14:34:23 n0ano: ok i'll review again today to see what you updatec 14:34:53 as an aside, I guess the scheduler rpcapi and cells rpcapi both need to become public APIs, similar issues 14:35:06 why cells? 14:35:06 btw, I'm not too worried about regenerating, now that I've done it once the second time should be easier, even if the history is a little convoluted 14:35:23 in the scheduler case, it's just the public API to gantt 14:36:21 cells, we wanted to make that public, so people can plug in there too 14:36:38 ah, i see ... 14:36:50 i think the public part would be the manager side, right? 14:36:58 but lets ignore that comment for now… its irrelevant really 14:37:02 ok :) 14:37:11 related: i want to talk about the future of cells next week 14:37:11 yes, possibly 14:37:29 #topic bugs 14:37:34 bug day tomorrow! 14:37:36 #link http://lists.openstack.org/pipermail/openstack-dev/2014-February/026320.html 14:37:49 thanks johnthetubaguy for putting message together 14:37:57 np 14:37:58 shall we just collaborate in #openstack-nova? 14:38:03 I think that works best 14:38:11 people spot its happening that way 14:38:11 agreed 14:38:45 196 new bugs 14:38:59 well, thats under the 200 we had before 14:39:09 yeah, someone(s) have been working on it apparently :) 14:39:18 email works, who knew 14:39:26 johnthetubaguy: :-) 14:39:38 i'll be very pleased if we can hit 150 tomorrow 14:39:43 and i'll buy cake if we hit 100 14:39:50 or something 14:39:53 beer 14:39:58 russellb: are you talking about triage or fixes? 14:40:00 a-gorodnev, +1 14:40:02 triage 14:40:09 fixes are great too of course 14:40:22 but we have to stay on top of triage to make sure we're catching the most critical reports and getting attention on them 14:40:23 ack 14:40:55 so we'll talk bugs tomorrow ... so on for now 14:40:57 #topic blueprints 14:41:06 o/ 14:41:10 #link https://blueprints.launchpad.net/nova/+milestone/icehouse-3 14:41:23 my biggest ask right now is to please make sure the status is accurate on your blueprints 14:41:34 because it's honestly too many for me to keep on top of and keep updated 14:41:35 jog0 has a tool scanning those now 14:41:48 mriedem: nice 14:41:51 script all the things 14:41:53 patches with unapproved/wrong blueprint links 14:42:07 cool, I should try get hold of that 14:42:08 time to bring out the -2 hammer 14:42:11 johnthetubaguy: :-) 14:42:24 status? 14:42:26 I was going to take on keeping the blueprints honest, I will try do a loop through those soon 14:42:45 johnthetubaguy: hugely appreciated 14:42:59 and this is another big topic for next week ... revisiting our process around blueprints 14:43:16 i wanted to point out that i've got my patches up for db2 support after working out the kinks locally, but of course it's going to be blocked on CI: https://review.openstack.org/#/c/69047/ 14:43:25 which is on hold until after the chinese new year... 14:43:27 would it be possible for you guys to review BP's next week - implementations that is 14:43:37 garyk: some i'm sure 14:43:41 mriedem: CI? 14:43:45 if there are a lot of people in the same room then it can be done in groups 14:43:48 3rd party CI with a db2 backend 14:44:09 that was contingent on the blueprint 14:44:12 mriedem: yeah, that's going to be a blocker ... but if it's running in time, ping 14:44:22 i don't have high hopes... 14:44:36 k, well saves me from having to break bad news to you then, you're fully aware :) 14:44:50 i've already been setting low expectations internally :) 14:45:02 the next cut on the icehouse-3 list is stuff still "not started" 14:45:16 PhilD: you have one "Not Started" 14:45:22 https://blueprints.launchpad.net/nova/+spec/tenant-id-based-auth-for-neutron 14:45:41 not sure assignees of the others are here 14:46:03 Code is up for review: https://review.openstack.org/#/c/69972/ 14:46:16 heh, k 14:46:25 * russellb sets needs review 14:46:34 OK, so I will try make sure we have all the stats straight on those, I see joe's tool now 14:46:55 thanks 14:46:56 there's probably some that are dead 14:47:03 abandoned patches that haven't been touched in weeks 14:47:06 also, abandoned patches, etc, should get detected 14:47:07 should just defer those too 14:47:10 yeah 14:47:19 need to aggressively prune this list toward reality 14:47:26 time to deploy some scripting! 14:48:06 wheeee 14:48:10 the blueprint API is not great. 14:48:18 I'm interested in one BP that hasn't be approved 14:48:23 better than screen scraping 14:48:31 a-gorodnev: which one is that 14:48:38 https://blueprints.launchpad.net/nova/+spec/glance-snapshot-tasks 14:48:49 there tons of text there 14:48:57 but nothing special 14:49:03 yeah that's not going to happen 14:49:19 I understand that it's already too late 14:49:41 we already delete snapshots that error out, so we are doing better now 14:50:04 of course, the target is ouf of i3 14:50:46 #topic open discussion 14:50:54 10 minutes to go, can talk blueprints or anything else 14:51:45 any major gate blockers? 14:51:58 this week seems quiet except for the one alexpilottie pointed out yesterday 14:52:14 russellb, https://review.openstack.org/#/c/49395/ it's been sitting there for months, and is a nice fix 14:52:16 which reminds me to check if that quieted another one down 14:52:32 probably needs a rebase tho 14:52:38 mriedem: yeah, we reverted it ... 14:52:59 i saw one gate bug related to shelving? though the timeframe sounded like it was probably caused by the same thing alex reverted yesterday 14:53:11 russellb: that's what i needed to correlate today 14:53:19 because yeah https://review.openstack.org/#/c/71105/ 14:53:38 the shelving fail and the nwfilter in use fail show up around the same time 14:53:49 k let me know what you find 14:53:54 yar! 14:54:12 ah cool, looks like that revert fixed the nwfilter in use fail 14:54:59 cool, makes sense 14:55:12 the patch that won't go away :) 14:55:42 :) 14:56:03 trends look the same with the shelving error, so must have been the root cause 14:56:06 jog0 was right! 14:57:02 ndipanov: i had some comments on https://review.openstack.org/#/c/49395/ 14:57:29 alright, well thanks for your time everyone 14:57:33 bye! 14:57:35 #endmeeting