21:02:06 #startmeeting nova 21:02:06 Meeting started Thu Jun 20 21:02:06 2013 UTC. The chair is russellb. Information about MeetBot at http://wiki.debian.org/MeetBot. 21:02:07 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 21:02:07 hello everyone! who's around to talk about nova? 21:02:09 The meeting name has been set to 'nova' 21:02:10 hi 21:02:12 here 21:02:15 hi! 21:02:20 hi 21:02:20 o/ 21:02:24 o/ 21:02:28 hi 21:02:34 #link https://wiki.openstack.org/wiki/Meetings/Nova 21:02:42 #topic blueprints 21:02:43 o/ 21:02:51 \o 21:02:53 \o 21:02:58 #link https://launchpad.net/nova/+milestone/havana-2 21:03:11 I think we're actually in decent shape given how much is on the havana-2 list 21:03:16 lots of stuff already up for review, which is good 21:03:31 but I figured it would be a good time to remind everyone of the schedule ... 21:03:44 we're roughly half-way through havana-2 (which means roughly half way through dev time for havana!) 21:03:52 #link https://wiki.openstack.org/wiki/Havana_Release_Schedule 21:05:03 sooner the better for putting some review time into the bigger stuff for these blueprints , so it's not so much of a rush at the end of this milestone 21:05:19 any comments/questions about release status / blueprints? 21:06:02 i am somewhat concerned about the API locale one I have up 21:06:10 mrodden: link? 21:06:14 https://blueprints.launchpad.net/nova/+spec/user-locale-api 21:06:21 and what's the concern? 21:06:27 its not very glorious of code 21:06:39 and hasn't had any feedback from any nova core's yet 21:07:05 yeah, don't think i've seen the nova code yet 21:07:31 the oslo stuff got merge yesterday (?) 21:07:38 so thats cool 21:07:40 well right, i saw that 21:07:51 is there a nova patch up? 21:07:57 yep... second 21:08:08 https://review.openstack.org/#/c/30459/ 21:08:19 yep that one 21:08:25 #help could use feedback on user-locale-api review 21:08:26 says it's merged 21:08:35 https://review.openstack.org/#/c/30479/ 21:08:36 oh 21:08:37 that one 21:08:38 yeah 21:09:11 i can't pull it up right now ... on edge internet :-/ 21:09:36 blueprint set to "Needs Code Review" ? 21:09:50 my main concern is that it took a lot of iterations to get it ready for g3 but then was too late to merge 21:09:56 yeah should be 21:10:00 ok 21:10:15 well i'll try to look soon 21:10:18 ok thanks 21:10:29 (and others too i hope) 21:10:34 i've been trying to do a better job of looking at older stuff first 21:10:43 using the next-review script, and looking at my review stats 21:11:05 cyeoh: how's the v3 api stuff coming? 21:11:27 I think its progressing fairly well, but nearly all of it is still waiting for review 21:11:35 (what we've submitted anyway) 21:11:38 and are review times causing you lots of pain? 21:12:04 i was afraid of that 21:12:05 I'm getting concerned because I know we are going to get merge conflicts because of setup.cfg changes 21:12:27 i like the new process you've started using, step 1/2 21:12:31 The v1 API patches are just copies of files to new places so they should sail through review; it's just getting 2 core people to look at them. 21:12:43 last time I checked the v3 api patches were about 30% of the review queue (if only taking into account changes waiting for reviewer rather than submitter) 21:12:50 yeah, but the step 1 patch shouldn't be approved before step 2 21:13:29 #help really need reviews on v3 API changes, they're easier to review than they seem, i promise :-) 21:13:52 that would be very much appreciated and lower my stress levels :-) 21:14:03 I think cyeoh just wants to race with dansmith on depleting all the devstack nodes :) 21:14:05 a lot of core reviewers seem busy with other things ... been slowing things down a bit i think 21:14:14 sdague: good luck to him 21:14:19 sdague :-) 21:14:57 all good stuff :) 21:15:31 i'll see what i can do to encourage more review time 21:15:33 maybe i'll bake cookies 21:15:48 mmm! 21:15:50 we can come back to blueprints and such in open discussion if needed 21:15:51 :-) 21:15:54 #topic bugs 21:16:04 so, bug triage 21:16:07 #link https://wiki.openstack.org/wiki/Nova/BugTriage 21:16:13 we have this handy process to split up the triage work 21:16:19 there are 68 new bugs right now, most of them tagged 21:16:34 so please check your queue if you signed up to help 21:16:43 and if you didn't sign up, please do 21:16:56 I triaged a bug as WONTFIX today and felt bad about it, but it was an SQLAlchemy bug so there's not much we can do. 21:17:14 68 is a bit too high 21:17:19 heh, you shouldn't feel bad :) 21:17:19 was it the ipv6 one? 21:17:23 yes 21:17:30 cool, i figured 21:17:49 i close bugs sometimes just because i think the bug report isn't good enough 21:17:53 "it doesn't work" 21:18:02 so i'm way more harsh :) 21:18:06 that's why I felt bad about that one; it was totally valid and well written, just not ours 21:18:17 ah, yeah ... well as long as you made all of that clear 21:18:41 is there any prereq for signing up for a tag i.e. is there additional permission needed to triage past tagging? 21:18:56 melwitt: you have to be a member of the nova-bugs team on launchpad 21:19:01 but it's an open team (anyone can join) 21:19:11 so, just willingness and time to give 21:19:15 russellb: ok, thanks 21:19:20 sure 21:19:37 it doesn't take *that* much time. it's like milking cows: it only takes a few minutes but you have to do it every day or the cows kick you 21:19:44 ha 21:19:51 yeah, doesn't take much to do a few 21:20:07 takes a lot if you're the one or two people trying to do most of it (which is what we kinda had before) 21:20:35 right, but if you're only triaging one tag it's not too bad. 21:20:55 i've been tagging, but not triaging as much, but happy to help on tricky ones 21:21:04 dripton: yeah hope so, spread the pain :) 21:21:32 let's talk subteams 21:21:35 #topic subteams 21:21:44 hi!! 21:21:46 devananda: what's up with baremetal / ironic ? 21:23:39 ok, can come back to that one 21:23:39 harlowja: 21:23:46 howdy 21:24:05 or hartsocks ? 21:24:09 (sorry if it's just my internet sucking here) 21:24:13 hey. 21:24:21 so one flow in cinder is working and awaiting the taskflow library to be released (so that it can be integrated, jenkins is right now complaining about missing dep() 21:24:23 who wants to give an update :) 21:24:35 cool. 21:24:42 sure, so we are also working with heat folks on there desired requirements 21:24:49 it'd be interesting to have a nova equivalent 21:24:53 *even if its just random thoughts* 21:25:02 russellb: sorry, looked away at an email for a minute... back now :) 21:25:03 #link https://wiki.openstack.org/wiki/Heat/TaskSystemRequirements 21:25:14 harlowja: k, i'd post to the ML 21:25:15 i'd be interesting to start collecting nova 'ideas/requirements' 21:25:26 russellb sounds good 21:25:32 anything else? 21:25:36 so otherwise, just heads down working 21:25:41 cool 21:25:42 thats about it :) 21:25:47 hartsocks: alright, you're up 21:25:51 hartsocks: i like the weekly emails 21:25:51 okee dokee 21:25:57 Thanks. 21:26:07 guess you can just link to those here, heh ... but highlights are good too 21:26:08 I appreciate the response we got off of that. 21:26:32 So… I'll just put out a list of stuff we think is good for core-review 21:26:42 ok 21:26:44 I'll send that around on fridays. 21:26:47 Otherwise... 21:26:48 So we VMwareAPI folks are drilling down on our Blueprints now. 21:27:01 We've got most of the critical/high priorty stuff assigned to people 21:27:14 There's only one outstanding bug that needs attention. 21:27:29 We've seen a potential problem with one of the blueprints slated for H2 21:27:41 We may have to move it to H3. 21:28:02 We're heads down now and H2 deadlines will be close I feel. 21:28:02 ok, assignee should be able to update that if needed 21:28:11 or ping me (or anyone on nova-drivers) otherwise 21:28:25 ok, to get stuff reviewed in time, try to have code up for review early 21:28:34 stuff that goes up in the last week will almost certainly slip 21:28:45 just based on past experience 21:28:53 sooner the better, of course 21:28:58 I've told folks to get there code up by July 11th if they really want it to get in. 21:29:09 Perhaps I should say the 4th 21:29:13 perfect 21:29:31 11th hard deadline, depending on size, there's still risk 21:29:34 We really only have 2 working weeks in this release then. 21:29:38 but it's not the havana feature freeze, so not a *hue* deal 21:29:45 Yep. 21:29:51 Just if you want it in H2. 21:29:56 yep. 21:30:06 and hopefully we can merge as much as we can in h2 21:30:09 so it's not so heavy in h3 21:30:15 I'm trying to push all the bug fix + put the fires out work… to H2 so we can save H3 work for new hotness. 21:30:25 heh 21:30:30 So we're rolling along. 21:30:38 cool, thanks for the update. 21:31:03 devananda: alright, how's ironic / baremetal ? 21:31:08 hi! 21:31:23 prepared a summary, pasting... 21:31:28 wooo 21:31:33 Many open bugs in baremetal still. I've continued to focus on ironic rather than fixing bugs, except for when the bugs are quick to fix. 21:31:41 dprince and pleia2 have been doing some interesting things testing bare metal with the TOCI project and fixing blocking bugs for their work. some bug fixes have trickled over to ironic from their work. 21:31:57 GheRivero seems to be making good progress porting image utils to glanceclient. Could probably use a few more reviews: 21:32:06 #link https://review.openstack.org/#/c/33327/ 21:32:19 Once that's done, we'll be able to move on implementing the PXE driver in Ironic 21:32:31 that's the last driver we're missing (IPMI and SSH/virtualpowerdriver are done) 21:32:40 I'm addressing the present lack of API and RPC functionality 21:32:50 and NobodyCam is working on integration with keystone, and scripting the installation of ironic services with diskimage-builder. 21:32:53 [EOF] 21:33:10 speaking of API, are you looking at pecan/wsme for your API? 21:33:15 yep 21:33:18 awesome 21:33:22 the basic components are already there and working 21:33:27 cool 21:33:30 i landed api unit tests today 21:33:39 need to actually implement all the handlers for things :) 21:33:42 and RPC layer 21:33:45 boris-42 was wanting to help get those used more throughout openstack, i recommended he check with you to see if you could use help 21:33:57 those == pecan/wsme 21:34:04 great 21:34:17 one of his guys, romcheg, has been working on porting dansmith's object code 21:34:20 so we are using that, too 21:34:25 awesome 21:34:37 which means once it settles we should be looking at oslo-ifying it 21:34:40 yep 21:35:27 cool, anything else on baremetal / ironic? 21:35:35 nope 21:35:37 any other subteams wanna give a report? 21:35:47 scheduler 21:35:54 n0ano: great, go for it 21:36:13 went over 2 things, scaling - I've started a thread on the mailing list, we'll see how that works out... 21:36:24 ah yes 21:36:44 sounds like consensus was to kill off all of the fanout_cast stuff, which i'm actually happy to hera 21:37:11 other issue was talked about BP for multiple default AZs (https://blueprints.launchpad.net/nova/+spec/schedule-set-availability-zones), looks like a reasonable idea, he'll be working on implementing it. 21:37:20 beyond scale, fanout also causes some problems with trusted messaging 21:37:44 yep, sounds reasonable to me 21:38:06 russellb, I still don't undersand, why do the fan out message have more latency than the DB updates, they should be faster 21:38:55 anything else? 21:38:56 #topic open discussion 21:39:08 I have something for open discussion 21:39:17 q on the multiple default AZs 21:39:17 https://review.openstack.org/#/c/33888/ 21:39:41 tox appears broken in anything but CI after a recent reqs change, that's the revert patch ^^ 21:39:50 I am a little confused on what the goal is there. 21:40:14 oops 21:40:14 n0ano: because the db updates are "instant" 21:40:15 n0ano: while the fanout stuff is periodic 21:40:17 dansmith: whose patch was it? 21:40:23 n0ano: ^ 21:40:26 russellb: geekinutah 21:40:54 dansmith: checking to see if that solves my tox issue... 21:41:07 I would think we just replate the DB updates with a fanout message, then they are no longer periodic 21:41:07 mrodden: if it's a failure to get oslo.config, then, yeah 21:41:18 ok, +2 21:41:19 it fails the nova-requirements test, which I don't know the content of 21:41:31 the DB can't be `instant', you have to sent a message to the DB server 21:41:33 is there a gate that prevents going backwards or something? 21:41:43 n0ano: right, we could, but that's a *ton* of messages 21:41:52 dansmith: take to the infra team about requirments test 21:41:57 versus a ton of DB update messages 21:42:07 pretty much 21:42:22 but thing is, right now we're using the db 21:42:30 the fanout stuff is just wasting resources 21:42:35 iiuc, db updates will grow linearly with # of nodes, whereas fanout will grow exponentially.... 21:42:54 (it's possible I am thinking of a different conversation, too) 21:43:11 devananda: basically just talking about how the scheduler gets the current state of compute nodes 21:43:18 resource usage and whatnot. 21:43:23 devananda, I think so, the fanout message is one from each node, adding a new node only adds one message (per state change) 21:43:23 devananda: AFAIK fanout grows with number of schedulers*compute-nodes and db is number of compute-nodes 21:43:36 jog0: right. 21:44:00 i don't think people are running many schedulers, but yeah. 21:44:06 n0ano: we already have the use DB for record keeping paradigm fanout generally breaks that. 21:44:34 russellb: one of the problems is single threaded scheduler processing all the fanouts, adding another scheduler doesn't make thigns better just the same 21:44:44 fanout also generally kills our ability to move to peer-to-peer messaging, which is much more scalable 21:45:37 wouldn't the message from the compute node to the scheduler be considered a peer-peer message, shouldn't that fit in? 21:45:42 and fanout isn't as good for trusted messaging, because you don't know the specific endpoint you're talking to 21:45:59 it's one message broadcasted to all schedulers 21:46:08 the end point is the scheduler, that 21:46:16 that's a trusted end point 21:46:38 but (at least with amqp) you're not talking to one thing 21:46:42 you're sending 1 message to N things 21:47:31 guess we should cover it more on the ML 21:47:35 easier to go into detailed discussion there 21:47:50 n0ano: if we didn't have a notion of a central DB already fanout has more benifits but we already have that concept, so why not use it 21:47:53 yeah, I think the ML is the proper place to discuss this 21:47:56 n0ano: question on https://blueprints.launchpad.net/nova/+spec/schedule-set-availability-zones how do propose having several default AZs? 21:48:09 isn't it a config option now? 21:48:11 jog0: because central db's are evil! evil! 21:48:39 lifeless: perhaps but that is the bigger debate that we should have 21:48:46 jog0: I was trolling. 21:48:47 jog0, we already have one default, the BP is about allowing mulitiple defaults if the user doesn't specify one 21:49:08 choose a host from any of these AZs, as opposed to just this one default AZ 21:49:13 * russellb shrugs 21:49:14 seems ok 21:49:16 n0ano: oh for where VMs go, not for where compute-nodes go? 21:49:18 russellb, +1 21:49:27 jog0, correct 21:49:32 that makes alot more sense 21:49:49 jog0: right :) 21:49:59 it took a long discussion to understand what the BP was proposing but we got it at the end :-) 21:50:00 that wasn't clear to me from the BP 21:50:05 jog0: but btw ... don't host aggregates allow you to technically put a compute node in more than 1 AZ? 21:50:19 russellb: yeah 21:50:20 n0ano: should update theblueprint then with clarification 21:50:37 russellb, that was the suggestion at the meeting 21:50:41 jog0: i don't think our API really deals with that 21:50:55 russellb: yeah, that is currently a don't do for deployers 21:51:01 russellb: they don't 21:51:08 * russellb nods 21:51:19 so you just have plenty of rope 21:51:26 yeah ... 21:51:30 be careful, don't hurt yourself 21:51:31 k 21:51:34 i guess we could put in a safeguard ... 21:52:04 block adding a host to an aggregate with an AZ set, if it's already in another aggregate with an AZ set 21:52:15 russellb: not against the idea, didn't seem worth it at the time though. took the deployers should be careful approach initially 21:52:26 yeah 21:52:37 i think it works from a scheduling point of view 21:52:44 russellb: taht isn't enough, because you can add a aggregate to become an AZ afterwards too 21:52:45 just reflecting data back through the APi doesn't account for it 21:52:51 ah, yeah. 21:53:12 well, guess we'll leave it alone for now then :) 21:53:20 russellb: API just lists all AZs with commans in between 21:53:30 oh it does? 21:53:41 well then, it's fine 21:53:42 last time I checked at least 21:53:52 i just misread it then 21:54:11 alrighty, coming up on time ... 21:54:21 any last minute comments/questions/concerns? 21:55:13 alright, feel free to stop by #openstack-nova any other time you want to chat. 21:55:15 thanks! 21:55:16 #endmeeting