21:04:42 #startmeeting nova 21:04:43 Meeting started Thu Nov 15 21:04:42 2012 UTC. The chair is russellb. Information about MeetBot at http://wiki.debian.org/MeetBot. 21:04:44 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 21:04:45 The meeting name has been set to 'nova' 21:04:58 #link http://wiki.openstack.org/Meetings/Nova 21:05:09 There's the agenda ... who's here? 21:05:09 hi huys 21:05:11 <-- here in body, at least 21:05:26 * jog0 says hi 21:05:29 Ahoy! 21:05:36 comstud: here? cells on deck first 21:05:42 <-- here 21:06:05 * devananda waves 21:06:05 k, we'll skip cells and wait for comstud 21:06:16 #topic state of the baremetal driver 21:06:29 devananda: hey, want to give an update on this one? 21:06:30 sorry 21:06:33 lunchies 21:06:39 russellb: sure 21:06:41 comstud: no problem, we'll do cells next 21:06:49 two more patches landed last night, total of 3 now 21:07:01 there are two more large-ish ones, and a few tiny ones, left 21:07:09 big one is replacing the baremetal driver code 21:07:18 after that is adding the pxe functionality to it 21:07:38 #link https://review.openstack.org/#/q/status:open+topic:bp/general-bare-metal-provisioning-framework,n,z 21:07:42 meanwhile i've also been working on devstack integration 21:07:52 and formulating a plan with the CI team for how to get it all going 21:08:08 to be shared shortly with the dev list (waiting on approval from the NTT guys) 21:08:10 has there been much feedback on the next big one? 21:08:31 little to none 21:08:31 #link https://review.openstack.org/#/c/11354/ 21:09:00 the big one is a little tough because there aren't any real users of the old baremetal code 21:09:02 #help need reviews on the next big baremetal driver patch 21:09:11 so it is hard to say if it is breaking anything 21:09:18 vishy: so shouldn't have qualms about removing it if nobody uses it ... 21:09:24 could do a post on openstack@ just in case 21:09:31 but perhaps breakage doesn't matter for that reason 21:09:38 at least we'd have something to point to and say "well, we tried to see if it mattered ..." 21:09:44 we can also clearly define it in release notes 21:09:50 vishy: i can at least say that the _new_ code works :) 21:09:52 yes i propose a quick message to openstack-operators 21:09:56 and/or openstack 21:10:06 great, who's going to send the message? 21:10:11 and reviews can stick to style / code-smell 21:10:14 I dont even think the old baremetal code works 21:10:20 devananda? 21:10:27 my guess is that it doesn't 21:10:37 it cant work, we pass args from compute it into it that it cant accept 21:10:42 vishy: i haven't heard anyone say anything about the ol code working 21:10:43 devananda: you good for putting out a post just making sure nobody freaks out that we rip out the old stuff? 21:10:59 well if it's that bad ... i guess no message needed 21:11:00 russellb: sure. destinations are just openstack@ and openstack-operators@ ? 21:11:05 yeah 21:11:11 a message still wouldn't hurt 21:11:11 I know this because we didn't update it a number of times when we changed driver calls for everything else 21:11:21 want to show the ops that we care about them :-) 21:11:26 message is always good but I'd say its very safe to rip it out 21:11:38 rmk: I updated it for no-db-virt :) 21:11:52 #action devananda to post to openstack/openstack-operators about removing old baremetal code, just to make sure everyone is good with it 21:12:09 alright, anything else on baremetal? lots of movement since last week. 21:12:27 that was the high level stuff 21:12:32 dansmith: cool -- do you know if all the entry points from compute match the required number of args? Unless someone changed it recently I dont think it does 21:12:43 yeah, anything deeper we should hit on the -dev list 21:12:46 rmk: no idea, but the tests seem to run 21:12:51 interesting 21:13:01 devananda: just want to hit status here to make sure it keeps moving 21:13:06 :) 21:13:07 alright, let's talk cells! 21:13:12 ok 21:13:12 #topic cells status 21:13:14 comstud: you're up 21:14:01 i spent a number of days breaking out the scheduling filter/weighting into code that was currently somewhat shared with the host scheduler. all of the base stuff landed including the host scheduler changes and new weighting plugin stuff 21:14:11 cells has been rebased against that stuff that landed (master, really)... 21:14:20 i broke out the filtering/weighting scheduler stuff in cells into its own review 21:14:23 #link https://review.openstack.org/#/q/status:open+topic:bp/nova-compute-cells,n,z 21:14:32 all of this lowers the main review down to < 4000 lines now (still large) 21:14:45 anyway... was in rebase hell for a while with all of the FLAGS -> CONF changes... 21:14:56 but I addressed a number of comments in the initial feedback 21:15:00 and those reviews are up 21:15:12 next is... trying to clarify some of the basic routing stuff. 21:15:24 and seeing if I can make the code make more sense there 21:15:29 w/ better doc strings 21:15:30 yeah, that would help me :) 21:15:50 code seems sane, but docstrings, and some overview on how the routing works would be useful for folks diving into it 21:15:56 yep 21:16:08 #action russellb to do another pass on reviewing the updated cells patches 21:16:14 I think I've thought of a way to clean it up 21:16:26 we're going to need at least one other person to do a deep dive into nova-cells to get these reviews done 21:16:27 then it becomes more obvious even without docstrings, but we'll see 21:16:32 how's cells with quantum integration btw? 21:16:44 cells works with a global quantum 21:16:59 and individual nova-network's in each cell 21:17:09 anyone else want to volunteer to start doing a deep dive in this code to help the review along? 21:17:11 except there's some stickiness if you use different networks in different cells right now 21:17:11 comstud: quantum v2 ? 21:17:14 that needs addressed 21:17:21 zykes-: ya 21:17:22 we use it 21:17:23 :) 21:17:25 ok 21:17:33 comstud: ? nova-network shouldn't be necessary with v2 21:17:42 it proxied to quantum in v1 21:17:47 #help need one more nova-core reviewer to volunteer to start reviewing the cells patches 21:17:57 but v2 the api nodes should be calling quantum directly.. 21:18:01 does network_api handle it ? 21:18:06 comstud: yeah 21:18:11 what's the diff though in cells and aggregates if I can ask ? 21:18:12 it should be fine then 21:18:19 configure network_api in each child cell appropriately 21:18:29 zykes-: cells each have their own db and queue 21:18:34 the only thing that will not work potentially is the API extension 21:18:37 aggregates are groups of hosts in a single installs 21:18:49 if you point API cells and child cells at same quantum, then you're ok 21:18:52 cells contain aggregates 21:18:53 but that's a limitation right now 21:19:11 vishy: ok, so aggregates are within that 21:19:14 would be good to keep a documented set of limitations somewhere 21:19:20 agree 21:19:21 not sure where ... 21:19:40 devref? 21:19:49 jog0: but it's a user doc, really 21:19:57 maybe wiki.openstack.org for now while it's still moving a lot 21:20:48 anything else on cells? 21:20:58 in any case.. that's about all i have. probably the largest issues (just understanding how the routing works) will be addressed shortly. 21:20:59 russellb: works for me, but long term openstack-manuals sounds like the right home 21:21:13 jog0: agreed 21:21:15 #topic bugs 21:21:24 #link http://webnumbr.com/untouched-nova-bugs 21:21:28 only 20 untouched bugs, very nice. 21:21:48 many thanks to those putting time into triage. if we all just get a few issues here and there, we'll stay on top of them 21:21:52 did mikal create his cookie report? 21:21:57 Yep, please hold 21:22:00 http://www.stillhq.com/openstack/nova/triage-20121116.txt 21:22:08 So, we're not as good as last week, but still ok 21:22:15 Mostly that's cause russellb has fallen off the wagon... 21:22:25 woohoo! 21:22:25 mikal steals all the cookies 21:22:28 mikal gets this week's triage gold star 21:22:28 Heh 21:22:36 Its still not all the nova-core people doing triage 21:22:37 Hint hint 21:22:54 i've got a focus right now, sorry :) 21:23:00 +different 21:23:16 Its ok 21:23:20 I shall just mock you 21:23:20 #help please help triage nova bugs! Even just a few would be awesome. 21:23:25 ok :) 21:23:29 cool, anything else on bugs? 21:23:40 i actually have one thing on bugs 21:23:48 We have 4 active nova security bugs (private) that need nova-core attention 21:24:21 please look at these bugs (and don't discuss them here): https://bugs.launchpad.net/nova/+bug/1074343 https://bugs.launchpad.net/nova/+bug/1070539 https://bugs.launchpad.net/nova/+bug/1069904 https://bugs.launchpad.net/nova/+bug/1058077 21:24:22 Launchpad bug 1074343 in nova "ec2 describe instances does not filter by project_id" [Undecided,Incomplete] 21:24:36 unless the bug is already public, heh 21:24:40 heh 21:24:44 #topic grizzly-1 21:24:53 Alrighty, grizzly-1 is one week out 21:24:53 uvirtbot: loose lips sink ships! 21:24:54 dansmith: Error: "loose" is not a valid command. 21:25:06 hrm 21:25:22 #link https://launchpad.net/nova/+milestone/grizzly-1 21:25:42 so, let's take a look at status of some of these things 21:25:45 we have a few things on the Hyper-V side that we'd like to have in for G-1 if possible 21:26:05 alexpilotti: ok, is there a blueprint or a bug? 21:26:21 russellb: we have a bp, let me fetch it 21:26:22 russellb: were you going to target no-db-virt for G-1? 21:26:26 I think we can call it done, right? 21:26:35 dansmith: yeah, we can, link? 21:26:35 russellb: the audit's going to be a couple weeks past G-1, I just moved it to G-2 21:26:44 is there a chance the remaining baremetal patches might land before G1? 21:26:44 sdague: ok great thanks 21:26:56 https://blueprints.launchpad.net/nova/+spec/no-db-virt 21:27:11 devananda: it's possible ... don't want to rush the reviews just to meet grizzly-1 though, since we're still early in grizzly 21:27:25 ack 21:27:26 https://blueprints.launchpad.net/nova/+spec/grizzly-hyper-v-nova-compute 21:27:58 dansmith: done 21:28:02 thanks 21:28:23 for the moment we have ConfigDriveV2 implemented 21:28:27 alexpilotti: so is everything for this blueprint up for review? 21:28:44 alexpilotti: really should split the features into individual blueprints 21:28:51 instead of 1 blueprint for all features 21:28:54 russellb: the BP is still generic, we have to update it, I know 21:29:14 vishy: I'm going to do it ASAP, in the next days 21:29:14 yeah, hard to target if it doesn't have a defined endpoint 21:29:38 alexpilotti: ok, well if you get a blueprint for config drive support, we can target that, that will probably go in for grizzly-1 21:29:40 anyway on the Nova side we don't have too much, we are targeting mainly Grizzly 21:29:49 for Quantum 21:30:02 russellb: ok tx 21:30:05 gotcha. 21:30:20 mikal: can I please ask you to re-approve the review? 21:30:29 mikal: https://review.openstack.org/#/c/15743/ 21:30:31 Sure 21:30:38 is Mate Lakat here? don't know the status of that xenapi volume driver blueprint 21:30:43 mikal: since you approved I had to rebase and fix 2 lines 21:30:55 Yep, I'll re-review this morning 21:31:22 xenapi-volume-drivers seems to be stalled 21:31:25 vishy: know anything on that one? 21:31:38 guess we can just bump early next week if needed 21:31:57 there are a bunch of reviews in 21:31:59 but not done 21:32:05 k 21:32:19 #topic Open Discussion 21:32:38 #help nova-core: review queue is longer than ideal, please help knock it down 21:32:53 ok i have an idea on that point 21:32:56 might be crazy 21:33:08 i like how this is starting 21:33:12 what if we allowed certain people to have +a access to subdirectories 21:33:33 so, if it only touched say ... nova/virt/libvirt/* ? 21:33:39 right 21:33:39 is that doable in the current ci system? 21:33:42 vishy: +1 :-) 21:33:46 sdague: no i asked 21:33:53 but we could do it through convention 21:34:07 does that seem useful? or just a crazy idea 21:34:24 sure, I suppose. It does bring up a question of creating consistent quality across the project 21:34:24 seems like we kinda already have that 21:34:29 Are there really that many extra reviewers that would add? 21:34:39 it isn't for extra reviewers 21:34:42 seems potentially useful ... but if you trust someone enough to approve nova/virt/libvirt/, we should be able to trust them to approve anything that is within their comfort zone 21:34:49 in that, if they hyperv guys propose something, everyone mostly trusts them that it's the right thing, as long as it doesn't break anything, right? 21:35:03 dansmith: yeah, that's what I do on those patches :-) 21:35:05 it is so alexpiloti can approve stuff in virt/hyperv/ without having to go track down two core members to +2 a 21:35:20 russellb: I keep adding beers for you ;-) 21:35:23 but his +1 is effectively that to me already 21:35:27 right 21:35:42 sure but we need a core member to take the time and go push the button 21:35:44 if I see that a domain expert has already reviewed it, i spend much less time on it 21:35:49 true 21:36:03 but it doesn't take that long :) 21:36:09 maybe we just need to get better about pushing the button 21:36:10 I dunno, I think that's opening the doors a bit, 21:36:12 :) 21:36:12 and it's good to have at least a 1 minute sanity check 21:36:24 fair enough, just wanted to throw the idea out 21:36:30 vishy: we do, but at the end of the day it's going to be a core member that might have to do an emergency fix. So it seems like they should have looked at it at some point :) 21:36:32 even if it's just to small stylistic things and conventions we avoid elsewhere but aren't in hacking.py 21:36:38 i have a general topic to bring up as well 21:36:41 sdague: ++ :) 21:37:03 cross tenant apis 21:37:21 currently we have things like all_tenants=True for admins 21:37:37 so some things an admin can do across tenants 21:37:49 is this correct / useful? 21:38:01 this sort of plays into the tenant_ids in urls for future versions of the api 21:38:11 and whether that is good or bad 21:38:40 i thought there was no concept of a tenant specific admin 21:38:51 and that an admin in keystone was admin of the world 21:39:28 * russellb must just not understand the question 21:39:52 yeah i'm discussing how we fix that 21:40:05 Well, we still want super-admins right? 21:40:12 I see that as useful functionality 21:40:17 oh, ok. 21:40:23 mikal: useful yes 21:40:35 presumably we want both tenant admin and operators(super-admin) 21:40:49 jog0: that would be good 21:40:56 but the question is should operators be using the normal urls 21:41:19 can an operator access http://api//servers 21:41:28 or just: 21:41:51 wouldn't using the normal urls make it harder to do cross tenant operations? 21:41:53 http://api//servers?all_tenants=True 21:42:10 jog0: yes 21:43:35 anyway something to chew on 21:43:49 according to the keystone readme, a role is "a first-class piece of metadata associated with many user-tenant pairs" 21:43:54 I am not sure where admin fits in to that 21:44:25 might be a good candidate for an openstack-dev thread to get more attention 21:44:32 russellb: +1 21:44:44 sounds like a good plan 21:44:54 on the subject of openstack-dev.... 21:45:00 new other topic - http://lists.openstack.org/pipermail/openstack-dev/2012-November/002889.html - was hoping to get some other eyes on this thread 21:45:18 it's i18n again 21:45:56 i18n in the API itself seems evil 21:45:59 but i'm an ignorant american 21:46:23 yeh... the problem is when you're an operator in .cn, assuming english, isn't often an option 21:47:06 anyway, just wanted to have people take a look and discuss out there. Not let it just wither on the vine 21:47:26 * russellb tends to just listen to the people that live elsewhere and this is a big issue for them 21:47:53 while talking about ML threads: http://lists.openstack.org/pipermail/openstack-dev/2012-November/002822.html 21:48:09 aw, how'd i miss that 21:48:54 i don't have a strong opinion on it. my gut reaction is to say leave a default quota in place, but i wouldn't fight it if enough people thought it made sense to have no quotas by default 21:49:50 anything else? or we'll #endmeeting 21:49:55 #topic ceilometer/nova interaction as discussed on the ML #link http://wiki.openstack.org/EfficientMetering/FutureNovaInteractionModel 21:50:02 russellb: while I would like to change it, I don't have a strong opinion on it either. was hoping to get more feedback 21:50:06 ooh, good one eglynn 21:50:20 #topic ceilometer/nova interaction as discussed on the M 21:50:21 we need to get closure/convergence on that thread 21:50:31 #topic ceilometer/nova interaction as discussed on the ML 21:50:34 #link http://wiki.openstack.org/EfficientMetering/FutureNovaInteractionModel 21:50:46 so the question for you nova folks, is either option #4a or #5 a runner? 21:50:57 eglynn: yeah ... i took a step back to see where everyone took it. there were a lot of different opinions flying aorund. 21:50:58 (both require buy-in from nova) 21:51:32 yep, lots of noise on the thread 21:51:42 i'm leaning towards #5 21:51:44 i guess i'd like to look closer at what 4a does, and figure out if it's something that could just be a part of nova-compute 21:52:20 that was the define a common interface that each driver could implement as a library right? 21:52:21 vishy: so we'd distribute 2 copies of the virt layer? 21:52:26 vishy: cool 21:52:28 russellb: my concern about 4a is the seperate daemon complicating deployment 21:52:47 eglynn: that's why i was saying we could look at adding whatever it does to the existing daemon 21:53:14 russellb: yeah, but the timeliness issue is a worry 21:53:21 i don't think i fully get 5 ... 21:53:41 (as in there seem to be just fairly weak guarantees on the period_tasks) 21:53:49 eglynn: correct 21:53:58 the only realy difference between 4 and 5 is nova-compute-pollster lives in ceilo instead of nova 21:54:02 russellb: any hazy around 5 may be my fault 21:54:04 the frequency they run is directly affected by how long they take 21:54:20 so, what does this library from nova have in it 21:54:25 in #5 21:54:44 an implementation of an interface defined by cielo 21:54:47 does it contain the virt drivers? 21:54:48 a cut-down subset of the hypersior driver API 21:54:53 like get_bw_for_period 21:55:00 instance_exists() 21:55:02 etc. 21:55:08 ah ok ... 21:55:15 and then nova-compute could consume this same library then? 21:55:16 yep, all read-only, non destructive 21:55:23 sure 21:55:29 for the bits it uses too 21:55:37 though the versioning might be oriented to external consumers 21:55:46 sure. 21:55:46 (slowly evolving, stable) 21:56:00 that doesn't seem so evil ... i just don't want to see duplicated virt code 21:56:14 cool 21:56:34 since that piece needs to exist regardless 21:56:46 yep 21:56:52 so, provisional good vibes around option #5? 21:56:56 we could define that api first before deciding where the daemon will live 21:56:58 my 2 cents: 5 seems to leave more control to Nova on the interface / mixin that needs to defined 21:57:16 eglynn: yeah, i think so ... 21:57:28 in Hyper-V we have quite cool APIs handled via WMI, so both 4a and 5 are not an issue 21:57:31 pending letting it simmer a bit, and going into more detail on what it would look like, etc 21:57:58 cool 21:58:05 eglynn: thanks for bringing that up 21:58:10 anything else? we have a whole 2 minutes 21:58:16 OK, so we'll look at fleshing out that API definition on the ceilo side to keep the ball rolling 21:58:47 eglynn: what's the deadline for the ceilo / nova integration? 21:59:13 alexilotti: probably G-2 21:59:21 (we're sitting out G-1) 21:59:33 seems reasonable 21:59:37 cool 21:59:48 wow, cool. Can't wait to have it done. 22:00:04 alright, then. thanks everyone 22:00:07 #endmeeting