17:00:03 #startmeeting nova_cells 17:00:04 Meeting started Wed Nov 9 17:00:03 2016 UTC and is due to finish in 60 minutes. The chair is dansmith. Information about MeetBot at http://wiki.debian.org/MeetBot. 17:00:05 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 17:00:08 The meeting name has been set to 'nova_cells' 17:00:10 o/ 17:00:12 cells peeps say wha? 17:00:17 \o 17:00:17 https://www.youtube.com/watch?v=JHYIGy1dyd8 17:00:19 o/ 17:00:19 is what i say 17:01:01 okay, I guess it's just us 17:01:06 #topic testing 17:01:12 anything on this? 17:01:20 we have some things pending for cellsv2 in grenade and devstack 17:01:22 well 17:01:24 yeah that 17:01:31 https://review.openstack.org/#/c/393441/ 17:01:34 is grenade ^ 17:01:43 https://review.openstack.org/#/q/topic:ocata-requires-cellsv2 17:01:53 i'm good with the nova change on top https://review.openstack.org/#/c/392227/ 17:02:00 so bauzas melwitt ^ 17:02:11 * bauzas looking 17:02:17 we kind of need sdague for the grenade change 17:02:22 heh, it was my question 17:02:23 yeah 17:02:24 but i'm not sure when he comes back 17:02:27 but https://review.openstack.org/#/c/392227/ resolved it 17:02:34 at least another week I think 17:02:40 i can ping him to see for sure 17:03:00 we have some things that can go in before the big ones in nova to get stuff done while we wait for that 17:03:04 so I'm not too concerned 17:03:28 the big issues is going to be, 17:03:33 okay, I'll dig into those changes 17:03:37 upgrading/starting api last in grenade/devstack for multicell 17:04:02 but before tackling that, we can probably deal with non-grenade multinode cells v2 17:04:08 mriedem: yeah, so I had some hacky patches up to try to do that, and have talked to sdague about it.. it's going to be a little tough because of when/where we setup the api endpoints in keystone, etc 17:04:17 yeah 17:04:22 also, 17:04:30 with https://review.openstack.org/#/c/392289/ we might just automatically get it... 17:04:37 because the neutron multinode job will now be running cells v2 17:05:28 so we'll see 17:05:40 sdague's orientation is the 29th 17:05:42 mriedem: that won't end up with old api I don't think 17:05:54 dansmith: no it would be full install multinode all ocata 17:06:14 wanted to make sure the scheduler changes are working in that case at least 17:06:24 oh I see what you mean 17:06:25 yep 17:06:50 we're kinda bleeding over already so let me just... 17:06:51 fudge....so, 17:06:55 #topic open reviews 17:07:05 sdague is basically out for the rest of the year... 17:07:14 maybe a week or two in december 17:07:16 just keep that in mind 17:07:20 I thought we'd get him for beginning of dec yeah 17:07:25 for like a week :) 17:07:31 and he'll be playing catchup 17:07:34 I'll probably drop off at some point too, haven't planned yet 17:07:50 alright 17:07:57 yeah, I know, but if we can at least get the grenade thing figured out, 17:07:59 there are a couple other grenade cores 17:08:04 if we know what he's okay with and not 17:08:13 there are other grenade cores?! :) 17:08:18 yeah man :) 17:08:21 treinish also being out forever 17:08:30 https://review.openstack.org/#/admin/groups/188,members 17:08:33 oh not even 17:08:35 so dean 17:08:54 we can move on 17:08:57 yeah 17:09:03 so, for open reviews, 17:09:06 interesting 17:09:13 I have the big set to work on moving scheduling to the conductor, 17:09:19 which is actually a lot of the work we have planned for this cycle 17:09:21 https://review.openstack.org/#/c/319379/ 17:09:38 it was pretty far off a week ago, but it's actually starting to work a bit now 17:09:58 we've discussed some of the sticking points like secgroups this morning 17:10:02 so still some work to do there, but it's going pretty well 17:10:05 glad to have melwitt back so she can find all the bugs 17:10:10 so after this morning we have at least 2 changes right? 17:10:14 yeah 17:10:14 1. store uuid in secgroup 17:10:16 heh 17:10:27 2. get the uuid from the name in the api and use that for #1 17:10:35 i'll take a crack at #2 17:10:38 I'm cool with that plan 17:10:44 mriedem: yeah, so trying to get #2 into its own patch is going to be a little hard I think 17:10:50 mriedem: because of the changeover that the last patch makes 17:11:02 mriedem: oh okay I was going to do both but that's fine 17:11:22 i need you to review specs :) 17:11:27 i have to tap out of specs for awhile 17:11:33 heh 17:11:37 okay 17:12:10 okay so aside from that set, 17:12:16 what else is on the blocks right now? 17:12:31 so, 17:12:40 instances going to cell0 is not a thing yet right? 17:12:45 it is 17:12:47 and we talked about working that in with the scheudler stuff 17:12:48 as part fo that set 17:12:49 orly 17:12:52 yup 17:13:03 there is a lot wrapped up in there 17:13:10 https://review.openstack.org/#/c/367557/ ? 17:13:17 that's in the series 17:13:27 mriedem: well, yes, but not until the last patch does it actually happen of course 17:13:27 "- in the case of failure method creates instance in cell0" 17:13:28 yup 17:13:47 ok, 17:13:50 hmm, 17:14:02 i might just mark that cell0 bp as complete then, or superseded by the scheduler interaction one 17:14:04 and after this set, we'll be scheduling across cells 17:14:07 so we kinda have the beginnings of multicell 17:14:08 since we're tracking them together 17:14:40 ok so those were the big things on my radar 17:14:55 melwitt is back and can start digging in on this or the quotas stuff if needed, 17:15:00 what I meant above was.. are there other big sets of changes up for review? 17:15:02 as noted i looked into the quotas counting thing a bit yesterday 17:15:06 no 17:15:11 okay 17:15:14 #topic open discussion 17:15:19 there are some tricky bits with counting rams/cores in the api 17:15:33 placement api and resource provider usage might help with that? 17:15:38 yeah 17:15:40 but that's not per-tenant is it... 17:15:42 although they don't know about tenants 17:15:44 right 17:15:52 so you'd have to do a pretty big query of them 17:16:02 i know a guy that likes to write big sql queries 17:16:08 well, 17:16:15 it can't be a sql query 17:16:34 we have to basically get a list of instance uuids and then query all the ram/cores taken by $uuids 17:16:38 you'd have to map instances to computes and computes to resource providers right? 17:16:41 which could potentially be yuuge 17:16:46 dansmith: talking about statistics ? 17:16:57 mriedem: no we have allocations by instance uuid (consumer) 17:16:58 bauzas: no? 17:17:24 mriedem: unfortunately, the data is still in our own db so we could do something super efficient, 17:17:41 mriedem: but the virtual wall between us an placement prevents it by decree 17:17:45 dansmith: http://developer.openstack.org/api-ref/compute/?expanded=show-hypervisor-statistics-detail ? 17:17:56 which is.. kinda the suck 17:18:03 bauzas: that's not tenant specific though 17:18:04 bauzas: I'm not sure what that has to do with what we're talking about 17:18:11 we need tenant specific ram/core usage 17:18:21 dansmith: okay, me not understanding what you're talking about :) 17:18:28 bauzas: quotas by counting 17:18:32 bauzas: as we discussed at summit 17:18:32 ack 17:18:39 there are some details in the ML 17:18:48 i had to talk to dan yesterday afternoon to understand more about that 17:18:51 i was a bit lost in the session 17:18:59 as well, I have to admit 17:19:24 so maybe we don't worry about that today 17:19:31 yeah, 17:19:39 let's let people (i.e. melwitt) stew on options a bit 17:19:42 we do have some other things that need to happen for cells v2, like the instance sort/filter spec alex_xu has up 17:19:54 https://review.openstack.org/#/c/388518/ 17:20:12 I will look at that when we're done here 17:20:18 ok 17:20:19 cool 17:20:23 that's all i have 17:20:28 there isn't anything in the allocations table yet, right? I don't know how that all works 17:20:28 mriedem: have you looked at it? 17:20:36 melwitt: there is 17:20:37 melwitt: and it's what we need 17:20:50 melwitt: but we're not supposed to query it directly from the api, even though we can 17:20:56 dansmith: not yet 17:20:57 okay. I set up a devstack yesterday and didn't find anything in there so there must be something else I have to do 17:20:58 it's on my growing list 17:21:05 mriedem: ack 17:21:12 melwitt: you didn't have placement enabled 17:21:14 melwitt: right 17:21:22 melwitt: which is required for ocata, but probably not defaulted in newton yet 17:21:22 placement-api needs to be in enabled services 17:21:26 er, defaulted in devstck 17:21:27 ah, okay. I'll redo that then 17:21:40 *enable it then 17:21:49 yeah, you need to enable placement-api service 17:21:51 melwitt: http://git.openstack.org/cgit/openstack-infra/project-config/tree/jenkins/jobs/nova.yaml#n51 17:21:54 is the placement job config if that helps 17:21:57 oh snap 17:22:07 thanks 17:22:30 which reminds me i need to get on the scheduler boyz 17:22:32 okay so sounds like that's most of the status.. good progress and inertia on the important things 17:22:35 about plans to make that required in ocata 17:22:42 anything else we need to discuss here 17:22:42 ? 17:22:45 nope 17:23:03 going once 17:23:15 there is the consoleauth open question still I guess (sorry) 17:23:27 melwitt: say more things 17:23:42 I don't remember seeing PaulMurray at the summit 17:23:52 I don't think he was there, at least that I saw 17:24:08 but basically for the console token auth there's an upcall that happens to the consoleauth service that's assumed to be running at the API cell 17:24:21 over rpc 17:24:48 and PaulMurray in the past had some plans to deprecate that service and replace it with some auth that happens in the DB 17:24:53 directly, I think 17:25:14 hmm, an upcall really? 17:25:17 so if that were to be deprecated we could avoid dealing with an rpc call up to the API cell 17:25:18 I don't think I know about that 17:25:38 I liked his auth-in-the-db thing anyway, 17:25:52 so that seems good 17:26:28 I'll re-dig up some details to put on one of our etherpads so you can see in detail what I'm talking about. I looked into it several weeks ago and don't remember everything now 17:27:03 okay cool 17:27:15 anything else? 17:28:09 okay then 17:28:12 #endmeeting