14:00:05 #startmeeting nova_scheduler 14:00:06 Meeting started Mon Jul 11 14:00:05 2016 UTC and is due to finish in 60 minutes. The chair is edleafe. Information about MeetBot at http://wiki.debian.org/MeetBot. 14:00:07 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 14:00:09 The meeting name has been set to 'nova_scheduler' 14:00:17 anyone here to talk about the scheduler? 14:00:18 <_gryf> o/ 14:00:19 o/ 14:00:23 o/ 14:00:36 o/ 14:01:14 Let's wait another minute 14:01:14 * bauzas refuels coffee with one hand and waves with the other 14:01:31 o/ 14:01:39 bauzas: terrible ending yesterday, no? 14:02:34 edleafe: graaah 14:02:56 morning everyone sorry for being late (in US west coast time zone for this and next week) 14:03:18 jaypipes: good segueway... 14:03:30 #topic Skip next week's meeting? 14:03:43 Many will be en route to the midcycle 14:03:45 * mriedem lurks 14:03:50 * edleafe included 14:04:02 Of course, someone else could run the meeting 14:04:08 or we could just skip it 14:04:13 thoughts? 14:04:22 ++ 14:04:27 +1 for skippinh 14:04:49 most meetings will skip next week, if not all 14:04:57 edleafe: yeah, skip that one I think. 14:05:34 #agreed Skip next week's meeting 14:05:38 Moving on 14:05:51 #topic Mid-cycle discussions 14:06:08 The only new one is Nested Resource Providers 14:06:11 #link https://etherpad.openstack.org/p/nested-resource-providers 14:06:26 jaypipes: anything to say about that now? 14:06:33 Or should we save it for F2F? 14:07:36 * edleafe realizes that was supposed to be in 'Opens'. 14:07:50 well, f2f is good to me 14:08:13 given we could first discuss on the current implementations and see where we are before discussing on the next items 14:08:15 midcycle etherpad is here for reference https://etherpad.openstack.org/p/nova-newton-midcycle 14:08:25 edleafe: f2f 14:08:35 i expect a lot of discussion at the midcycle will be about what needs to get done in newton, 14:08:37 like the placement api 14:08:40 #link https://etherpad.openstack.org/p/nova-newton-midcycle 14:08:49 mriedem: yeah 14:08:52 and looking toward host capabilities for ocata 14:09:28 The other was Separating qualitative requirements from quantitative in a request 14:09:31 #link https://review.openstack.org/313784 14:09:59 That I think ties into alex_xu's post: 14:10:01 ResourceProviderTags 14:10:01 #link http://lists.openstack.org/pipermail/openstack-dev/2016-July/099032.html 14:10:09 edleafe: thanks 14:10:17 mriedem: I'm preparing content for next week to discuss all those things as well as the Ironic aspects of the resource-providers work and nested resource providers (which handle the NUMA stuff) 14:10:31 edleafe: ++ 14:10:45 I and Yingxin works on this propose, as part of solution for Capabilities. Propose using Tags to manage Capabilities with ResourceProvider 14:11:02 mriedem: got alaski, cdent, tdurakov and rpodolyaka here in CA this week for a sprint. 14:11:16 mriedem: we will all be up in Hillsboro next week 14:11:26 alex_xu: ++ 14:11:26 Cool, so let's make sure that we are familiar with the issues before we meet 14:11:42 edleafe: agreed. lots of reading and writing to do this week. 14:11:48 jaypipes: would be happy to help review / add to content before midcycle 14:11:51 jaypipes: looking for your comment :) 14:12:21 alex_xu: I did read your proposal, looks interesting 14:12:23 yes, the main idea is to manage capabilities under the scope of resource providers 14:12:24 jaypipes: if you want everyone to level set before the midcycle, i'd suggest a ML info thread later in the week 14:12:25 edleafe, mriedem: before Thursday, I will send a ML post with summary of latest resource-providers (quantitative) progress. Can you do same for qualitative side? 14:12:32 bauzas: thanks 14:12:36 mriedem: heh, beat me to it :) 14:12:37 jaypipes: you read my mind 14:12:58 mriedem: yeah, will do that after some meetings today 14:13:02 jaypipes: sure. I'll coordinate with alex_xu 14:13:12 jaypipes: alex_xu made a nice proposal about using the existing resource_providers API for the qualitative resources too 14:13:16 * johnthetubaguy looks forward to the email 14:13:30 mriedem: I finally figured out NUMA resources and modeling all that jazz while in the shower this morning (yes I know, weird...) 14:13:31 devil is in the details of course, but I like the idea 14:13:34 <_gryf> jaypipes, the NUMA case is similar to FPGA case - where whole FPGA can be resoucre, and accelerator which it contains might be one at the same time 14:13:39 #action jaypipes to post summary of resource provider state later this week 14:13:42 edleafe: cool 14:14:07 #action edleafe and alex_xu to post to ML about qualitative status 14:14:11 bauzas: is that the ResourceProviderTags proposal? 14:14:24 indeed 14:14:51 I and Yingxin will try to work on the spec and work out more detail, to figure out more devil :) 14:14:53 _gryf: perhaps, will continue to thnk about the FPGA action plans in relation to resource providers. 14:15:00 aren't resource providers made of aggregates which have metadata (tags)? 14:15:03 bauzas: cool, will read today. 14:15:24 basically, a resource provider can expose tags 14:15:27 alex_xu: link to that patch? 14:15:47 jaypipes: we didn't have spec yet, just about email http://lists.openstack.org/pipermail/openstack-dev/2016-July/099032.html 14:15:58 gotcha 14:16:00 <_gryf> jaypipes, I can share with you what I already know, although I'm not going to participate midcycle… 14:16:28 _gryf: sounds good. or better yet email the ML and get the conversation going there, yeah? 14:16:30 mriedem: yea, but aggregate for a group host 14:16:42 <_gryf> jaypipes, yup, that would be great 14:16:53 _gryf: or pass on info to pawel who I will see on Friday in Portland ;) 14:16:55 _gryf: well, I would certainly wait for a bit of the resource providers infrastructure in place before including FPGA, tbh 14:16:57 mriedem: and aggregate metadata is not very easy to manage 14:17:06 <_gryf> jaypipes, I will :) 14:17:43 _gryf: no, really, an ML post would be better :) 14:18:20 mriedem: that means if we want to attatch a capability to a host, we don't need to create a host aggregate 14:18:54 <_gryf> bauzas, that's the plan, but I'd like to be sure, that we don't close for "special cases". 14:19:10 _gryf: gotcha 14:19:11 Yingxin: yeah, aggregates of 1 don't make a lot of sense :) 14:19:14 Yingxin: ++ 14:19:28 <_gryf> jaypipes, sure. 14:19:29 so would we have resource providers of 1? 14:19:31 yup, we can also attatch a capability directly to a resource pool 14:19:36 One-man Wolfpack. 14:19:50 i'll just shut up and read specs eventually 14:19:57 mriedem: lol 14:20:08 mriedem: sure. A host is a RP of RAM, CPU, disk, etc 14:20:12 BTW, alex_xu and Yingxin will you be in Portland? 14:20:27 edleafe: not necessarily disk ;) 14:20:27 i will be there. 14:20:34 well, 14:20:46 there are a class of things i'd expect the virt driver and nova to provide as capabilities, not the operator 14:20:51 e.g. i support file injection 14:20:52 better yet: who here will *not* be at the midcycle? 14:20:52 I'm supporting him in china :) 14:20:59 mriedem: yep, zactly. 14:21:05 Yingxin: :) 14:21:06 Yingxin: ok! :) 14:21:08 e.g. i support volume multiattach, or device tagging 14:21:18 ok 14:21:18 mriedem: yeah, more auto configuring of somethings is going to be really handy 14:21:27 <_gryf> edleafe, i'm not going to participate in midcycle 14:21:45 mriedem: yep, those are the "standardized" host capabilities I refer to in this spec: https://review.openstack.org/#/c/309762/ 14:22:00 edleafe: sorry, got late 14:22:04 mriedem: of course, there's many more non-standardized capabilities, as edleafe would remind you ;) 14:22:18 jaypipes: :) 14:22:31 I am not sure if I will be there yet (to be precise, I have nothing booked yet), still hoping I can get passport etc sorted in time 14:22:41 johnthetubaguy: :( 14:22:50 johnthetubaguy: ugh. 14:23:15 yeah, getting messed around by house builders too, but slowly getting sorted 14:23:31 oh grah 14:24:12 * johnthetubaguy makes secret British anger face 14:24:44 johnthetubaguy: that's indistinguishable from normal British face, right? :-) 14:25:02 <- is British 14:25:15 mgould: possibly slightly different eyebrows I guess? 14:25:36 johnthetubaguy: I wasn't expecting the Brexit to apply so fast :p 14:25:58 OK, anything else we need to discuss in advance of the midcycle? 14:25:58 hahaha 14:26:02 bauzas: heh 14:26:10 edleafe: not from my end. 14:26:25 nothing urgent to review ? 14:26:38 I'm still in the weeds of reviewing the placement API 14:26:46 i was going to say, placement API 14:26:49 bauzas: nothing added to the agenda 14:26:55 i only spent some time on that a couple of weeks ago 14:27:15 * mgould would like to ask about extra_specs matching; should I wait for AOB? 14:27:19 but if we're going to have a placement API in newton, we should make sure we're happy with what cdent has proposed by the time we leave the midcycle 14:27:31 mriedem: +1 14:27:33 yeah, and having the placement a bit different from the other bits I know makes my review time a bit longer :) 14:27:50 but I guess, no surprise here :) 14:28:15 mriedem: that's why I'm tempted to only review the bits and make sure we can fast iterate during the sprint if needed 14:28:20 mgould: Opens is next 14:28:22 I'd certainly love some consensus on those bits 14:28:32 edleafe: thanks! 14:28:44 bauzas: yeah, I have the same issue 14:28:55 triggering +W on such huge changes like a whole new WSGI stack is chilling me 14:29:12 bauzas: I'm assuming that it won't be perfect on the first try. :) 14:29:25 yeah, but that's an API, right ? :) 14:29:36 microversions FTW 14:29:39 bauzas: the good thing is that it's not in use yet, so we won't break anything 14:29:40 like the ones we contractually assume as being stable ? :) 14:30:03 sure, I'm not saying it's impossible, just that's it's a big event :) 14:30:15 Can we start with microversion 0.1, indicating that it *isn't* stable yet? 14:30:26 do we have the framework for this ? 14:30:36 I think while its an internal API, I feel happier about raising the min_version of the placement API, wibble upgrade worries 14:31:21 mgould: let's discuss that in #openstack-nova after meeting, yeah? 14:31:36 jaypipes: cool 14:31:40 Let's move on 14:31:42 jaypipes: sure, thanks 14:31:47 #topic Opens 14:31:56 Other services directly accessing the Scheduler and/or request database 14:31:59 Watcher has updated their spec for integrating with Nova: 14:32:01 #link https://review.openstack.org/#/c/329873 14:32:08 Other services directly accessing the Scheduler and/or request database 14:32:11 Watcher has updated their spec for integrating with Nova: 14:32:13 #link https://review.openstack.org/#/c/329873 14:32:24 * edleafe fails copy/paste 14:32:40 They plan to submit a Nova backlog spec before the midcycle 14:32:40 We need to clarify the projected availability of the Placement API, and strategies to use until that is available. 14:33:02 Anyone from Watcher here to add to this? 14:33:16 Watcher will have its midcycle in Hillsboro next week 14:33:32 Yep, same time and place as Nova 14:33:53 not the same building tho 14:34:08 we hope to discuss with you guys about this spec 14:34:31 honestly i want to spend a very minimal amount of time in hillsboro talking about this 14:34:33 edleafe: honestly, I can try to get to reading that Watcher spec but its going to be a stretch. 14:34:38 it would be great that you had a look to this before the midcycle 14:35:08 wow, two "honestly" replies :) 14:35:17 honestly for serious 14:35:21 mriedem : do you think we could submit a Nova backlog spec instead of this ? 14:35:30 but we could spend all of tuesday about scheduler and RP 14:35:39 oh gosh 14:35:49 I dont want to take your midcycle time, just trying to find a way to explain our issue 14:35:51 spending one full day about scheduler <3 14:35:52 acabot_: submit whatever for review, but i'm just saying, we already have a full list for next week https://etherpad.openstack.org/p/nova-newton-midcycle 14:35:55 edleafe: they need the placement API not only to expose all the resource providers stuff (our current Newton priority) but also to have it offer a select_destinations() replacement. That ain't gonna happen in Newton for sure. 14:35:59 and a lot that's not close to getting done for newton 14:36:05 * edleafe thought we would spend all three days on scheduler and RP :) 14:36:09 * bauzas was ironic 14:36:32 we also have cinder/neutron stuff on wednesday, and i figured a hodge podge on thursday 14:36:48 mriedem : no pb, I fully understand 14:36:49 mriedem: isn't scheduelr a hodge-podge? ;P 14:36:53 anyway, my point is, i don't want the watcher spec to be a distraction for newton 14:36:59 jaypipes: oh, for sure. The big question is the guesstimate of when a placement API for that would be available 14:37:17 edleafe: we won't know that until after the midcycle probably 14:37:18 edleafe: you know how I feel about guestimates :) 14:37:20 mriedem : its definitely not a newton priority by the way ;-) 14:37:25 mriedem: well, I wouldn't certainly hope more than just a brief discussion on the Watcher issue, but a short overlook would be interesting tho 14:37:45 bauzas: sure, unconference at the midcycle 14:37:54 mriedem: that sounds good to me 14:37:56 bauzas : +1 14:38:00 a perfect slot for this 14:38:01 mriedem: yeah, unconference sounds right 14:38:20 acabot_: I can be bribed with cookies. just FYI. 14:38:25 mriedem : any agenda available for unconference ? 14:38:29 acabot_: and bauzas can be bribed with cheese. 14:38:39 acabot_: Or sushi. jaypipes LOVES sushi 14:38:41 jaypipes : ;-) 14:38:44 lol 14:38:49 acabot_: no, i wasn't being literal about unconference 14:38:53 our etherpad of items is here https://etherpad.openstack.org/p/nova-newton-midcycle 14:38:59 sushi, cheese and sushi would be great ! 14:39:03 we do loose scheduling of the agenda for the midcycle 14:39:08 which is the beauty of a midcycle 14:39:35 mriedem : thx 14:39:46 Next up 14:39:55 mgould: extra_specs matching? 14:40:31 oh, I thought we were discussing that in the normal channel afterwards 14:40:38 but I can ask now 14:40:46 go for it 14:41:30 comments in the implementation suggest should match capabililties which are lists 14:41:44 I think mgould wants to discuss about http://lists.openstack.org/pipermail/openstack-dev/2016-July/098936.html 14:41:47 I've proposed patches to test this 14:41:49 yes, exactly 14:42:09 honestly, that sounds an implementation detail, but I could be wrong 14:42:12 since then, rloo found places in nova wehre the argument is force-stringified before matching 14:42:20 which means you get substring matching rather than sublist matching 14:42:29 this is a bug, right? 14:42:32 mgould: so the thing is, you want to define an oslo lib, right? 14:42:40 see final coment on https://review.openstack.org/#/c/339596/ 14:42:43 bauzas: yes 14:43:01 the oslo.lib people are saying "this doesn't enforce input types, we don't like it" 14:43:12 mgould: so, my take on this is that yes, strings as arguments are doom 14:43:19 mgould: for a lib 14:43:29 OK 14:43:37 but within nova, you don't care? 14:44:01 mgould: given we only have a few callers, it's certainly a nice move, but not really useful 14:44:17 see also https://github.com/openstack/nova/blob/90ec46c87fbf572805c7758377431e26c93622a4/nova/scheduler/filters/compute_capabilities_filter.py#L87, which stringifies the capability, giving substring matching 14:44:37 mgould: I'm pretty sure you could clean that up :) 14:44:43 OK 14:44:57 mgould: yeah, that looks wrong 14:45:01 mgould: but like I said, here I think you should focus on the lib 14:45:01 I'll submit a patch to fix that, then 14:45:23 right, OK 14:45:30 Any other opens? 14:45:43 mgould: I will respond on the ML thread. need to read your proposal fully... 14:45:50 jaypipes: thanks! 14:45:54 jaypipes: bauzas: Is there still a plan to move the claims from RTs to schedulers? 14:46:08 Yingxin: ssssssht :p 14:46:31 I'm still thinking of the bp eventually-consistent stuff 14:46:44 Yingxin: :) 14:46:52 Yingxin: there are patches up for having contextmanagers (call it "claims' if you will) in the scheduler codepath 14:46:53 Yingxin: yes. :) 14:46:57 And there's a long post in ML :P 14:47:11 bauzas: which patches are those? 14:47:25 Yingxin: I saw your post, haven't had a chance to digest it yet :( 14:47:31 jaypipes: sec 14:48:17 jaypipes: but that's not the idea of "claiming a resource', that's just for making sure we have a correct infrastructure for distributed writes on a synchronized section 14:48:59 bauzas: which patch is it? 14:49:01 https://review.openstack.org/#/c/262938/ 14:49:08 no real surprise 14:49:53 bauzas: thanks 14:50:07 OK, anything else? 14:50:15 Yingxin: I added a block on the eventual consistent and scheduler claims thing to the agenda. Long-term discussion... 14:50:24 yup, I agree 14:50:55 (on putting it to the agenda if we have time and if mriedem is not killing us for eating all the time) 14:51:11 bauzas: oh that's also my patch 14:51:21 Yingxin: hence my "no surprise here" 14:51:45 hehe 14:52:45 if we're done with this, I'd like to advocate for a long-standing patch 14:53:11 I'd love some reviews with https://review.openstack.org/#/c/244489/ 14:53:30 before calling out of moving claims, we need to make sure we do claim on 14:53:36 *all* our operations 14:53:42 which is not the case atm, see ^ 14:53:56 that reminds me 14:54:05 moshele has a thread in the dev list about cleaning up some RT stuff for pci/sriov 14:54:06 jaypipes, mriedem: I'd love some feedback on this ^ 14:54:14 so they can get a multinode job going for resize with sriov/pci 14:54:24 mriedem: that's certainly a good idea :) 14:54:44 just a reminder though, i haven't dug into it - i reviewed one of the patches awhile back 14:54:51 the "if pci:" idea wasn't probably the best idea we had 14:54:53 also reminder that non-priority FFE is wednesday 14:55:07 for the things that got FFEs 14:55:13 that's a reminder to the cores here 14:55:37 :) 14:55:38 5 minutes 14:55:45 Anythign else? 14:56:11 edleafe: I have setup ready, anything I can test 14:56:22 Yingxin: https://github.com/cyx1231st/nova-scheduler-bench used for setup 14:56:38 Yingxin: thanks for your help 14:56:58 diga_: np! does it work? 14:57:05 bauzas: ++ will do this mronig. 14:57:09 guh, morning. 14:57:13 4th script is not working for me 14:57:29 need some changes in that 14:57:42 will ping error after this meeting 14:57:50 diga_: we can discuss it in openstack-nova later 14:58:00 sure 14:58:05 OK, looking forward to seeing most of you in Hillsboro next week! 14:58:09 #endmeeting