21:00:13 #startmeeting nova 21:00:14 Meeting started Thu Mar 15 21:00:13 2018 UTC and is due to finish in 60 minutes. The chair is melwitt. Information about MeetBot at http://wiki.debian.org/MeetBot. 21:00:15 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 21:00:17 The meeting name has been set to 'nova' 21:00:20 o/ 21:00:24 hello peeps 21:00:26 o/ 21:00:27 \o 21:00:40 @/ 21:00:43 o/ 21:01:02 #topic Release News 21:01:19 #link Rocky release schedule: https://wiki.openstack.org/wiki/Nova/Rocky_Release_Schedule 21:01:26 next milestone is r-1/spec freeze on Apr 19 21:01:53 I was thinking we'll do a spec review day, will mention that in reminders 21:02:06 (it's on the agenda, I mean) 21:02:17 anyone else have anything for release news? 21:02:39 okay 21:02:44 #topic Bugs (stuck/critical) 21:02:55 no critical bugs in the query 21:03:04 #link 47 new untriaged bugs https://bugs.launchpad.net/nova/+bugs?search=Search&field.status=New 21:03:33 untriaged bug count is up, which I think is not too surprising because of the PTG and everyone being busy with that 21:03:56 need to get back to triaging those down 21:04:01 #link bug triage how-to: https://wiki.openstack.org/wiki/Nova/BugTriage#Tags 21:04:26 ^ instructions on how to do bug triage. first tag with a category, then set importance/validity 21:04:35 #link untagged untriaged bugs: https://bugs.launchpad.net/nova/+bugs?field.tag=-*&field.status%3Alist=NEW 21:04:54 we have 12 untagged bugs, so people have been good about applying categories to bugs 21:05:17 but we need more people to dive into an area and set validity and importance on bugs in the area they're familiar 21:05:25 #help tag untagged untriaged bugs with appropriate tags to categorize them 21:05:31 #help consider signing up as a bug tag owner and help determine the validity and severity of bugs with your tag 21:05:52 if you're a bug tag owner, please do take a look and see if there's anything new to triage for your bug tag 21:06:12 #link check queue gate status http://status.openstack.org/elastic-recheck/index.html 21:06:49 today there were problems with the latest zuul security fixes, so zuul was having POST_FAILUREs 21:06:58 that has recently been resolved, so it's safe to recheck your patches now 21:07:06 #link 3rd party CI status http://ci-watch.tintri.com/project?project=nova&time=7+days 21:07:26 that link isn't opening for me 21:08:03 but from anecdotal experience, it seems like 3rd party CI has been okay for the most part 21:08:09 anything else on bugs or gate status? 21:08:23 so, 21:08:29 I haven't seen the xen driver vote in a while, 21:08:41 including on those xen patches for live migration (I don't think they have a live test, but still) 21:09:11 hm, good point 21:09:22 being that we have xenapi driver work slated for rocky 21:09:30 I mean, there are some green checks there, 21:09:32 so I dunno 21:09:46 but it might be worth figuring out if there's something going on there 21:09:56 those guys aren't around during these hours though 21:10:03 anyway, just another gut data point 21:10:18 i tend to only look for virt driver specific CI on virt driver specific patches 21:10:32 but i will hold up a +1/+2 until i see it 21:10:42 yeah, let me look into it and reach out to the xenapi peeps to make sure we have CI going for it 21:10:51 right, I noticed it on the xen patches 21:10:56 or rather, didn't notice it 21:11:18 #action melwitt to look into xen driver CI and reach out to the xen people if needed 21:11:31 cool, thanks for bringing that to attention 21:11:45 anyone have anything else on bugs or gate status or 3rd party CI? 21:12:06 #topic Reminders 21:12:13 #info Rocky PTG nova session summaries are being sent out to the dev ML this week and will continue the following week if needed 21:12:42 I'm writing up the non-placement PTG session summaries and sending them to the dev ML. will continue next week if I don't finish them this week 21:12:52 thanks for doing this 21:13:15 big thank you to jaypipes, mriedem, and cdent and anyone else who contributed to the placement summary 21:13:41 #help Need one or two volunteers to help with the 40 minute Nova on-boarding session at the summit 21:13:55 I did it the last two times, 21:14:05 the most recent being not as useful it seemed 21:14:09 so maybe it's me :) 21:14:34 it was probably me since that was my first time helping with an onboarding session :| 21:14:42 no no no, 21:14:46 you were all equally bad 21:14:51 nah, it had to be me 21:14:53 aw thanks mriedem 21:15:06 edleafe: i was implicitly including you 21:15:32 so maybe we should have some newer people do it 21:15:45 I'll be glad to be in the room 21:15:51 I'll be one of the speakers, so if you'd like to help out with onboarding, please let me know and I'll add you to the list so I can tell kendall 21:16:10 I'll be busy prepping for the placement talk with efried 21:16:25 Was this the session where sdague showed slides with log extracts and talked through them? 21:16:45 yes 21:16:45 and mriedem made fun of the PowerVM driver? 21:16:46 in boston 21:16:51 i did? 21:16:57 I did 21:17:04 I don't think mriedem was in there, IIRC 21:17:11 heh, i was for part of it 21:17:18 our collective memories suck 21:17:20 right but not up front? 21:17:21 I was only in the room for the last 10 or 20 minutes so I missed that 21:17:22 anyway 21:17:26 I definitely made fun of powervm 21:17:30 once I knew efried was in there 21:17:31 Anyways... If the timing works out, I don't mind being in the room to field questions. 21:17:45 Not sure how much use I'll be, but maybe able to give a noob's perspective. 21:17:50 i volunteer efried to be up front 21:17:52 +1 for efried being up front 21:17:53 vay 21:18:04 +gibi 21:18:07 there 21:18:10 it's a party 21:18:12 that would be a good combo 21:18:17 if gibi is in 21:18:20 Hah, poor gibi_vacation 21:18:23 cool. well, fwiw I think it would help to have newer perspectives when giving an onboarding session 21:18:51 alright 21:18:52 Based on the probably-faulty assumption that I actually understand anything at this point. 21:19:04 efried: I'll be there to point out anything you get slightly wrong 21:19:05 don't worry 21:19:06 pff, you understand stuff 21:19:17 oh, THANK you dansmith. Then I'm in. 21:19:21 "actually efried, that's a CAPITAL F" 21:19:38 cool, I think it will go well :) 21:19:39 I'll give you a capital F... 21:19:42 Heh, the people in the audience won't know if you screw up 21:19:53 'cept when dansmith points it out 21:20:00 Okay, okay. 21:20:19 * efried has thick ears and can take it. 21:20:40 I'm gonna try to prepare some stuff this time. apparently there's a format that the foundation and upstream institute would like if we used, so I'm gonna learn about that and see if/how we can incorporate some of it to have on-hand depending on what a poll of the room decides they'd like to hear 21:21:08 anything else on onboarding? 21:21:31 #info Rocky spec freeze is about a month out, Apr 19. Need to be thinking about when to have a Rocky spec review day. melwitt will post to the dev ML to gather input about what date we should schedule it 21:22:12 we've had spec review days in the past, so I was thinking we should do one for rocky. only thing not sure about is the date we should pick, so I'll send email to the dev ML to gather input on when we should do it 21:22:31 #link Rocky Review Priorities https://etherpad.openstack.org/p/rocky-nova-priorities-tracking 21:23:24 ^ looks like people have been using the priorities etherpad, which is great. please keep using it as your place to find things to review and add your items where applicable 21:23:50 there are sections for non-priority approved blueprints, trivial bug patches, the various virt drivers, subteams, osc-placement, and so on 21:24:01 which leads into ... 21:24:06 #info We're going to be experimenting with a new non-priority blueprint tracking a reviewing process this cycle: "runways" and melwitt is writing up a rough draft on how it will work and gather input, then document it and communicate it to the dev ML 21:24:40 I feel like if we're going to do this, we need to be doing it like yesterday 21:24:40 the "runways" idea has been talked about in the past, many cycles ago, and this cycle we're going to try it out 21:24:45 I dunno how other people feel, 21:24:48 dansmith: ++ 21:24:51 but the longer we wait the harder it'll be 21:25:01 I kinda hoped to have more closure on the idea at ptg 21:25:08 like, a plan 21:25:19 I agree. I'm sorry it didn't get started earlier but I was on PTO last week and this week I'm writing a bunch of session summaries 21:25:41 If there are open questions, perhaps we can talk through during Open Discussion here? 21:25:52 oh, I see. well, I hope it will be pretty straightforward. I'm basically going to write up my understanding of what's on the PTG etherpad about it 21:26:20 I don't want it to get stuck in committee but just wanted to keep it open to some amount of feedback before just going ahead, if that makes sense 21:26:26 right, so, 21:26:35 I think we just need to pick some details and run with them 21:26:46 we can discuss some more, but we either need to do it or punt to next cycle I think 21:26:58 agreed, it's not going to be perfect and I'm 100% expecting this will be, just do something and adjust it as we go 21:27:30 I didn't mean for it to sound like I want to spend a lot of time designing it. I was thinking a week at most 21:28:00 I just heard dansmith volunteer to have it written up by EOD. 21:28:09 no, melwitt already said she was doing it 21:28:12 which is cool, 21:28:32 if dansmith is willing to do it, that would be great. I just figured no one else wanted to do it 21:28:44 * efried puts away pokeystick and stfu 21:29:26 so if someone else wants to write up an etherpad, please feel free :) else, I'll do it and send it out and get quick feedback and then we roll with the plan and adjust along the way 21:30:13 anything else on runways? 21:30:30 don't run-away from reviewing them? :D 21:30:40 heh 21:30:41 claudiub|2: you had one good one today, don't push it 21:31:02 #topic Stable branch status 21:31:10 #link stable/queens: https://review.openstack.org/#/q/status:open+project:openstack/nova+branch:stable/queens,n,z 21:31:49 several backports there but looks like they're getting reviews 21:31:58 https://review.openstack.org/#/q/status:open+project:openstack/nova+branch:stable/pike,n,z 21:32:11 same with pike ^ 21:32:17 #link stable/ocata: https://review.openstack.org/#/q/status:open+project:openstack/nova+branch:stable/ocata,n,z 21:32:34 #link EOL is EOL https://review.openstack.org/#/c/548916/ 21:32:49 I was about to ask about that 21:33:01 more like EOL is dumpsterfire 21:33:37 okay, so stable branches are to remain open for as long as reasonably possible 21:33:57 yar 21:34:13 people will come help as a result, 21:34:16 and everyone will be happy 21:34:29 well, there ya go 21:34:30 and you get a pony 21:34:30 and you get a pony 21:34:31 and you get a pony 21:35:21 okay, well, that seems okay. sounds like it means things go EOL when things are broken and too high cost to keep the branch anymore 21:35:21 we'll cut another release on stable, 21:35:25 once mnaser's reverts are merged 21:35:55 except we can't delete branches that are broken and in between branches that work, right? 21:35:56 cool 21:35:59 Ill keep following them through 21:36:22 dansmith: I would think not but AFAIK that hasn't happened before. (has it?) 21:36:34 it can't happen with the current policy 21:37:15 anyway, I'm just throwing stones, move along :) 21:37:20 if Pike stops working and we delete it, we'd have to delete ocata too 21:37:23 well, what I mean is when we've had queens/pike/ocata I have only seen a job break on ocata, not anywhere in the middle. but I see your point that the window for it being observable has been really small 21:37:52 mm, I think it's broken in the middle before, but it certainly *can* happen 21:37:54 more releases, more potential 21:38:03 fair 21:38:32 okay, so something for everyone to be aware of 21:38:38 #topic Subteam Highlights 21:38:47 dansmith: cells v2 21:39:01 we had a meeting, discussed a bunch of in-progress stuff, 21:39:21 including bugs and features, lots of stuff up from tssurya, pretty good consensus on a new bug that got filed 21:39:28 everything's good I think 21:39:38 cool, thanks 21:39:42 edleafe: scheduler 21:39:48 Some brief discussion about the spec for supporting traits in Glance to give operators more control over where their images are run. 21:39:51 The 'member_of' API change was holding up the Request Filter series; that patch is now complete and awaiting Tempest recheck. 21:39:54 We got somewhat risqué and talked about Forbidden Traits. 21:39:57 cdent updated us on the patch moving ResourceProvider objects to placement. 21:40:00 We ended the meeting deciding what should be authoritative about traits. The issue was that the virt driver could set traits, and then the operator could unset them. The driver would then re-set them, and this could go on forever. 21:40:04 The discussion spilled over into -nova, where we decided that there would be a set of traits defined that a compute node would have domain over. Operators can then set any others without worrying about them getting overwritten. 21:40:08 That's it 21:40:23 cool, thanks 21:40:33 there aren't any notes for the notifications subteam from gibi 21:40:42 anything else on subteams? 21:40:55 #topic Stuck Reviews 21:41:07 nothing on the agenda. does anyone have anything for stuck reviews? 21:41:29 #topic Open discussion 21:41:40 there are several specless blueprints on the agenda 21:41:48 first up, dansmith for CPU weigher 21:41:58 yeah, so stephen wants this approved, 21:42:09 a new weigher for spreading due to cpu allotment 21:42:18 weighers are good, I don't imagine this is controversial 21:42:21 so I imagine we should approve it 21:42:51 yeah, seems like the addition of a new weigher should be fine 21:43:39 anyone else have opinions about a new CPUWeigher? 21:44:12 okay, I'll ask about it in -nova later when more people are around 21:44:25 next is from mriedem, adding granular policy code to placement 21:44:27 -ops list would be more useful, but 21:44:44 we agreed at the ptg to do granular placement policy 21:44:47 just don't know if i need to write a spec 21:44:48 mriedem: we have ops that want it, in case it's not obvious :D 21:44:53 psh, 21:44:57 i don't trust red hat customer ops 21:45:15 i thought stephen just wanted it for his own pet scheduler 21:45:38 the bp is really old, from 2014 and it was from someone else, Charlotte Han 21:45:39 this placement granular policy thing, all rules would default to admin-only as today 21:46:23 i personally don't care about the cpu weigher 21:46:25 but, 21:46:25 I think there was wide agreement about this, 21:46:33 i could see a bunch of people wanting to add more weighers 21:46:35 I don't think we need a spec for it.. it's mechanical 21:46:35 just cuz 21:47:01 mriedem: weighers that work on our fundamental data structures are good right? 21:47:03 I guess I don't see what's wrong with adding more weighers 21:47:15 weighers to do crazy things, totally agree 21:47:22 they are all on by default aren't they 21:47:31 by i guess the default weight factor isn't? 21:47:32 wanting to land on the least-overcommitted node is pretty common 21:47:49 so by default, the weigher doesn't do anything 21:47:51 rihgt? 21:47:53 are they all on by default? 21:47:58 pretty sure 21:48:28 weight_classes = nova.scheduler.weights.all_weighers 21:48:48 but you tweak them on or off 21:48:55 whether you want to pack or spread with this for example 21:49:00 by the weight option i think 21:49:12 e.g. ram_weight_multiplier = 1.0 21:49:18 yeah 21:49:30 and we already have weighers for ram and disk 21:49:33 this exact point by the way.. we have the for things like ram 21:49:33 so sure, add cpu 21:49:34 right 21:50:33 okay, so sounds like we agree CPUWeigher is cool to approve, and so is the granular policy, we agreed about it at the PTG and the question was whether it needs a spec, and it doesn't need a spec 21:50:35 left a comment in the bp, i'm good with it 21:51:13 okay, next on the agenda is from gibi, specless bp for including traceback in versioned error notifications 21:51:20 I think we agreed on this at ptg too, 21:51:22 https://blueprints.launchpad.net/nova/+spec/add-full-traceback-to-error-notifications 21:51:32 or at least I don't think I heard anyone say anything about why we shouldn't have this right? 21:51:58 yeah, I don't remember any disagreement 21:52:23 gibi was going to find out why it wasn't in the versoined notifications to start, 21:52:32 which was discussed in the original versoined notification spec, 21:52:46 and people were just worried about tracebacks and unstructured blobs, 21:52:50 but we have that in the rest api today so meh 21:52:54 and it's in the legacy notificatoins 21:52:57 seems like a non-issue 21:53:05 it's a for-humans text field 21:53:07 it seems fine to me 21:53:22 yeah, I think he explains why it wasn't originally in the notifications in the bp 21:53:22 typos are not intentional, just typing fast 21:53:51 okay, so we're cool to approve that one too then 21:54:15 next is specless blueprint for xenapi aggregate removal (and related alternate pool implementation) from dansmith 21:54:22 https://blueprints.launchpad.net/nova/+spec/live-migration-in-xapi-pool 21:54:23 this also was discussed at ptg, 21:54:30 and I think we should obviously approve it 21:54:39 it will end up removing the aggregate upcall from the xen driver, 21:54:46 and moves to an actually-usable pool model for them 21:54:54 so, fixes workiness and removes an upcall, 21:54:59 all with just driver maintenance 21:55:03 yeah. the host aggregate upcall thing is broken today IIUC, so they need to do this work to fix it 21:55:10 they don't have the up-call removal patch up yet right? 21:55:16 so is the actual aggregate pool as far as I understand 21:55:18 mriedem: they don't 21:55:27 mriedem: they're implementing the new scheme and will follow that with removing the other stuff 21:55:35 makes for less churn anyway 21:55:39 so I think it's appropriate 21:55:41 yeah multiple patches is good 21:55:46 i'm happy with this 21:55:50 jianghuaw_ and co do nice work 21:56:02 okay, so we will approve that one 21:56:14 next is specless blueprint for versioned notification when removing members from a server group from takashin 21:56:22 https://blueprints.launchpad.net/nova/+spec/add-server-group-remove-member-notifications 21:56:34 mriedem and I have commented on the bp already 21:57:30 we agreed it's not useful to notify for the quota recheck removing a server group member 21:57:54 i want to see the PoC before we approve the bp 21:58:05 "As for the other scenario (deleting a server) that is more straight-forward but I'd like to see the proposed code for that, since it would likely involve needing to lookup the group that the server is in *before* we delete it, since there is no specific method to remove a server group member when an instance is deleted, it's all implicit." 21:58:31 yes, so we'll hold off on that one and let takashin know to propose the PoC 21:58:39 1.5 minutes 21:58:41 last is SpecLess blueprint (not sure if it's required, but I filed one anyway): libvirt-cpu-model-extra-flags from kashyap 21:58:45 yeah 21:59:09 yeah this is important for several reasons, but meltdown specifically 21:59:21 there's agreement on the code already 21:59:26 tests notwithstanding 21:59:38 okay, cool. so that one should be approved 21:59:55 we have like 30 seconds left so have to wrap up now. thanks everyone 21:59:56 if it's a bp, 21:59:59 #endmeeting