14:00:13 <mriedem> #startmeeting nova
14:00:15 <openstack> Meeting started Thu Jul 13 14:00:13 2017 UTC and is due to finish in 60 minutes.  The chair is mriedem. Information about MeetBot at http://wiki.debian.org/MeetBot.
14:00:17 <openstack> Useful Commands: #action #agreed #help #info #idea #link #topic #startvote.
14:00:19 <openstack> The meeting name has been set to 'nova'
14:00:24 <edleafe> \o
14:00:25 <alex_xu> o/
14:00:25 <efried> \o
14:00:28 <takashin> o/
14:00:29 * ildikov is lurking
14:00:36 <bauzas> \o
14:01:00 <mriedem> alright let's get started
14:01:03 <mriedem> #link agenda https://wiki.openstack.org/wiki/Meetings/Nova#Agenda_for_next_meeting
14:01:09 <mriedem> #topic release news
14:01:14 <mriedem> #link Pike release schedule: https://wiki.openstack.org/wiki/Nova/Pike_Release_Schedule
14:01:21 <mriedem> #info July 27 is feature freeze (2 weeks away)
14:01:33 <mriedem> #info July 17 (next week) is the release freeze for non-client libraries (oslo, os-vif, os-traits, os-brick): https://releases.openstack.org/pike/schedule.html#p-final-lib
14:01:50 * gibi sits in a bit late
14:02:07 <mriedem> i haven't gone digging yet, but if we have any changes to os-vif or os-traits that are going to be dependencies for other work within nova, please let me know asap so i can monitor for a final release
14:02:22 <bauzas> none I saw for os-traits
14:03:03 <mriedem> ok
14:03:10 <efried> I have a question
14:03:13 <mriedem> shoot
14:03:17 <efried> "feature"
14:03:27 <efried> meaning things that affect functionality or change APIs or whatnot?
14:03:34 <efried> Not just "you can't drop code for a blueprint after this date"
14:03:43 <edleafe> efried: as in "bug fixes still ok"
14:03:51 <efried> Cause the service catalog endponit stuff may be tight.
14:04:05 <mriedem> efried: doesn't that rely on a ksa release?
14:04:16 <efried> Yes, we're still waiting on a big pile of stuff from ksa, sta, and the newly-minted os-service-types
14:04:21 <mriedem> ksa is a non-client lib so that would have to get released next week
14:04:33 <mriedem> mordred: ^
14:04:41 <mriedem> efried: which blueprint is that holding up?
14:04:54 <efried> service catalog for endpoints - lemme get the name...
14:05:02 <mriedem> https://blueprints.launchpad.net/nova/+spec/use-service-catalog-for-endpoints
14:05:10 <efried> yuh
14:05:58 <mriedem> so how about we talk with mordred about the state of things after the meeting?
14:06:18 <efried> Sure.  Been working closely with mordred, and I know we're close.  But yes.
14:06:21 <mriedem> #action mriedem and efried to talk to mordred about the final ksa release for next week
14:06:36 <mriedem> remind me if i don't follow up :)
14:06:41 <efried> rgr
14:06:48 <mriedem> #info Blueprints: 65 targeted, 63 approved, 29 completed (+3 from last week)
14:06:55 <mriedem> so we're making some progress,
14:07:20 <mriedem> and there are at least about 5 other bps i'm watching that are very close to done
14:07:41 <mriedem> maybe i'll send those to the ML after the meeting to focus our attention on pushing those over the line
14:07:48 <bauzas> yeah good idea
14:07:59 <bauzas> if some are close to be merged, gtk
14:07:59 <gmann_> mriedem: this one can be closed or after pythonclient change - https://blueprints.launchpad.net/nova/+spec/fix-quota-classes-api
14:08:01 <mriedem> #action send list of close to done bps to ML after the meeting
14:08:14 <mriedem> gmann_: after the client change
14:08:18 <gmann_> ok
14:08:22 <mriedem> i don't consider the api microversion ones done until the novaclient patch is also merged
14:08:46 <mriedem> ok any other questions about the release?
14:08:56 <mriedem> i know it's tight and we have a lot of work yet to do
14:09:02 <mriedem> so just put a pot of coffee on
14:09:19 <mriedem> #topic bugs
14:09:32 <mriedem> there are no critical bugs
14:09:37 <mriedem> #help Need help with bug triage; there are 115 (+6 from last week) new untriaged bugs as of today (July 13).
14:10:17 <mriedem> gate status is mostly ok
14:10:25 <mriedem> there were some things yesterday with zuul but infra sorted it out
14:10:35 <mriedem> no news for 3rd party ci
14:10:51 <mriedem> #topic reminders
14:10:58 <mriedem> #link Pike Review Priorities etherpad: https://etherpad.openstack.org/p/pike-nova-priorities-tracking
14:11:10 <mriedem> i don't know how realistic that is at this point
14:11:22 <mriedem> #link Consider signing up for the Queens PTG which is the week of September 11 in Denver, CO, USA: https://www.openstack.org/ptg
14:11:44 <mriedem> #topic stable branch status
14:11:51 <mriedem> #link stable/ocata: https://review.openstack.org/#/q/status:open+project:openstack/nova+branch:stable/ocata,n,z
14:11:54 <cdent> is there ptg planning etherpad yet?
14:11:56 <mriedem> #link stable/newton: https://review.openstack.org/#/q/status:open+project:openstack/nova+branch:stable/newton,n,z
14:11:58 <mriedem> cdent: nope
14:12:04 <cdent> should we intentionally wait?
14:12:12 <cdent> (so as not to be distracted)
14:12:14 <bauzas> that's very early
14:12:21 <mriedem> cdent: that's why i haven't created one yet
14:12:24 <edleafe> +1 on too early
14:12:27 <cdent>14:12:30 <bauzas> for stable branches, I did a bit of review for ocata
14:12:31 <mriedem> my focus is about hour to hour right now :)
14:12:33 <gmann_> yea lot of time
14:12:46 <mriedem> bauzas: yeah i saw that, thanks
14:12:49 <bauzas> do we need some point release soon ?
14:12:57 <mriedem> i wasn't planning one
14:13:01 <mriedem> we've done several this release already
14:13:13 <mriedem> we'll do another one around the rc1 i suppose
14:13:13 <bauzas> okay, so maybe after pike-3 I guess
14:13:18 <bauzas> yeah ok
14:13:21 <bauzas> so nothing urgent
14:13:29 <mriedem> i don't really have other news except they are finally killing off mitaka
14:13:35 <mriedem> infra is removing jobs, etc
14:13:45 <bauzas> yeah there was a thread
14:13:47 <mriedem> and the stable/mitaka branch is gone
14:14:08 <mriedem> ok moving on
14:14:14 <mriedem> #topic subteam highlights
14:14:28 <mriedem> i'll cover cells v2 as dan is not around this morning
14:14:36 <mriedem> #help Need reviews on quotas https://review.openstack.org/#/c/416521/ and fixing server external events https://review.openstack.org/#/c/445142/
14:14:52 <mriedem> i've spent a good chunk of 2 days reviewing the bottom quotas change
14:14:56 <mriedem> and talking through things with melwitt
14:15:01 <mriedem> so she has some stuff to update in there
14:15:13 <bauzas> the quota thing is a bit frightening :(
14:15:19 <mriedem> that bottom one is the change to start counting instances, which is the last reservable resource we have left
14:15:25 <mriedem> yes so it's complicated
14:15:35 <mriedem> and melwitt has done an awesome job working through this and being patient
14:16:03 <mriedem> the other patch there is another dependency for multi-cell,
14:16:05 <bauzas> sure, my main question is how we could raise our confidence in that patch
14:16:05 <mriedem> and a bug fix,
14:16:09 <mriedem> looks like that needs to be rebased
14:16:16 <mriedem> bauzas: you could review it for one :)
14:16:22 <bauzas> mriedem: I began
14:16:35 <bauzas> I have some comments but nothing uploaded yet
14:16:35 <mriedem> and get it merged so we have time to bake it in and flush out issues
14:17:03 <mriedem> i would love to see someone do a rally comparison before and after that patch to see if performance is impacted much
14:17:23 <mriedem> dtp was looking for things to help with, maybe he could do that
14:17:30 <mordred> mriedem, efried (yes to ksa discussion)
14:17:40 <mriedem> #action mriedem to talk to dtp about running rally against the bottom quotas change
14:17:50 <bauzas> mriedem: could we see the quota patch being tested live ?
14:18:00 <mriedem> bauzas: what does 'live' mean?
14:18:10 <mriedem> it's tested in our ci like everything else
14:18:11 <bauzas> some end-to-end testing
14:18:14 <bauzas> anyway
14:18:24 <bauzas> let's not derail the meeting
14:18:32 <mriedem> the thing that worries me the most about quotas test coverage,
14:18:47 <mriedem> is that tempest uses a separate tenant for every test
14:19:02 <mriedem> so if we start leaking quota, we won't find that in a normal tempest dsvm run
14:19:10 <mriedem> especially for complicated things like resize
14:19:14 <bauzas> :/
14:19:16 <mriedem> so,
14:19:34 <mriedem> something someone could do is pull that change down and do like 20 resizes in a row or someting
14:19:37 <mriedem> and find out if we blow up quota
14:19:59 <mriedem> or, push a DNM patch that does that with tempest
14:20:08 <bauzas> that's the end-to-end testing I had in mine
14:20:10 <bauzas> mind*
14:20:10 <mriedem> just run the resize tests 10 times in a row with the same tenant
14:20:12 <gmann_> mriedem: separate tests you mean separate quota tests?
14:20:23 <bauzas> but I could try to devstack it
14:20:39 <mriedem> gmann_: no the tenant isolation in tempest
14:20:46 <mriedem> quotas are counted against the project
14:20:52 <mriedem> so each test has it's own project and quota
14:21:14 <mriedem> so if we're leaking quota, we won't notice because we aren't using a single tenant through the entire tempest run
14:21:36 <mriedem> gmann_: are you aware of any ci jobs that run tempest without tenant isolation? i know there used to be a periodic job that did that
14:21:54 <mriedem> the novaclient functional tests also do that, fwiw
14:22:08 <gmann_> it used to be a serial job which use single tenant
14:22:09 <mriedem> so we could point a dnm novaclient change at that series and see what happens
14:22:22 <mriedem> the novaclient functional test is a serial job with a single tenant
14:22:25 <mriedem> so we can start with that
14:22:48 <mriedem> #action run a DNM novaclient patch against dan's fleetify devstack change which depends on the quotas series to see if we're leaking quota at all
14:23:07 <mriedem> i still think we need to do something like resize 20 times in a row
14:23:10 <gmann_> mriedem: but we have deprecated that things and we can do via account file which has pre defined tenants to run tempest against
14:23:36 <mriedem> but let's move on - if we're worried about this stuff, we have things we can do, like reviews, rally testing and endurance testing with a single tenant
14:23:37 <bauzas> mriedem: writing a novaclient change that would do 20 resizes ?
14:23:53 <mriedem> bauzas: yes maybe
14:23:57 <bauzas> mmm
14:23:57 <mriedem> moving on
14:24:10 <mriedem> i will rebase and address comments on the services/hypervisors API uuid changes
14:24:24 <mriedem> thanks to alex_xu for the thorough reviews there
14:24:30 <alex_xu> np
14:24:32 <mriedem> #info Started an etherpad to track remaining TODOs for Pike (like docs gaps): https://etherpad.openstack.org/p/nova-pike-cells-v2-todos
14:24:45 <mriedem> i should throw the test ideas up in ^
14:25:10 <mriedem> ok moving on to scheduler
14:25:11 <mriedem> edleafe:
14:25:15 <edleafe> We briefly discussed the spec amendment regarding picking up the Ironic flavor migration work that jroll was going to do. In that discussion, dtantsur brought up his work to make devstack use custom resource classes in Ironic by default. Patches to use Traits with AllocationCandidates have also been pushed by alex_xu.
14:25:20 <edleafe> We then spent the next half hour discussing jaypipes's patch for claiming in the scheduler, and the implications of those changes. We managed to achieve harmony, then held hands and sang folk songs afterwards.
14:25:24 <edleafe> Finally there was a discussion about only implementing half of the proposed change to the scheduler for claiming, and that until that is made, complex resource providers cannot be supported. Since that's not happening until Queens, it was agreed that it was fine to change the method signatures twice to reflect that support for complex RPs would not happen in Pike.
14:25:29 <edleafe> EOM
14:26:26 <mriedem> ok
14:26:58 <mriedem> for those wondering,
14:27:07 <mriedem> the current focus for that series is starting here https://review.openstack.org/#/c/482381/
14:27:14 <mriedem> #link claims in scheduler series is here now https://review.openstack.org/#/c/482381/
14:27:39 <mriedem> ok moving on to api
14:27:41 <bauzas> not really nope ?
14:27:46 <mriedem> alex_xu: anything to bring up here?
14:27:49 <alex_xu> Talk about 'PUT /os-services' API, agree with the API should be idempotent
14:27:57 <alex_xu> #link https://review.openstack.org/#/q/status:open+project:openstack/nova+branch:master+topic:bp/api-no-more-extensions-pike
14:27:58 <bauzas> mriedem: https://review.openstack.org/#/c/482382/ is the bottom one now
14:28:02 <bauzas> AFAIK
14:28:12 <bauzas> anyway
14:28:42 <alex_xu> need review help for the api-no-more-extension-pike, they are easy, just need one more core reviewer
14:28:54 <alex_xu> that's all
14:28:54 <edleafe> bauzas: the dependencies are mixed up for that series
14:29:04 <jaypipes> and then jaypipes almost committed suicide last night attempting to deal with the awfulness of the scheduler.
14:29:34 <mriedem> #action mriedem to create a scheduler hotline for people working on it
14:30:06 <mriedem> ok moving on to notifications
14:30:08 <mriedem> gibi: you're up
14:30:14 * edleafe has an idea for a new market for Valium
14:30:39 <gibi> We discussed the last pieces of the additional-notification-fields-for-searchlight bp and as a result mand
14:30:42 <gibi> atory part of the BDM piece has already been merged. The tag support at instance create is still ongoing and in focus.
14:30:45 <gibi> Another patch in focus is a bugfix for https://bugs.launchpad.net/nova/+bug/1684860 https://review.openstack.org/#/c/475276/
14:30:46 <openstack> Launchpad bug 1684860 in OpenStack Compute (nova) "Versioned server notifications don't include updated_at" [Undecided,In progress] - Assigned to Takashi NATSUME (natsume-takashi)
14:30:48 <gibi> Most of the notification transformation patches need a rebase due to the BDM addition mentioned above.
14:30:53 <gibi> that is all
14:31:11 <mriedem> takashin: fyi on https://review.openstack.org/#/c/475276/
14:31:35 <takashin> mriedem: Thank you.
14:31:50 <mriedem> yeah and https://review.openstack.org/#/c/469800/ is on my short list of bps
14:31:57 <mriedem> that's the tags at server create one
14:32:14 <mriedem> i know alex_xu was +2 on it at one point
14:32:26 <mriedem> ok moving on
14:32:29 <mriedem> cinder stuff
14:32:36 <mriedem> needs reviews: https://review.openstack.org/#/q/status:open+project:openstack/nova+topic:bp/cinder-new-attach-apis+branch:master
14:32:51 <mriedem> specifically the swap volume change has a +2 https://review.openstack.org/#/c/456971/
14:33:00 <mriedem> so it would be cool to see another core get to that one
14:33:11 <ildikov> Cinder dependencies are all merged for the patches in the chain
14:33:18 <mriedem> stvnoyes is working on some grenade testing
14:33:27 <mriedem> hyper-v live migration is passing with the new flow
14:33:39 <mriedem> need testing from xenapi team at Citrix
14:33:40 <ildikov> XenAPI is confirmed to work as well
14:33:46 <mriedem> ok, yeah i saw they were reviewing
14:33:54 <mriedem> the live migration patch is https://review.openstack.org/#/c/463987/
14:34:03 <mriedem> i just started back on that before the meeting
14:34:31 <mriedem> but, i posted a change last week that runs the new flow through with the live migration changes and with the volume-backed live migration tests in tempest and things were all passing
14:34:53 <mriedem> so i'm fairly confident in the functionality, just need to dig into edge cases
14:35:10 <mriedem> moving on
14:35:13 <ildikov> we will have the Cinder-Nova meeting soon, so can sort those out hopefully
14:35:13 <mriedem> #topic stuck reviews
14:35:26 <mriedem> nothing on the agenda - is there anything to bring up here?
14:35:41 <mriedem> #topic open discussion
14:35:49 <mriedem> anything anyone wants to bring up?
14:35:50 <stephenfin> I've got one
14:35:57 <stephenfin> https://review.openstack.org/#/q/topic:doc-migration+project:openstack/nova+status:open
14:36:14 <stephenfin> ^ they need +2s if we ever want to have documentation again
14:36:15 <mriedem> #link nova docs migration starts here https://review.openstack.org/#/c/478472/
14:36:52 <mriedem> "It'll be a great day in the parish"
14:36:54 <mriedem> oh come on
14:37:02 <mriedem> your irish catholicism is bleeding through
14:37:09 <stephenfin> personally, I think if we ignore stylistic things and merge anything that's not actually broken, we can come back and revisit after p-3
14:37:24 <stephenfin> mriedem: Never forget where you came from ;)
14:37:37 <edleafe> Can we discuss the ironic migration? https://review.openstack.org/#/c/481748/
14:37:54 <mriedem> edleafe: sure if jaypipes is around for it
14:37:58 <mriedem> i know dan isn't
14:38:06 <edleafe> I'm still not clear on the order that things will be changing for this
14:38:11 <jaypipes> I'm here.
14:38:31 <mriedem> step 1 is getting the custom resource class reported into allocations
14:38:39 <mriedem> we report the inventory, but not the actual allocation
14:39:08 <mriedem> and i believe the vehicle for doing that is putting the resource class info in the flavor extra spec on startup of the driver
14:39:20 <jaypipes> mriedem: step 1 is getting the custom resource class added to the flavor so we can tell that it's being requested.
14:39:27 <jaypipes> mriedem: yes, bingo.
14:39:28 <mriedem> because then the scheduler report client can pull the resource class data off the extra spec in the update_available_resource periodic
14:39:39 <jaypipes> ++
14:39:49 <mriedem> that's the order i have in the work items section anyway
14:40:06 <mriedem> https://review.openstack.org/#/c/481748/1/specs/pike/approved/custom-resource-classes-in-flavors.rst@165
14:40:23 <edleafe> So what about existing ironic nodes? Will the extra_spec change show up in their instance.flavoe?
14:40:35 <edleafe> s/flavoe/flavor
14:40:41 <mriedem> yes
14:40:45 <mriedem> that's the whole poin
14:40:47 <mriedem> *point
14:40:56 <mriedem> on startup of nova-compute, the driver gets it's existing instances,
14:41:04 <jaypipes> right
14:41:07 <mriedem> and for their corresponding ironic node,
14:41:07 <edleafe> ok, just checking whether this would be a new flavor or an update
14:41:10 <bauzas> that's in the spec, nope ?N
14:41:19 <mriedem> get the node.resource_class and shove that into the embedded instance.flavor.extra_specs
14:41:29 <mriedem> bauzas: it was never clear in the original spec
14:41:33 <mriedem> which is why i was amending it
14:41:39 <bauzas> ok, I'm unclear about the problem
14:41:42 <bauzas> ok
14:41:52 <mriedem> edleafe: to be clear, this isn't updating existing global flavors,
14:41:57 <mriedem> it's updating the embedded flavor in the existing instances
14:42:11 <jaypipes> right
14:42:16 <jaypipes> re: edleafe
14:42:24 <mriedem> because that's what the scheduler report client looks at when putting allocations
14:42:26 <edleafe> so the first step is adding the resource class to existing nodes?
14:42:31 <mriedem> no
14:42:33 <jaypipes> 's question about should be zero out cpu/ram/disk, etc... I think the aNSWER TO THAT IS YES.
14:42:38 <jaypipes> oh ffs, typing sucks.
14:42:57 <mriedem> if the ironic node.resource_class is not None, then do data migration steps
14:42:59 <mriedem> else skip
14:43:03 <edleafe> ok, compute starts up and looks at node.resource_class. Where/how does that get populated?
14:43:12 <mriedem> it's populated in ironic
14:43:17 <mriedem> when the node is created in ironic
14:43:18 <jaypipes> edleafe: in the ironic get_inventory()
14:43:24 <jaypipes> here:
14:43:39 <edleafe> so all existing ironic nodes will be updated to return that class?
14:43:47 <mriedem> if we're going to step through this, can we do it in -nova rather than hold up the meeting?
14:43:50 <jaypipes> https://github.com/openstack/nova/blob/master/nova/virt/ironic/driver.py#L615
14:43:57 <edleafe> mriedem: sure.
14:43:59 <jaypipes> edleafe: that's already been done.
14:44:01 <mriedem> ok anything else?
14:44:12 <edleafe> jaypipes: ok, that wasn't clear
14:44:29 <mriedem> alright i'm going to end it, thanks everyone
14:44:31 <mriedem> #endmeeting