13:00:56 #startmeeting powervm_driver_meeting 13:00:57 Meeting started Tue Jul 25 13:00:56 2017 UTC and is due to finish in 60 minutes. The chair is esberglu. Information about MeetBot at http://wiki.debian.org/MeetBot. 13:00:58 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 13:01:00 The meeting name has been set to 'powervm_driver_meeting' 13:01:10 o/ 13:01:11 o/ 13:01:45 #topic In Tree Driver 13:01:48 \o 13:01:56 #link https://etherpad.openstack.org/p/powervm-in-tree-todos 13:02:07 Oh and here's the agenda 13:02:08 #link https://etherpad.openstack.org/p/powervm-in-tree-todos 13:02:18 #link https://etherpad.openstack.org/p/powervm_driver_meeting_agenda 13:02:39 No topics for in tree. 13:03:13 yeah, I don't have anything for IT 13:03:13 Reminder that this week (thursday) is the pike 3 milestone (feature freeze) 13:03:37 https://review.openstack.org/#/c/476945/ 13:04:13 Keeps failing CI. I imagine it's for whatever reason everything is failing CI, and we'll get to that later? 13:04:31 efried: Yeah. I'll update on that status later 13:04:58 efried should we open a bug with which to associate that change? 13:05:18 Wouldn't think so. It's not a bug. 13:05:27 k 13:05:33 and we don't need to "backport" it. 13:05:40 I thought it fixed something, but must be wrong 13:05:49 even if it doesn't make Pike, we don't need to backport it to Pike. 13:05:55 fine 13:06:44 edmondsw: efried: I've been seeing errors about power_off_progressive in CI, assuming this is related somehow? 13:06:58 oh? 13:07:05 Yes, that would be related. 13:07:08 I haven't seen those. 13:07:10 Well, shit. 13:07:18 Guess we spend some time on that soon. 13:07:20 But maybe not too soo. 13:07:23 Because feature freeze. 13:08:30 efried: I don't have logs handy atm, I'll link them when I get to CI status 13:08:58 Thought we ran that bad boy through when it was first proposed. But mebbe not. 13:09:14 Twill suck if power_off_progressive is STILL broken. 13:09:27 Hate that code. 13:09:36 efried: The related error is not hitting every run so we may have ran it and just not hit the issue 13:10:13 We would have needed to run with a develop pypowervm and in-flight nova[-powervm] at the time, which I guess we might have done. But anyway... 13:10:55 I don't have anything else in-tree. Moving on? 13:11:01 Anyone have anything OOT or ready to move straight to PCI passthrough? 13:11:27 NVRAM stuff 13:11:41 efried: Discuss that later perhaps? 13:12:14 thorst_afk is home with sick kid, would like him involved in both of those discussions. 13:12:30 So the NVRAM improvement was what, again? 3.4%? 13:12:37 Mhm 13:12:43 That's... 13:12:52 Annoying 13:12:53 #topic Out Of Tree Driver 13:13:01 ...not nothing, but pretty marginal. 13:13:13 I'm here. 13:13:22 slowly 13:13:54 mdrabe And this was run through a pvc suite that actually exercises the NVRAM/slotmgr code paths? 13:14:15 Has it been run through pvc fvt kind of thing that hammers those code paths for functionality? 13:14:18 thorst: This is about https://review.openstack.org/#/c/471926/ 13:14:42 It can't run through fvt until it merges 13:14:44 that is ... quite the change set. 13:14:53 Yeah, that's the problem. 13:14:57 mdrabe: it'd have to be cherry picked. 13:15:16 thorst: Not necessarily within nova-powervm 13:15:17 It completely swizzles the way nvram and slot management is stored & retrieved in swift. 13:15:31 for a 3.4% improvement in deploy path? 13:15:32 Backward compatible, right mdrabe? Upgrade-able? 13:15:39 overall deploy path 13:15:44 Yes backward compatible 13:15:59 It doesn't affect the slot manager path greatly 13:16:21 By which I mean: if you have pvc running the old code, so objects are stored in swift the old way, then you upgrade pvc, it still works on the old objects? 13:16:53 and if you have some nova computes running at old level and new nova computes running at new level 13:16:56 does it all just work... 13:17:03 cause agents don't get upgraded all at once. 13:17:03 efried: The objects aren't stored any differently 13:17:28 I think I need to spend some time reviewing this (I'm just asking naive questions) 13:17:41 It's still keyed off the instance UUID regardless 13:17:43 Oh, right, they were always keyed by UUID in swift - the difference is the way they were passed around nova-powervm. 13:18:53 So yeah, thorst, that's the debate. This is a nontrivial change across code that we chucked together at the last minute during that one TCC to cover the RR emergency, so I'm kinda scared of touching it. 13:19:04 Actually efried this isn't that code 13:19:17 yeah, that nightmare is in pypowervm mostly 13:19:18 Most of the NVRAM stuff was already there, we slapped together the slot manager stuff 13:19:38 this nvram stuff was from kyleh long ago 13:20:05 so the basic change here is - move from list to set, and allow passing in the uuid or the instance, instead of just the instance. 13:20:14 thus reducing the number of instance lookups I assume 13:20:24 That's the gist 13:20:46 thorst: Yes, the performance improvement comes from avoiding getting the instance object in the NVRAM event handler 13:20:47 let me review in more detail, but I think its a solid change. The net though is, yeah, it needs some good test pre-merge 13:21:04 so we've got to figure out a plan to do that....patch onto some PVC's I assume 13:21:08 since PVC always has this turned on 13:21:30 It might be worth noting there's also performance improvement in idle compute time 13:21:41 ? 13:21:51 ? to me or thorst? 13:22:00 I think he's asking for an explanation 13:22:02 and if he isn't, I am 13:22:19 Any NVRAM events that come in won't grab the instance object 13:22:30 that's ... nice 13:22:30 NVRAM events aren't exclusive to deploys 13:22:39 right right... 13:23:53 thorst: I'm not the biggest fan of the caching that's done in the change at the moment 13:23:54 Okay, so the plan is to move forward carefully, get some good pvc fvt (somehow) 13:24:08 +1 13:24:18 I'd be curious on your review thoughts for that bit 13:24:35 I'll review today 13:25:45 Move on to PCI? 13:25:57 I think so 13:26:00 #topic PCI Passthrough 13:26:00 k 13:26:14 No progress on my end since last week. 13:26:17 I don't have anything new here, but I wanted to make sure this is a standing item on the agenda going forward 13:26:41 so that's all for today then, probably 13:27:15 Okay on to CI 13:27:17 #link https://etherpad.openstack.org/p/powervm_ci_todos 13:27:27 #topic Devstack generated tempest.conf 13:27:43 I've got 3 changes out in devstack, need a couple more 13:27:47 https://review.openstack.org/#/c/486629/ 13:27:57 https://review.openstack.org/#/c/486671/ 13:28:05 https://review.openstack.org/#/c/486701/ 13:28:57 Then just need to backport those as needed and add the new options to the local.conf files 13:29:18 At that point we should be ready to use the generated tempest conf 13:29:24 sweet 13:29:32 #topic CI Status 13:30:01 The last couple weeks the networking issues have been affecting the CI, taking down systems and leading to poor performance 13:30:26 Now that all of that seems to have settled down (fingers crossed) I'm gonna take inventory of tempest failures 13:30:28 zat what's been causing the timeouts? 13:31:12 efried: Seems like they have been way up when the network has been poor. Anecdotal though 13:31:36 k, well, we'll keep an eye on it. 13:31:41 Progress on the mystery 500? 13:32:07 Anyhow I'll let you guys know what I find when looking through the failures. 13:32:40 efried: Haven't looked back into it 13:33:43 And the power_off_progressive thing - you'll pop me links when you see it again? 13:33:59 efried: Yep 13:34:01 That's prolly all we can do here for CI, then. 13:34:49 mdrabe: Any progress on the jenkins/nodepool upgrade stuff? Haven't touched base on that in a while 13:35:20 esberglu Nah been entirely pvc of late 13:35:23 but I wanted to ask 13:35:45 How tough would it be to go to all zuul? 13:35:56 It's zuulv3 right? 13:36:23 mdrabe: Last time I checked zuul v3 wasn't ready for 3rd party CI's. But that may have changed 13:36:42 Once zuul v3 is ready I want to move to it 13:37:11 I shouldn't be too hard. Mostly just moving our jenkins jobs definitions over to zuul (where they would be defined with ansible instead) 13:38:29 Alright let's move on 13:38:36 #topic Driver Testing 13:38:40 Developments here? 13:40:05 jay1_ chhavi How's iSCSI going? 13:40:29 thorst wanted to make sure we got an update here. 13:40:34 Yeah, Last week encountered the volume attach issue 13:41:06 efried: ravi and jay is looking into it, i am not currently working on it. 13:41:32 and by the time Chhavi/Ravi debug it further system UI was not loading might be bcz of the NW issue we got last week 13:42:05 so, I reconfigured the system and again facing the Volume creation issue 13:42:20 Ravi is looking into it 13:45:06 from talking to ravi, he has very limited time to spend on this 13:45:42 so we're going to have to figure that out 13:49:03 next topic? 13:49:20 #topic Open Discussion 13:49:34 #subtopic Test subtopic 13:50:20 Just curious if that command is a thing or not 13:50:34 I don't have anything further 13:50:44 Thanks for joining 13:50:56 #endmeeting