20:00:22 #startmeeting heat 20:00:23 Meeting started Wed May 21 20:00:22 2014 UTC and is due to finish in 60 minutes. The chair is zaneb. Information about MeetBot at http://wiki.debian.org/MeetBot. 20:00:24 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 20:00:26 The meeting name has been set to 'heat' 20:00:55 #topic Raise your hand if you're not here 20:00:59 good afternoon 20:01:02 o/ 20:01:07 o/ 20:01:09 ,o 20:01:12 o/ 20:01:19 hello 20:01:24 o/ 20:01:28 hello all 20:01:34 Hi 20:01:39 o/ 20:01:39 Hey there 20:01:46 Hi 20:01:59 Hello all - nice meeting a big part of the heat dev community at the summit last week. 20:02:08 yep, it was awesome :) 20:02:15 o/ I'm here. 20:02:17 stevebaker: about? 20:02:19 hi all 20:02:22 o/ 20:02:27 hey 20:02:39 hey jpeeler, welcome back 20:02:50 o/ 20:02:52 thanks 20:02:57 #topic Adding items to the agenda 20:03:08 #link https://wiki.openstack.org/wiki/Meetings/HeatAgenda 20:03:18 big turnout and a full agenda today :) 20:03:37 anybody got anything else? 20:04:37 #topic Alternate meeting time 20:04:47 Oooh! Just in time to troll! :) 20:04:47 I sent a mail out to the list about this 20:04:57 #link http://lists.openstack.org/pipermail/openstack-dev/2014-May/035480.html 20:05:18 I want to try having the alternate meeting at 1200 UTC, starting next week 20:05:43 +1 obviously :) 20:05:44 if anybody wishes to enter into correspondence, let them do so now 20:06:19 #topic Rackspace 3rd-party CI job 20:06:20 +1 for 1200 UTC 20:06:21 I shall abstain, this meeting isn't for me 20:06:24 +1 20:06:29 + 20:06:37 o/ just running out the door, but wanted to let everybody know that there's a spec in-progress for Convergence here https://blueprints.launchpad.net/heat/+spec/convergence and https://wiki.openstack.org/wiki/Heat/Blueprints/Convergence . 20:06:38 +1 20:06:40 +1 20:06:41 * SpamapS disappears 20:06:55 SpamapS: hooray 20:06:57 thanks SpamapS 20:06:59 hey 20:07:17 andrew_plunk: ah, perfect timing :) 20:07:21 ;) 20:08:02 so, I am strongly biased toward +1 for the Rackspace Jenkins to post 3rd party CI stuff to Gerrit 20:08:09 only one question really 20:08:18 See, if it was an hour later, I'd totally join in during my morning tea, but I'm mostly here to troll. :) 20:08:30 zaneb, there may be false negatives if downstream services are having problems 20:08:35 is the community ok with that? 20:08:39 what percentage of failures are likely to be upstream problems vs. config problems 20:08:49 andrew_plunk: I assume it will be non-voting, at least initially? 20:08:59 andrew_plunk, that may well be the norm for 3rd-party CI 20:09:04 shardy: I would hope so 20:09:12 +1 for non-voting 20:09:18 * stevebaker shakes fist at XenServer CI 20:09:25 sounds good to me shardy 20:09:27 otherwise it can block upstream 20:09:27 randallburt: then I'm +1 on getting it going and seeing how it works out 20:09:37 failures will most likely be downstream failures, but we could build some padding for that in the tests themselves 20:10:06 well, it can post a -1 on failure, voting isn't really a thing for 3rd party 20:10:39 yeah, it'd be under the Code-Review column 20:10:40 Yeah, I have the shudder of experience at flaky downstreams causing test failures. It's a shudder unique to the layer at which we work. 20:10:42 stevebaker: Ok, so we can overrule it with +2/+A 20:10:52 if it's obviously wrong that is ;) 20:11:10 IIRC its mostly to make sure the v2 shim works, yes? 20:11:16 shardy, yeah, and some form of recheck triggering would be needed eventually 20:11:19 or are we testing more than that? 20:11:38 randallburt, what cloud does it test against? a devstack? 20:11:48 Rackspace public cloud currently 20:11:58 randallburt, in standalone mode? 20:12:19 I don't think so. andrew_plunk? 20:12:21 in normal mode with an admin user stevebaker 20:12:43 you named your admin user stevebaker? :) 20:12:48 lol 20:12:49 haha 20:12:54 no, just steve. 20:12:58 why not? 20:12:58 Are the tests publicly available? 20:13:10 therve: not currently. 20:13:23 I think one of our QE folks is working on that, though 20:13:40 Okay. Ultimately it'd be cool to know what's going on if it fails 20:13:47 therve: agreed 20:13:48 andrew_plunk: I think you mentioned we need an official vote for the infra people to enable this? 20:13:59 to enable voting zaneb 20:14:02 +1 20:14:09 If our jenkins is going to just comment then I don't think it matters 20:14:22 I will take an action item to get it set up today zaneb 20:15:02 #action andrew_plunk to set up a non-voting Rackspace CI Jenkins job against Heat 20:15:18 capital 20:15:22 thanks andrew_plunk 20:15:34 anytime :) 20:15:41 indeed, chap 20:15:41 #topic Blueprint process 20:16:13 most projects seem to be switching away from the wiki and toward a blueprint repo 20:16:22 do we want to follow? 20:16:40 +0. We do have some horrible outdated stuff in the wiki 20:16:52 launchpad is still peripherally involved btw 20:16:52 it seems like a good way of collaborating on the specifics of a spec 20:16:53 it seems that would at least get more scrutiny on the blueprints 20:17:12 but Gerrit is a pretty good tool for collaboration 20:17:23 *cough* 20:17:41 therve: not *great* 20:17:57 It make sense to align on the global process, and it seems to have some benefits 20:18:00 but, you know, decent 20:18:05 The downside is more reviews though, right? 20:18:06 better than wiki/etherpad/mailinglist for certain things 20:18:15 * shardy looks at the review queue 20:18:23 I don't think we were specifically broken on that front, that's all 20:18:34 shardy, more reviews upfront, which means less reviews when the features land ;) 20:18:39 shardy, Possibly better future review on features discussed though 20:18:50 therve: agreed. The bp allows for a spec from the wiki and that's collaboratively edited. 20:18:54 I think it definitely makes sense for stuff like API specs and other interfaces 20:19:03 and aren't there comments/discussions on the wiki? 20:19:11 I don't want this to turn into a more formal process like some other projects have 20:19:21 zaneb: +1 20:19:41 I don't feel like I'm collaborating when I edit someone elses wiki 20:19:46 but I do think gerrit is a good tool 20:19:48 zaneb, +1 20:19:51 I never use the wiki 20:20:45 zaneb: +1 i think people are good about asking for feedback on a spec when it's needed and just showing up with code when it's not 20:20:49 The sad state of our wiki should be handled separately too 20:20:56 The wiki makes me sad. 20:20:58 jasond: ++ 20:21:28 sounds like we're mostly +1 on it then 20:21:30 I am actually a zillion times more fond of wikis than our present blueprint process. Although I understand that certain aspects — the semantically enabled chain of blueprint relationships, for example — are really awesom 20:21:44 I once saw someone on ask.openstack citing a 2-year-old wiki page as an answer 20:22:08 We should organize a wiki cleaning day or something 20:22:12 wirehead_, our present blueprint process is wikis 20:22:22 wirehead_, and launchpad 20:22:29 wirehead_: launchpad doesn't go away, but basically only I have to deal with it ;) 20:22:35 stevebaker: there are a lot of bp's without wiki specs, though 20:22:51 I think the launchpad is the part I like less. :) It seems to be a home of crushed dreams. 20:23:04 lol 20:23:09 wirehead_: +1 20:23:23 Hopefully that'll change soon 20:23:35 therve: storyboard? 20:23:37 so ML/Wiki and then make zaneb deal with LP if we think its a thing to do? 20:23:40 zaneb, speaking of which, there are a lot of blueprints which are yet to experience the benign gaze of the PTL 20:23:40 pas-ha, Yeah 20:24:21 stevebaker: Thierry isn't giving me grief about them until we have resolved this question ;) 20:24:36 lol, ok 20:24:51 so who wants to set up the specs repo? 20:25:01 what I'd like to propose is that we move to a specs repo, but keep it extremely lightweight 20:25:13 zaneb, +1 20:25:20 Aren't we founded on rough consensus? 20:25:22 ttx is talking about a script to push them into lp 20:25:26 zaneb: +1 20:25:31 +1 20:25:33 that would be handy 20:25:33 we can crank up the process when we have 400+ developers like nova 20:25:33 zaneb: +1 20:25:38 zaneb, +1 20:25:39 randallburt: ^ how say you? 20:25:47 zaneb: +1 20:26:04 one sec 20:26:19 oh, ok. +1 20:26:30 z+1 20:26:41 +1 20:26:47 +1 20:26:52 I think we might have a complete circle of love. Wow. 20:27:04 #agreed Heat will move to a lightweight blueprint process using a specs repo 20:27:15 ok, and now... 20:27:24 who want to volunteer to set it up? ;) 20:27:53 ... crickets ... 20:27:57 tumbleweeds 20:28:03 #info repo name should be orch-specs or orchestration-specs, per decision of the cross-project meeting yesterday 20:28:11 what's involved in setting it up? 20:28:36 cyli, Talking to our lovely infra people I think 20:28:36 cyli: find another project's one, clone it, request that the infra team add it to openstack/ 20:28:36 cyli, probably just a infra config gerrit change 20:29:03 cyli: I can add a task to our sprint. :) 20:29:09 we also need to decide on what will be validated, I think 20:29:11 or start with an empty repo and we can all add the initial files 20:29:14 i.e. as little as possible ;) 20:29:26 I'll do it, so long as no one minds me asking stupid questions about how to do so here 20:29:48 #action cyli to investigate setting up a specs repo 20:29:51 cyli, thanks. ask in #heat and #openstack-infra 20:29:54 many thanks cyli 20:30:02 :) 20:30:18 #topic Feature proposal freeze 20:30:23 I just patted cyli on the back. 20:30:35 do we want to do a Feature proposal freeze again? 20:30:38 (IMO we do) 20:30:41 stevebaker: will do, thanks 20:30:56 hey, I'm late to the party.. .if launchpad stops getting used, how do non technical folks submit feature requests/suggestions? 20:31:10 kebray: launchpad will still be used with a spec repo 20:31:14 zaneb, Definitely. 20:31:25 radix ok.. thx. 20:31:35 radix: well, there's a danger that no-one will look at it though 20:31:36 It went well last time AFAICT 20:31:51 agreed 20:31:54 that danger is very high and very real 20:31:54 zaneb: well, surely the PTL will continue? ;) 20:31:55 kebray, launchpad will be used until storyboard is ready. the specs repo is just to collaborate and store the text of the spec 20:31:57 it sets the right expectations 20:32:23 radix: maybe not if a script manages it all ;) 20:32:37 auto-PTL 20:32:40 .sh 20:32:57 Is it the case that blueprints currently in the New state still have a shot of getting in? Or they need to be in the Approved state? 20:33:16 kebray: blueprints are about implementation, so we'd expect implementers to be able to use Gerrit 20:33:32 iqster, You mean in for Juno? 20:33:38 cyli, a single change to http://git.openstack.org/cgit/openstack-infra/config/ should be all that is required 20:34:15 iqster: everything has a shot of getting in if you can write it ;) 20:34:17 stevebaker: thanks 20:34:23 maybe not a good shot... 20:34:25 neat :) ty 20:34:53 iqster: but if it's a significant change, you should be looking for early feedback 20:35:32 zaneb, +1 for continuing the FPF, but it only slightly mitigates the problem of many features landing at once 20:35:39 iqster: i.e. don't do more work than you're prepared to throw away until you think there's consensus 20:36:24 stevebaker: agreed, but I think it does a good job of communicating expectations 20:36:39 is anyone against the FPF? 20:36:43 yep 20:37:01 cool ... we have a patch/blueprint/spec for stack lifecycle plugpoints .. we wanted to get it in today but BillArnold was having some issues with git review 20:37:05 Your agreement was badly timed I think :) 20:37:13 heh heh 20:37:23 zaneb: I think it's a good idea, we have to draw a line somewhere, although I actually think it makes the late-cycle feature rush worse if anything 20:37:37 shardy, Worse a bit sooner 20:37:50 I think it'd be the same without freeze except before the actual release 20:37:52 therve: ha, true ;) 20:37:56 we're discussing a freeze for *proposals* to be accepted, right, not *implementations*? 20:38:06 iqster: what kind of issues was he having? 20:38:07 (or proposals to be proposed, perhaps) 20:38:28 radix: yes, typically a few weeks before the freeze for implementation to be merged 20:38:29 we're stuck with something like this until all of openstack moves to a CD 20:38:36 so we can review the glut of patches ;) 20:38:37 based model 20:38:40 radix: proposals means patches proposed to Gerrit 20:38:53 oh 20:38:54 radix: i.e. implementations, but not necessarily merged yet 20:39:04 ok, I completely misinterpreted "proposal" to mean something like "blueprint" 20:39:24 #link https://wiki.openstack.org/wiki/Juno_Release_Schedule 20:39:32 so that ^ is the release schedule 20:39:38 the aim of the FPF is to ease the load on the core reviewers at feature freeze time, because some of the reviewing will be done already 20:39:40 radix: The blueprints should be proposed much earlier for major stuff, but sometimes minor stuff can be added near FPF if there is code ready 20:39:47 FPF is the week of August 21st 20:39:57 #info FPF is the week of August 21st 20:40:01 sounds reasonable 20:40:18 I don't see why FPF would make a rush worse 20:40:19 whats FPF? 20:40:25 erecio: Feature Proposal Freeze 20:40:28 erecio: see #topic 20:40:57 #link https://wiki.openstack.org/wiki/FeatureProposalFreeze 20:41:13 I think we have a lazy consensus 20:41:21 #agreed Heat will observe the project-wide Feature Proposal Freeze date again in Juno 20:41:48 Heh. 20:41:53 #topic Autoscaling roadmap 20:42:00 is shadower here? 20:42:01 no 20:42:04 thanks. 20:42:08 radix: any deadline creates a mad rush when folks realize they've not finished their feature and it's targeted at the next release, which results in attention from release managers and PTLs 20:42:25 shardy: but wouldn't there be a rush at the actual feature freeze then? 20:42:40 zaneb: is there a particular question behind this agenda item? 20:42:43 and god, and ttx 20:42:48 I have been curious about other use cases for autoscaling 20:42:54 ok, I'd like to set up a discussion between radix, shadower, shardy and anyone else interested 20:43:01 who's shadower? 20:43:14 to discuss what our plans are for autoscaling in Juno 20:43:17 Thomas Sedovic 20:43:26 ah ok 20:43:34 radix: Tomas Sedovic. ex-Heat and now Tuskaer developer 20:43:38 Tuskar* 20:43:57 moving on... 20:44:04 *my* plans right now is to just do some refactoring of Heat autoscaling (along with cyli) and also help out with the convergence effort 20:44:13 which I think plays into practical use of autoscaling 20:44:26 radix: ok 20:44:26 is a separate API still a thing we're doing? 20:44:44 Tuskar also has some immediate issues with quiescing and victim-selection 20:44:49 on scale-down 20:44:56 ok, that sounds like some good stuff 20:44:58 one shortcoming we have observed in heat autoscaling is that if vms die, the auto-scaling doesn't react to it 20:45:04 I discussed with wirehead_ and radix about Tuskar (and other's) requirement for evacuation on scaledown and parameters to choose victims 20:45:15 iqster: yep, that's what I mean about convergence being necessary for practical autoscaling 20:45:15 i presume this is a known defect? 20:45:17 I may look at those unless anyone gets to them first 20:45:24 iqster: convergence is the long-term plan to fix that 20:45:26 iqster: also for practical heat usage in general ;-) 20:45:31 * shardy not planning to get to them anytime soon 20:46:00 stevebaker: I think... there is less urgency behind that idea 20:46:07 radix: I am also keen to see refactoring continue to happen 20:46:12 stevebaker: and more behind getting heat to be more usable :) 20:46:15 we can't wait for convergence to come to us 20:46:18 at the summit, i didn't get a sense that convergence was defn in Juno 20:46:19 zaneb: agreed, yeah 20:46:26 we have to start pushing towards it now 20:46:36 zaneb: cyli and I have a sprint task to work fix up the factoring of AutoScalingResourceGroup 20:46:44 I'm off to do the school run \o 20:46:45 iqster: convergence will certainly not be complete in Juno 20:47:05 iqster: but autoscaling isn't going to be able to react to servers randomly dying until convergence happens 20:47:05 Have some faith! 20:47:15 doh! 20:47:29 iqster: maybe in L 20:47:30 so ... no reactive autoscaling in Juno it seems :( 20:47:33 iqster: It's a huge amount of work, we'll have to take it in several steps and it depends on who shows up to help 20:47:53 iqster, Please help by reviewing and fixing bugs! 20:47:54 That i don't doubt 20:48:07 but things will continue to get better 20:48:18 we're making continuous progress 20:48:30 will do ... we'd like to help... speaking for BillArnold and myslef here :) 20:48:42 zaneb, stevebaker: anyway I don't think there are other big pushers for the autoscaling API other than the rackspace-autoscale team, if there *are* I would love to know about them 20:49:09 radix: I think the API is the icing on the cake 20:49:13 yep 20:49:17 radix, ok, lets park it until someone cares enough to do it then 20:49:23 radix: all of the preceding steps are the important ones 20:49:27 radix: most folks I've spoken to care primarily about the features, not the API per-se 20:49:31 at that point the API is easy 20:49:33 right 20:49:53 (also more people would probably be interested if they knew about it) 20:49:56 I was just confirming that :) it's something we want so we can basically stop maintaining our own service 20:50:46 I'd like to see it happen because autoscaling is useful independently of Heat, just like other OpenStack services 20:51:12 #topic Critical issues sync 20:51:19 lifeless: o/ 20:51:31 shardy suggested we add this as a recurring agenda item 20:51:33 * stevebaker *really* goes now. just when its getting interesting 20:51:45 can it be at the start of the agenda? 20:51:54 So I found this yesterday 20:51:56 so that we can all get in sync on critical bugs 20:51:58 #link https://bugs.launchpad.net/heat/+bug/1321303 20:52:04 Launchpad bug 1321303 in heat "engine broken with multiple workers w/qpid" [High,In progress] 20:52:07 stevebaker: will there be time for anything else if we do that? 20:52:17 yeah... I gotta run too, sorry ;- ) 20:52:18 multiple workers are broken with qpid 20:52:48 (this is particularly for the benefit of TripleO) 20:52:51 shardy: is this just when we fork multiple processes on a box or does it affect multi-engine in general? 20:52:53 I *think* I have a fix, but will need review feedback, particularly from jasond as it changes how the EnginListener works 20:53:06 randallburt: only forked workers, multi-engine works OK 20:53:09 zaneb, the meeting nazi can time-limit it 20:53:14 shardy: k, thanks 20:53:21 shardy, why heat engine workers and not other workers e.g. nova conductor? 20:53:30 * zaneb looks around for the meeting nazi 20:53:31 although see my comment about the watch threads, which I am also looking at fixing 20:54:03 shardy: no problem, will keep an eye out for your fix 20:54:06 BillArnold: because we wrote it wrong, AFAICT, but I'm still digging into why 20:54:38 anyway, any other issues we need to know about, e.g for TripleO or anything else? 20:55:15 sdf 20:55:25 shardy, I wonder how that forking thing will change with oslo.messaging 20:55:57 We'll see how it goes I guess 20:56:00 therve: yeah I was wondering the same thing, not had time to properly look at sdake's patch yet 20:56:12 zaneb: we're still trying to get the bottom of this 2 minute delay between deployments being created 20:56:20 that ^ was posted in #heat 20:56:35 therve: my plan it to fix the immediate problem, then potentially look at a bigger refactor 20:56:48 shardy, The patch has been reverted fwiw 20:57:20 therve: I mean the problem that you can't specify multiple workers with qpid 20:57:31 zaneb: do we have a bug reference? 20:57:44 not that I know of 20:57:50 but stevebaker might know more 20:58:24 something about large numbers of calls to Nova for get_attr 20:58:46 2 minutes 20:59:54 zaneb, is this extra keystone work? using the fake hypervisor driver i'm seeing like a factor of 5 slowdown before the heat engine returns from a create call relative to havana 21:00:12 and keystone is being hammered 21:00:20 BillArnold: interesting 21:00:32 BillArnold: bug with details please :) 21:00:40 zaneb same problem maybe? 21:00:49 can't be helping 21:00:55 please do raise a bug for that 21:00:59 k 21:01:09 #endmeeting