20:01:14 #startmeeting heat 20:01:15 Meeting started Wed Apr 23 20:01:14 2014 UTC and is due to finish in 60 minutes. The chair is zaneb. Information about MeetBot at http://wiki.debian.org/MeetBot. 20:01:16 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 20:01:18 The meeting name has been set to 'heat' 20:01:30 #topic roll call 20:01:31 o/ 20:01:32 o/ 20:01:34 o/ 20:01:34 \o 20:01:35 hi all 20:01:36 o/ 20:01:40 o/ 20:01:42 hi 20:01:44 o/ 20:01:45 hi 20:01:48 o/ 20:01:48 o/ 20:01:48 o/ 20:02:11 o/ 20:02:13 hey 20:02:30 rakhmerov: welcome :) 20:02:42 tnx! 20:02:45 good turnout today :) 20:03:05 wrong channel :) o/ 20:03:10 #topic Review last meeting's actions 20:03:23 :) 20:03:24 #link http://eavesdrop.openstack.org/meetings/heat/2014/heat.2014-04-17-00.02.html 20:03:37 zaneb to add deprecation of RouterGateway to Release Notes 20:03:44 stevebaker: I did that 20:03:55 ta 20:03:57 zaneb to propose options for alternate meeting times on Mailing List 20:04:02 I also did that 20:04:07 just now :D 20:04:18 I saw it :-) 20:04:26 of course none of the relevant people are here, so not much point discussing it 20:04:37 reply on the ML if you have a preference 20:04:52 * everyone to submit design summit session proposals by the end of the week 20:04:56 o/ 20:04:59 did everyone do that? 20:05:06 well done everyone 20:05:13 :) 20:05:22 #topic Adding items to the agenda 20:05:32 #link https://wiki.openstack.org/wiki/Meetings/HeatAgenda#Agenda_.282014-04-23_2000_UTC.29 20:05:49 any additional items? 20:06:29 I'll take that as a no 20:06:55 #topic Design Summit planning 20:07:06 so I had an idea of how to handle this 20:07:21 ...and then lifeless proposed another session 20:07:35 hello all, sorry for the late 20:07:46 heh 20:08:09 so, there is an etherpad up to collaborate on this 20:08:09 so, action item... hassle lifeless? 20:08:16 #link https://etherpad.openstack.org/p/heat-juno-design-summit-schedule 20:08:46 only randallburt, therve and asalkeld have commented on it so far 20:09:41 radix, lifeless: both of your sessions touch on convergence and healing 20:09:58 yeah, it's something that a lot of people care about :) 20:10:06 I'm not sure whether to merge those into one session or schedule them back-to-back 20:10:11 thoughts? 20:10:15 its a topic which almost deserves 2 sessions 20:10:17 btw, shardy, I only added the v2 api session so it got on the list; I'd like to collaborate with you on that or let you run it if you prefer 20:10:37 zaneb: I think it's more important to give thorough focus to convergence + healing, informed by use cases such as autoscaling and tripleO 20:10:40 randallburt: sure sounds good 20:10:51 merge makes more sense 20:10:59 randallburt: not had time to look at that lately so will be good to kick-start the effort again ;) 20:11:24 zaneb: A session isn't really needed for http://summit.openstack.org/cfp/details/388 if we're only doing the reorg 20:11:36 shardy: k 20:11:48 stevebaker: I totally plan to reject that one ;) 20:12:34 zaneb: though I won't say anything about the tripleO session, I am pretty far removed from that topic so maybe it should still have its own 20:12:42 * stevebaker goes to get some Rope 20:12:42 fwiw I don't think it's necessary in general for the person who proposed the session to run it 20:12:48 by which I mean https://github.com/python-rope/rope 20:12:59 do you guys have an item to discuss things like integration with taskflow/mistral/whatever it is at the summit? 20:13:14 the person who proposed the session will be accountable for finding _someone_ to run it though ;) 20:13:15 there is a separate session that ttx setup that is approve3d 20:13:25 tspatzier: I'm getting to replying to your software-config email. Do you mind if I run the session? 20:13:28 rakhmerov: yes, it's in the cross-project track 20:13:30 for cross project integration with emerging stackforge projects 20:13:52 this one http://summit.openstack.org/cfp/details/1 20:14:00 yes #1 - must be important :) 20:14:02 rakhmerov: there's a session proposed for the cross-project sessions: http://summit.openstack.org/cfp/details/335 20:14:04 first post! 20:14:05 stevebaker: sounds good; would be happy to drive this together with you 20:14:17 rakhmerov: though maybe that's not exactly what you're getting at 20:14:32 wow, we got 2 time slots 20:14:49 huh? 20:15:05 no, I think that's fine. Basically I'm here to figure out how we can communicate at the summit 20:15:16 radix: 1 and 335 20:15:21 oh 20:15:34 and they're already scheduled: http://junodesignsummit.sched.org/ 20:16:23 yup, as usual russellb is making the rest of us look bad with his ruthless efficiency 20:16:34 o/ 20:17:08 shardy: how important is the auth model thing, and could we combine it with the operator session? 20:17:14 Where did this idea that any one person should run a session come from? 20:17:34 If you have something to say, you come to the front and say it. 20:17:41 zaneb: probably, I just wanted to see if there is still a need for discussion on that topic 20:17:44 run = facilititate. 20:18:00 SpamapS: I agree, but somebody has to start :) 20:18:02 wirehead_: disagree. We always designate somebody to take notes, but that is not the same as facilitating. 20:18:23 zaneb: Sure, whoever proposes sessions will likely be the first with anything to say. :) 20:18:23 taking notes != facilitating 20:18:32 I just don't like the sessions where one person "takes the lead" 20:18:33 zaneb: feel free to drop or combine as needed 20:18:36 thats another job 20:18:42 is moderator a better word? discussion conductor? 20:18:46 I prefer that the idea dominates. 20:18:59 shardy: ok, thanks. could you leave a comment to that effect? 20:19:08 anyway, it's a format issue 20:19:26 Anybody should feel empowered to take the floor. All should be respectful of whoever has the floor. That is all. 20:19:51 zaneb: sure 20:19:52 SpamapS: ++ 20:20:17 That said, I want stevebaker to run the software-config session. 20:20:18 ;-) 20:20:26 lol 20:20:36 :-) 20:20:40 so he can teach us how it works? :) 20:20:42 lol 20:20:45 anybody know anything about #15? 20:20:49 need m0ar youtube 20:20:57 also, come to my main conference talk 20:21:05 zaneb: the Swift one? 20:21:18 http://summit.openstack.org/cfp/details/387 20:21:21 #15 on what list? 20:21:23 that one ^ 20:21:29 stevebaker: when is it? 20:21:52 zaneb smells like w orkflow 20:21:56 "holistic scheduling" sounds like mspatzier would know something 20:21:57 zaneb: I think that's the internal service version of the lifecycle callbacks. 20:22:13 radix: you mean mspreitzer I guess 20:22:13 radix: mspreitzer 20:22:19 er, yeah, mspreitzer 20:22:21 sorry :) 20:22:26 Holistic scheduling isn't workflow. 20:22:27 notifying OpenStack in a way that an operator or other OpenStack services can respond to 20:22:28 btw, good news. we will have a table assigned to us for discussions throughout the conference 20:22:37 nice 20:22:39 It is actually quite a refined form of orchestration. 20:22:46 not as good as a meeting room, but considerably better than nothing 20:22:47 zaneb: nice 20:22:51 tspatzier: 3:40, Tuesday 20:22:54 #link http://openstacksummitmay2014atlanta.sched.org/event/dae9536f6bb9ad61b3b2ccf39a18515f#.U1ghBXWSzKQ 20:23:00 thanks stevebaker 20:23:04 Declare the things you want, and the engine deploys them optimally. 20:23:11 radix: so perhaps you me and SpamapS hangout and do joint prep work for one session ? 20:23:37 radix: which is your session proposal ? 20:23:38 (I also think it is _WAY_ over reaching for what Heat and OpenStack are capable of at this time) 20:23:43 lifeless: are you talking about the convergence thing? 20:23:49 yes, this sessions is related to a patch that got recently posted by Bill Arnold, he also opened a BP after sdake requested it on the review 20:23:53 lifeless: http://summit.openstack.org/cfp/details/144 20:24:11 lifeless: I'm suggesting that convergence + healing have its own session, but I didn't say that tripleo shouldn't have its own 20:24:23 well as I recall after looking at the review tspatzier, I was certain there would be popcorn on that idea - so may be best to have a session on it 20:24:39 btw, I did remove the -2 :) 20:24:46 radix: we're suggesting that convergence and healing are paramount to operating an OpenStack cloud with Heat. 20:24:54 SpamapS: for a moment there I thought you were talking about lifeless's proposal 20:24:57 ;) 20:25:00 agree sdake, this is a bigger topic to be sorted out 20:25:05 radix: I think your session and the one I've proposed overlap a lot 20:25:08 sdake: I'm tempted to add my own −2 then. 20:25:09 spamaps: ok, if that's the main point of your proposal then that sounds like it should all just be merged together 20:25:21 radix: I'd hate to have the same discussion twice 20:25:23 I -2 because there was no blueprint and its popcorn material 20:25:24 lifeless: I would suggest to get randallburt in that hangouts too 20:25:27 zaneb: I do thin holistic scheduling will be far simpler once we've moved to distributed state machine and convergence. 20:25:35 think its worth having a design discussion around however 20:25:51 radix: sure, I'm game 20:25:57 it might be a good idea, I just dont know what the idea is in detail :) 20:26:06 radix: we have a specific design in mind (of course) - and HP is ponying up people to implement it (its somewhat bold in nature) 20:26:16 interesting, ok :) 20:27:00 radix: but we're also interested in scale 20:27:08 radix: e.g. we want to deploy 10K node clusters 20:27:26 thus, no polling statuses every 30 seconds ;-) 20:27:27 radix: (or more - think moonshot / seamicro style rack densities) 20:27:52 radix: indeed, so http://summit.openstack.org/cfp/details/428 20:28:02 longpoll would be fine. ;) 20:28:05 this is definitely something not contemplated when we started heat :D 20:28:16 lol :) 20:28:47 pretty soon the heat of heat will be generating sea salt! 20:28:49 we can put > 1K machines in a rack... 20:29:03 lifeless: so I'm pretty busy the rest of the day, can you suggest a time for us to get together? (randallburt and I are in the same timezone, CST) 20:29:04 its scary density 20:29:43 ok, so we're supposed to have a schedule by Friday(!) 20:29:47 radix: I'm free for the next 4 hours; SpamapS ? 20:29:55 lifeless: is this yet another thing pinned to taskflow or is that being used as a generic "distributed asynchronous workers" 20:29:57 I have 2.5 hours 20:30:06 radix, lifeless: sounds like y'all should get together and let me know what you decide 20:30:11 randallburt: the latter 20:30:23 SpamapS: oh good. 20:30:36 randallburt: but given taskflow's origins and future... no sense using something else. 20:30:44 #topic Heat/Mistral collaboration 20:30:52 rakhmerov: go :) 20:30:53 randallburt: taskflow as a design paradigm is uninteresting in the heat context - but 'run this bit of code with guaranteed completion semantics somewhere arbitrary' is very important - and joshua and I did a mini-deep-dive on that in sunnyvale a couple months ago 20:30:58 yep 20:30:58 SpamapS: I remain to be convinced. 20:31:03 radix: feel free to reschedule planned meetings if necessary. :) 20:31:11 wirehead_: ok :) 20:31:17 randallburt: and so rather than reinvent things I'd like to use a couple of key slices from within taskflow 20:31:22 lifeless, SpamapS: ok, I have time after this then :) 20:31:25 ok, so for now I have a couple of questions 20:31:27 randallburt: ? 20:31:33 randallburt: are you free after this meeting? 20:31:51 1. Are you guys still interested in discussion about integrating with something like Mistral? 20:32:12 2. If yes, how could we plan this collaboration further? 20:32:12 radix, he's free for 15-20 minutes, then I'm going to steel him. 20:32:17 rakhmerov: define 'integrating' 20:32:19 radix: sorry, not for another few hours 20:32:45 guys, let's take the post-meeting scheduling to #heat ;) 20:33:11 well, in HK we briefly touched that you might want to use Mistral once it gets mature internally for internal Heat processes automation 20:33:17 using workflow engine 20:33:31 So, for reasonable auto scaling workflows, we need scheduling, perhaps like Mistral provides, to trigger scaling events. 20:33:36 ok 20:33:48 I recognize it may sound tricky because it wasn't clear what it was going to be and so on and so forth 20:33:51 rakhmerov: so we're talking about doing away with the large-scale workflow-ish parts of Heat and focusing on a convergence model. 20:33:54 so the options there as I see them are to use (1) taskflow, or (2) mistral 20:34:07 ok 20:34:10 rakhmerov: we'll still have workflow needs.. 20:34:17 wirehead_: aren't scaling events triggered by alarms? 20:34:23 zaneb: I'm not entirely sure (1) is a realistic near-term option 20:34:25 good to hear that 20:34:32 rakhmerov: but I think they'll be entirely internal and thus will likely just leverage taskflow. 20:34:44 shardy: "Hey, we've got a PR blast going out at 9 AM. Let's make sure we have enough servers up" 20:34:56 I was personally never convinced that using an externally-facing API (mistral) would be better for the job than a library (taskflow) 20:35:05 wirehead_: Ok, the cron-aas use-case, got it 20:35:07 zaneb: agreed there as well 20:35:10 ok, I see :) So getting back to quetion #2 20:35:30 some people were convinced though, so they should speak up ;) 20:35:31 wirehead_: wouldn't you just use mistral, and schedule a workflow which pokes more servers into your autoscaling group? 20:35:49 do you think it makes sense to discuss separately this topic more detailed? 20:35:52 wirehead_: Heat wouldn't need any special Mistral sauce for that use case. 20:36:09 mostly, I'd like to get more details on how exactly you see using taskflow/mistral 20:36:30 rakhmerov: so there's 3 possible integration points 20:36:32 zaneb: My thinking about using Mistral was that we'd move to a distributed workflow model instead of having the engines be directly responsible for managing the workflow. 20:36:49 the point is: I am not sure everyone clearly understands the key differences between taskflow and mistral 20:36:52 zaneb: I'm now convinced that we should do away with the large scale workflows, which would mean Mistral would be overkill. :) 20:37:02 #1 is what we discussed just now. I think we're leaning toward taskflow for that 20:37:27 #2 is implementing Mistral resources in Heat. I think we would probably do that 20:37:33 rakhmerov: I think that is true, I certainly don't 20:38:32 ok, so given that we don't have much time right now (not a suitable format) I'd suggest we have a separate discussion/meeting about that. 20:38:41 I want to see some Mistral resources in heat, but for pretty boring reasons (just to expose scheduling stuff) 20:39:15 rakhmerov: maybe a ML post with a summary and links to most appropriate docs? 20:39:16 #3 is integration points for external workflows within the Heat workflow 20:39:24 we finally got a lot implemented and are about share the results with the community so we can tell in details 20:39:28 rakhmerov: I think #1 is dependent on the outcome of our design sessions on the convergence shift, and so can't really be discussed seriously right now. #2 is a no brainer.. we should have resources for all OpenStack API's. 20:39:49 rakhmerov: I think most folks understand the goals in a broad sense, but not the details, current status or roadmap 20:40:07 SpamapS: agree that it can't be discussed seriously now 20:40:22 zaneb: I think #3 is something we can do so generically that Mistral doesn't need special sauce. 20:40:38 ok, so let me use ML to plan further activities on that 20:40:59 so we decide where and how Mistral might be useful for you 20:41:29 SpamapS: agree, but there's also the question of making a higher-level tool like Murano that combines declarative (HOT) and workflow (mistral) elements of an application 20:41:54 zaneb: Let's call that one Sauron. 20:41:55 rakhmerov: +1, mailing list is the best place to have that discussion imo 20:42:32 all right, I think we're good to move on? 20:42:39 yesplz 20:42:43 zaneb: ok, sure. Just wanted to have a brief "live contact" with you guys and see if I can help with this 20:43:04 rakhmerov: np, thanks for coming along :) 20:43:08 #topic Pending heat-templates patches for software config 20:43:16 stevebaker: I assume this is you 20:43:18 so i added this one 20:43:22 oh, ok 20:43:46 I realised that there are a lot of good changed pending in heat-templates for the software config stuff 20:43:56 all gated basically by one change: https://review.openstack.org/#/c/79758 20:44:11 I reviewed all of these IIRC and my comments have not been addressed 20:44:12 I need to refresh those changes based on feedback, but once that is done some reviews would be good, since software-config doesn't do much without them ;) 20:44:19 most of them look good to go 20:44:41 I've been distracted by other things 20:45:14 stevebaker: yeah, that's what I thought. If you like I can take this one and post a new patch set. but wanted to talk to you first. 20:45:41 This one and then the next dependent one ... and after that a lot of stuff can land. most of it already approved. 20:46:44 tspatzier: I'll let you know. If I don't get to them today I won't be able to until Monday 20:46:52 tspatzier: be aware that you'd probably have to resubmit all of them 20:46:59 anyone test this patches accordingly to readmes? I was trying to do that and I had some issues with pulling deployments list 20:47:02 otherwise they will have outdated dependencies 20:47:19 zaneb: good to know 20:47:51 also there have been some fixes in os-*-config which needed to be released. Images you build won't work without those fixes 20:48:01 bgorski: I did test them. And they seem to work pretty well. Except the one issue on the first in chain. 20:48:07 SpamapS: I haven't checked if you did that yet 20:48:21 stevebaker: _just_ now 20:48:26 \o/ 20:48:49 tspatzier: I'll ping you if a handover is needed 20:48:51 stevebaker: it was 5 minutes ago, so it might even be on pypi 20:49:05 #topic Free-for-all 20:49:09 https://pypi.python.org/pypi/os-collect-config/0.1.16 indeed it is 20:49:40 stevebaker: sounds good. just wanted to make sure we get all this really great stuff in which is depending on the first patch. 20:50:35 I'm not opposed to wrapping up 10 minutes early, for the record ;) 20:51:01 If anyone wants to take a look, we have a nasty VolumeAttachment regression: 20:51:04 zaneb: well, we could look at lolcats in lieu of wrapping up 10 minutes early. 20:51:05 https://review.openstack.org/#/c/89796/ 20:51:51 evidently the cinder volume attachment API doesn't actually, uh, attach the volume 20:52:00 shardy: any evidece that doesn't re-introduce the race? 20:52:09 just FYI if anyone feels like taking a look 20:52:28 stevebaker: I'm not 100% sure but I tested locally and it worked for attach and detach/delete 20:52:47 shardy: looks like this depends on actual cider driver 20:52:48 stevebaker: the commit message *claims* it's considered the race, but yes that is a concern 20:52:55 shardy: ok, could you reproduce the race before? 20:53:07 reviews welcome, but volume attachements just silently break on current master 20:53:14 so would be good to fix :) 20:53:32 FYI I'm working on a tempest test which would catch any future regression like this 20:53:33 * SpamapS eyes tempest 20:53:39 * pas-ha facepalms himself 20:53:45 shardy: odd, do you know how long ago the original change landed? 20:53:56 SpamapS: patch should be up for review tomorrow ;) 20:54:01 shardy: my hero! 20:54:22 I thought my test was broken, then I realized it was *actually* broken 20:54:52 SpamapS: have we talked the scaling bug ? 20:54:54 <3 20:55:03 randallburt: https://review.openstack.org/#/c/86638/ 20:55:08 SpamapS: cause stevebaker's patch set didn't make enough difference 20:55:22 that broke it two days ago AFAICT 20:55:35 stevebaker: do you think running multiple engines would give enough headroom? We're testing on 24 core machines .... 20:55:43 shardy: cool, got it. I wondered why we weren't seeing it on our cloud. We're a few days behind master so we missed the breakage 20:56:05 randallburt: ah, cool 20:56:57 lifeless: that might help. I've since added to the patchset but am still working on an optimisation which would make the biggest difference to the number of queries per stack load 20:57:45 randallburt: because you don't have provisioned servers polling heat every 30 seconds, because you have no mechanism for those servers to authenticate ;) 20:59:18 stevebaker: so rackspace have no dynamic config via heat? 20:59:26 stevebaker: that seems to rather devalue it :) 20:59:42 stevebaker: randallburt was talking about the cinder bug 20:59:47 oh! 20:59:58 but actually that was a question we wanted to ask 20:59:59 lifeless: currently rackspace chef config is triggered by sshing *from* heat-engine 21:00:12 has rackspace seem the same performance/scale concerns we have 21:00:25 time's up 21:00:26 stevebaker: to end user instances? 21:00:34 -> #heat 21:00:39 #endmeeting