15:00:27 #startmeeting massively_distributed_clouds 15:00:27 #chair ad_rien_ 15:00:27 #info agenda 15:00:27 #link https://etherpad.openstack.org/p/massively_distributed_ircmeetings_2017 15:00:27 Meeting started Wed Mar 1 15:00:27 2017 UTC and is due to finish in 60 minutes. The chair is ad_rien_. Information about MeetBot at http://wiki.debian.org/MeetBot. 15:00:28 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 15:00:31 The meeting name has been set to 'massively_distributed_clouds' 15:00:33 Current chairs: ad_rien_ 15:00:45 Hi 15:00:47 HI 15:01:31 #link https://etherpad.openstack.org/p/massively_distributed_ircmeetings_2017 (line 251) - Agenda for the meeting 15:01:35 o/ 15:01:47 hi o/ 15:01:52 hi 15:01:54 Hi o/ 15:01:57 As usual Guys, may I ask you to add your name in the pad. line 253 15:01:58 thanks 15:02:23 let's wait one more minutes before starting 15:02:57 #topic Announcement 15:03:10 So from my side, nothing special, 15:03:32 I just put a link toward some discussions the Scientific WG had on identity federation challenges. 15:04:05 so that's all. These last few days were completely crazy from my side ;) 15:04:21 Anything else from your side before diving into the agenda 15:04:26 ansmith: jamemcc ? 15:04:59 ok looks not…. 15:05:12 BTW I did receive the scores for the presentations 15:05:27 does anyone know when this should be notified? 15:05:38 s/I did/I didn't 15:05:42 sorry 15:06:14 ok so todays for the agenda we planned to discuss two points: 15:06:20 lol - I'm not following adrien - ask it again - are you talking about the Summit seelction process 15:06:39 jamemcc: yes 15:06:39 ;) 15:07:01 sorry I should probably try to be less enthusiast 15:07:11 I didnt' think the selection was over - just the initial public voting 15:07:18 and leave time for getting answers 15:07:20 ok 15:07:42 do you have an idea about the notification of acceptance date? 15:08:15 Though I'm not on the selection committee this time. Lets keep going and I'll look as I am interested - it's a common question. 15:08:28 jamemcc: ack 15:08:29 ok 15:08:55 so as mentioned today we planned to discuss two points mainly: ecomp (ATT solution) and QPID related discussions. 15:09:10 "After track chairs make their decisions, speakers will be informed by mid March 2017" 15:09:12 so who wants to start ? 15:09:21 thanks Menthos 15:09:53 #info speakers will be informed by mid March 2017 15:10:06 I can present (it should be quick)... 15:10:18 ok please go ahead kgiusti 15:10:18 https://docs.google.com/presentation/d/1ghwinrArfoCw1qIsNGxWSrd8NTynYomThERGUpN4f0U/edit#slide=id.p 15:10:42 sure - the slide deck provides a very high level 15:10:57 introduction to the various messging backends 15:11:03 supported by oslo.messaging 15:11:33 We'll talk about zeromq, rabbit/qpid/ and qdrouterd 15:11:40 So Page 2 15:12:00 This is the traditional message bus deployment for OS 15:12:22 ok 15:12:24 Well understood, much tribal knowledge, used by 99.9999999 deployments 15:12:45 I think we all understand this - so... slide 3 15:13:10 (the pro/con slides were just added ... certainly incomplete) 15:13:25 this is a valuable starting point at least ;) 15:13:30 Let's move forward please 15:13:46 So the limitations and pros should be obvious - any q's 15:13:47 ? 15:14:17 On page 4 - this is the logical topology that can be acheived using zeromq for RPC 15:14:32 The broker is still present, but used only for notifications 15:14:55 This has two immediately obvious advantages - less load on the broker, and direct (non-queued) RPC 15:15:13 I though that zeroMQ was fully P2P based. 15:15:34 Why should we go through the router? 15:15:36 Yes it is, but doesn't support notificiation traffic 15:15:56 Slide 4 - there is no router used in this scenario 15:16:05 ok sorry 15:16:09 was on slide 6 15:16:09 thanks 15:16:27 Just broker for Notifications, Zmq for RPC (dual backend - supported in ocata+) 15:16:52 "(dual backend - supported in ocata+)" did someone test it ? 15:16:58 Slide 5: The biggest issue is that zeromq has very little adoption 15:17:04 yes - ansmith! 15:17:14 good job I didn't know that 15:17:17 is there any link? 15:17:22 He's adding it to Triple O at the moment.... 15:17:34 ansmith: ? 15:17:39 #topic AMQP bus discussion 15:18:03 dual backend in tripleo and across puppet modules 15:18:03 So the biggest issue here is that zeromq is largely unproven. 15:18:35 I'm not sure if anyone is actually deploying it in production.... may be wrong here. 15:18:35 #info zeroMQ support will be available in ocata+ versions (ansmith is working on that) 15:18:57 specifically working on "enabling" dual backends 15:19:04 ansmith: I believe it will be qdrouter first.... 15:19:15 We'll get to that in a moment 15:19:21 primary target is hybrid with qdrouterd for rpc and rabbit for notify 15:20:01 The zeromq driver is very complex when compared to the other driver codebases since the driver has all the intelligence for message topology control embedded 15:20:19 IOW: complexity "at the edge" 15:20:36 ok 15:20:59 And it is unknown how much of a issue the large # of TCP connections required will be 15:21:08 (think TCP resource limits). 15:21:22 So and Q's for 2a? 15:21:32 once available it will be rather simple to evaluate that point. 15:21:38 +1 15:21:54 #action make a comparison with Enos of RabbitMQ, ZeroMQ and QPid for scenario 1 15:21:55 The driver seems well supported (mostly mirantis folks) 15:22:22 rcherrueau: do you know whether kolla provides support for using zmq instead of rabbit? 15:22:23 So bugs should be addressed relatively quickly I would think 15:23:03 ad_rien_: let me check 15:23:04 on to slide 6... 15:23:26 This is a variant of the last slide - just replacing zeromq with qdrouterd 15:23:58 So any "single broker for notification" downsides/upsides also apply here 15:23:59 ad_rien_: No, there is no support in kolla-ansible 15:24:10 rcherrueau: ack 15:24:24 could you please check what will be mandatory to provide such a feature? 15:24:51 However, there is much lower TCP resource usage - one connection to the local router per client. 15:25:02 ad_rien_, rcherrueau : the effort will be akin to what what we did with puppet and tripleo 15:25:04 And the ability to add redunancy 15:25:10 path redundancy 15:25:20 for failover (availability) 15:25:32 ansmith: ack 15:25:34 thanks 15:26:01 kgiusti: could you please just clarify why/when you have to go through the dispatch router? 15:26:10 There is a degree of parallelism in the RPC traffic flows, though less than zmq 15:26:34 ad_rien_: for RPC traffic only. 15:26:59 not sure I correctly understood. 15:27:00 each client would also need a connection to the single centralized broker for notifications 15:27:09 ok like zmq 15:27:22 Each client connects to it's qdrouterd "onramp" 15:27:32 I'm trying to understand the != between the zmq rpc stuff and the qpid one? 15:28:17 Fault recovery via redundant routers, the ability to cross unroutable IP subnets (firewall) 15:28:38 All the routing intellegence and policy is controlled by the routers 15:28:47 ok thanks 15:28:49 the driver is very simple 15:29:05 so page 7 15:29:24 I think we already touched on that stuff - q's? 15:29:49 Slide 8 15:30:04 This simply puts the broker "behind" the router 15:30:20 But this allows broker redundancy 15:30:33 e.g. have a "higher cost link" to the backup broker 15:31:02 Switch on failover, but some message loss of course.... unless clustering/HA is used by the broker 15:31:32 msimonin is absent today unfortunately (and is the AMQP expert one :( ) 15:31:36 but this is moving towards more of a "messaging as a service" model where the message bus is simply a black box 15:31:40 :( 15:32:11 so we have a couple of questions but maybe they are naive 15:32:14 For example there is an Openshift based project (enmass) which we hope to integrate Oslo.messaging with.... futrue stuff 15:33:08 if you are considering a large number of site, let's say 1000 , each composed of 50 servers 15:33:30 if you have router/broker functions on Site 0, would it scale? 15:33:57 in the zeromq (please don't get me wrong I do not say zmq is the solution to select) 15:34:14 only the notification would be sent to Site0 is it correct? 15:34:44 Be aware I don't have actual working experience with such a deployment - so take this as an educated guess :( 15:34:55 ok 15:35:04 Yes w/zmq only notifications goto site 0 15:35:13 ok 15:35:44 The biggest issue is how stable zmq will be given the number of TCP connections necessary to interconnect all servers 15:35:52 that need to communicate via RPC 15:36:10 can be an issue indeed. 15:36:57 In the same case, but with qdrouterd the issue will be how long will the topology take to stablize 15:37:04 can the router block be a distributed one? 15:37:18 (HA/scalability considerations?) 15:38:13 you can deploy as many routers as you deem necessary, but each hop across a router adds some latency (depends on the load on the router) and 15:38:48 more routers == longer time to recover the topology should a failure occur 15:39:17 so each site could have multiple routers if there is a benefit 15:39:42 ok 15:39:59 any other Q's? 15:40:23 Feel free to followup later if you want - my email and nick are in the notes 15:40:23 ok 15:40:30 if we consider the scenario 1 15:40:36 (i.e. the one we selected 15:40:57 where we have all control services in Site 0 and only compute nodes on remote sites) 15:41:20 if we correctly understood, compute nodes will not communicate each other, right? 15:41:52 I believe that is correct - it would be a centralized hub and spoke topology 15:42:07 so the use of zmq or qpid is valid only for scenarios 2/3+ where we will have several services geographically spread 15:42:13 ok 15:42:21 So it is 45. 15:42:37 May I propose to prepare some questions for the next meeting. Can all please agree on that point ? 15:42:44 well, you _could_ try zeromq for hub and spoke if the load on the broker becomes critical 15:43:12 #link https://docs.google.com/presentation/d/1ghwinrArfoCw1qIsNGxWSrd8NTynYomThERGUpN4f0U/edit#slide=id.g1b037c92e2_0_269 AMQP discussion 15:43:34 #action all - please see the amqp slides and prepare questions for the next meeting. 15:43:50 kgiusti: do you want to add something else? ansmith? 15:44:16 specifically, update to kolla-ansible would be to use oslo.messaging transport url inplace of deprecated rabbit host, port, auth as configuration options 15:44:32 on the list to do 15:44:54 this will enable separating the messaging backends 15:45:00 ansmith: you are right 15:45:59 ok 15:46:10 may I ask to move to the following point in the agenda? 15:46:20 (we can come back later if we have time) 15:46:47 +1 15:46:52 thanks 15:47:05 #topic ecomp 15:47:14 jamemcc? 15:47:15 Thanks - I'm here still. I first want to status that I didn't get a commitment from one of the more estblished eComp architects at AT&T to join us. 15:47:32 Just a quick check though: Anyone else joined to talk about eComp? 15:48:16 jamemcc: if you want we can postpone the discussion next time? 15:48:47 I have better leads on who might be appropriate - but perhaps we could instead check to see if anyone read the materials sent out and then talk about where the investigation/collaboration might go. 15:49:22 guys ? 15:49:39 What do you prefer? To talk now or to discuss next time? 15:50:15 (from my side, I would rather prefer the second option as I did not have time to give a deeper look in the different materials) 15:50:20 Yeah - next time - though if there are general quesions - like what level of functionality or whats the strategy for the project - I think I could answer that. 15:50:29 +1 15:50:58 Btw, the slot ends in 10 min, so next time is maybe more appropriate ... 15:51:15 #action all - give a look to the content sent by jamemcc and prepare questions for next meeting. 15:51:39 Answering to the ML ahead of time will be helpful as well: 15:51:40 http://lists.openstack.org/pipermail/user-committee/2017-February/001722.html 15:52:11 #link http://lists.openstack.org/pipermail/user-committee/2017-February/001722.html ML message regarding ecomp 15:52:37 #topic scenario deployment. 15:53:04 ok right now we are just performing the first experiments with Scenario 0. we expect to gather preliminary results by next meeting 15:53:27 nothing more so unless you have question I proposed to move on the next point. 15:54:35 #topic AMQP stresser 15:54:44 I'm not sure Orange folks are there? 15:55:01 I saw mabderrahim but didn't see jfpeltier nor mrohon 15:55:01 ? 15:55:03 guys ? 15:55:06 are you there? 15:55:12 I am here 15:55:44 ok I proposed to keep that on the agenda 15:55:56 next point OpenDiscussion 15:56:01 #topic open discussion 15:56:24 apart what I put in the pad I have nothing more 15:56:33 anything you would like to discuss from your side Guys ? 15:56:58 three minutes before ending the meeting? 15:57:41 ansmith: jamemcc kgiusti anyone else ? 15:57:51 I'm good :) 15:58:00 good 15:58:06 So I propose to conclude the meeting. thanks for being there. 15:58:13 No - but perhaps we can chat a little offline afterwards - cross working group 15:58:13 #end meeting 15:58:23 #endmeeting