15:00:18 #startmeeting massively_distributed_clouds 15:00:19 Meeting started Wed Feb 15 15:00:18 2017 UTC and is due to finish in 60 minutes. The chair is ad_rien_. Information about MeetBot at http://wiki.debian.org/MeetBot. 15:00:20 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 15:00:23 The meeting name has been set to 'massively_distributed_clouds' 15:00:24 #chair ad_rien_ 15:00:31 Current chairs: ad_rien_ 15:00:43 #link https://etherpad.openstack.org/p/massively_distributed_ircmeetings_2017 agenda line 185 15:01:13 Hi Adrien 15:01:17 Hi all 15:01:21 o/ 15:01:36 ping ad_rien_, msimonin, denaitre, dsantoro, menthos,  serverascode, jfpeltier, kgiusti, ansmith 15:01:47 o/ 15:02:06 o/ 15:02:11 ok let's wait one more minute 15:02:11 o/ 15:02:12 o/ 15:02:12 o/ 15:03:53 ok 15:04:23 so let's start (I was waiting additional folks from Orange but they will join us during the meeting hopefully) 15:04:47 #topic Announcement 15:05:10 So from our side, only one news regarding the patch for Kolla that has been accepted 15:05:28 So we should be able to deploy multi regions by leveraging kolla soon. 15:05:46 #link https://review.openstack.org/#/c/431588/ kolla multi region support 15:06:03 There is one more patch to be accepted (please rcherrueau could you put the link) 15:06:34 * rcherrueau looking for the link 15:07:07 #link https://review.openstack.org/#/c/431658/ 15:07:18 Thanks 15:07:53 finally we are working on the experiment plan for the first series of WAN evaluations (but we will discuss that point later). 15:08:00 So that's all from Inria side. 15:08:09 Anything else from your side guys ? 15:08:25 From Orange Rabbit Noise generator, 15:08:32 ok 15:08:39 do you want to discuss it now? 15:08:53 (@all please remind/feel free to contribute to the pad) 15:08:57 I have a tool allowing to send 400+ msg/s but no results with OpenStack 15:09:13 I don't have a presentation ready 15:09:44 I had to multi-thread it for performance and got some issues 15:09:57 jfpeltier: is the tool available - open source? 15:10:12 It will soon be, yes 15:10:16 links? 15:10:22 jfpeltier: sweet! 15:10:33 I did not put it on my github yet 15:10:54 It is based on the tool made by Ayeb, I'll send his link 15:11:02 ok 15:11:16 If I'm right we pushed the link last week. 15:11:49 yes https://github.com/abousselmi/osnoise 15:11:56 ok not that one ;) 15:11:57 sorry Ayoub 15:12:11 #link https://github.com/abousselmi/osnoise Orange contribution to stress the AMQP bus 15:12:47 ok 15:12:56 do you want to add something else? 15:13:15 nothing else from me 15:13:22 ok 15:13:34 anyone before switching to the next topic? 15:14:29 seems not 15:14:40 #topic Deployment scenarios / WAN evaluations 15:15:02 As discussed our tool is now ready to emulate different network topologies 15:15:22 #link https://enos.readthedocs.io/en/latest/network-emulation/index.html Traffic shaping with ENOS 15:15:42 so we are currently working on the experiment methodology 15:16:01 I put some notes in the pad (with the support of rcherrueau and Menthos) 15:17:04 so we would like to perform two kinds of experiments the first one will be based on scenario 1 (i.e. control services deployed on one site and computes notes deployed on remote locations) 15:17:27 we are wondering whether you have some feedbacks/remarks regarding such an evalution? 15:17:42 What should be the latency/bandwith metrics we should emulate? 15:17:51 How many computes nodes per site? 15:18:00 is there a particular rally scenario to execute? 15:18:09 Any remark would be more than welcome ;) 15:18:30 jamemcc: jfpeltier ? 15:18:53 well any measurement is interesting... 15:19:04 Why not just perform all of them and look at what fails? :-) 15:19:05 latency from 10ms to 200ms 15:19:36 (That's what I'm doing with other scenarios, it allows to turn on some red lights) 15:20:08 I don't think bandwidth is so much an issue 15:20:51 ok 15:21:20 because we are not a cloud provider/network operator, we are just wondering what can be the representative topologies? 15:21:45 so any remarks/feedbacks from cloud providers/telcos would be highly appreciated. 15:21:58 I will try to discuss this point next week with telcos. 15:22:36 Ok the second experiment will focus on the multi regions use-case 15:23:12 although this will be addressed after the first series of experiments, we would like to start the braimstorming sessions and get ideas/feedbacks as well? 15:24:02 jamemcc: I know that ATT solution is based on this idea of having independent Openstack on each site and a glue that provides a unified view. 15:24:12 ad_rien_, do you still use the nova fake driver? 15:24:27 yes we still use it 15:24:51 While preparing this meeting, we discussed whether this is the right way to go or not. 15:25:02 Do you think it's a big issue for the tests we envision? 15:25:28 ok let's make a pause on the multi region discussion. 15:25:43 the fake driver is not using nova-compute to neutron communications 15:25:45 matrohon: what do you have in mind? 15:25:46 AFAIR 15:26:18 I'm afraid we're missing some RPC call made by real drivers 15:26:31 not sure about that, need to check 15:26:38 matrohon: I think you're right 15:26:48 ok so this is an interesting point. 15:26:52 Have you looked at our presentation at the last summit? 15:26:57 matrohon: Good remark 15:27:03 Mirantis used a modified libvirt driver instead 15:27:12 Maybe we can use remove the fake drivers from this picture? 15:27:30 Menthos, which eumulates real calls? 15:27:30 I mean since we are not targetting the scalability in this experiment, we can use real nodes? 15:27:55 I mean the number of remote sites is probably the main issue? 15:28:18 matrohon: from what I understood, they took the libvirt driver and just removed the calls to the hypervisor 15:28:30 So I think it was vanilla Neutron 15:28:43 ok, so neutron did the plumbing job 15:28:56 matrohon: Mirantis solution uses a modified libvirt that doesn't start vms 15:29:11 rcherrueau, sounds more realistic 15:29:11 * Menthos looking for a link 15:29:57 cinder should be impacted too if VMs have volume attached 15:30:08 yes 15:30:22 so we need to evaluate a scenario that include cinder volumes 15:30:30 Foud it: https://youtu.be/XURkQ3biF6w?t=10m6s 15:30:56 Menthos, thanks 15:30:58 #link https://youtu.be/XURkQ3biF6w?t=10m6s Mirantis explain the fake driver issue 15:31:53 getting back to cinder, we can have a cinder per site 15:32:07 but this is true that right now we never dive into cinder considerations 15:32:08 rcherrueau: notifies me that my link broke his Emacs, so he can't speak anymore 15:32:43 So that leads us to my initial question: what can be the right scenarios to evaluate? 15:32:49 rcherrueau says Inria haven't had a look at Cinder yet (only Nova, Keystone and Neutron) 15:32:59 There are a lot of components/services, I don't know whether we can evaluate all of them. 15:33:15 ok 15:33:17 agreed 15:33:22 anything else. 15:33:32 Can we move to the region use-case? 15:33:59 I think we should stick to simple scenario first : VMs attached to providers net with no volumes 15:34:12 yes that was indeed the initial idea 15:34:21 then complexifies the picture by adding cinder volumes 15:34:29 +1 15:34:35 by checking live migrations from one remote compute to another one (intra and inter sites) 15:35:00 Yes the rally scenarios for live migrations are a good way to start testing cinder 15:35:45 So getting back to the multi region scenario 15:36:06 we are not convinced that this is the right direction 15:36:31 We are wondering actually why the region abstraction has been proposed? 15:36:42 This is another way to segregate an infrastructure 15:36:58 but there is the cell abstraction 15:37:29 every big cloud provider (amazon ahead) is providing multi-region 15:37:57 matrohon: you mean availability zone, don't you? 15:38:49 i.e. with the multi region there is a strong segregation. For instance you should run a unique/common keystone 15:39:05 if you want your users to be able to run VM on every site 15:39:09 every region (sorry) 15:39:19 i don't know the name at amazon... but I think one need to choose a regino to get endpoint it will work with am i wrong 15:39:34 to link both aspects, you need additional filters in nova-cheduler 15:39:37 so this mean that either you have a global keystone or you have several keystone and a mean to federate them. 15:40:37 jfpeltier: that's exactly my point, I have the feeling that on the first hand you are segregating your infrastructure into several regions (i.e. independent openstack) but then you need additional pieces of software to federate them 15:40:44 adrien in your use-case, do you have unique database? 15:40:47 ….. this looks weird, doesn't it? 15:40:52 not yet 15:41:00 for the moment, we only work with the vanilla code 15:41:25 we would like to identify all issues that prevent the use of one technology/solution 15:42:08 jfpeltier: not yet was the answer of your question (sorry for not beeing clear) 15:42:58 We tried both configurations, unique database allows more fetures but also has some drawbacks 15:43:17 if you look at ongoing deployments with multi-region, they all have a piece of software to aggregate those regions 15:43:26 I guess the normal one is to segragate 15:43:31 matrohon: such as? 15:43:41 ATT is using ecomp for instance 15:43:52 yes that was the question to jamemcc? 15:43:54 an ochetsrator on top of regions 15:44:07 but this mean you have to developp/maintain such a component 15:44:16 indeed 15:44:28 in somehow, you use openstack as libvirt 15:44:52 so it means that you have to redevelop most of services (for instance a scheduler, ...) 15:45:03 this looks like the initial TriCircle proposal 15:45:15 yep 15:45:32 The question that would be great to answer is what are the right arguments to justify such an approach 15:45:47 I guess that they think about it before implementing such an orchestrator 15:45:52 independant failure domain 15:45:58 in the sense ? 15:46:14 is there a ecomp presentation somewhere available? 15:46:26 jamemcc, still there? do you have such a pointer? 15:46:28 workloads in one failing openstack can move to another openstack 15:46:59 Yeah 15:47:05 ok maybe we can put this question on the pad and discuss it next time 15:47:23 I'll look and see if I can find a link before end of meeting. If not will send e-mail. 15:47:27 jamemcc? What do you prefer? 15:47:51 from my point of view, the advantages is that ecomp can operate whatever cloud systems 15:47:56 ecomp should be opensource source, jamemcc, is it already? 15:48:06 s/source/soon 15:48:12 Next meeting would be great - I can probably arrange to have appropriate architect 15:48:13 (i.e. not just OpenStack but all clouds that are openstack compliants) 15:48:18 great 15:48:30 I think this is an important point to clarify because 15:48:48 it will help us to justify either a top/down or a bottom/up approach 15:48:54 I'll have to loko for specific status - which license Mathieu. I'll include that as well. 15:49:01 great 15:49:19 (ten minutes before the end of the meeting, I propose to switch to the next point) 15:49:20 Certainly is the intent and we've announced the intent. 15:50:41 #topic F2F meeting in Boston 15:51:26 I will open an etherpad so we can keep up-to-date informations related to Boston. Who will be there? What are the points we would like to discuss? etc.. 15:51:48 Since I didn't do this pad yet, I propose to discuss it next week. 15:51:48 Ok ? 15:52:04 #action ad_rien_ Create a pad for Boston F2F meeting 15:53:00 ok so let's move to the open discussion moment 15:53:04 #topic open discussions 15:53:11 so please guys the floor is yours 15:54:11 we have looked on the different solutions used to deploy openstack 15:54:35 more particularly kolla, enos, juju, kuberneted and tripleo 15:55:39 * denaitre looking for a link 15:56:47 I'd like to discuss https://docs.google.com/presentation/d/1ghwinrArfoCw1qIsNGxWSrd8NTynYomThERGUpN4f0U/edit?usp=sharing *before* boston if possible 15:56:53 https://goo.gl/7IksQY 15:57:02 next meeting? Just a real quick run through. 15:57:15 #link https://goo.gl/7IksQY how deploying OpenSTack in a massively distributed context 15:57:25 kgiusti: yes I will add it to the agenda 15:57:29 thanks 15:57:31 as one item 15:58:03 [ ] ad_rien_ put the AMQP discussion as a main point for the next meeting. 15:58:13 #action ad_rien_ put the AMQP discussion as a main point for the next meeting. 15:58:19 Ok I propose to close the meeting 15:58:28 ok 15:58:35 ok so talk to you in two weeks 15:58:39 thanks for attending the meeting 15:58:47 #endmeeting