15:00:12 #startmeeting manila 15:00:13 Meeting started Thu Nov 5 15:00:12 2015 UTC and is due to finish in 60 minutes. The chair is bswartz. Information about MeetBot at http://wiki.debian.org/MeetBot. 15:00:14 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 15:00:16 The meeting name has been set to 'manila' 15:00:39 hello all 15:00:41 hello 15:00:41 hi 15:00:43 \o 15:00:44 hi! 15:00:44 hi 15:00:45 hello 15:00:48 welcome back! 15:00:58 hi 15:01:07 Hi 15:01:41 #agenda https://wiki.openstack.org/wiki/Manila/Meetings 15:01:55 welcome back from tokyo 15:02:05 (for those of you who went) 15:02:15 It was a great summit 15:02:31 A lot of what we discussed was captured on the etherpads for the sessions 15:03:00 I will probably send out my own email summary 15:03:15 bswartz: Could you link the etherpads? Or just send the link in the email? 15:03:23 but we decided that a number of things needed followups 15:03:28 https://wiki.openstack.org/wiki/Design_Summit/Mitaka/Etherpads#Manila 15:03:33 dustins: https://wiki.openstack.org/wiki/Design_Summit/Mitaka/Etherpads#Manila 15:03:39 gouthamr, bswartz: Thanks! 15:03:43 oh thanks gouthamr 15:04:01 you're welcome bswartz dustins 15:04:15 okay so one topic from before the summit 15:04:23 #topic Decide on read-only access rules as a required feature 15:04:50 we were discussing this one before the summit, and it also came up duing the session about export location metadata 15:05:13 as far as I know, there is only one driver that has problems with read-only access, and that is the generic driver 15:05:16 hello all 15:05:43 the problem is limited to CIFS and there is a proposed workaround involving an additional export location 15:06:13 so I propose that we make read-only support mandatory for all drivers and document that 15:06:23 bswartz: +1 15:06:30 I had a problem with it on 3PAR NFS but verified a work-around exists. 15:06:40 * bswartz tries to remember how to do an IRC vote 15:06:44 Also on CIFS I would need to stop using IP in addision to User 15:06:57 addition 15:07:02 markstur_: IP-based access control is not required for CIFS 15:07:16 so I'd remove and existing feature 15:07:41 bswartz: but is the workaround related to export location metadata? 15:08:10 ganso: yes 15:08:33 well export location metadata would help you figure out which export location to attach to depending on if you wanted read or write access 15:08:46 ganso: problem of generic driver that it can not have both RW and RO rules for one export 15:09:10 okay lets register official stances 15:09:13 but most drivers would still use one export location for both RO and RW 15:09:16 #startvote Should read-only access be a mandatory feature for all drivers? Yes, No 15:09:17 Begin voting on: Should read-only access be a mandatory feature for all drivers? Valid vote options are Yes, No. 15:09:18 Vote using '#vote OPTION'. Only your last vote counts. 15:09:24 #vote yes 15:09:32 #vote yes 15:09:34 #vote yes 15:09:37 #vote yes 15:09:41 #vote yes 15:09:45 #vote yes 15:09:45 #vote yes 15:09:49 #vote yes 15:09:51 #vote yes 15:09:57 30 more seconds 15:10:02 #vote yes 15:10:25 #endvote 15:10:26 Voted on "Should read-only access be a mandatory feature for all drivers?" Results are 15:10:28 Yes (10): bswartz, ganso, gouthamr, JayXu, cknight, vponomaryov, markstur_, mkoderer, xyang1, dustins 15:10:36 okay that was easy 15:10:44 boring 15:10:47 lol 15:10:59 okay ganso you can update the minimum required features doc 15:11:26 #action ganso: ganso you can update the minimum required features doc 15:11:29 sure, I was waiting for this definition 15:12:24 so I think we will go forward with proposed workaround in the generic driver 15:13:04 and after we have export location metadata we will discuss exactly how we want to communicate to users about read-only export locations 15:13:18 #topic Manila DR - Design/Code feedback 15:13:26 gouthamr: you're up 15:13:38 okay 15:13:49 bringing up the DR design.. 15:14:08 I'm hoping some of you have had a chance to look at the design document 15:14:20 #link https://wiki.openstack.org/wiki/Manila/design/manila-mitaka-data-replication 15:14:24 #link https://review.openstack.org/#/c/238572/ 15:14:28 #link https://review.openstack.org/#/c/235448/ 15:14:42 I was hoping to get some feedback on the code and some situations where we have no clear answers as of now.. 15:14:56 thanks bswartz.. 15:15:25 gouthamr: is there a list of open questions you want to get answers to? 15:15:25 The questions are here: https://wiki.openstack.org/wiki/Manila/design/manila-mitaka-data-replication#FAQs_and_Unanswered_Questions 15:16:14 some of them have assumptions we made while proposing the implementation : https://review.openstack.org/#/c/238572/ 15:16:49 gouthamr: it's confusing to mix the answered and unanswered questions 15:17:11 gouthamr: Looking at the design doc, I'd change the name of the "replication" extra-spec 15:17:14 bswartz: ah true. Will seperate them out.. 15:17:15 I'll update the doc to split them 15:17:23 Perhaps to something like "replication-type" 15:17:24 thank you bswartz 15:17:34 * dustins makes a note to reflect that in Gerrit 15:17:34 gouthamr: you are only targeting on AZs replication? What about regions? 15:17:54 mkoderer: yes, currently AZs.. 15:17:57 mkoderer: inter-region replication is not in scope for this feature 15:18:15 bswartz: fine, just for my understanding 15:18:17 mkoderer: it's clearly a feature we want, but it would look pretty different from intra-region replication 15:18:25 bswartz : you will notice that the code currently does not stop replication from being initiated within the same AZ.. 15:18:54 gouthamr: we need to add that as a question -- Should we support intra-AZ replication? 15:18:56 dustins: Sure.. I agree. 15:18:57 I say now 15:19:00 I say no* 15:19:03 bswartz: exactly... geo region replication would be an awesome feature for us :) but anyway first step AZs are ok 15:19:09 but I've heard others make cases for why we should 15:19:39 bswartz: Perhaps it's something that we can plan on for later, but not for this particular instance 15:20:03 And we should make it clear that this spec is only for inter-AZ replication? 15:20:14 bswartz: sure adding that question. 15:20:18 dustins: well according to gouthamr intra-AZ is currently allowed 15:20:27 I would like to disallow it 15:20:34 bswartz: Ah, I gotcha now 15:20:36 unless we can come up with a clear use cases 15:21:03 we're talking about replicating to another place in the same AZ 15:21:18 for testing its clearly simpler to have 1 AZ instead of 2 15:21:29 but for real use cases, I'm not sure what the point would be 15:21:41 bswartz: true, the original objective to allow it was testing. 15:21:43 bswartz: do we have a clear definition of AZ? do we check if it the same as nova AZ, etc 15:21:50 bswartz: Well, if your cloud only has 1 AZ… 15:22:09 so 1 AZ is only for testing purpose, and real case is to support different AZ? 15:22:12 xyang1: they're supposed to be the same as nova AZs, but that's not actually enforced 15:22:34 xyang1: the only thing we enforce is that each m-shr instance is in exactly 1 AZ 15:22:55 ok 15:23:05 and the cinder folks have even discussed the possibility of a c-vol service spanning AZs so we me may need to consider something similar 15:23:31 our plan is to enforce the AZ across all the projects (nova, cinder,..) using a Congress policy btw 15:23:31 cknight: valid point. bswartz, What if there is only one AZ? 15:23:49 bswartz: right, in cinder AZ is just a string currently 15:23:58 gouthamr: if there's only 1 AZ then you can't use replication 15:24:25 bswartz: that seems pretty limiting for smaller deployers. 15:24:26 xyang1: it's a string, but it's stored in 3 places and used for decision making in 2 places 15:24:51 bswartz: yes, there is plan to make it really meaningful 15:25:14 cknight, +1 for small use cases. 15:25:21 cknight: smaller deployers won't really have a DR solution then -- which isn't surprising because it's hard to do proper DR at small scale 15:25:30 bswartz: i think having replication support within 1 AZ is good for testing 15:25:54 of course we when add other forms of replication, such as inter-region or inter-cloud, then we may address those smaller use cases better 15:26:23 bswartz: or poc type of deployment 15:26:37 AZ are used for many different things in many deployments.. I would suggest to allow replication within the same AZ too 15:26:45 okay so let me try to explain my objection to single-AZ replication 15:26:57 the AZ is part of the UI for replication 15:27:26 you create a share, and you specify the AZ at creation time, and also that you want the share to be replicated (by selecting a share type with replication) 15:27:43 then you add a replica of that share by specifying the AZ for the replica 15:28:18 now you have a share with 2 replicas, in 2 AZs, with different export locations in each AZ (because there are actually 2 instances under the covers) 15:29:32 if the second AZ is the same as the first AZ, then you just have a bunch of export locations in the same AZ 15:29:57 and it's tough to test a real failure scenario in that case 15:30:11 how would you even simulate a failure for testing purposes? 15:31:19 bswartz: You could take down a single backend to simulate a failure. 15:31:29 testing with generic driver or vendor storage? 15:31:40 cknight: yes and not the entire az bswartz 15:31:45 Hi, everyone. I am from Nexenta and we are considering making a Manila driver for our backend. I have a question: what is the mandatory requirement for access to share? Is it enough to grant access by user/group or access by IP is mandatory too? 15:31:45 jayxu: both cases matter 15:32:08 qeas: please wait until we finish this topic 15:32:14 for me all depends on the definition what AZ really means.. because it is sometime just used to group compute hosts together 15:32:27 bswartz: sure, sorry 15:32:59 you could have two backends in the same AZ 15:33:18 mkoderer: yeah I get that different people use AZs differnetly 15:33:45 xyang1, you mean manila backends? 15:33:50 mkoderer: however I believe that AZs are *supposed* to represent a failure domain, which is why we limit volume attachment across AZs 15:33:53 JayXu: yes 15:34:21 I meant you can failover from one backend to another in the same AZ 15:35:09 so the use case we're thinking about is a single system failure (of the storage system) 15:35:18 where the compute and network resources are all fine 15:35:33 bswartz: +1, but anyway the questions is if Manila wants to enforce this defintion for this feature :) I am uncertain 15:35:39 bswartz : yes, failure can happen at the specific storage backend 15:35:50 in that case replication between 2 systems in the same AZ makes sense 15:36:24 bswarstz, in vendor storage, it may allow to trigger failover in array side. 15:36:40 but I think you have to have a pretty low opinion of your storage system to expect it to fail by itself (in the absence of a real disaster) 15:36:43 failover a share to the replica 15:37:11 we have the command to do the unplanned failover 15:37:21 sorry, planned failover 15:37:46 planned failovers may be a more compelling use case 15:37:53 bswartz, as you mentioned, it would be unplanned failover 15:38:02 because then we're not talking about disaster recovery, but perhaps maintenance 15:39:26 well I guess there is broad support for the idea of intra-AZ replication 15:40:02 bswartz : +1 15:40:04 let's put it to a vote 15:40:08 in my experience storage fails alot (hpc background). this sounds good 15:40:41 #startvote Should Manila share replication allow 2 replicas in the same AZ? Yes, No, Undecided 15:40:42 Begin voting on: Should Manila share replication allow 2 replicas in the same AZ? Valid vote options are Yes, No, Undecided. 15:40:43 Vote using '#vote OPTION'. Only your last vote counts. 15:40:49 #vote yes 15:40:50 #vote yes 15:40:51 #vote yes 15:40:53 #vote undecided 15:40:54 #vote yes 15:40:56 #vote yes 15:41:03 #vote yes 15:41:05 #vote undecided 15:41:08 #vote yes 15:41:12 #vote yes 15:41:18 #vote yes 15:41:23 #vote yes 15:41:24 mkoderer, make it less boring? 15:41:30 :) 15:41:35 jasonsb: white boxes fail a lot, and individual nodes in a clustered system fail a lot, but a correctly designed system should be tolerant to such failures 15:41:52 should being the operative word :P 15:41:53 30 seconds 15:42:22 #endvote 15:42:23 Voted on "Should Manila share replication allow 2 replicas in the same AZ?" Results are 15:42:24 Yes (10): ganso, gouthamr, JayXu, cknight, vponomaryov, jasonsb, markstur_, mkoderer, xyang1, dustins 15:42:25 Undecided (2): bswartz, u_glide 15:42:38 okay I think the people have spoken 15:42:51 #action: gouthamr will update documentation 15:43:00 gouthamr keep going in the direction you're going with supporting replicas in the same AZ 15:43:08 also make the design doc clear on that point 15:43:30 gouthamr: anything else for this week? 15:43:31 bswartz: good stuff.. 15:43:57 I have a question: what is the mandatory requirement for access to share? Is it enough to grant access by user/group or access by IP is mandatory too? 15:43:58 at this point, nothing else.. i hope i can get some reviews to continue to enhance this 15:44:09 okay great 15:44:14 #topic open discussion 15:44:37 okay qeas now we'll answer your questions 15:44:48 qeas: IP for NFS, user/group for CIFS is fine. 15:45:11 yes, we don't require IP-based access control for CIFS shares -- only user/group based 15:45:20 cknight: okay, thanks 15:45:28 and for NFS - IP is required, right? 15:45:33 qeas: you should review ganso's doc and make sure it is clear 15:45:38 the docs might not be clear on that point because we initially did require IP-based access for CIFs shares and then later we decided that was a bad idea 15:45:52 qeas: yes 15:45:52 ganso: can you provide link 15:45:53 qeas: yes for NFS shares IP-based access control is mandatory 15:45:54 xyang1: could you give me a link? 15:46:17 xyang1: sure just a sec 15:46:17 xyang1: is his doc merged yet? 15:46:27 not yet 15:46:38 http://docs-draft.openstack.org/14/235914/9/check/gate-manila-docs/7c45126//doc/build/html/devref/driver_requirements.html 15:47:06 ganso: I thought it is still in review 15:47:16 xyang1: it is 15:47:42 docs-draft.. never heard about it.. cool 15:47:56 mkoderer: New one for me too 15:47:57 Neat! 15:48:07 ganso: nice 15:48:15 mkoderer: that's just the output of a jenkins test job 15:48:37 another question I haid is: does nova have cli command to attach share to instance, like it has for cinder volumes? 15:48:39 bswartz: yeah.. but anyway helpful :) 15:48:47 qeas: not yet, but it may be added in mitaka 15:49:04 bswartz: ok thanks 15:49:20 qeas: the current design is that the guest has direct network access to the share and doesn't need the hypervisor to be involved at all 15:49:31 bswartz: do you have anyone working on it? 15:49:45 this allows us to support VMs, containers, and bare metal all with a single solution 15:49:50 bswartz: I saw that coming up in Sage's talk, good idea 15:50:11 xyang1: sage said he'd find someone at redhat to work on it 15:50:18 bswartz: cool 15:50:41 I hope he's able to do that quickly -- I know the nova guys and if there isn't a spec up in the next few weeks it's not happening in mitaka 15:50:43 bswartz: makes sense, I was just thinking of end-user comfort 15:51:05 qeas: yes, automating the mount process is very important from and end-user perspective 15:51:15 bswartz: ya, getting stuff in nova is not trivial 15:51:23 but we have multiple approaches for that, not limited to automation through nova 15:52:18 alright anything else for today? 15:52:26 i have topic 15:52:48 attended magnum weekly on tuesday 15:52:54 can i go ahead? 15:53:03 jasonsb: yeah we have about 7 minutes 15:53:11 this is short 15:53:35 i mentioned that manila had an interest in seeking requirements from magnum for mount automation 15:53:51 adrian responded first with this: one thing I'm a bit worried about is filesystems as an attack vector 15:54:06 adrian: so I'd rather not allow filesystems to be user supplied 15:54:17 adrian: I'd like a way for a service provider to "sign" a filesystem 15:54:26 adrian: but I'm not even sure if what I want is possible/practical yet 15:54:49 so i think kernel vulnerabilities is something on their minds 15:55:04 then we ran out of time 15:55:13 i'll keep attending and gather more requirements 15:55:18 so obviously they're thinking about access to non-empty file systems 15:55:48 the core manila use case is we give you an empty file system and you store data in it 15:55:54 non empty and filesystems perhaps that have been specially crafted to exploit the kernel 15:56:11 sharing filesystems between tenants does create opportunities for malice 15:56:26 of course the specially crafted only comes into play if you have genesha etc on the host 15:56:41 so it is pertinent to the automation strategy 15:57:15 bswartz: yes, if you have host with single tenant then this problem goes away i think 15:57:29 I think our core use case is people mounting their own filesystems 15:57:39 that needs to be automated and there is no danger in that case 15:58:20 the risks of exploits comes from sharing across tenants, not from automating the mount process 15:58:24 perhaps. i think his comment speaks to the concern about containers in a multi-tenant environment 15:58:45 I think that's more of a corner case 15:58:47 maybe manila doesn't have a problem, but the scrutiny is there 15:58:58 but I realize that containers are bit "special" when it comes to storage 15:58:59 i can see it 15:59:13 I need to spend more time playing with container solutions and filesystem mounts 15:59:18 you have a host and its running ganesha on host and sending manila into container 15:59:41 somebody could craft things to go after the genesha and compromise the kernel 15:59:50 IMO ganesha wouldn't be running on the same box as a guest container 16:00:14 one of sage's examples did show it this way iirc 16:00:24 all of our ganesha-related use cases involve a separate bridge machine or running it on the hypervisor with full virtualization 16:00:31 it would be pointless on a container host 16:00:44 oh 16:00:49 maybe it was just knfsd 16:00:53 yeah... 16:00:53 my bad 16:00:58 anyways we're past time 16:01:01 thanks everyone 16:01:08 #endmeeting