14:00:00 #startmeeting cinder 14:00:01 Meeting started Wed Dec 11 14:00:00 2019 UTC and is due to finish in 60 minutes. The chair is rosmaita. Information about MeetBot at http://wiki.debian.org/MeetBot. 14:00:02 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 14:00:05 The meeting name has been set to 'cinder' 14:00:07 o/ 14:00:09 o/ 14:00:10 #topic roll call 14:00:19 \o 14:00:43 o/ 14:00:45 o/ 14:01:59 guess that's everyone 14:02:24 just out of curiosity, can everyone state their UTC offset? 14:02:26 -5 14:02:39 -6 14:02:53 +8 14:02:54 -6 14:03:02 +5:30 14:03:17 smcginnis: and I agree. That is a good sign. :-) 14:03:29 :) 14:03:30 +1 14:03:34 ;) 14:03:49 geguileo: i think you are +1 ? 14:04:15 +1 to what? 14:04:16 he is (unless he is travelling) (even though he is physically on +0, but that's a long story) 14:04:25 rofl 14:04:33 oh TZ 14:04:52 yup +1 now 14:04:56 i will have to get that story some time 14:04:59 ok, thanks 14:05:04 #topic updates 14:05:39 this week is milestone-1, but since we don't have to release any more, it's not that big a deal 14:05:58 but, it's time to do our every-2-month release from stable branches 14:06:15 #link https://etherpad.openstack.org/p/cinder-releases-tracking 14:06:29 so, basically, cinder and os-brick could use releases 14:06:53 but, there is a pep8 failure happening in brick backports 14:07:05 you can see an example here: https://review.opendev.org/#/c/697116/1 14:07:38 it's over W503 and W504 14:07:56 which have to do with whether an operator should end or begin a continued line 14:08:06 pep8 has gone back and forth on that 14:08:20 and master and train for us ignore them both 14:08:45 anyway, i think we should backport the ignore to the stable branches that don't have it 14:09:05 because i'd rather not have to edit cherry picks for something silly like this 14:09:13 save that for serious conflicts 14:09:32 so i have proposed 14:09:35 #link https://review.opendev.org/#/c/698471/ 14:09:48 if everyone is cool with that plan, what i'd like to do is 14:09:59 get the pep8 ignore change backported 14:10:19 then complete the backport of the actual fix, which is "iscsi: Add _get_device_link retry when waiting for /dev/disk/by-id/ to populate" 14:10:28 and then do point releases of os-brick 14:10:39 Obviously the right answer is it should be end of line, but best to ignore. :D 14:11:17 maybe it's because of my eyeglass prescription, but i don't see that well all the way out to column 79 14:11:24 so i prefer it at the front! 14:11:36 but yeah, this is a contentious issue 14:11:54 but i guess the key point here is we should do it the same from master on back 14:12:01 until we take a definitive stand 14:13:02 ok, so if everyone is OK with that plan, please keep an eye out for the patches so we can get it merged 14:13:26 i have not looked at cinder, python-cinderclient to see what they have for this 14:13:34 probably should 14:13:50 #action rosmaita check pep8 ignores in all cinder projects 14:13:58 ok 14:14:09 one more thing: cinderlib is ready for its train release 14:14:15 thanks to geguileo 14:14:20 this will be 1.0.0 ! 14:14:35 awesome!! XD 14:14:52 and i think since we remove py2 support in ussuri, the first ussuri release will be 2.0.0 14:14:58 that project is moving right along! 14:15:13 yeah, sort of off-topic (sowwy), but I'd like to thank geguileo for his work on cinderlib - it already exposed two (very minor) bugs in the StorPool driver 14:15:39 Roamer`: nice!! 14:15:44 Roamer`: thanks to geguileo are always on topic 14:15:50 lol 14:16:33 that's good news though, glad to see that it's helping keep the cinder drivers stable 14:16:48 ok, the final thing: branch closure situation 14:17:05 #link http://lists.openstack.org/pipermail/openstack-discuss/2019-December/011378.html 14:17:29 as we discussed last week, i sent a proposal to the ML to EOL the driverfixes branches, but keep o and p in EM 14:17:50 from the absence of responses, it sounds like that's acceptable to the community 14:18:11 so, i'll revise my EOL patch and resubmit it early next week 14:18:25 (give the release team time to handle all the M-1 releases this week) 14:18:42 so if anyone wants to save driverfixes/newton, this is your last chance 14:19:27 i guess the only other announcement is a reminder to everyone to review open cinder-specs 14:20:26 #topic the FCZM situation 14:20:59 #link https://review.opendev.org/#/c/696857/ 14:21:21 ok, so the situation is that we currently have 2 zonemanagers for fibre channel 14:21:38 brocade has announced EOL for its driver 14:21:46 last supported release for them is Train 14:22:08 unfortunately, this got announced the week before train release 14:22:20 so we did not deprecate the driver in Train 14:22:32 They have also reached out directly and asked us to clean things up since they are not able to. 14:23:00 yes, they are dropping and running 14:23:11 rosmaita: lol 14:23:19 so the first issue is: 14:23:54 can we consider their EOL announcement to customers as a deprecation in Train, so we could remove the driver now in ussuri? 14:24:12 my reason for this is that we don't know if it will run under py3 14:24:23 (the py3 compat seems to have been the final straw for them) 14:25:19 I would hate it for us to have only 1 FCZM driver 14:25:27 even if we mark it as unsupported 14:25:51 hemna feels that way, too 14:26:00 geguileo: ++ 14:26:25 would the community be willing to get leave it as unsupported if I can confirm it works for Py3? 14:26:41 I don't like it, but if the vendor isn't going to support it, then we can only provide the one that does. 14:26:43 Or if I can fix it and make it work for py3? 14:27:16 i am not against that, at least not short-term 14:27:36 if we confirm that it works on py3, i will withdraw my proposal to remove it it ussuri 14:27:44 I would prefer if we keep it unsupported in Ussuri, but I don't think it should fall on the community to support it and keep it around past that. 14:27:59 yeah, we don't want to set a bad precedent 14:28:12 So even if we fix any py3 issues, I think it's safer if we clearly announce it is going away in V. 14:28:40 I think not having that as a FCZM will be a big problem for OpenStack customers 14:29:12 well, we can state "subject to removal in V as per the openstack deprecation policy" 14:29:25 that gives us some room to keep it around longer if we decide it's a good idea 14:29:25 rosmaita: and what are customers going to do? 14:29:26 geguileo: ++ 14:29:38 rosmaita: ok 14:29:52 Too bad hemna isn't here. He thought maybe we could help leverage HPE. 14:30:07 geguileo: i really don't know -- i would hope that customers would lean on brocade to keep it around 14:30:09 I will look for a cheap brocade FC switch I can use to test/fix the driver at home 14:30:12 Well, if Brocade isn't going to support it, I think that's a pretty strong statement. 14:30:36 Customers can complain to them and get them to bring it back if it's that badly needed. 14:30:43 smcginnis: it is, but it will have a big effect on the perception of Cinder/OpenStack on customers 14:31:06 i think that's already happened, though, by their announcement 14:31:16 Only with FC. I don't think that's too big of a perception issue. :D 14:31:21 rosmaita: ++ 14:31:21 as a customer I would complain, and if that gets me nowhere I may decide to go to other platform... :-( 14:31:34 right 14:32:01 ok, let's go with rosmaita's suggestion and see if I can get this fixed tested 14:32:05 ok, well the key issue i wanted to sort out today was our community attitude 14:32:11 and I can bring the topic back once I do that 14:32:11 so sounds like: 14:32:34 we deprecate it now "subject to removal in V as per policy" 14:32:53 (2) geguileo figures out the py3 situation 14:33:13 and then we discuss again 14:33:24 sounds good to me 14:33:41 Great. 14:33:46 ok, thanks geguileo -- you are addressing my major concern 14:34:07 geguileo: To the rescue again. 14:34:19 in the mean time, i don't know if there's anything we can do to encourage more FC participation? 14:34:50 rosmaita: maybe the effort we discussed in the PTG 14:34:57 of making it easy to have a CI 14:35:03 that would lessen the burden 14:35:10 true 14:35:23 ok, that gives us a reason to keep that as a priority 14:36:00 it's weird -- everyone wants excellent performance, but they also want to use commodity hardware 14:36:19 but that's off topic 14:36:22 I believe we discussed the software factory thingy that the RDO folks have to do the 3rd party CI 14:36:30 and I think tosky started documenting some of it 14:37:42 i'll check in with tosky and see what i can do to help 14:38:02 It would be so nice if someone could make 3rd Party CI easy. 14:38:16 amen, you are singing to the choir 14:38:20 I collected some information; the softwarefactory people are going to release a new version these days, with more documentation 14:38:30 great 14:38:33 the idea is that following their documentation should be enough to get started 14:38:54 tosky: that's awesome!! 14:39:16 and we can always submit clarifying patches 14:39:22 that's really good news 14:39:35 #topic support volume local cache 14:39:46 LiangFang: that's you 14:40:07 I work for the poc code change 14:40:18 and the poc is almost finished 14:40:30 will push to review in coming days 14:40:44 poc code is almost finished 14:40:52 one question is 14:41:08 the cache software will have some configuration 14:41:11 e.g. cache id 14:41:35 normally if we use one fast ssd as cache, the the cache id will be 1 14:41:52 the use two ssd, it may be 2 14:42:37 then if we use this ssd to cache for a remote volume, we need to specify the cache id (means the fast ssd) 14:43:03 but we can specify the cache id as a big number, e.g. 9999 14:43:37 choice1: the cache id can be set in nova-cpu.conf 14:44:29 choice2, the cache id can be hard coded in os-brick, but the number should be large enough, e.g. 9999, so this number will not be used by others 14:45:06 choice 2 is easy for coding 14:45:35 choice 1 need the configuration change in nova 14:45:52 I prefer choice 1. 14:46:10 LiangFang: I am not following sorry 14:46:17 LiangFang: where do we use this cache ID? 14:46:32 Is this ID something that matches a hardware ID of the cache? 14:46:41 is the issue that there may be multiple caches available, and we need to know which one to use? 14:46:42 what if there are multiple caches? 14:46:47 when caching a remote volume, should specify which cache to use 14:46:57 rosmaita: yes 14:47:31 normally one cache in system, but it can be multiple caches 14:47:34 what if we want to use multiple caches? 14:47:51 I just need one cache 14:48:15 but what if you have 2 and you want to use both? 14:48:33 we don't need to solve it now, but the solution should be able to support it in the future 14:48:38 without having to change everything 14:48:46 I plan to use only one 14:49:10 the cache id is increase one by one by default if not specified 14:49:11 but someone else may want to use 2, right? 14:49:25 LiangFang: is increased by the Linux system? 14:49:38 by the opencas software 14:50:14 and that ID may change when you reboot the system? 14:50:22 it can add more than one cache. the first cache's id if 1 by default, the second id would be 2 by default 14:50:35 what happens if you reboot? 14:50:44 can they have different IDs then? 14:50:52 it can be configured in opencas configuration file, or use the opencas admin tool 14:51:16 don't caches have a unique ID somewhere? 14:51:22 I mean, os-brick don't know which cache to use 14:51:43 one way is let nova to tell os-brick 14:52:04 yes, but if nova tells os-brick use 1 14:52:07 another way is os-brick just use a hard coded cache id, e.g. 9999 14:52:11 and the system have 1 and 2 14:52:14 then we reboot 14:52:26 and now the cache that was 1 before is now using id 2 14:52:34 then we won't be using the cache we want 14:53:10 the cache id will not change after reboot 14:53:18 oh, ok 14:53:21 then we are good 14:53:30 it is configured in opencas configuration file 14:53:38 I prefer the nova conf, same as you 14:53:54 thing is: how os-brick know which cache to use 14:54:02 one is : nova tell it 14:54:17 another is hard coded in os-brick 14:54:41 the nova approach seems like the right one 14:54:44 second way is easy, but need to document: cache id is 9999 14:55:10 geguileo: ++ 14:55:11 ok, that asnwers my question ... the operator would have to configure a cache with id 9999 14:55:31 Hardcoded values always seem dangerous to me. 14:55:37 rosmaita: yes, if 9999 is hard coded in os-brick 14:55:46 ok... 14:56:04 so with choice 1, you would define a dedicated cache id for brick to use in nova.conf ? 14:56:09 so nova configuration file need to be changed 14:56:25 rosmaita: yes 14:56:43 ok 14:56:50 but 9999 would be safe for opencas forever 14:56:50 I think that's the best approach 14:56:59 ok 14:57:12 I don't like the hardcoding magical numbers in our code 14:57:19 probably everyone will use 9999, but making this configurable is better 14:57:32 ok 14:57:51 so I will change my poc code 14:57:51 ok, we are down to 2.5 minutes 14:58:04 LiangFang: did you get all the answers you need? 14:58:12 (for now) 14:58:32 I'm write a white paper about this with our telecom customer these 3 weeks, will let you know when we finished 14:58:33 LiangFang: thanks 14:58:44 LiangFang: awesome!! 14:58:47 rosmaita: yes, thanks 14:59:02 yes, it will be helpful to see how all this works for real 14:59:09 ok, 1 minute for 14:59:12 #topic open discussion 14:59:47 ok, well thanks for a productive meeting everyone! 14:59:56 Thanks rosmaita ! 15:00:00 see you next week 15:00:06 don't forget to review cinder-specs! 15:00:07 see you, thanks 15:00:12 #endmeeting