14:01:12 #startmeeting cinder 14:01:12 Meeting started Wed Nov 13 14:01:12 2024 UTC and is due to finish in 60 minutes. The chair is jbernard. Information about MeetBot at http://wiki.debian.org/MeetBot. 14:01:12 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 14:01:12 The meeting name has been set to 'cinder' 14:01:16 o/ 14:01:17 #topic roll call 14:01:19 hey 14:01:24 o/ heya everyone 14:01:26 o/ 14:01:30 o/ 14:01:41 Hi 14:01:49 o/ 14:02:04 Hi everyone 14:02:08 o/ 14:02:29 o/ 14:02:49 o/ 14:03:07 o/ 14:05:08 hello everyone, thanks for coming 14:05:13 lets get started 14:05:33 #topic annoucements 14:05:42 M1 is this week 14:05:54 #link https://releases.openstack.org/epoxy/schedule.html 14:07:05 I was in a TC meeting yesterday disucssing the state of our documentation for 3rd party CI 14:07:30 rosmaita reminded me of this one: https://etherpad.opendev.org/p/cinder-drivers-documentation 14:08:17 the general idea in the meeting is that it's difficult to find up-to-date docs on how to configure zuul to run devstack, test a local feature, and report that to gerrit 14:09:06 i agree. it was very painful setting up SoftwareFactory - very little or poor documentation for Zuul configuration of SF 14:09:11 i suspect that's true, CI is not my area of comfort 14:09:46 so I'm wondering, what are our toughts, do we agree? Is there a doc/s that isnt' well known? 14:10:14 simondodsley: how did you finally get it to work, just trial and error? 14:10:15 I remember we had a presentation in some PTG about setting up CI with software factory, not sure how in-depth was it but at least a reference 14:10:51 for the TC meeting reference, here's the backlog: https://meetings.opendev.org/meetings/tc/2024/tc.2024-11-12-18.00.log.html 14:11:15 > test a local feature, and report that to gerrit 14:11:15 What does a local feature in the context of a Zuul configured as a third-party CI mean? 14:11:15 > I remember we had a presentation in some PTG about setting up CI 14:11:15 that was me presenting :) 14:11:20 someone from Pure did a virtual presentation at one of the midcycles (i think), and someone else did a presentation at the last vancouver summit 14:11:33 sp-bmilanov: was that you at vancouver? 14:11:44 they're mostly concerned about how potential contributers may be turned away if 3rd party CI is both required and nearly un-navigatable 14:12:08 This is all the documentation out there for SoftwareFactory - https://softwarefactory-project.io/docs/contributor/index.html. It is currently only supported on RHEL 9 these days as well. 14:12:22 oy 14:12:34 well, should be ok on Rocky 9, right? 14:12:38 The Pure person is no longer with us, hence why I had such issues rebuilding a dead CI 14:12:47 sp-bmilanov: as in, testing code that requires local hardware 14:13:03 sp-bmilanov: cinder is relatively unique to the other projects in that respect 14:13:09 No idea- it comes from Red Hat so unlikely they will help with Rocky 9 issues 14:13:25 manila also requires 3rd party CIs as well 14:14:21 jbernard: sure, but what I would usually do submit a WIP to upstream which will get detected as an event at our Zuul, that's why I got confused by the "local" part 14:14:54 I got what you mean now 14:14:55 sp-bmilanov: ahh, understandable 14:15:40 i think i'll try to compile a list of our various docs to start 14:16:03 sp-bmilanov: you should really be running all the main tox tests locally and passing before submitting an upstream patch 14:16:09 having a single living setup/configuration document would be ideal, but starting with a survey of what we have seems like a good starting point 14:17:02 simondodsley: oh yeah, definitely, it's after those that I send the WIP upstream 14:17:34 jbernard: it might be worth setting up a working group to try and build a CI from scratch and document it. 14:17:47 simondodsley: that's an excellent idea 14:18:04 +1 14:18:09 I have some SF documentation from my build attempts. Happy to join this 14:18:16 given that I have nearly zero experience in doing that, i have no idea how much time and effort might be required 14:18:23 simondodsley: what did you end up doing with your CI? Zuul from scratch or a new SF? 14:18:31 one thing to ask here though, is how vendors feel about third party CI in general 14:18:34 new SF 14:18:47 i think 3rd party is critical for cinder 14:18:56 becasue the alternative is no CI and no in-tree non-community drivers 14:19:09 rosmaita: my sense is that they are interested in complying, as long as the hoops are not too numerous and time consuming 14:19:20 there are so many potential changes that could affect other drivers, especially if vendors make changes to os-brick 14:19:49 getting the CI setup is the annoying part - but once it is up and running it doesn't need much care and feeding 14:19:51 rosmaita: but i have no empirical evidence, just a hunch 14:20:12 ok, how about this: 14:21:07 i will distill the TC meeting backlog into a summary, collect our current scattered docs into a single reference 14:21:38 and we can see what we have, and try to create a POC project that demonstrates a simple CI that works 14:21:46 a starting point 14:21:59 both for potential new drivers, and a reference for existing ones 14:22:38 it would be good if we could get a "tiger team" of vendor driver maintainers to work together on this 14:22:47 i'll help 14:22:47 rosmaita: i agree 14:22:55 simondodsley: ++ 14:22:59 I can help too 14:23:05 thank you simondodsley and sp-bmilanov 14:23:06 sp-bmilanov: ++ 14:23:19 ok, that's enough to start 14:23:28 it would be good to have someone from the Red Hat Software Factory team assist as well... 14:23:45 i can try to arrange that 14:23:53 or at least reach out and see if it's possible 14:23:56 they can then see how their stuff is being used in real life and see how they can make their docs better 14:24:09 TBH I am not sure SF is viable anymore since it requires an OKD cloud (I need to fact-check the specifics) 14:24:31 you can use RHEL 9 for SF - OpenShift is also an option 14:24:43 We tried deploying SF from scratch on a centos box few months back, and it worked after lot of attempts. 14:24:44 sp-bmilanov: the TC backlog (https://meetings.opendev.org/meetings/tc/2024/tc.2024-11-12-18.00.log.html) has some discussion about different approaches 14:25:19 msaravan: the issues you faced and what you ended up with would be really helpful as we try to capture this information 14:25:30 msaravan: that is the issue - it's the multiple attempts due to poor support and documentation that is annoying 14:25:39 jbernard: thanks 14:25:49 Sure, we have our notes.. and we can share 14:25:55 i'm glad to see that a key takeaway from this discussion is that vendors think that the third-party CI requirement is worthwhile 14:26:15 sp-bmilanov: there was significant discussion after the meeting ended, i have that in a local log i can make available 14:26:36 sp-bmilanov: but ill try to summarize it in the thing I start 14:26:56 jbernard: nice, can you send it over at #openstack-cinder ? 14:27:21 rosmaita: i think the value is understood, it's the time, trial-and-error, and poor docs that are frustrating 14:27:47 sp-bmilanov: can do 14:28:11 simondodsley: sure, and I guess that's not an issue for many vendors, but the minimal base for a cluster is IIRC three machines and 96GB ram.. Zuul can be run on a smaller machine 14:28:58 I have SF running my CI on a single VM (ironically in OpenStack) 14:28:59 getting the new SF to run now includes getting the undercloud to run 14:29:27 you don't need all those resources 14:29:46 is that an SF version that's still not EOL? 14:29:55 yes 14:30:17 nice, I need to read thru the docs again then 14:30:20 i ended up speaking directly with the SF developers 14:31:09 ok, i think we have a starting point 14:31:25 i will have updates in next week's meeting, if not sooner 14:31:54 #topic multi-transport protocol volume driver support 14:31:56 sp-bmilanov: ^ 14:32:24 #link https://review.opendev.org/c/openstack/cinder/+/847536 14:33:11 thanks, that's a continuation of last week's discussion: how can we enable a single backend driver to handle multiple transport protocols 14:34:19 i think the model followed by all other vendors is to have seperate drivers for different protocols. 14:34:22 right now, in initialize_connection(), we try to detect what the host wants and return an appropriate driver_volume_type 14:35:31 yep, but there are use cases with hypervisors connecting over different protocols 14:36:31 that is OK - you can define multiple backend stanzas in the config file pointing to the same backend, but with a different driver. 14:36:44 then you just use volume_types 14:39:04 but in that case, how would Cinder work if the volume is requested from a client that does not support the protocol of the backend instance that manages the volume? 14:39:17 if you have a situation where there are multi hypervisors in the cluster that don't all have the same dataplane connectivity, you can also cater for that with nova flavors 14:40:37 so a Nova flavor would map to another volume type? 14:41:01 s/another volume type/another volume type -> another volume backend/ 14:42:15 you set a flavor to have a requirement of a hypervisor of a specific device availablility, eg an FC card. I can't remmeber the exact setup, but we use that in our CI ironically 14:43:24 it is usual though tht production clusters will all have hypervisors with the same connectivty so cater for failover or upgrsde events 14:44:57 usually yes, but not always 14:46:45 sp-bmilanov, I'm not 100% clear on the use case, the consumer of cinder volumes (mostly Nova) sets up the appropriate initiator related packages in the compute node right? doesn't this look like a deployment/consumer issue than a cinder issue? 14:48:03 the issue is that a volume managed by one backend instance (that defines the transport protocol as well) will not be accessible to part of the clients that do not support that protocol 14:48:53 i just had a quick look - we use flavor metadata parameter of `pci_passthrough:alias`. In this case , This flavor is then reflected in the zuul config. So maybe this is only good for a CI system. 14:50:31 sp-bmilanov: i think your issue is because you are a SDS solution, rahter than a full external storage platform. If that is the case, this must be catered for in the was Ceph deals with the issue??? 14:51:44 what is the client that we are referring to here? is it a compute node or a different consumer? 14:52:40 simondodsley: I see Ceph exporting only over iSCSI in the codebase? 14:53:02 whoami-rajat: yes, compute nodes 14:53:08 so what are the protocols SP support? 14:53:27 the StorPool block protocol and iSCSI 14:54:02 whoami-rajat: we can extend this to how Cinder attaches volumes to itself as well 14:54:09 but it's mainly compute nodes 14:55:09 so SP has it's own connector for your block storage protocol? 14:55:56 yes, but the SP driver in Cinder can also expose volumes over iSCSI, and in that case, os-brick uses the iSCSI connector/initiator 14:55:58 looks like it https://github.com/openstack/os-brick/blob/master/os_brick/initiator/connectors/storpool.py 14:58:14 my concern is that is you start merging protocols into a single driver, this will potentially cause issues in the future. What happens if/wne you start to support NMVe 14:59:16 the way it's proposed now, that would be a case of extending initialize_connection() 14:59:31 we're nearly out of time 14:59:47 mhen, Luzi: i see your updates, regarding your questions 14:59:49 and then what happens when you add FC support - it will get too complicated. Better to split 15:00:01 sp-bmilanov: can you weigh in on https://etherpad.opendev.org/p/cinder-epoxy-meetings#L86 ? it is storpool ci related 15:00:12 mhen, Luzi ill look at the IBM CI situation 15:00:44 jbernard: thanks, I will have a look 15:00:46 sp-bmilanov: is there a technical reason that it cannot be split? 15:01:21 lets continue this one in the cinder channel 15:01:25 or in next week's meeting 15:01:31 very quickly 15:01:39 let's continue next week 15:01:51 this friday is the 3rd friday, so it's the festival of reviews 15:02:07 i propose we add spec reveiw to that 15:02:12 there are only 2 or 3 15:02:24 raised by whoami-rajat in the agenda 15:02:40 ++ thanks 15:02:47 and with that I think we've covered everything (at least in some part) 15:03:07 I hope to have the dm-clone driver spec ready by friday as well 15:03:13 #topic last thoughts 15:03:16 jhorstmann: awesome 15:03:19 just wanted to raise awareness so we don't forget about specs (I also have one there) + anyone who want to propose spec for features, this was a reminder 15:03:30 thanks jhorstmann , looking forward to it 15:03:49 ok, thanks everyone! 15:03:52 #endmeeting