Wednesday, 2025-05-14

*** mhen_ is now known as mhen01:56
jbernard#startmeeting cinder14:03
opendevmeetMeeting started Wed May 14 14:03:47 2025 UTC and is due to finish in 60 minutes.  The chair is jbernard. Information about MeetBot at http://wiki.debian.org/MeetBot.14:03
opendevmeetUseful Commands: #action #agreed #help #info #idea #link #topic #startvote.14:03
opendevmeetThe meeting name has been set to 'cinder'14:03
jayaanand_hi14:03
simondodsleyo/14:03
agalicao/14:03
Luzio/14:03
jbernardo/14:04
hvlcchao1o/14:04
eharneyhi14:04
gireesho/14:04
jbernard#link https://etherpad.opendev.org/p/cinder-flamingo-meetings14:04
rosmaitao/14:04
whoami-rajathi14:05
jungleboyjo/14:06
jbernardwelcome everyone, lets get started14:06
jbernardrosmaita: you have some annoucements14:07
jbernard#topic annoucements14:07
rosmaitayeah, looks like maybe possibly the gates will get under control today14:07
rosmaitainfo is on the agenda14:07
simondodsleyexciting - been waiting for this one14:08
rosmaitakeep your fingers crossed14:08
rosmaitathe other CI news is about the NFS CI job14:08
simondodsleythanks rosmaita for sorting this out14:08
rosmaitatosky has a patch up to add the cinder-tempest-plugin tests to the devstack-plugin-nfs job14:09
rosmaitathere's a question on his patch about whether the swap size change will work14:09
Saio/14:09
rosmaitaso i put up a patch to run the job 10 times14:09
rosmaitait's still running14:09
rosmaitabut that should give us some info ... key thing is, we need to be running the cinder-tempest-plugin tests on NFS14:10
rosmaitabut hopefully we can get https://review.opendev.org/c/openstack/devstack-plugin-nfs/+/898965 merged soon14:10
simondodsleyis there a simple way 3rd party CIs can force this to run on our side?14:10
rosmaitaby "this' do you mean the NFS job, or the cinder-tempest-plugin tests?14:11
simondodsleyNFS14:11
simondodsleyI'm happy to run the NFS generic plugin on our CI but need to understand how to configure Zuul so it uses NFS shares I provide14:12
simondodsleyThis would also provide some validation on using the generic driver on 3rd party paltforms14:12
rosmaitathat would be nice ... but to answer your question, i'm not sure14:13
eharneydefinitely possible, if you have NFS shares, you'd just need to configure the generic driver and its related options in cinder14:13
rosmaitawhat Eric said14:13
eharneyhttps://opendev.org/openstack/devstack/src/branch/master/lib/cinder_backends/nfs    shows what needs to be set14:14
simondodsleygret - I'll look into that14:14
rosmaitasounds good!14:15
rosmaitathat's all from me14:15
whoami-rajatI've proposed some patches to better address the new location API issue but it needs attention from glance rather than cinder https://review.opendev.org/q/topic:%22fix-location-apis%2214:16
rosmaitawhoami-rajat: those would allow us to run our tests with 'do_secure_hash = true" ?14:17
whoami-rajatwith the tempest change merged, YES14:18
rosmaitaok, but for now, we can just happily run with  'do_secure_hash = false' and the ceph job will pass14:18
whoami-rajatyep, I also have a patch to disable it in glance since it might cause issues in deployments as well14:19
rosmaitathat secure hash seems to be a performance hit14:19
eharneythe test failures aren't because it's a perf issue14:20
eharneythe test fail b/c the image is in "active" while there are still tasks happening that interfere with other operations14:22
rosmaitayeah, but isn't the "tasks still happening" taking up time & cpu ?14:23
eharneyyes14:24
rosmaitai mean, if they were instantaneous, we wouldn't see this14:24
rosmaitaok14:24
whoami-rajatif we disable hashing then we are not really testing it anywhere14:24
eharneyin short i think there are design problems that that tempest test exposed, and we need to document those (as a bug) before we start merging workarounds into tempest tests14:24
whoami-rajatbug: https://bugs.launchpad.net/glance/+bug/211058114:27
whoami-rajatfix: https://review.opendev.org/c/openstack/glance/+/94966814:27
whoami-rajatdoc: https://review.opendev.org/c/openstack/glance/+/94967214:27
eharneyi think there is another bug in that it is unclear what the expected behavior is for an image that's in "active" (and therefore usable) but that still has ongoing hash calculation tasks running -- it seems like it shouldn't be in the active state14:29
eharneybut we don't have to sort all that out here14:29
whoami-rajatit was by design this way but the cases like delete were not considered so turned out to be problematic14:31
whoami-rajatyep, let us meet with glance and we can discuss14:31
eharneydelete is not the scariest part of this design IMO14:31
jbernardok, thanks for getting all of this sorted14:33
jbernardthe glance meeting is tomorrow, maybe this is a good topic for their agenda14:34
jbernarddo we have any specific plans to introduce this at that meeting, or are the relevant persons already aware?14:35
eharneythere still isn't a bug fully describing the problem so i'm not assuming that all the relevant material is being discussed yet14:36
jbernardok, unless someone wants to volunteer, i can jump in there tomorrow and at least raise the issue14:38
hemnamep14:39
jbernardok, next on the list14:40
jbernardtobias-urdin has an rbd data pool patch up14:41
jbernardhttps://review.opendev.org/c/openstack/cinder/+/91493014:41
jbernardasking if it should honor the configured data pool when cloning, or use the one from the original volume14:42
eharneyin the situation where you've configured a data pool on an existing deployment that didn't have one?14:42
jbernardor if the configured data pool doesn't match the data pool of the volume being cloned14:43
eharneygood question14:44
jbernardyeah14:44
eharneyit's also not obvious to me how this impacts our capacity stats collection14:44
eharneybecause the patch seems to ignore that currently14:44
jbernardi lean toward honoring the configuration defined, but ive only been thinking about it for a minute14:44
jbernardeharney: that's a good questiion, that patch definately needs some input14:45
jbernardeharney: ive added us as reviewers, so i should show up in your queue14:46
jbernards/i/it14:47
whoami-rajatI'm missing the use case of the patch, is the purpose to have RBD image data in a separate data pool and the metadata in the volumes pool?14:48
eharneyyes, it doesn't really spell out the use case very well, but one of the reasons that's desired is to enable use of erasure-coded pools14:49
jbernardyep14:49
whoami-rajatok, don't we also need to document the performance impact of using erasure coded pool? I'm sure the parity calculation should be pretty CPU intensive14:50
eharneyit will definitely have a perf impact (on the ceph cluster side)14:51
jbernardrelated to tobias-urdin, his microversion patches are all posted and ready for review14:52
jbernardhttps://etherpad.opendev.org/p/cinder-flamingo-reviews#L1814:52
whoami-rajatyes, so deployers/tooling needs to configure more resources to the nodes deploying the OSDs hosting the data pool14:52
whoami-rajatto me this patch is very abstract given the impact it can have on deployments14:53
jbernardwhoami-rajat: these would be good review comments ;)14:54
eharneyit's geared toward someone who already decided they wanted to setup erasure-coded pools, but yeah, more documentation would definitely help14:54
jbernardi want to get this one in before the bell,14:55
jbernardmhen, Luzi have some cycles to work on image encryption14:55
jbernardhttps://review.opendev.org/q/topic:%22LUKS-image-encryption%2214:56
whoami-rajatnot for a specific customer but a lot of deployments adapt new features thinking it's suited for their use case without knowing the full context so I'm more inclined towards documenting this better14:56
Luziwell starting with next week we will have a few hours per week to work on the image encryption again14:56
Luziit would be nice to finally come to terms with this :D14:57
jbernardthose patches need review (as do others), if anyone is looking for things in immediate need of attention14:57
jayaanand_we have multiple customers looking for in-use NFS volume extension. there is a proposal to fix generic NFS issue related bug https://bugs.launchpad.net/cinder/+bug/1870367. can we implement NetApp driver in similar line.14:57
jbernardhttps://etherpad.opendev.org/p/cinder-flamingo-reviews14:57
jbernardeharney: i meant to relay jayaanand_'s quetions from earlier, i think you might know best this answer14:58
eharneythere is a patch up currently implementing this in the netapp nfs driver, right?14:59
eharneyhttps://review.opendev.org/q/topic:%22bp/extend-volume-completion-action%2214:59
eharneyin there ^14:59
jayaanand_it didn't went pass review14:59
agamboahey rosmaita, we took a look at your comments and addressed them for this merge: https://review.opendev.org/c/openstack/cinder/+/901318/14:59
eharneyi'm not sure i understand the question. this area is still being worked on15:00
rosmaitaagamboa: ok, will take a look15:00
jayaanand_https://review.opendev.org/c/openstack/cinder/+/87388915:00
jayaanand_also bluprint is re-proposed https://review.opendev.org/c/openstack/cinder-specs/+/92237415:01
eharneyi think the intent is to complete 873889 and land it15:02
jayaanand_ok15:02
jayaanand_thank you!15:02
jbernardwe're just a bit over time, last call everyone15:03
agalicaI have a quick question if we're in the open questions section15:04
agalicaIf we were to support a new type of storage array which in turn requires a new volume type for one of the required features do we require two patches or just one?15:04
agalicaMy suspicion is 2, but I wanted to confirm since the new storage device requires said feature15:04
jbernardagalica: you mean a new volume driver?15:05
jbernardagalica: do you have an code up that we can look at?15:05
agalicanot at this time as it's a future.  I mean a storage device that we currently do not support in the driver.  We are adding support for it, but it requires a volume type feature as well.  15:05
whoami-rajatadded my comment to the RBD patch15:06
jbernardwhoami-rajat: thank you!15:06
tobias-urdinsorry late to the meeting15:06
jbernardtobias-urdin: no worries15:06
jbernardagalica: i personally am not sure, i say post the code and we can discuss15:07
whoami-rajatagalica, volume types can have driver specific attributes in extra specs so a special volume type should not require code changes but would be better to understand your use case more in detail15:07
agalicaSo, for example, supporting storage type X requires feature Y which has not been implemented.  In order to support X, we must support Y.  Does this mean that Y requires its own patch and then we can have a patch for storage type X?  Or can both of these things happen at once?  If it helps, most likely the only changes required will be these new supporting features15:07
agalicaok, thanks guys. 15:08
jbernardok, lets wrap up, thank you everyone!15:10
jbernard#endmeeting15:10
opendevmeetMeeting ended Wed May 14 15:10:23 2025 UTC.  Information about MeetBot at http://wiki.debian.org/MeetBot . (v 0.1.4)15:10
opendevmeetMinutes:        https://meetings.opendev.org/meetings/cinder/2025/cinder.2025-05-14-14.03.html15:10
opendevmeetMinutes (text): https://meetings.opendev.org/meetings/cinder/2025/cinder.2025-05-14-14.03.txt15:10
opendevmeetLog:            https://meetings.opendev.org/meetings/cinder/2025/cinder.2025-05-14-14.03.log.html15:10
whoami-rajatthanks everyone!15:10

Generated by irclog2html.py 4.0.0 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!