*** whoami-rajat__ is now known as whoami-rajat | 10:36 | |
*** dviroel_ is now known as dviroel | 11:19 | |
*** TusjarTgite is now known as TusharTgite | 13:46 | |
*** dasm|off is now known as dasm | 13:52 | |
whoami-rajat | #startmeeting cinder | 14:00 |
---|---|---|
opendevmeet | Meeting started Wed May 18 14:00:04 2022 UTC and is due to finish in 60 minutes. The chair is whoami-rajat. Information about MeetBot at http://wiki.debian.org/MeetBot. | 14:00 |
opendevmeet | Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. | 14:00 |
opendevmeet | The meeting name has been set to 'cinder' | 14:00 |
whoami-rajat | #topic roll call | 14:00 |
rosmaita | o/ | 14:00 |
abishop | o/ | 14:00 |
TusharTgite | Hi | 14:00 |
eharney | hi | 14:00 |
mnaser | o/ | 14:01 |
ricolin | o/ | 14:01 |
caiquemello[m] | 0/ | 14:01 |
fabiooliveira | o/ | 14:02 |
whoami-rajat | good turnout, let's wait few more minutes for others to join | 14:02 |
jungleboyj | o/ | 14:02 |
sfernand | hi | 14:03 |
whoami-rajat | guess that's all, let's get started | 14:03 |
whoami-rajat | #topic announcements | 14:03 |
whoami-rajat | there are a bunch of announcement so i won't explain anything in depth but there are links if people want to know more | 14:03 |
whoami-rajat | Updates from TC Meeting | 14:04 |
whoami-rajat | #link http://lists.openstack.org/pipermail/openstack-discuss/2022-May/028541.html | 14:04 |
whoami-rajat | most of the announcements are from the TC updates which i thought would be good for awareness of the cinder team | 14:04 |
whoami-rajat | first is, FIPS goal is selected community-wide goal now | 14:04 |
whoami-rajat | #link https://governance.openstack.org/tc/goals/selected/fips.html | 14:04 |
whoami-rajat | second, Release naming change | 14:05 |
whoami-rajat | #link https://review.opendev.org/c/openstack/governance/+/841800 | 14:05 |
whoami-rajat | We are moving away from release names and will use release number which is the year of release followed by 1 or 2 (6 months release model) | 14:05 |
whoami-rajat | Eg: 2022.2 for AA, 2023.1 for BB | 14:06 |
whoami-rajat | Release naming will be handed over to OpenInfra Foundation and will be used for marketing purposes (TC won't be involved now) | 14:06 |
whoami-rajat | ^ that's what is written in the proposal and TC updates | 14:06 |
whoami-rajat | next, SLURP (Skip Level Upgrade Release Process) | 14:06 |
whoami-rajat | #link https://review.opendev.org/c/openstack/governance/+/840354 | 14:06 |
whoami-rajat | As per OpenStack legal team, it is not OK to use tick-tock naming so we will be shifting to SLURP (previously tick) and not-SLURP (previously tock) names | 14:07 |
whoami-rajat | again, this is not a major change but you can visit the review of the change proposed to know more | 14:07 |
whoami-rajat | next, Releasenote handling in the new release candence (SLURP now) | 14:08 |
whoami-rajat | Brian has a patch up for 3 approaches on how to handle it | 14:08 |
whoami-rajat | the details of the 3 approaches are in the commit msg | 14:08 |
whoami-rajat | #link https://review.opendev.org/c/openstack/cinder/+/840996 | 14:08 |
whoami-rajat | i think TC will resolve to one and that we will follow with the new SLURP release model | 14:08 |
whoami-rajat | last announcement is regarding cinderlib release | 14:09 |
whoami-rajat | Cycle trailing release for cinderlib (yoga) | 14:09 |
whoami-rajat | #link http://lists.openstack.org/pipermail/openstack-discuss/2022-May/028522.html | 14:09 |
whoami-rajat | The deadline is 23rd June which is far away but i already discussed this once with geguileor | 14:09 |
whoami-rajat | and release team had a mail regarding it so good to mention | 14:10 |
whoami-rajat | that's all the announcements from my side, anything else from anyone? | 14:10 |
tobias-urdin | forgot to wave o/ everybody | 14:11 |
whoami-rajat | hey | 14:11 |
whoami-rajat | ok, we can move on to topics then | 14:11 |
whoami-rajat | #topic Add image_conversion_disable config | 14:11 |
whoami-rajat | rosmaita, mnaser ricolin that's you | 14:11 |
mnaser | henlo | 14:12 |
mnaser | so the main reason behind this is that we have environments where we allow both qcow2 and raw images being uploaded | 14:12 |
mnaser | reason being that if a user uses qcow2, they can use local storage on the hvs and its easier to use qcow2 | 14:12 |
mnaser | however, if they use qcow2 + volume, then bad times for your control plane | 14:12 |
mnaser | which is why we would like to allow disabling conversion for these environments | 14:13 |
mnaser | and the user can get a note saying they are using a bad image and they need to upload the right image | 14:13 |
mnaser | aka a raw format image in order to do bfv | 14:13 |
rosmaita | just to have it in the meeting log, here's the links to what we're talking about: | 14:14 |
rosmaita | #link https://review.opendev.org/c/openstack/cinder/+/839793 | 14:14 |
rosmaita | #link https://launchpad.net/bugs/1970115 | 14:14 |
whoami-rajat | thanks rosmaita | 14:14 |
whoami-rajat | i went through the replies in the patch and i think this is a good use case | 14:14 |
whoami-rajat | currently what we're doing is rejecting a format that we think is bad (not matching the volume) and asking users to manually upload a new image satisfying the criteria | 14:15 |
whoami-rajat | i was thinking if we could do something in glance to automate this to avoid the manual work | 14:15 |
mnaser | yep, that's the idea/goal. it's also allowing to close a potential for a end user to effectively dos a cloud's control plane | 14:15 |
rosmaita | yeah, i don't know that we can push it to glance | 14:15 |
mnaser | my "bigger" long term goal would be to optimize the conversion to be backend specific | 14:16 |
rosmaita | you can restrict what disk_format glance will allow, but mnaser's use case is that you want multiple formats | 14:16 |
mnaser | for example, qemu-img can write directly to the ceph cluster | 14:16 |
ricolin | consider it as a two step plan?:) | 14:16 |
rosmaita | you just don't want to convert them | 14:16 |
mnaser | so instead of convert to local disk, then upload to ceph -- you can convert straight to ceph, which can maybe potentially allow us to re-enable this | 14:16 |
mnaser | since the disk i/o is really what hurts our control planes really bad | 14:17 |
mnaser | but that is a _faaaar_ bigger project, and this is a good interim option | 14:17 |
mnaser | (imho) | 14:17 |
whoami-rajat | ack, i wasn't aware of the idea of handling this better in the coming future (hence this discussion is useful) | 14:18 |
rosmaita | so as far as end users go, they'll get a 400 if they upload a volume as an image requesting a format that would require conversion | 14:18 |
rosmaita | on the download side, we have create-volume and reimage-volume | 14:18 |
rosmaita | for both of those, we don't know the volume format until after the REST API has returned a 202 | 14:19 |
mnaser | #link https://bugs.launchpad.net/cinder/+bug/1970114 | 14:19 |
mnaser | ^ this is the 'bug' for the more 'efficent' way of doing things | 14:19 |
rosmaita | but the patch adds user messages for when the volume goes to error | 14:19 |
whoami-rajat | was just going through the bug report, looks like 2 is something we should discuss about in the midcycle | 14:22 |
rosmaita | i agree | 14:23 |
whoami-rajat | but i also remember a discussion that concurrent requests wait for image cache to be created by the first one and then use the cache | 14:23 |
whoami-rajat | or maybe not | 14:23 |
abishop | yes, it does wait for the cache to be seeded | 14:23 |
whoami-rajat | so I don't have much objections given this is a short term solution/workaround for a problem that is being worked upon | 14:24 |
rosmaita | yeah, i thought we had addressed that, not sure if the conversion means it's handled differently | 14:24 |
whoami-rajat | ok, then 2 in the bug should not be true since the image cache is the first volume entry we create (after conversion) and clone from it | 14:25 |
rosmaita | i guess my feeling is that it's a good idea not to kill your cinder-volume nodes, and as long as you communicate to your users what their expectations should be about volume-image format relations, this is ok | 14:26 |
whoami-rajat | agreed and since default is False (which is current behavior) this won't harm anyone not opting for it | 14:27 |
whoami-rajat | any other concerns or observations on this? | 14:30 |
rosmaita | i guess for the record, for the upload use case you could push this to glance using image import, but then you could kill the glance node instead of cinder-volume | 14:30 |
rosmaita | and we'd have to modify cinder to use image import | 14:31 |
rosmaita | on the download sides, i don't think this could be handed off to glance | 14:31 |
rosmaita | because the use case is to have multiple formats stored in glance, and only a subset of those would be accepted by cinder for create-volume or reimage-volume | 14:32 |
mnaser | rosmaita: could be an idea for glance to have multiple types the same way it has multiple backends | 14:32 |
rosmaita | so the end user could always pick the wrong format image by mistake | 14:32 |
mnaser | so then cinder requests a specific format and glance provides it, but yeah, that's also a 'perfect world' scenario | 14:32 |
whoami-rajat | that's a good motivation to adapt to import plugin but we might not have enough bandwidth to work upon it | 14:33 |
rosmaita | mnaser: that was what Glare was for (remember that?) | 14:33 |
mnaser | openstack lore :) | 14:33 |
mnaser | glare was a bit ahead of its time | 14:33 |
mnaser | now we have OCI | 14:34 |
whoami-rajat | maybe I'm wrong but a use case of glance image carrying multiple locations is to have different formats of the same image linked with a single image? | 14:35 |
whoami-rajat | or not, maybe it's more of a redundancy feature | 14:36 |
mnaser | multistore is multiple locations for the same format for hte same image | 14:36 |
rosmaita | whoami-rajat: well, that was what Glare was for | 14:36 |
mnaser | i.e.: edge cloud where you have 1 image identifier, but it can be stored in a ceph cluster in each site | 14:36 |
rosmaita | glance is supposed to work with immutable images | 14:36 |
mnaser | so there would be multiple locations, but all hosting the same exact data | 14:36 |
whoami-rajat | ack, thanks for the clarification | 14:37 |
rosmaita | ok, as far as the patch goes | 14:37 |
rosmaita | i would feel better with functional tests, but not sure we can do those with something that involves the image service | 14:37 |
tobias-urdin | which fwiw is a real world issue when storing for example glance images in a ceph cluster and then needing to copy that to another ceph cluster in another availability-zone | 14:38 |
tobias-urdin | sorry kinda semi-unrelated to cinder tho | 14:38 |
rosmaita | whoami-rajat: i am thinking that we don't have tempest coverage for reimage yet? | 14:39 |
whoami-rajat | rosmaita, I've proposed a patch in tempest which I'm working on | 14:40 |
whoami-rajat | it's failing for unknown reason as of now as I've already tested the flow manually multiple times and it works | 14:40 |
rosmaita | mnaser: are you already running a version of this patch? | 14:40 |
whoami-rajat | but i will try to get that fixed and merged in Zed | 14:40 |
rosmaita | whoami-rajat: yeah, you definitely need more stuff to work on :) | 14:41 |
mnaser | rosmaita: we are not, no | 14:41 |
mnaser | ricolin has been putting in most of the work | 14:41 |
rosmaita | ok, i think the unit test coverage is good, i'm just worried about it blowing up in an unexpected way | 14:41 |
ricolin | the test was only run against my local test environment | 14:42 |
rosmaita | for no particular reason (my being worried, i mean) | 14:42 |
rosmaita | well, it will ship with the option "off" | 14:42 |
rosmaita | so that makes me feel a bit better | 14:43 |
whoami-rajat | would be good to get assurance if it's running in a functional cluster (but maybe that's a lot to expect) | 14:43 |
ricolin | rosmaita: yeah, so it should not make difference to current use cases | 14:43 |
mnaser | If the team is +2 on this | 14:44 |
mnaser | Then we can run it as a patch on top of your env and see if it blows up | 14:44 |
mnaser | Since we were gonna cherry pick it back internally anyways | 14:44 |
rosmaita | yeah i think we have enough test coverage that we can be confident of no regression when the option is "off" | 14:44 |
mnaser | That way we don’t have to land follow ups out of the box | 14:44 |
rosmaita | so as far as the discussion has gone, it sounds like there's no objection to the concept? | 14:45 |
rosmaita | eharney: you should like this patch, it adds a new user message | 14:46 |
eharney | i like the concept | 14:46 |
whoami-rajat | not from my side | 14:46 |
whoami-rajat | looks like the major concerns on this are resolved and we are short on time so let's move on to the next topic | 14:46 |
rosmaita | thanks mnaser and ricolin | 14:47 |
whoami-rajat | thanks mnaser ricolin and rosmaita for the discussion | 14:47 |
whoami-rajat | #topic Getting pylint job into shape | 14:47 |
ricolin | thanks rosmaita whoami-rajat | 14:47 |
whoami-rajat | eharney, that's you | 14:47 |
eharney | well, i looked at our pylint results, and found broken code with it | 14:47 |
eharney | i think it would be useful to get this integrated back into the review process, but the job is kind of noisy | 14:48 |
eharney | so i'm submitting patches to clean up all the code it complains about | 14:48 |
eharney | review them and look at the pylint output, especially on large patches | 14:48 |
eharney | https://review.opendev.org/c/openstack/cinder/+/842149 | 14:48 |
eharney | that's about it | 14:48 |
eharney | (there's more to clean up yet, but it's a start) | 14:49 |
rosmaita | eharney: is it the pylint job or the mypy job that doesn't post html output? | 14:49 |
rosmaita | (i forgot to follow up on that) | 14:49 |
eharney | mypy | 14:49 |
rosmaita | ok | 14:49 |
eharney | the pylint job shows a bunch of red errors: https://zuul.opendev.org/t/openstack/build/ccc4a1165ab94f689984f49b417a24ed | 14:49 |
whoami-rajat | cool, thanks eharney for keeping an eye on the pylint job | 14:49 |
whoami-rajat | moving on | 14:51 |
whoami-rajat | #topic Open discussion | 14:51 |
whoami-rajat | we've a bunch of review requests on the etherpad (looks like it's beginning to be a trend for open discussions) | 14:51 |
whoami-rajat | just a mention that we won't have a bug squad meeting today since Sofia is not around but she has posted a mail covering the bugs this week | 14:53 |
whoami-rajat | anything else for open discussion? | 14:54 |
whoami-rajat | looks like everyone is busy reviewing the patches so we can end 5 minutes early | 14:55 |
whoami-rajat | Thanks everyone! | 14:55 |
whoami-rajat | #endmeeting | 14:56 |
opendevmeet | Meeting ended Wed May 18 14:56:02 2022 UTC. Information about MeetBot at http://wiki.debian.org/MeetBot . (v 0.1.4) | 14:56 |
opendevmeet | Minutes: https://meetings.opendev.org/meetings/cinder/2022/cinder.2022-05-18-14.00.html | 14:56 |
opendevmeet | Minutes (text): https://meetings.opendev.org/meetings/cinder/2022/cinder.2022-05-18-14.00.txt | 14:56 |
opendevmeet | Log: https://meetings.opendev.org/meetings/cinder/2022/cinder.2022-05-18-14.00.log.html | 14:56 |
jungleboyj | Thank you! | 14:56 |
tobias-urdin | thanks! | 14:56 |
*** dviroel is now known as dviroel|lunch | 15:39 | |
*** dviroel|lunch is now known as dviroel | 16:31 | |
*** dviroel is now known as dviroel|out | 20:28 | |
*** dasm is now known as dasm|off | 22:32 |
Generated by irclog2html.py 2.17.3 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!