16:02:07 #startmeeting nova 16:02:07 Meeting started Tue May 6 16:02:07 2025 UTC and is due to finish in 60 minutes. The chair is Uggla. Information about MeetBot at http://wiki.debian.org/MeetBot. 16:02:07 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 16:02:07 The meeting name has been set to 'nova' 16:02:14 \o 16:02:17 Hello everyone 16:02:27 awaiting a moment for people to join. 16:02:50 o/ 16:02:54 o/ 16:03:17 o/ 16:04:03 thanks bauzas for last week meeting. 16:04:11 np 16:04:11 o/ 16:04:47 #topic Bugs (stuck/critical) 16:04:53 #info No Critical bug 16:05:01 #topic Gate status 16:05:11 #link https://bugs.launchpad.net/nova/+bugs?field.tag=gate-failure Nova gate bugs 16:05:19 #link https://etherpad.opendev.org/p/nova-ci-failures-minimal 16:05:33 #link https://zuul.openstack.org/builds?project=openstack%2Fnova&project=openstack%2Fplacement&branch=stable%2F*&branch=master&pipeline=periodic-weekly&skip=0 Nova&Placement periodic jobs status 16:05:39 #info Please look at the gate failures and file a bug report with the gate-failure tag. 16:05:44 #info Please try to provide meaningful comment when you recheck 16:06:13 If I understood well, the gate was blocked by this https://review.opendev.org/c/openstack/nova/+/948392 last week. 16:06:27 jepp 16:06:30 fix is landed 16:06:31 It is landed now. 16:06:48 It the gate looks good. Please tell me if I'm wrong. 16:06:54 the cyborg and barbican jobs are still broken i think 16:06:54 I have hit by https://bugs.launchpad.net/glance/+bug/2109428 multiple times recently. Not a blocker but definitly a source of rechecks 16:07:05 sean-k-mooney: yepp those non-votings are broken 16:07:08 bugs are filed 16:07:25 i may try and find time to go fix them 16:07:35 they are both broken on the lack of a pyproject.toml 16:07:40 yepp 16:07:47 good to know thx gibi 16:07:51 https://bugs.launchpad.net/barbican/+bug/2109584 16:07:54 https://bugs.launchpad.net/openstack-cyborg/+bug/2109583 16:07:56 if i can fix it without having to install the service locally i might give it a try 16:08:01 i can refence thos bugs 16:08:07 thanks 16:08:18 I think we need to make pyproject.toml for all projects in openstack otherwise they will break slowly at some point 16:08:30 gmaan: I agree 16:09:21 should we track this ? 16:09:41 gmaan: yes we will 16:09:56 Uggla: nova is mostly doen stephen and i tried to do this 2 years ago 16:10:01 to get ahead of things breaking 16:10:12 so nova and placment are done 16:10:17 i need to check os-* 16:10:31 but for the libs we are responible for it trivial 16:11:03 ok os-vif is still pending 16:11:09 any bug / blueprint to refer to this work ? 16:11:15 ill start working on them and ping folks to review 16:11:26 ++ 16:11:51 sean-k-mooney++ 16:12:05 anything else ? 16:12:23 moving on to next item. 16:12:31 #topic tempest-with-latest-microversion job status 16:12:39 #link https://zuul.opendev.org/t/openstack/builds?job_name=tempest-with-latest-microversion&skip=0 16:12:56 I have just discussed with gmann about it. 16:13:30 gmann is progressing on this periodic job. 16:13:59 gmaan, I let you give a quick status if you wish. 16:14:05 sure 16:14:21 it is not worst than what i suspected, only 23 tests failing 16:14:30 I started fixing those one by one 16:14:32 #link https://review.opendev.org/q/topic:%22latest-microversion-testing%22 16:14:46 hypervisor are fixed, I will fix a few more today and this week 16:15:01 do you know if some of them are invalid failure 16:15:07 that is ^^ topic I am adding all changes to, feel free to review/comment 16:15:13 i.e. the test is depending on an older microverion behavior 16:15:25 and should be skipped when using latest 16:15:30 sean-k-mooney: not invalid, either we need to fix schema or cap the test with min/max microversions 16:15:46 ok ya that actully what i was wondering 16:15:49 sean-k-mooney: yeah, for example hypervisor uptime test should run till 2.87 16:15:51 can we express in tempst 16:16:03 that this test has a max rather then just min verion requirement 16:16:05 yes with 'max_microversion' 16:16:09 ack 16:16:26 ah i see that how your fixing it in https://review.opendev.org/c/openstack/tempest/+/948490 16:16:28 cool 16:16:38 yeah ^^ 16:17:12 sometime we need to refactor test but not bug deal. 16:17:46 that is all on these, fell free to review, as I am only active core in tempest I will merge them but keep it open for sometime if anyone would like to review 16:18:14 most probably, i will keep all fixs open till job is green 16:18:14 thx gmaan 16:19:12 #topic Release Planning 16:19:19 #link https://releases.openstack.org/flamingo/schedule.html 16:19:24 #link https://releases.openstack.org/flamingo/schedule.html 16:19:56 The patch about nova deadlines has been merged so I think it is ok. 16:20:41 Please let me know if it is not the case or if something is wrong. 16:20:59 #topic Review priorities 16:21:06 #link https://etherpad.opendev.org/p/nova-2025.2-status 16:22:00 I'd like to progress on openapi and I'll try to check with stephenfin about it. 16:22:52 #topic Stable Branches 16:23:09 elodilles, the mic is yours 16:23:19 thanks Uggla , so 16:23:22 #info stable/2023.2 (bobcat) is End of Life, branch is deleted (tag: bobcat-eol) 16:23:31 #info maintained stable branches: stable/2025.1, stable/2024.2, stable/2024.1 16:23:49 down to 3 maintained branches again ;) 16:24:02 #info nova stable release from stable/2024.1 is out (29.2.1) 16:24:15 #info not aware of any stable gate failure 16:24:30 #info stable branch status / gate failures tracking etherpad: https://etherpad.opendev.org/p/nova-stable-branch-ci 16:24:54 we had a broken gate on stable/2025.1 osc-placement, 16:25:07 but it is now fixed 16:25:15 thanks all for the help :) 16:25:30 and i think that's all from me 16:25:36 Uggla: back to you 16:25:39 elodilles, fyi 947847: nova: Release 2024.2 Dalmatian 30.0.1 | https://review.opendev.org/c/openstack/releases/+/947847 should be ok now. 16:25:52 As I have done 948811: Add uc check alternative method | https://review.opendev.org/c/openstack/requirements/+/948811 16:26:04 Uggla: thanks for working on that! 16:26:29 the path is on master with a backport to the stable branch. 16:26:30 i've just added a comment on your patch o:) 16:26:53 so it needs reviews. 16:27:22 but at least the verify is +1 16:27:51 sean mooney proposed openstack/os-vif master: add pyproject.toml to support pip 23.1 https://review.opendev.org/c/openstack/os-vif/+/899946 16:28:19 #topic vmwareapi 3rd-party CI efforts Highlights 16:28:35 #info No updates 16:28:46 Uggla: Back to you 16:28:52 fwiesel, btw welcome back, I hope you are good. 16:29:20 #topic Gibi's news about eventlet removal. 16:29:21 Thanks! Just still catching up with everything. 16:29:24 o/ 16:29:31 #link Series: https://gibizer.github.io/categories/eventlet/ 16:29:48 so last week we saw nova-scheduler running with native threading. There is a blogpost about it 16:29:58 had no time to review those patches yet :( 16:29:59 this week I started testing the non happy path of the scatter-gather 16:30:05 * bauzas hardly crying 16:30:25 good news both mysql server and pymysql client can be configured with timeouts to avoid handing gather threads 16:30:47 I will add a new doc about it in tree 16:31:02 bad news, the oslo.service threading backend uses forks 16:31:23 which does not play nice with out global threadpools initialized before the service workers are forked off 16:31:32 we have workarounds 16:31:41 details are in https://review.opendev.org/c/openstack/oslo.service/+/945720/comments/fb5d6632_f0eaf102 16:31:51 there are two patches to look at 16:32:09 https://review.opendev.org/c/openstack/nova/+/948064?usp=search now with test coverage 16:32:15 and 16:32:16 https://review.opendev.org/c/openstack/nova/+/948437?usp=search 16:32:40 this week I will work on cleaning up the long service to make more patches ready to review 16:32:43 that is all 16:32:46 Uggla: back to you 16:33:16 gmaan: Uggla: just a quick update on the pyproject.toml change if i may 16:33:28 https://etherpad.opendev.org/p/pep-517-and-pip-23 is my old ether pad to track that work 16:33:33 thx gibi 16:33:37 sean-k-mooney, sure go ahead 16:33:40 and only the os-vif change above is not mergd for nova 16:33:47 so i rebased and appvoed that 16:33:55 just saw, ++ 16:33:57 once that is landed we shoudl be good 16:34:07 \o/ 16:34:43 anyway that was all on that topic 16:35:09 gibi, just a question you have said : the oslo.service threading backend uses forks it means real processes and not threads ? 16:35:24 Uggla: no we want worker processes with thread pool 16:35:26 pools 16:35:44 but it uses os.fork to create the worker from the main process 16:35:44 context is oslo service 16:35:55 allows you to have multiple workers 16:36:00 fork copies the state of the process 16:36:01 we use that in the schduler and conductor 16:36:12 but not nava-comptue or the api 16:36:17 we would need os.spawn to get a totally new worker process 16:36:22 without inheriting state 16:37:04 ok I think I have understood the pb. 16:37:04 the wa is to reset the problematic state 16:37:13 after the fork 16:37:25 but I think the real solution would be os.spawn 16:37:48 but changing to that is not super simple as oslo.config has non pickleable lambdas :/ 16:37:56 details are in the linked gerrit comment 16:38:07 and are up in the today IRC log 16:38:16 we can reset after the fork (as you're doing) or reset before the fork, as I suspect will also work and be less janky 16:38:23 yepp 16:38:30 I will try dansmith's suggestion next 16:38:40 while we wait for the oslo folks to respond 16:38:47 https://github.com/openstack/oslo.config/blob/d6e5c96d6dbeec0db974dfb8afc8e508b74861e5/oslo_config/cfg.py#L1387 16:38:52 I agree the real solution is spawn, but we probably don't want to wait 16:38:55 that is there only use of lambda i think by the way 16:39:23 dansmith: yepp we won't wait, we will go with the WA, and adapt when oslo makes the move 16:39:52 +1 16:40:39 moving on 16:40:49 #topic Open discussion 16:41:30 fwiesel, wants to propose something about Cross hypervisor resize. 16:41:45 I guess that was last week's topic 16:42:07 yep but I understood you wanted to discuss it more. 16:42:09 or maybe a continuation 16:42:21 Yes, but there was limited feedback. I mean, if no one feels strongly about it, then I can go forward with a blueprint. 16:43:39 We would like to allow mobility between two hypervisors, and I was thinking the cross-cell migration might already cover it to a large degree. 16:44:05 The question is though, if that is a use-case you would feel worthwhile supporting or rather not. 16:44:50 I think it _has_ to be more like cross-cell migration than regular 16:45:10 I'm a bit mixed on whether or not I think this is worth it, because I suspect there will be a lot of gotchas in the image properties 16:45:24 and because you can pretty much do this with snapshot yourself now 16:45:37 and because we won't really be able to test it regularl I don't think 16:46:27 yeah testing this will be painful 16:46:32 Okay, got it. That's fine. 16:46:55 dansmith: specicly like updating the hw_vif_model for say the vmware one to virtio if they dont happen to have common values already 16:46:55 yeah, all that kind of stuff 16:47:21 Well, we would set those in the flavours and that would override that. 16:47:39 so general question. is there anyting we know off that would prevent you form shelving, modifying them in glance and unshleving? 16:47:54 No, it is more a usability issue. 16:48:02 For our users. 16:48:08 sean-k-mooney: if the flavor you booted from was vmware-specific you can't unshelve to a libvirt-y one right? 16:48:18 fwiesel: so historically flavors are for amounts and images propereis are for changing how the devices are presented 16:48:31 And we were thinking of using shared NFS shares to avoid going through image upload, etc... 16:48:36 fwiesel: so in the past the precendet was dont put device models in flavors 16:48:49 dansmith: correct 16:49:01 dansmith: what im thinkign is we dicussed allowing resize in the sheleved state 16:49:08 to me, snapshot, tweak, boot fresh is the best pattern here 16:49:18 so the workflow would be shelve. resize , update image properties and then unshleve 16:49:22 sean-k-mooney: that's a whole other conversation 16:49:48 But I got it... hard to test so, no takers... Perfectly understandable. 16:49:49 ya im just confirming if there is a way to do that today with the api we have and i think the gaps are resize while shleved 16:50:23 Balazs Gibizer proposed openstack/nova master: DNM:Run nova-next with n-sch in threading mode https://review.opendev.org/c/openstack/nova/+/948450 16:50:36 Thanks for the feedback. 16:50:36 but ya snapshot. detach volumes/ports and boot new vms with thos is a valid workflow too 16:52:49 Shall I say it is ok for fwiesel to propose something that goes in that direction ? 16:53:44 My understanding is, I won't propose things as it is doable with the existing API and implementing it within Nova is hard to test and maintain. 16:53:48 I think anyone can propose anything they want :) 16:54:02 fwiesel: ++ 16:54:32 oh ok good. 16:56:11 We are almost at the top of the hour. Are you ok triaging a couple of bugs or would you prefer to do it next week ? 16:58:00 sean-k-mooney, dansmith ^^ 16:59:48 ok no answer so you might be busy. So closing the meeting. 16:59:59 thanks all 17:00:02 Uggla: thanks 17:00:06 thanks everyone 17:00:10 #endmeeting