16:01:26 #startmeeting nova 16:01:26 Meeting started Tue Jun 27 16:01:26 2023 UTC and is due to finish in 60 minutes. The chair is bauzas. Information about MeetBot at http://wiki.debian.org/MeetBot. 16:01:26 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 16:01:26 The meeting name has been set to 'nova' 16:01:31 hey folks \o 16:01:33 o/ 16:01:34 o/ 16:01:39 o\ 16:01:40 #link https://wiki.openstack.org/wiki/Meetings/Nova#Agenda_for_next_meeting 16:02:54 ok let's start while the folks are arriving :) 16:03:02 #topic Bugs (stuck/critical) 16:03:10 #info No Critical bug 16:03:14 #link https://bugs.launchpad.net/nova/+bugs?search=Search&field.status=New 31 new untriaged bugs (+4 since the last meeting) 16:03:19 o/ 16:03:22 I only had time to look at two bugs this week 16:03:24 sorriy 16:03:37 I can try to look at the bugs this new week 16:05:24 okay, nobody is against 16:05:32 #info Add yourself in the team bug roster if you want to help https://etherpad.opendev.org/p/nova-bug-triage-roster 16:05:39 #info bug baton is being passed to bauzas 16:05:51 I would be super suprised if somebody want to stole the baton :D 16:05:52 do people want to discuss on bugs ? 16:05:59 gibi: :) 16:06:29 looks not 16:06:34 #topic Gate status 16:06:40 #link https://bugs.launchpad.net/nova/+bugs?field.tag=gate-failure Nova gate bugs 16:07:12 it looks to me we have a new bug :( 16:07:14 #link https://bugs.launchpad.net/nova/+bug/2025096 16:07:29 it's been skipped for now, 16:07:43 but that appears to just have not been tested on ceph and something is wrong with it in that environment 16:08:03 I'm not sure about the non-ceph assertion in the comment, I'll have to look 16:08:12 dansmith: ack 16:08:13 but that's a test we should have landed a year ago 16:08:16 and it just got merged 16:08:19 I was looking at Zuul 16:09:19 anyway, like you say, the job is not failed https://zuul.openstack.org/builds?job_name=nova-ceph-multistore&skip=0 16:10:34 dansmith: https://b932a1446345e101b3ef-4740624f0848c8c3257f704064a4516f.ssl.cf5.rackcdn.com/883557/2/check/nova-ceph-multistore/d7db64f/testr_results.html 16:10:48 the test is not skipped 16:11:23 https://review.opendev.org/c/openstack/devstack-plugin-ceph/+/887003 16:11:42 we should be inheriting from that afaik 16:12:04 ahah 16:12:06 ok 16:12:30 why do you say it's not skipped? 16:12:45 gmann: could you add a new change for skipping test_rebuild_volume_backed_server on nova-ceph-multistore 16:12:47 ? 16:13:14 dansmith: oh wait, verifying 16:13:42 the patch specifically calls out the nova job 16:14:02 yeah, you're right, the test is now skipped 16:14:07 gmann: nevermind what I said 16:14:17 https://6354f24cae23db064c15-540aa9666e0f6140cd4132fd113878d1.ssl.cf5.rackcdn.com/876075/14/check/nova-ceph-multistore/617bfe5/testr_results.html I don't see the test 16:14:46 dansmith: sorry, I misunderstood what you said 16:15:28 ..okay 16:15:55 dansmith: sometimes my brain doesn't run correctly 16:16:04 :( 16:16:25 anyway, moving on 16:16:32 #link https://zuul.openstack.org/builds?project=openstack%2Fnova&project=openstack%2Fplacement&pipeline=periodic-weekly Nova&Placement periodic jobs status 16:16:43 nothing to relate 16:16:50 #info Please look at the gate failures and file a bug report with the gate-failure tag. 16:16:56 #info STOP DOING BLIND RECHECKS aka. 'recheck' https://docs.openstack.org/project-team-guide/testing.html#how-to-handle-test-failures 16:17:06 (I could remove this note next week) 16:17:32 moving on then, unless folks want to discuss on some other CI failures 16:17:59 #topic Release Planning 16:18:04 #link https://releases.openstack.org/bobcat/schedule.html 16:18:10 #info Nova deadlines are set in the above schedule 16:18:15 #info Nova spec review day today (June 27th) 16:18:24 I started to review a few specs 16:18:27 I approved one 16:18:43 I still have two other specs to look at 16:19:09 if I don't have time to look at them today, I'll continue to review the specs tomorrow 16:19:31 and tomorrow, I'll add an etherpad about the already merged specs and their code 16:19:58 as an important reminder, 16:20:00 #link https://releases.openstack.org/bobcat/schedule.html#b-nova-spec-freeze Nova Spec Freeze on July 6th ! 16:20:08 which is next week 16:20:26 most of the specs I've seen already have implementation changes 16:20:41 so thanks folks for that 16:21:11 anything anyone that wants to discuss on either specs or our release planning ? 16:21:52 - 16:22:01 moving on then 16:22:09 #topic Stable Branches 16:22:16 elodilles: wants to discuss anything ? 16:22:26 bauzas: nope, actually 16:22:31 cool cool 16:22:38 not much happened recently 16:22:40 around stable 16:22:46 #info stable gates should be OK (from stable/2023.1 to stable/train) 16:22:51 #info stable branch status / gate failures tracking etherpad: https://etherpad.opendev.org/p/nova-stable-branch-ci 16:22:57 elodilles: no news is good :) 16:23:08 :) 16:23:08 #info train-eol patch proposed https://review.opendev.org/c/openstack/releases/+/885365 16:23:14 elodilles: I've seen your comment 16:23:23 yepp, it needs an update 16:23:51 so you're also asking to EOL placement, os-vif, os-resource-classes, and the clients ? 16:24:04 I can work on it 16:24:22 yes, i guess those would be good to EOL together 16:24:24 with nova 16:24:36 as no reason to keep them in EM, right? 16:24:40 without nova... 16:24:45 elodilles: also, my question is, shall we abandon the changes in the stable branches, or would it be done automatically ? 16:24:58 os-traits and os-resource classes should be tight to the placement eol 16:25:12 they need to be abandoned, there's no automation 16:25:14 elodilles: for os-vif and novaclient, sounds ok 16:25:23 for placement dependencies, I'm quite torn 16:25:29 but os-traits might also be used be neutron 16:25:51 gibi: good call, then let's stick to os-vif and novaclient 16:26:01 i guess if nova eol's it's train branch then maybe other projects will do the same 16:26:06 we could discuss about EOLing placement and its deps on another change 16:26:20 bauzas: os-vif is also used by neutron for sure 16:26:21 (and i won't insist to keep open neutron either in that case o:) 16:26:38 elodilles: I've heard other projects that would EOL *all* their EM branches but I haven't seen yet any change :) 16:26:49 python-novaclient is used by cinder and neutron for the external events no? 16:26:54 bauzas: yes, i only seen the cinder patch yet 16:27:02 gibi: good point, let's then be conservative and only EOL novaclient 16:27:02 also heat probably uses nova-pythonclient 16:27:28 elodilles: what have we done when we EOL'd Stein ? 16:27:30 can we delete all the train branches at once? :D 16:27:36 I thought nova did it in advance 16:27:47 heat is already EOL'd up to ussuri 16:28:27 bauzas: if you ask about the open patches: i ussually abandon the patches to be able to delete the branch :) 16:28:57 so i can abandon them whenever there is a decision 16:28:58 elodilles: so I need to click on *every* open patch to abandon it ? 16:29:27 disclaimer, my office room is 28°C now, my brain is absolutely fried. 16:29:31 bauzas: i can do that (there aren't many open patches) 16:30:09 so, what's the outcome of this long conversation , 16:30:11 ? 16:30:19 here the temperature is quite moderate now 21°C :) 16:30:19 I guess only nova and novaclient can be EOL now 16:30:43 elodilles: it's 28°C *inside* my room (I don't have any A/C) 16:30:47 bauzas: I'm not sure about python-novaclient 16:31:01 okay, so let's be strict 16:31:05 and only EOL nova 16:31:11 EOLign should not break others as we are just closing the branch and tagging it 16:31:13 elodilles: fancy revisiting your comment ? 16:31:22 and libs like os-vif are installed form pypi by default 16:32:17 at most we would need a overdied-checkout for os-vif to poitn to the eol tag unless i missed something 16:32:19 bauzas: i still think it's better to EOL all the other nova projects in the same go 16:32:32 ya same ^ 16:32:59 gibi: agree then ? 16:33:02 elodilles: sure, but then somebody need to handle the fallout 16:33:24 gibi: by fallout you mean? 16:33:46 other project's job is failing to check out the deleted branch 16:34:05 as an exmple 16:34:06 example 16:34:34 yeah, that probably needs to be fixed (and honestly, more like EOL'ing everything else 16:34:43 I guess we need to work on this closely 16:35:15 AFAIK I tried to look at the jobs to find the ones that use stable/train branch of nova 16:35:29 btw, was there any result of the PTG forum session about the EM story? 16:35:33 and unfortunately, I have no crazy idea on how to get a list of such jobs 16:35:50 i haven't seen news on ML about that 16:36:12 since the job definitions are in the project repos, I would have to discover every single git repo in order to ensure this isn't the case :( 16:36:29 elodilles: the result was to follon on the ML 16:36:31 elodilles: about that, the conclusion is mostly that the TC will discuss this 16:36:42 and yeah, follow on the ML 16:37:07 bauzas: yes, looking for those jobs is a bit tricky, the "easiest" maybe to compare with ussuri's .zuul.yaml and what jobs were removed there 16:37:21 tl;dr: no agreement was there to stop the EM experience, but a clarification was needed. 16:37:46 elodilles: you mean stein ? 16:38:07 bauzas: OK, so we either rush forward with the train EOL patch or wait then until tc discussion happens? 16:38:10 * bauzas is lost in translation 16:38:26 (and again, the room temperature doesn't help) 16:38:36 the TC discussion is about EM as a whole and would affect everythin before ZED 16:38:57 elodilles: not sure any conclusion would hit our EOL of train 16:38:59 i think we can EOL train regardless of what happens for the newer EM branchs 16:39:04 righjt 16:39:07 for sure 16:39:09 bauzas: not stein, ussuri. the job, that is removed in ussuri probably was last used in train, thus its definition can be deleted (if it is outside of nova repo) 16:39:19 and IIRC, EM will continue to exist in some form 16:40:09 maybe the M of EM carries too much context, but the idea is that we'll continue to have other branches 16:40:14 I agree to drop train regardless of the TC discussion 16:40:23 those branches could be somewhere else, but that's an alternative to discuss yet 16:40:44 elodilles: again, lost in translation but I'll ping you tomorrow about it 16:40:58 OK, anyway, i think the 'fallout' will happen regardless we EOL only nova. EM / EOL'ing would be confusing if even within a project would be mixed 16:41:23 bauzas: sure, we can check it outside of the meeting :) 16:41:47 on a morning, with a fresh room temperature, a rested brain and a coffee, please. 16:42:15 bauzas: ++ :) 16:42:17 anyway, doesn't sound there is a change in direction 16:42:34 I guess we're done on that topic anyway 16:42:41 ++ 16:42:46 #topic Open discussion 16:42:49 I have one thing 16:42:55 (bauzas) Specless blueprint approval request for https://blueprints.launchpad.net/nova/+spec/num-instances-weigher 16:43:33 tl;dr: some operators said during the PTG that they were having some problems for spreading between heteregenous hardware 16:43:48 so we found we missed a weigher 16:44:23 we already have a filter for making sure a host can't have more than a max number of instances 16:44:32 +1 to have this 16:44:38 it is a quick and easy win 16:44:38 but it doesn't spread the hosts 16:45:11 yup 16:45:20 there is a current discussion on the already-provided code change https://review.opendev.org/c/openstack/nova/+/886232 16:46:11 the concern is that we have a negative default value for spreading 16:46:15 whihc is not realted to if we shoudl have the weigher. im ok wigher jsut dislike that we have to default to -1 to spread when positive multipes are normally spread since we spread by default 16:46:48 yep so after some other discussion i woudl just prefer you retrined the number of instance * -1.0 16:46:54 and default to 1.0 16:47:11 sean-k-mooney: yeah and I understand your concern but my point is that I don't want to have a max-instances value that would be the same for all hosts 16:47:14 that is a prefernce not a blocker 16:47:34 yep which is why i provided two suggestion orginally 16:47:52 hm, how the -1 relates to the max-instance? 16:48:05 it does not 16:48:11 i provided 2 suggestsions 16:48:22 return max_instace- current 16:48:32 sean-k-mooney: okay, so you would be okay if _weigh_object() would just do -1 * host_state.num_instances ? 16:48:32 or return current * -1.0 16:48:42 yep 16:48:50 tottaly fine with that 16:48:52 The -1.0 one works for me. I would not complicate it with the max-instance 16:49:23 sean-k-mooney: as I said in https://review.opendev.org/c/openstack/nova/+/886232/comment/565abda0_70ab7fde/ we already have negative values for spreading 16:49:29 https://docs.openstack.org/nova/latest/configuration/config.html#filter_scheduler.io_ops_weight_multiplier 16:49:43 yep i pointed that out to you in the review 16:50:02 and noted that its unforturneate because its teh opisctie behavior of the cpu ram and disk weighers 16:50:17 and i didnt want to proliferate that by adding more useage 16:50:33 but as i also said its just a prefrence if other dont mind then fine 16:50:37 okay, if really you want it to be positive, then I'll just multiply by -1.0 16:51:01 but I'm not sure it's really a problem to have a negative value for spreadinh 16:51:23 anyway, looks to me an implementation nit 16:51:24 it just makes it harder to reason about 16:51:35 if it would not break peopel i would change the io_ops_weigher 16:51:39 do people agree on having https://blueprints.launchpad.net/nova/+spec/num-instances-weigher to be specless ? 16:51:40 but its really not worth it 16:51:41 I wasn't aware we had any sort of weigher convention where one value means pack or spread 16:51:47 bauzas: yes, specless 16:51:55 ya specless is good with me 16:52:00 cool, then we'll revisit this conversation in gerrit 16:52:09 ack 16:52:09 I've heard you say that, but I've never heard that before 16:52:10 anyone fancy to review the code itself 16:52:18 I guess I'll have to go look at the existing weighers and see 16:52:23 dansmith: I can't remember it too 16:52:42 in general our weighers have different default values 16:52:45 +1 on specless 16:52:59 like https://docs.openstack.org/nova/latest/configuration/config.html#filter_scheduler.build_failure_weight_multiplier 16:53:03 bauzas: yeah I thought they were all pretty much individually contextual 16:53:09 which defaults to 1000000.0 16:53:10 in generall they all spread by default and positive measn spread. those are the two ruels of tumb i learend early on 16:53:21 anyway 16:53:28 if we broke that at some poin thats unfortunet btu its what we are stuck with 16:53:39 #agreed https://blueprints.launchpad.net/nova/+spec/num-instances-weigher is approved as a specless bp 16:53:47 \o/ 16:53:55 I guess we can continue this in gerrit :!) 16:53:57 :) 16:54:34 okay, so I don't see any other item 16:54:42 anyone wanting to discuss on something ? 16:55:09 * gibi sends the nice cool weather towards France 16:55:41 thanks :) 16:55:50 :) 16:56:01 I guess I'll need to stop sooner than later 16:56:06 anyway 16:56:09 thanks all 16:56:11 #endmeeting