16:00:43 #startmeeting nova 16:00:43 Meeting started Tue Oct 11 16:00:43 2022 UTC and is due to finish in 60 minutes. The chair is bauzas. Information about MeetBot at http://wiki.debian.org/MeetBot. 16:00:43 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 16:00:43 The meeting name has been set to 'nova' 16:00:45 heyho 16:00:50 o/ 16:01:09 who's around ? 16:02:08 we can start if needed 16:02:20 hopefully, should be a quick one, as we have the PTG next week 16:02:43 ++ 16:02:50 #topic Bugs (stuck/critical) 16:02:56 #info No Critical bug 16:03:01 #link https://bugs.launchpad.net/nova/+bugs?search=Search&field.status=New 4 new untriaged bugs (-1 since the last meeting) 16:03:01 o/ 16:03:05 o/ 16:03:10 #link https://storyboard.openstack.org/#!/project/openstack/placement 26 open stories (+0 since the last meeting) in Storyboard for Placement 16:03:20 elodilles: thanks for having looked at the bugs 16:03:24 np 16:03:30 anything you would want to discuss ? 16:03:36 #info Add yourself in the team bug roster if you want to help https://etherpad.opendev.org/p/nova-bug-triage-roster 16:03:48 maybe one bug 16:03:52 (or two) 16:04:14 */ 16:04:19 related to the well-know volume timeout failure 16:04:33 a new general bug was open: https://bugs.launchpad.net/nova/+bug/1992328 16:04:52 i know that there are a ton of similar bugs already 16:05:04 some are more specific and some are general 16:05:18 elodilles: do you want to close it as a duplicate ? 16:05:34 if we have the exact same somewhere, then we could 16:05:44 otherwise we can keep it open... 16:06:11 also saw that maybe somewhat coupled with this 30 days old bug: https://bugs.launchpad.net/nova/+bug/1989232 16:06:30 i mean this is also opened around the same issue 16:07:50 otherwise i didn't have time to go deep into them 16:08:05 well, I dunno 16:08:09 so any idea about what to do with these bugs are welcome 16:08:38 these need to be trobule shooted to find the root cause 16:08:47 but I don't have time for that 16:09:05 timeout is a the visible fault but there is deeper reasons why that happnes 16:09:28 ok, so let's continue to have this bug report then 16:09:43 and if someone wants to fix it, he/she could duplicate it if needed 16:09:57 ack 16:10:59 ok, any other bug report to look at ? 16:11:11 nothing else from me 16:12:52 ok, continuing then 16:13:09 gibi: can you be the bug baton for next week ? 16:13:15 lets see 16:13:17 even if next week it will be the PTG ? 16:13:30 I will take it 16:13:32 ok 16:13:34 thanks 16:13:41 #info bug baton is being passed to gibi 16:13:58 (sorry, passing it to you, as I was having the baton for 2 weeks :) ) 16:14:13 moving on 16:14:15 #topic Gate status 16:14:20 #link https://bugs.launchpad.net/nova/+bugs?field.tag=gate-failure Nova gate bugs 16:14:26 #link https://zuul.openstack.org/builds?project=openstack%2Fnova&project=openstack%2Fplacement&pipeline=periodic-weekly Nova&Placement periodic jobs status 16:14:28 heh :) 16:14:44 as you see, we have a timeout with the centos9-fips job 16:15:20 but I dunno for how long it was having this timeout 16:16:20 https://zuul.openstack.org/builds?job_name=tempest-centos9-stream-fips&project=openstack%2Fnova&project=openstack%2Fplacement&pipeline=periodic-weekly&skip=0 16:16:28 it seems always? 16:16:33 yes 16:17:53 I guess the owner of the job was adalee, right?. 16:18:40 this is bizarre, I don't see where the job run is timing out 16:20:01 well the default timeout i think is 2 hours 16:20:14 so it might just need a little longer sicne the fips job does a reboot 16:20:24 yeah but the job seems to have done 16:20:30 be* done 16:20:38 anyway, nothing urgent 16:20:53 this is just the fact we can't get the result 16:21:00 even if it looks tempest works fine 16:21:11 its posible it has addtional test in a post playbook 16:21:24 I don't know who wrote the job 16:21:34 but I'll try to find it 16:21:46 should be easy to find 16:22:24 https://zuul.openstack.org/build/5f6e6a2f65ee4a5e90754d07142fab9f/log/job-output.txt#23945-23946 shows where it is timing out 16:23:24 found who added it https://review.opendev.org/c/openstack/nova/+/831844 16:23:54 thanks clarkb, will look 16:24:31 ok its tempest full 16:24:47 so its two tempest runs 16:25:05 | RUN END RESULT_TIMED_OUT: [untrusted : opendev.org/openstack/tempest/playbooks/devstack-tempest.yaml@master] 16:25:09 the first one completed but the second that runs the slow test and senarior tests timed out 16:25:10 hmpf 16:25:24 yup 16:25:45 so ya it just need anther say 30 mins added to the tiemout to be safe 16:26:05 it also could have just been a slow node 16:26:10 and maybe double check logs to see that something didn't get stuck there due to fips 16:26:19 there is a bit of a time delta between the timeout and tempest reporting anything 16:26:38 oh its only 15 seconds nevermind 16:26:42 yep 16:26:47 there isnt really a break in the logs 16:26:53 I mathed wrong the first time 16:27:08 yes, this is just a slow test 16:27:42 but should we modify the timeout for that job to be larger ? 16:27:46 i dont know how much time the reboot adds for fips but 2 hours is proably borderlien 16:28:01 we can DNM this at least 16:28:16 and verify if adding more timeout time helps 16:28:17 i would add 30mins and monitor it and see roughly how long it takes over the next few weeks 16:28:23 right 16:28:46 I can propose a patch 16:28:51 ack 16:29:21 #action bauzas to add a 30min more timeout for the centos9-fips periodic job so we will see whether it fixes the timeout 16:29:27 moving on 16:29:38 #info Please look at the gate failures and file a bug report with the gate-failure tag. 16:29:51 bauzas: its here https://opendev.org/openstack/tempest/src/branch/master/zuul.d/integrated-gate.yaml#L325 just incase your wondering where to do that 16:31:19 ack, thanks but I found it 16:31:40 hence the DNM patch in nova 16:31:51 to test it 16:32:07 with experimental 16:32:10 anyway 16:32:36 nothing to tell more about blind rechecks, I haven't looked at the numbers 16:32:38 moving on 16:32:48 #info STOP DOING BLIND RECHECKS aka. 'recheck' https://docs.openstack.org/project-team-guide/testing.html#how-to-handle-test-failures 16:32:53 #topic Release Planning 16:32:59 #link https://releases.openstack.org/zed/schedule.html 16:33:04 #info Zed GA was last week, kudos to the team. 16:33:10 we're officially done with Zed :) 16:33:21 so... 16:33:23 #link https://releases.openstack.org/antelope/schedule.html 16:33:28 #info Antelope-1 is planned in 5 weeks 16:33:38 nothing to tell it, the crux is for the next topic 16:33:50 #topic PTG planning 16:33:55 #link https://etherpad.opendev.org/p/nova-antelope-ptg Antelope PTG etherpad 16:34:00 #link https://ptg.opendev.org/ptg.html PTG schedule 16:34:09 #info Attempted schedule in the etherpad 16:34:21 yeah I did a bit of a topics scrambling 16:34:38 but honestly, we don't have a large agendat *yet* 16:35:04 I'll just move the nova-otherproject topics 16:35:29 for the moment, we have nova-neutron (Thur) and nova-ironic (reminder : *Monday*) 16:35:44 do people want to talk with cinder or other projects ? 16:35:52 I haven't got any ask from the cyborg project 16:36:06 or oslo or QA ? 16:36:25 or manila, or glance... :) 16:36:57 nothing on my side 16:37:04 i think the ironic topic will fill any space we give it 16:37:14 but im not sure if we will be able to make progress remotely 16:37:57 hah 16:38:31 sean-k-mooney: as I already said a couple of times, I don't think we will find a solution 16:38:33 but, 16:38:55 at least I want our community to understand why we can't just modify an instance .host DB value 16:39:25 we also shoudl not have the hashring in the driver as we do today 16:39:35 if we're able to explain it even after all the comments dansmith and you did, well, it will be nice then 16:40:03 i look forward to your summary 16:40:16 anyway, I don't want to discuss this spec by now, we'll have time on Monday :) 16:40:21 * sean-k-mooney has flushed most of that out of my brain but will re read the spec 16:40:31 sean-k-mooney: /o\ 16:40:46 but yeah, I'll summarize this session after the PTG :) 16:40:55 moving on 16:41:12 looks we can find other projects to discuss if we want during the PTG anyxway 16:41:32 as a reminder too, we could get pain points or feature requests from operators 16:41:51 #link https://etherpad.opendev.org/p/oct2022-ptg-operator-hour-nova Operator specific etherpad for our nova operator hours 16:41:59 i hope we do. even if they are ones we have gotten before 16:42:23 #topic Review priorities 16:42:29 #link https://review.opendev.org/q/status:open+(project:openstack/nova+OR+project:openstack/placement+OR+project:openstack/os-traits+OR+project:openstack/os-resource-classes+OR+project:openstack/os-vif+OR+project:openstack/python-novaclient+OR+project:openstack/osc-placement)+(label:Review-Priority%252B1+OR+label:Review-Priority%252B2) 16:42:55 I need to add some changes I have as a review-prio 16:43:05 apart from this, nothing to tell 16:43:17 next topic ? 16:43:27 (we'll discuss the review-prio flag at the PTG anyway) 16:43:42 #topic Stable Branches 16:43:51 elodilles: your dog. 16:43:53 o/ 16:43:56 #info from stable/zed back till stable/ussuri branches' gates should be OK 16:44:07 #info stable/train seems broken again: devstack-gate's test-matrix playbook fails ("No module named yaml") 16:44:26 and it seems we already have a workaround for this: https://review.opendev.org/c/openstack/devstack-gate/+/860961 16:44:46 and for some information this came with Ansible 6 16:45:03 the rest are the same: 16:45:06 #info stable/stein (and older) are blocked: grenade and other devstack based jobs fail with the same timeout issue as stable/train was previously 16:45:13 #info stable branch status / gate failures tracking etherpad: https://etherpad.opendev.org/p/nova-stable-branch-ci 16:45:18 and that's all 16:45:46 cool thanks 16:46:01 can't wait for next week for discussing about stein :) 16:46:12 stein and older :) 16:46:35 oh you're right btw. 16:46:42 we can now backport again \o/ 16:46:46 given Zed is released 16:46:49 huzzah 16:46:56 \o/ 16:47:00 :) 16:47:15 ok, anything to add ? 16:47:33 looks not 16:47:40 #topic Open discussion 16:47:47 just two paperwork items on my side 16:47:58 (bauzas) Cancel next week's meeting due to PTG 16:48:10 I forgot the question mark, but is this really needed ? 16:48:27 anyone having problems about cancelling ? 16:48:42 like, I can only attend by IRC, zoom is terrible to me 16:49:13 or, I can show my face, I just prefer us to continuously discussing over IRC 16:49:28 looks not 16:49:57 no need for a meeting next week 16:50:00 on IRC 16:50:01 #action bauzas to communicate (email and wiki) on Nov 18 nova meeting be CANCELLED 16:50:14 last item from me 16:50:23 (bauzas) Nov-1 is an holiday here, anyone wants to run the meetingĀ ? 16:50:31 all said ^ 16:51:00 I'm usually on perpetual PTO, but this time I'm asked to be on holiday 16:51:38 I guess most of the Europe is also having holiday at that day, if I trust my company calendar 16:51:52 shall we just cancel it too ? 16:52:11 the fun is, Nov-8 is ALSO an holiday for me the week after 16:52:13 I'm of on Nov-1 too 16:52:16 off 16:52:23 same for me 16:52:27 cancel it 16:52:53 yeah, was just opening the door to anyone fancy running the meeting if he/she wanted 16:53:07 I don't see much hands in the air 16:53:08 :) 16:53:10 ++ cancel 16:53:12 gibi: on holiday nova-8 as well or not ? 16:53:17 shit 16:53:19 nov-8 16:53:58 oh nevermind 16:54:01 I'm stupid 16:54:09 I'm not good at knowing our holidays 16:54:17 I confused myself with may 16:54:35 nov-8 is a workday, that's nov-11 which is holiday 16:54:42 * bauzas feels ashamed 16:54:52 anyway, sold. 16:54:57 I will be in Brno on nov-8 16:55:15 so working but not on upstream nova :) 16:55:26 #undo 16:55:26 Removing item from minutes: #action bauzas to communicate (email and wiki) on Nov 18 nova meeting be CANCELLED 16:55:33 Nova 18? 16:55:36 #action bauzas to communicate (email and wiki) on Oct 18 nova meeting be CANCELLED 16:55:37 Nov 18? 16:55:45 #action bauzas to communicate (email and wiki) on Nov 1st nova meeting be CANCELLED 16:55:54 getting better :) 16:56:01 I realized 16:56:02 and Oct 18 16:56:13 oct 18 and nov 1 ACK 16:56:14 sorry, there it is 16:56:18 :X 16:56:23 after this and my confusion on holiday dates, that means it's time for me to end the meeting 16:56:24 calendars are hard :D 16:57:08 thanks all and kudos for the hard work you made on Zed, btw. 16:57:20 #endmeeting