15:00:20 #startmeeting neutron_ci 15:00:20 Meeting started Tue Jan 10 15:00:20 2023 UTC and is due to finish in 60 minutes. The chair is slaweq. Information about MeetBot at http://wiki.debian.org/MeetBot. 15:00:20 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 15:00:20 The meeting name has been set to 'neutron_ci' 15:00:21 o/ 15:00:25 o/ 15:00:32 amorin, I'll check it after this meeting 15:00:33 o/ 15:00:34 hi 15:00:51 * amorin will be quiet during the meeting, thanks ralonsoh 15:01:02 o/ 15:01:25 I think we can start as we have quorum 15:01:32 Grafana dashboard: https://grafana.opendev.org/d/f913631585/neutron-failure-rate?orgId=1 15:01:32 Please open now :) 15:01:37 #topic Actions from previous meetings 15:01:49 lajoskatona to check dvr lifecycle functional tests failures 15:02:17 I saw the same (probably) issue this week too 15:02:55 yes, I still can't reproduce it locally so today I tried to add some extra logs, but seems I touched some shaky part as due to the extra logs 3 unit tests are failing :-) 15:03:19 https://review.opendev.org/c/openstack/neutron/+/869666 15:03:54 so I have to check why my logs break those tests 15:04:35 that's all for these issue from me 15:04:52 I just commented there 15:04:58 thanks 15:05:01 I think You missed "raise" after logging of error 15:05:10 as You now silently catching all exceptions there 15:05:40 I check it, the first ps reraised it but the result was the same, but I will check it again 15:05:50 ok 15:05:52 next one 15:05:57 lajoskatona to check networking-odl periodic failures 15:06:44 no time for it, but I saw the red results, so on my list 15:07:33 #action lajoskatona to check networking-odl periodic failures 15:07:38 ok, lets keep it for next week 15:07:43 atleast one error i show was related to tox4 in odl 15:08:21 ykarel: thanks I will check, perhaps it is just to add the magic words to tox.ini as for other projects 15:08:42 yeap 15:10:39 ++ 15:10:45 thx lajoskatona 15:10:49 and ykarel 15:11:20 ok, next one 15:11:25 slaweq to check logs in https://55905461b56e292f56bb-d32e9684574055628f247373c3e6dda1.ssl.cf1.rackcdn.com/868379/2/gate/neutron-functional-with-uwsgi/60b4ea3/testr_results.html 15:11:38 I was checking it and I proposed patch https://review.opendev.org/c/openstack/neutron/+/869205 15:12:04 I was trying to reproduce the issue in https://review.opendev.org/c/openstack/neutron/+/869225/ but so far I couldn't 15:12:15 I run 4 times 20 neutron-functional jobs and all of them passed 15:13:23 please review this patch when You will have some time 15:13:30 and thx haleyb for review 15:13:49 next one 15:13:51 slaweq to talk with hrw about cirros kernel panic 15:14:05 I talked with hrw last week and he told me to try new cirros 15:14:10 slaweq: np, i'll take a look if there are any updates 15:14:23 and if that will be still happening, increase memory in the flavor to 192 or 256 MB 15:14:50 slaweq, so they have seen similar kernel panic with older cirros versions? 15:15:11 i mean it was a known issue and newer version fixing it? 15:15:15 so I proposed https://review.opendev.org/c/openstack/neutron-tempest-plugin/+/869152 and https://review.opendev.org/c/openstack/neutron/+/869154 15:15:34 just curios on what was the fix 15:15:38 ykarel he didn't confirm it for sure but his advice was "try newer cirros first" 15:16:20 uhhk i think we have seen that issue very rarely so just trying new cirros we can't confirm until it reproduces 15:16:33 personally /me seen only once 15:16:39 didn't we have problems with nested virt and this cirros version? 15:16:53 I think I saw something similar once this week too 15:17:01 https://28f5318084af7eb69294-d7da90a475a01486cfcea9707ed18dfb.ssl.cf2.rackcdn.com/864000/5/check/tempest-integrated-networking/10b07aa/testr_results.html 15:17:25 slaweq, ^ is different 15:17:35 and workaround for that was to use uec image 15:18:03 ok, this one happened in tempest-integrated-networking which isn't using uec image 15:18:16 that would explain why it happened this week :) 15:18:27 this cirros 0.6.1 is not the one on which frickler is working to replace ubuntu minimal am I right? 15:18:48 lajoskatona I'm not sure on which he was working on 15:18:55 ok 15:19:15 ok, last one 15:19:16 yatin to check status of ubuntu minimal as advanced image in ovn tempest plugin job 15:19:17 ralonsoh, yes there were issues with cirros 0.6.1 and nested virt 15:19:32 i pushed upated to not use host-passthrough and it worked 15:19:53 but there were still some failures in stadium project jobs but /me not checked 15:20:07 for ubuntu minimal pushed https://review.opendev.org/q/topic:ubuntu-minimal-as-adv-image 15:20:45 noticed taas too was using regular image so updated those too to use minimal and as per tests working there too 15:21:02 I just approved taas one 15:21:10 thx 15:21:28 thanks for it, good catch 15:21:40 thx ykarel 15:21:48 ok, I think we can move on to the next topic now 15:22:17 you were thinking about https://review.opendev.org/c/openstack/neutron-tempest-plugin/+/854910 I guess 15:22:33 #topic Stable branches 15:22:36 bcafarel is not here but do You have anything related to stable branches ci? 15:22:38 since that's auto-abandoned, I guess I'll leave it at that 15:23:44 frickler we can restore it if You want to continue 15:23:48 frickler: yes that was it, the abandon I guess was just the usual cleanup process, we can check if it worth to continue 15:25:17 ok, if there's nothing regarding stable, lets move on 15:25:19 #topic Stadium projects 15:25:35 lajoskatona any updates, except odl which we already discussed earlier 15:25:48 I hope all of them are now safe from the tox4 issue ecept odl 15:26:00 that's it from me 15:26:47 lajoskatona I saw many other projects red this week in periodic weekly jobs 15:27:06 I didn't check results so I'm not sure if that was still tox4 issues or something else 15:27:13 yes those were tox4 issues 15:27:19 ahh, ok 15:27:29 so next week it should be much more green I hope 15:28:01 I hope :-) 15:28:11 LOL 15:28:16 :) 15:28:19 next topic then 15:28:20 let's hope for greener pastures 15:28:21 #topic Grafana 15:28:26 https://grafana.opendev.org/d/f913631585/neutron-failure-rate 15:29:53 TBH there is not much data from last few days there and I don't know exactly why 15:30:34 people returnign from holidays until this week maybe 15:30:59 there was some spike of failures during the weekend but now seems that things are getting back to normal 15:31:41 I think we can move on 15:31:48 #topic Rechecks 15:32:03 regarding number of rechecks, we are back to the much better numbers 15:32:14 | 2023-1 | 0.29 | 15:32:14 | 2023-2 | 0.0 | 15:32:29 and I hope it will stay like that for longer time 15:32:53 regarding bare rechecks, it also seems good: 15:33:13 3 out of 17 were bare, which is about 18% 15:33:37 thx for checking ci issues before rechecking 15:33:50 any questions/comments regarding rechecks? 15:34:05 nope 15:34:22 thanks for keeping track of it 15:35:23 ok, so lets move on 15:35:24 next topic 15:35:25 #topic fullstack/functional 15:35:26 for functional I have 2 issues this week 15:35:27 neutron.tests.functional.agent.l3.test_keepalived_state_change.TestMonitorDaemon.test_read_queue_change_state 15:35:30 https://c50fdb7f046159692f4d-3059cf1890ea1358c70d952067d56657.ssl.cf2.rackcdn.com/869388/1/check/neutron-functional-with-uwsgi/1e50279/testr_results.html 15:36:00 anyone wants to check it? 15:36:29 o/ 15:37:15 thx mlavalle 15:37:29 #action mlavalle to check failed neutron.tests.functional.agent.l3.test_keepalived_state_change.TestMonitorDaemon.test_read_queue_change_state 15:38:01 is that the only instance so far? 15:38:03 second one is (probably) the same issue as lajoskatona is already checking: 15:38:08 https://storage.bhs.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_b1f/869163/3/check/neutron-functional-with-uwsgi/b1f4063/testr_results.html 15:38:16 yes mlavalle, I found it only once 15:38:22 ack 15:38:28 mlavalle, slaweq it is a timing issue 15:38:34 the text is in the file 15:38:42 but populated just before the timewait 15:38:44 yes at least it is from the DVR ones, but the tracebeck is not exactly the same 15:38:46 (it is in the logs) 15:39:08 ralonsoh so maybe we should increase the timeout slightly then? 15:39:15 yes, a couple of secs 15:39:28 ok, I'll try that 15:39:32 mlavalle so it should be easy patch for You :) 15:39:44 hope so :-) 15:39:52 :) 15:40:47 now fullstack 15:40:58 neutron.tests.fullstack.test_agent_bandwidth_report.TestPlacementBandwidthReport.test_configurations_are_synced_towards_placement(Open vSwitch agent) 15:41:00 https://7a456b090239dc19e21e-66179e131883a8ab832a0afb9e9b5999.ssl.cf5.rackcdn.com/869388/1/check/neutron-fullstack-with-uwsgi/d4f2039/testr_results.html 15:41:31 I check this one 15:41:39 only one occurance? 15:41:40 anyone wants to check if this is maybe some timing issue or something worth to report and investigate? 15:41:46 thx lajoskatona 15:41:54 yes, it also happened only once so far 15:42:00 ok 15:42:16 #action lajoskatona to check fullstack failure neutron.tests.fullstack.test_agent_bandwidth_report.TestPlacementBandwidthReport.test_configurations_are_synced_towards_placement(Open vSwitch agent) 15:42:18 and second one 15:42:30 IP or gateway not configured properly (again): https://de87c0b256f64d4fa9ad-627fb04945741dffd55f8af38c253b04.ssl.cf2.rackcdn.com/869613/1/check/neutron-fullstack-with-uwsgi/bc5e466/testr_results.html 15:43:04 this one I will check 15:43:18 as it's something what I was hope to be fixed few weeks ago already 15:43:28 #action slaweq to check IP or gateway not configured properly (again): https://de87c0b256f64d4fa9ad-627fb04945741dffd55f8af38c253b04.ssl.cf2.rackcdn.com/869613/1/check/neutron-fullstack-with-uwsgi/bc5e466/testr_results.html 15:43:48 and with that I reached end of my list for today :) 15:43:57 #topic On Demand 15:44:10 anything else You want do discuss today? 15:44:16 related to the CI of course ;) 15:44:36 all good 15:44:50 not from me 15:45:38 ok, so I think we can end the meeting now 15:45:42 thx for attending 15:45:47 o/ 15:45:48 have a great week and see You online 15:45:48 see you 15:45:50 o/ 15:45:53 o/ 15:45:53 o/ 15:45:54 #endmeeting