*** haleyb_ is now known as haleyb | 01:58 | |
*** haleyb is now known as haleyb_away | 01:59 | |
thuvh | IDENTIFY | 05:04 |
---|---|---|
*** akekane is now known as abhishekk | 05:36 | |
*** blarnath is now known as d34dh0r53 | 06:52 | |
*** ralonsoh__ is now known as ralonsoh | 07:35 | |
congnt | Hi everyone, I have a new compute with CPU (Intel(R) Xeon(R) Gold 5320 CPU @ 2.20GHz) Icelake Intel. But libvirt recognize model is Broadwell-noTSX-IBRS. Anyone know this bug? I'm use OpenStack Victoria deploy by kolla-ansible, libvirt version 6.0.0-0ubuntu8.8 | 08:08 |
obre | congnt: Hi! I guess that a "cat /proc/cpuinfo | grep mpx" returnes no results? | 08:15 |
congnt | Yes | 08:15 |
congnt | No result with this command | 08:16 |
obre | The issue is basicly that libvirt expects that flag to call a CPU "IceLake". | 08:17 |
obre | But Intel does not ship mpx-support for quite a few of their newer CPU-lines. | 08:17 |
obre | But you can configure libvirt with your own custom cpu-models. | 08:17 |
obre | We do for example use this file on our IceLake nodes: https://github.com/ntnusky/profile/blob/master/files/libvirt/cpu/x86_Icelake-Server-NTNU.xml | 08:18 |
obre | Add that file to /usr/share/libvirt/cpu_map/ and update /usr/share/libvirt/cpu_map/index.xml to include a link to it. Then after a restart of libvirt you should have an ICE-Lake CPU available. | 08:19 |
congnt | obre: thank you so much, i will research it. | 08:32 |
*** bhagyashris is now known as bhagyashris|sick | 09:26 | |
*** kopecmartin is now known as kopecmartin|sick | 10:45 | |
sean-k-mooney | congnt: ya so this is a common issue with older release of libvirt. | 10:52 |
sean-k-mooney | congnt: there are 3 ways to work around it if you cant us a newer libvirt, 1 is create a custom model, possibly by copying the one form a newer release, the next is to use the model that matches closed by setting cpu_mode=custom and cpu_model=<whatever> the final way to work around this is set cpu_mode=host-passthrough | 10:54 |
sean-k-mooney | for 2 i forgot to say you can add the missing cpu flage with cpu_model_extra_flags | 10:55 |
sean-k-mooney | those config options are all in the libvirt section fo the nova.conf | 10:55 |
songwenping | sean-k-mooney: hi, i remember the live migration will check source node's resource capacity right? | 11:09 |
songwenping | sean-k-mooney: hi, i remember the live migration will also check source node's memory capacity right? | 11:20 |
sean-k-mooney | i would have to check if live migration is using 2 seperate allcotions or just one. i know for some move operations we use the migration context to hold the allcoation for one of the nodes and the vm for the other | 11:27 |
sean-k-mooney | we have not ported all move operations to use that workflows. cold migration does if i recall correctly, evacuate does not i think live migration will use 2 allocations but again i would need to look at the code to confirm | 11:28 |
sean-k-mooney | gibi: bauzas do ye recall off the top of ye're heads what we do for live migration ^ | 11:32 |
gibi | live migration uses the migration uuid to hold the source node alloc | 11:56 |
gibi | only evacuate is an exception | 11:56 |
sean-k-mooney | ack that is what i tought | 11:59 |
sean-k-mooney | we should fix evacuation sooner rather then later... im going to go add that to our downstream backlog | 11:59 |
sean-k-mooney | maybe i can push for that in B or C it would be nice to get that finally fixed | 12:00 |
sean-k-mooney | although the placement allocation explostion issue might be more imporant | 12:00 |
sean-k-mooney | we also still need to start using the consumer types feature right | 12:01 |
sean-k-mooney | to actully start marging the migration allcoations as migrations | 12:01 |
gibi | yes and yes | 12:02 |
sean-k-mooney | by the way i will shortly be resuming the pci series review stephenfin are you on pto from today? | 12:03 |
sean-k-mooney | stephenfin: im hoping to finish reviewign the rest of the pci seriese today but if not it will be my goal to get it completed the first week of january when im back | 12:04 |
songwenping | sean-k-mooney,gibi: it seems not reasonable if live migration uses 2 seperate allocations, the vm cannot be migrate if the source node have not enough resouce. | 12:42 |
sean-k-mooney | songwenping: it can we wont stop the migration if the source is over commited | 12:50 |
sean-k-mooney | that will fail for evacuate but live migration shoudl work | 12:51 |
sean-k-mooney | that said you shoudl never get into that situration unless you change something in a way that was unsupproted or hit a bug | 12:51 |
sean-k-mooney | for example reduced the memory in the source node either intentiollay or due to a dim failure or change the allcoation ratio | 12:52 |
songwenping | we use the Rocky code, and placement is integrated with nova. | 12:52 |
sean-k-mooney | multiple allocation was intoduced in Queens | 12:53 |
sean-k-mooney | so it shoudl be there in rocky | 12:53 |
songwenping | but we encounter the problem, the vm cannot migrate because the source node's memory over commit | 12:54 |
sean-k-mooney | we may have a bug in rocky or master then but you shoudl not be in an over commit senario | 12:55 |
sean-k-mooney | its fine to oversubscibe provided the total amoutn adn allocation ratio align | 12:55 |
sean-k-mooney | the temporay fix woudl be to increase the allocation ratio to enabel the migration | 12:55 |
sean-k-mooney | and then restore it when done | 12:56 |
sean-k-mooney | althernitivly you coudl try cold migration. | 12:56 |
songwenping | yes, we try to increase the allocation ratio now, and analyse why the memory is over commit | 12:58 |
gibi | yepp it is a know behavior that you cannot migrate from an overallocated node. remove the overallocation by temporarily increasing allocation ratio, move the instance, restore the allocation ratio. And separately investigate how you ended up in an overallocated scenario | 13:12 |
sean-k-mooney | gibi: i tought that only affected migrtions that uses a single allcoation | 13:18 |
gibi | sean-k-mooney: I think it is the other way around. Evacuation is the only move that does not use migration allocation on the source. It only extend the instance allocation to cover both the source and the dest node. So I think evac is not effect by this | 13:20 |
gibi | all the other move operators move the instance allocation from the instance_uuid to the migration_uuid on the source node | 13:20 |
gibi | that move allocation is what triggers the situation | 13:21 |
gibi | becuase placement does not have a move semantic | 13:21 |
sean-k-mooney | hum maybe | 13:22 |
gibi | https://github.com/openstack/nova/blob/36091a7ed7ad553d5cbb5dcfde0090e1e762bc34/nova/scheduler/client/report.py#L2023-L2043 | 13:23 |
sean-k-mooney | evacuate is broken for overcommited case too however | 13:23 |
sean-k-mooney | the extention failes if the souce is over allcoated | 13:23 |
gibi | OK then I was mistaken on that part. then all allocation manipulation is reject by placement if there is at least one overallocated RP in the change | 13:23 |
sean-k-mooney | proably yes | 13:24 |
gibi | for some reason I assumed that if the allocation on the overallocated RP does not change then it is OK and evac only adds new alloc on new RPs. | 13:24 |
sean-k-mooney | i filed a downstrema tracker for this and its a know issue upstream so hopefully this will eventually get resovled | 13:24 |
sean-k-mooney | although it might require placemtn changes | 13:24 |
gibi | the resolution probably needs a new placement microversion either to change the semantic of the existing POST /allocations or to add a new API that understands move semantic | 13:25 |
gibi | but I agree to do something as it is a common issue from deployers | 13:26 |
*** dasm|off is now known as dasm | 13:36 | |
opendevreview | Jorge San Emeterio proposed openstack/nova-specs master: Review usage of oslo-privsep library on Nova https://review.opendev.org/c/openstack/nova-specs/+/865432 | 13:40 |
opendevreview | Ruby Loo proposed openstack/nova stable/yoga: Ironic nodes with instance reserved in placement https://review.opendev.org/c/openstack/nova/+/867912 | 13:49 |
opendevreview | Ruby Loo proposed openstack/nova stable/xena: Ironic nodes with instance reserved in placement https://review.opendev.org/c/openstack/nova/+/867913 | 14:03 |
opendevreview | Ruby Loo proposed openstack/nova stable/wallaby: Ironic nodes with instance reserved in placement https://review.opendev.org/c/openstack/nova/+/867914 | 14:07 |
gibi | bauzas, sean-k-mooney: fyi I filed two bugs in the last two days about gate instabilties as I'm hitting them https://bugs.launchpad.net/glance/+bug/1999800 https://bugs.launchpad.net/tempest/+bug/1999893 | 14:08 |
bauzas | shitty shit | 14:09 |
bauzas | gibi: thanks | 14:09 |
gibi | these are infrequent ones but I see both more than once so I reported them | 14:09 |
*** akekane is now known as abhishekk | 14:11 | |
*** umbSubli1 is now known as umbSublime | 14:16 | |
gibi | I'm a magnet of bugs these days | 14:52 |
gibi | the latest, this is from my local env | 14:52 |
gibi | Dec 16 15:51:21 bedrock kernel: traps: flake8[1268565] general protection fault ip:55f5a5ed6e83 sp:7ffdf4a39d50 error:0 in python3.10[55f5a5dba000+2a3000] | 14:52 |
gibi | I cannot even run tox -e pep8 as flake8 fails all the time | 14:52 |
ykarel | Hi can someone look into https://bugs.launchpad.net/nova/+bug/1949606 | 14:54 |
ykarel | libvirt-8.0.0 now provides option to set tb-cache | 14:55 |
ykarel | without it it's difficult to run multiple guests vm together in CI on jammy hosts | 14:56 |
gibi | ykarel: can we default tb-cache size globally via some libvir configuration? | 15:07 |
gibi | kashyap: ^^ | 15:07 |
ykarel | gibi, no idea, but if that's possible then would be helpful as can be set outside of nova too | 15:08 |
gibi | ykarel: yep, it would be convinient otherwise we need to create a nova feature just for our CI usage | 15:08 |
ykarel | yes | 15:09 |
gibi | i.e. a new nova compute host level config variable in [libvirt] section set to some small value applied blindly to all emulated domains by the nova-compute service | 15:09 |
ykarel | it used to be 32MiB before it was raised to 1GiB | 15:10 |
ykarel | so that should be good for CI atleast | 15:10 |
kashyap | gibi: Hmmm, good question | 15:13 |
kashyap | gibi: It rings a faint bell as I looked at it in the past, but I forget | 15:14 |
* kashyap looks | 15:14 | |
kashyap | I'm in a hurry as I need to take a train shortly, but I'll take a quick look | 15:14 |
gibi | kashyap: no worries, it is not super urgent :) | 15:14 |
kashyap | gibi: Good news: yes! libvirt does allow it | 15:15 |
kashyap | LOL, I tested it even upstream libvirt myself and totally forgot: | 15:15 |
kashyap | gibi: ykarel: https://listman.redhat.com/archives/libvir-list/2021-November/224873.html | 15:15 |
ykarel | kashyap, yeap i tested that and it works, now we are looking if we can set it globally by some libvirt conf | 15:16 |
gibi | kashyap: with my limited understanding it only show that it is allowed via the domain xml, can we also set it via some hypervisor level global config? | 15:16 |
ykarel | so we don't have to change nova code just to support CI usecase | 15:16 |
kashyap | gibi: ykarel: I don't think global config is possible - near as I know | 15:19 |
gibi | kashyap: thanks | 15:19 |
kashyap | ykarel: But just shoot an email to libvirt-users@redhat.com list and ask there. | 15:19 |
kashyap | People are friendly :) | 15:19 |
ykarel | kashyap, Ok Thanks | 15:20 |
ykarel | will send a mail | 15:20 |
kashyap | ykarel: A quick tip: Ask them to keep you explicitly in Cc you on responses, as you're not subscribed to that list (I guess) | 15:23 |
opendevreview | Balazs Gibizer proposed openstack/nova master: Split ignored_tags in stats.py https://review.opendev.org/c/openstack/nova/+/867978 | 15:24 |
gibi | sean-k-mooney: I did the split as we discussed ^^ | 15:24 |
ykarel | Thanks kashyap, yes right /me not subscribed | 15:24 |
ykarel | kashyap, gibi sent https://listman.redhat.com/archives/libvirt-users/2022-December/013844.html | 16:17 |
rloo | hi sean-k-mooney, these should be ready to approve (zuul is happy anyway!): https://review.opendev.org/c/openstack/nova/+/867912, https://review.opendev.org/c/openstack/nova/+/867913 & https://review.opendev.org/c/openstack/nova/+/867914 (thanks!) | 16:26 |
sean-k-mooney | rloo: ack | 17:18 |
opendevreview | Edward Hope-Morley proposed openstack/nova stable/yoga: ignore deleted server groups in validation https://review.opendev.org/c/openstack/nova/+/867989 | 17:29 |
sean-k-mooney | rloo: the older backports are not quite right | 17:31 |
rloo | sean-k-mooney: gahhhh. did you comment? I'll take a look. | 17:32 |
sean-k-mooney | the content is fine but it looks like you cherry picked form master in all cases instead of form the previous cherry pick | 17:32 |
sean-k-mooney | yep its pretty minor | 17:32 |
rloo | yes, i cherry picked from master. do you want to do it from previous cherry pick? | 17:32 |
sean-k-mooney | just the commit is wrong | 17:32 |
sean-k-mooney | rloo: yep you should cherry pick form the previous cherry prick | 17:33 |
sean-k-mooney | i think ironic does this slightly differntly due to how ye do bugfix branches | 17:33 |
rloo | geez. i thought if i used the UI to do the cherry pick, it'd do the right thing. the reason i didn't do from previous, was cuz things looked messier, heh. | 17:33 |
sean-k-mooney | for nova the backport go form newest to oldest branch and you cherry pick form the previosu branch | 17:33 |
rloo | i haven't been doing upstream stuff, so i don't even recall how ironic does it... i did try to find doc about it but gave up. | 17:34 |
sean-k-mooney | rloo: i acttully care about the cherry-pick lines less then other but i knwo melwitt and elodilles do like them to be done a specific way | 17:34 |
sean-k-mooney | for me i just do a git reset --hard origin/stable/<whatever> then git review -X <previous version> | 17:35 |
rloo | no worries. should i create new PRs, the 'right' way? | 17:35 |
sean-k-mooney | well they dont have to be new reviews just need to fix the commit message with the cherry pick lines | 17:36 |
rloo | well, if i manually do that -- there won't be a conflict in the wallaby one (if i recall) cuz the change was similar to the xena one. but i didn't tell you that, i'll fix the commit messages... | 17:38 |
sean-k-mooney | right so i do not normllay remove the confit bit in that case although i know other do | 17:39 |
sean-k-mooney | i do if others ask | 17:39 |
rloo | (and if someone had time to fix that UI so it doesn't allow cherry picking from master to n-2+ stable branches, heh) | 17:39 |
sean-k-mooney | one thing i have not tested is if the behvior change if the patch is merged | 17:39 |
rloo | sean-k-mooney: ahh, yes, you're right. if i had cherry picked from xena (which mentions the conflict), the wallaby one would have the same commit msg so. | 17:40 |
sean-k-mooney | i think it does | 17:40 |
sean-k-mooney | basically if its merged and you cherry pick it i think it addes the line properly | 17:41 |
sean-k-mooney | i think it only doesnt if you do it to an open reivew. this has changed in differnt gerrit versions | 17:41 |
opendevreview | Ruby Loo proposed openstack/nova stable/yoga: Ironic nodes with instance reserved in placement https://review.opendev.org/c/openstack/nova/+/867912 | 17:57 |
opendevreview | Ruby Loo proposed openstack/nova stable/xena: Ironic nodes with instance reserved in placement https://review.opendev.org/c/openstack/nova/+/867913 | 17:58 |
opendevreview | Ruby Loo proposed openstack/nova stable/wallaby: Ironic nodes with instance reserved in placement https://review.opendev.org/c/openstack/nova/+/867914 | 18:00 |
opendevreview | Ruby Loo proposed openstack/nova stable/xena: Ironic nodes with instance reserved in placement https://review.opendev.org/c/openstack/nova/+/867913 | 18:06 |
opendevreview | Ruby Loo proposed openstack/nova stable/wallaby: Ironic nodes with instance reserved in placement https://review.opendev.org/c/openstack/nova/+/867914 | 18:07 |
sean-k-mooney | those all look good bauzas if you are around the next few days can you review them and babysit those through the gate | 18:10 |
rloo | sean-k-mooney: thx for reviewing them. now i feel like i should do more upstream stuff before i forget. ha ha. (I might backport https://review.opendev.org/c/openstack/nova/+/842478 just for fun, we don't have a need for that. yet.) | 18:12 |
sean-k-mooney | so my understandign is taht should not be needed with the fix you have backported | 18:12 |
sean-k-mooney | rloo: well it would be good to have if you disable the fix you backported | 18:13 |
sean-k-mooney | so i guess if you dont have cleaning and dont want the extra time | 18:13 |
sean-k-mooney | then having both might make sense | 18:13 |
sean-k-mooney | so looking at it quikly it shoudl be backportable too so if you want too go for it | 18:14 |
rloo | we have cleaning and we don't put nodes in maint often. but i could see that being useful for others, and who knows, we might want it. The trick is getting my downstream stuff done so I have time to do some upstream stuff ;) | 18:15 |
sean-k-mooney | i know that feeling right now my upstream time is 99% reviews currently | 18:16 |
sean-k-mooney | well and irc | 18:16 |
rloo | wow, i appreciate that and I'm sure others do to sean-k-mooney! Just don't burn out on that. | 18:16 |
sean-k-mooney | well its how i can best supprot the rest of the team | 18:17 |
rloo | ++++ | 18:17 |
sean-k-mooney | i could write a bunch of code but i know it wont get reviewed quickly so while my upstream time is limited im puting it to reviews to enabel other to land there fixes | 18:18 |
* sean-k-mooney does not want to be a manager but does like to help other fix things | 18:18 | |
sean-k-mooney | rloo: by the way if the ironic folks ever want to turn there json rpc impl into an oslo messaging dirver so we can deploy nova without rabbit... i would not be upset | 18:19 |
* sean-k-mooney hopes the nats driver actully becomes a thing | 18:20 | |
sean-k-mooney | zigo: are you still persuing ^ | 18:22 |
sean-k-mooney | i ocationally look at https://review.opendev.org/q/topic:asyncio-nats but you know time | 18:22 |
rloo | sean-k-mooney: that is a great idea, might be good if you mentioned it in the ironic channel. problem is so few people, so many things we'd like to do. but worth it if we can get rid of rabbit.... | 19:14 |
opendevreview | Ruby Loo proposed openstack/nova stable/zed: Ironic: retry when node not available https://review.opendev.org/c/openstack/nova/+/867924 | 19:32 |
opendevreview | Ruby Loo proposed openstack/nova stable/yoga: Ironic: retry when node not available https://review.opendev.org/c/openstack/nova/+/868010 | 21:27 |
opendevreview | Ruby Loo proposed openstack/nova stable/yoga: Ironic: retry when node not available https://review.opendev.org/c/openstack/nova/+/868010 | 21:30 |
opendevreview | Ruby Loo proposed openstack/nova stable/xena: Ironic: retry when node not available https://review.opendev.org/c/openstack/nova/+/868011 | 21:30 |
opendevreview | Ruby Loo proposed openstack/nova stable/xena: Ironic: retry when node not available https://review.opendev.org/c/openstack/nova/+/868011 | 21:33 |
opendevreview | Ruby Loo proposed openstack/nova stable/wallaby: Ironic: retry when node not available https://review.opendev.org/c/openstack/nova/+/868012 | 21:33 |
opendevreview | Ruby Loo proposed openstack/nova stable/wallaby: Ironic: retry when node not available https://review.opendev.org/c/openstack/nova/+/868012 | 21:36 |
*** dasm is now known as dasm|off | 21:54 | |
*** osmanlicilegi is now known as Guest0 | 22:33 | |
*** ChanServ changes topic to "This channel is for Nova development. For support of Nova deployments, please use #openstack" | 22:40 |
Generated by irclog2html.py 2.17.3 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!