*** hamalq has quit IRC | 00:12 | |
*** hamalq has joined #openstack-nova | 00:12 | |
*** rchurch has quit IRC | 00:23 | |
*** ociuhandu has joined #openstack-nova | 00:26 | |
*** mlavalle has quit IRC | 00:26 | |
*** rchurch has joined #openstack-nova | 00:26 | |
*** mlavalle has joined #openstack-nova | 00:27 | |
*** ociuhandu has quit IRC | 00:30 | |
*** LinPeiWen has joined #openstack-nova | 00:33 | |
prometheanfire | melwitt: thanks :D | 00:45 |
---|---|---|
*** bbowen has joined #openstack-nova | 01:02 | |
*** k_mouza has joined #openstack-nova | 01:30 | |
*** k_mouza has quit IRC | 01:35 | |
*** aarents has quit IRC | 01:36 | |
*** aarents has joined #openstack-nova | 01:36 | |
*** sapd1 has joined #openstack-nova | 01:39 | |
*** martinkennelly has joined #openstack-nova | 01:58 | |
*** xinranwang has joined #openstack-nova | 01:59 | |
*** martinkennelly has quit IRC | 02:00 | |
*** martinkennelly has joined #openstack-nova | 02:00 | |
*** tbachman has quit IRC | 02:02 | |
*** dviroel has quit IRC | 02:13 | |
*** macz_ has quit IRC | 02:15 | |
*** rcernin has joined #openstack-nova | 02:39 | |
*** hamalq has quit IRC | 02:45 | |
*** spatel has joined #openstack-nova | 02:47 | |
*** LinPeiWen has quit IRC | 03:02 | |
*** khomesh24 has joined #openstack-nova | 03:26 | |
*** martinkennelly has quit IRC | 03:26 | |
*** macz_ has joined #openstack-nova | 03:33 | |
*** macz_ has quit IRC | 03:38 | |
*** mkrai has joined #openstack-nova | 03:43 | |
*** dklyle has quit IRC | 04:06 | |
*** LinPeiWen47 has joined #openstack-nova | 04:07 | |
*** LinPeiWen47 has quit IRC | 04:17 | |
*** spatel has quit IRC | 04:25 | |
*** links has joined #openstack-nova | 04:36 | |
*** nweinber has joined #openstack-nova | 04:37 | |
*** nweinber has quit IRC | 04:43 | |
*** hemanth_n has joined #openstack-nova | 05:16 | |
*** ratailor has joined #openstack-nova | 05:16 | |
*** vishalmanchanda has joined #openstack-nova | 05:19 | |
*** whoami-rajat__ has joined #openstack-nova | 05:23 | |
*** LinPeiWen has joined #openstack-nova | 05:25 | |
*** xinranwang has quit IRC | 05:39 | |
*** k_mouza has joined #openstack-nova | 05:54 | |
*** yonglihe has quit IRC | 05:57 | |
*** k_mouza has quit IRC | 05:58 | |
*** pmannidi has joined #openstack-nova | 06:32 | |
openstackgerrit | sean mooney proposed openstack/nova master: [WIP] harden image metadata storage https://review.opendev.org/c/openstack/nova/+/774044 | 06:34 |
*** mkrai has quit IRC | 07:01 | |
*** mkrai_ has joined #openstack-nova | 07:01 | |
*** zzzeek has quit IRC | 07:07 | |
*** khomesh24 has quit IRC | 07:07 | |
*** zzzeek has joined #openstack-nova | 07:08 | |
*** lpetrut has joined #openstack-nova | 07:11 | |
*** slaweq has joined #openstack-nova | 07:15 | |
*** mkrai_ has quit IRC | 07:20 | |
*** khomesh24 has joined #openstack-nova | 07:28 | |
*** pmannidi has quit IRC | 07:31 | |
*** pmannidi has joined #openstack-nova | 07:34 | |
*** ralonsoh has joined #openstack-nova | 07:36 | |
*** rpittau|afk is now known as rpittau | 07:39 | |
*** xek has joined #openstack-nova | 07:51 | |
gibi | good morning | 08:02 |
*** mkrai_ has joined #openstack-nova | 08:16 | |
*** tesseract has joined #openstack-nova | 08:17 | |
*** mkrai_ has quit IRC | 08:18 | |
*** mkrai_ has joined #openstack-nova | 08:18 | |
openstackgerrit | Yongli He proposed openstack/nova master: smartnic support https://review.opendev.org/c/openstack/nova/+/758944 | 08:21 |
*** mkrai_ has quit IRC | 08:23 | |
*** mkrai_ has joined #openstack-nova | 08:29 | |
*** cgoncalves has quit IRC | 08:31 | |
*** cgoncalves has joined #openstack-nova | 08:33 | |
*** haleyb has quit IRC | 08:34 | |
*** haleyb has joined #openstack-nova | 08:37 | |
*** pmannidi has quit IRC | 08:43 | |
*** pmannidi has joined #openstack-nova | 08:45 | |
*** tosky has joined #openstack-nova | 08:46 | |
*** andrewbonney has joined #openstack-nova | 08:50 | |
*** LinPeiWen has quit IRC | 08:53 | |
*** tesseract has quit IRC | 08:59 | |
*** tesseract has joined #openstack-nova | 09:01 | |
*** tobias-urdin has joined #openstack-nova | 09:05 | |
*** LinPeiWen has joined #openstack-nova | 09:05 | |
*** rcernin has quit IRC | 09:07 | |
*** rcernin has joined #openstack-nova | 09:17 | |
*** derekh has joined #openstack-nova | 09:22 | |
*** rcernin has quit IRC | 09:25 | |
*** rcernin has joined #openstack-nova | 09:43 | |
*** pmannidi has quit IRC | 10:00 | |
*** pmannidi has joined #openstack-nova | 10:03 | |
*** sapd1 has quit IRC | 10:09 | |
*** ociuhandu has joined #openstack-nova | 10:37 | |
*** mkrai_ has quit IRC | 10:43 | |
*** mkrai_ has joined #openstack-nova | 10:43 | |
*** dviroel has joined #openstack-nova | 10:47 | |
*** ociuhandu has quit IRC | 10:48 | |
*** rcernin has quit IRC | 10:48 | |
*** ociuhandu has joined #openstack-nova | 10:49 | |
*** rcernin has joined #openstack-nova | 10:53 | |
*** ociuhandu has quit IRC | 10:55 | |
*** ociuhandu has joined #openstack-nova | 10:57 | |
*** rcernin has quit IRC | 10:57 | |
*** dtantsur|afk is now known as dtantsur | 11:01 | |
*** pmannidi has quit IRC | 11:14 | |
*** pmannidi has joined #openstack-nova | 11:18 | |
*** k_mouza has joined #openstack-nova | 11:18 | |
*** mkrai_ has quit IRC | 11:21 | |
lyarwood | stephenfin: morning, do you know if there's a standard way of checking the config of an instance in the libvirt func tests before I go looking or hacking around? | 11:39 |
* lyarwood assumes a pass-through mock saving the config object somewhere would do | 11:39 | |
stephenfin | wdym the "config"? | 11:39 |
stephenfin | As in the XML generated by nova or attributes of the instance? | 11:40 |
lyarwood | stephenfin: well that or the GuestConfig (?) objects we generate that in turn create the XML | 11:40 |
lyarwood | stephenfin: I don't really want to assert things in the XML if we have objects that work just as well tbh | 11:40 |
lyarwood | LibvirtConfigObjects is what I mean sorry | 11:41 |
lyarwood | np if there's no prior art, I'll hack something up now | 11:41 |
stephenfin | Hmm, I don't think so. The only example I can think of is the NUMA live migration tests, but that intercepts the live migration API call to validate the XML | 11:41 |
lyarwood | cool yeah I think that's the one I saw yesterday, I'll hack something up now | 11:42 |
lyarwood | _get_guest_config should work tbh | 11:43 |
*** ociuhandu_ has joined #openstack-nova | 11:46 | |
*** rcernin has joined #openstack-nova | 11:49 | |
*** ociuhandu has quit IRC | 11:50 | |
*** ociuhandu_ has quit IRC | 11:50 | |
*** ociuhandu has joined #openstack-nova | 11:57 | |
*** ociuhandu has quit IRC | 12:02 | |
sean-k-mooney | lyarwood: there isnt really one | 12:14 |
sean-k-mooney | unless you have a reference to the driver | 12:15 |
sean-k-mooney | normally the funct test start the services | 12:15 |
sean-k-mooney | and there is no rpc call that you can make to teh compute manager to get it so you need to use teh service instance ot get teh driver an then call functions on it. | 12:15 |
sean-k-mooney | but what you can do is use the logs | 12:16 |
sean-k-mooney | and you can technically get teh domain xmls form that if you had too | 12:16 |
lyarwood | yeah I've got it working | 12:19 |
lyarwood | we keep a reference to the driver so it's a normal passthrough mock | 12:19 |
sean-k-mooney | cool | 12:20 |
*** hemanth_n has quit IRC | 12:22 | |
*** pmannidi has quit IRC | 12:24 | |
*** pmannidi has joined #openstack-nova | 12:27 | |
*** k_mouza has quit IRC | 12:31 | |
*** k_mouza has joined #openstack-nova | 12:38 | |
*** ratailor has quit IRC | 12:42 | |
lyarwood | sean-k-mooney: so https://review.opendev.org/c/openstack/nova/+/774044/1/nova/utils.py would still let hw_machine_type through right | 12:46 |
* lyarwood pushes his WIP stuff to show the func tests | 12:46 | |
sean-k-mooney | yes | 12:47 |
sean-k-mooney | but i know how to fix that im debating if i want to go with a simpler fix before this one | 12:47 |
*** gryf is now known as _gryf | 12:47 | |
sean-k-mooney | basically i notice we never included ramdisk_id and kernel_id when we converted image metadata to ovos | 12:48 |
*** _gryf is now known as gryf | 12:48 | |
sean-k-mooney | so this became much more involed of a change involving object changes | 12:48 |
lyarwood | kk | 12:48 |
sean-k-mooney | but it was quite late/early this moring when i figured that out and didnt want to start agin | 12:48 |
*** gryf is now known as niech_co_krwawy_ | 12:49 | |
*** niech_co_krwawy_ is now known as gryf | 12:49 | |
sean-k-mooney | it will include hw_machine_type because you are using the same key as the image proerty - a sufix | 12:49 |
sean-k-mooney | lyarwood: if you add a prefix it would not be included | 12:49 |
lyarwood | kk, well I think I might leave it, the actual behaviour means I don't need to make any other changes to the util method fetching the machine type | 12:50 |
sean-k-mooney | what im going to do in the simpler version is remove key = key[len(SM_IMAGE_PROP_PREFIX):] | 12:50 |
lyarwood | kk | 12:50 |
sean-k-mooney | and build the proerties differnetly so it only contianes the one with image_ | 12:50 |
sean-k-mooney | lyarwood:hehe | 12:51 |
*** songwenping_ has joined #openstack-nova | 12:51 | |
lyarwood | so that would cause me to also add a system_metadata lookup in that util method | 12:51 |
lyarwood | that isn't the end of the world, it's what I had originally | 12:51 |
lyarwood | in a seperate change | 12:52 |
sean-k-mooney | well i wanted you to use image_ orginally as a prefix to avoid the extra lookup | 12:52 |
sean-k-mooney | we can do it either way | 12:52 |
sean-k-mooney | i was going to look at wriing the other patch to have someting backportable | 12:52 |
sean-k-mooney | assumeing we wanted too | 12:52 |
sean-k-mooney | if we dont then i can just finish fixing this patch | 12:53 |
lyarwood | that's also something I could do, always storing it as SM_IMAGE_PROP_PREFIX_hw_machine_type | 12:53 |
lyarwood | and that wouldn't bork me | 12:53 |
lyarwood | with this fix in place | 12:53 |
sean-k-mooney | ya | 12:53 |
*** swp20 has quit IRC | 12:54 | |
sean-k-mooney | do you have an opipion on what way i should go with the fix. should i also do the version without the ovo change | 12:54 |
lyarwood | I don't at the moment tbh | 12:59 |
lyarwood | Let me post this and grab something to eat and then I'll think it through once I'm back | 12:59 |
sean-k-mooney | cool its more dod we need a backportable fix or not | 12:59 |
sean-k-mooney | my current fix uses the ovo field but we forgot to add at least 2 when we created the object | 13:00 |
sean-k-mooney | so it cant be backported but i can fix it without relying on the fields too | 13:00 |
sean-k-mooney | which would be | 13:00 |
sean-k-mooney | i can do both too so its not either or | 13:01 |
openstackgerrit | Lee Yarwood proposed openstack/nova master: WIP libvirt: Record the machine_type of instances in system_metadata https://review.opendev.org/c/openstack/nova/+/767533 | 13:02 |
openstackgerrit | Lee Yarwood proposed openstack/nova master: WIP nova-manage: Add commands for managing instance machine type https://review.opendev.org/c/openstack/nova/+/769548 | 13:02 |
openstackgerrit | Lee Yarwood proposed openstack/nova master: WIP nova-status: Add hw_machine_type check for libvirt instances https://review.opendev.org/c/openstack/nova/+/770643 | 13:02 |
lyarwood | sean-k-mooney: https://review.opendev.org/c/openstack/nova/+/767533/5/nova/tests/functional/libvirt/test_machine_type.py - comments on the tests here would be appreciated btw | 13:04 |
*** nweinber has joined #openstack-nova | 13:04 | |
sean-k-mooney | sure. i think i owe bauzas a review of his routed stuff first but i now have your open in front of me :) | 13:05 |
*** sapd1 has joined #openstack-nova | 13:06 | |
*** mtreinish has joined #openstack-nova | 13:06 | |
lyarwood | thanks | 13:08 |
* lyarwood -> lunch | 13:08 | |
bauzas | sean-k-mooney: I'll upload a new revision https://review.opendev.org/c/openstack/nova/+/773976 today later | 13:09 |
sean-k-mooney | bauzas: ok ill review lyarwood patch then now and ill look at yours when you push it | 13:10 |
bauzas | thanks | 13:11 |
*** dtantsur is now known as dtantsur|brb | 13:14 | |
openstackgerrit | Balazs Gibizer proposed openstack/nova master: Remove __unicode__() from nova unit test Exception https://review.opendev.org/c/openstack/nova/+/769894 | 13:26 |
*** rcernin has quit IRC | 13:43 | |
*** ociuhandu has joined #openstack-nova | 13:45 | |
*** sapd1 has quit IRC | 13:45 | |
*** ociuhandu has quit IRC | 13:48 | |
*** spatel has joined #openstack-nova | 13:49 | |
*** ociuhandu has joined #openstack-nova | 13:49 | |
*** pmannidi has quit IRC | 13:49 | |
*** pmannidi has joined #openstack-nova | 13:51 | |
*** ociuhandu has quit IRC | 13:54 | |
*** ociuhandu has joined #openstack-nova | 13:55 | |
*** songwenping__ has joined #openstack-nova | 14:01 | |
*** songwenping_ has quit IRC | 14:05 | |
openstackgerrit | Lee Yarwood proposed openstack/nova master: Add regression test for bug #1908075 https://review.opendev.org/c/openstack/nova/+/766976 | 14:05 |
openstack | bug 1908075 in OpenStack Compute (nova) "Nova allows a non-multiattach volume to be attached to multiple instances *if* its volume state is reset by an admin" [Undecided,In progress] https://launchpad.net/bugs/1908075 - Assigned to Lee Yarwood (lyarwood) | 14:05 |
openstackgerrit | Lee Yarwood proposed openstack/nova master: api: Reject volume attach requests when an active bdm exists https://review.opendev.org/c/openstack/nova/+/768472 | 14:05 |
openstackgerrit | Lee Yarwood proposed openstack/nova master: fup: Merge duplicate volume attachment checks https://review.opendev.org/c/openstack/nova/+/773380 | 14:05 |
lyarwood | stephenfin: ^ updated, would you mind hitting the changes below that fup as well? | 14:05 |
*** khomesh24 has quit IRC | 14:08 | |
*** ociuhandu has quit IRC | 14:09 | |
*** ociuhandu has joined #openstack-nova | 14:09 | |
*** dtantsur|brb is now known as dtantsur | 14:16 | |
*** ociuhandu has quit IRC | 14:16 | |
*** Underknowledge has quit IRC | 14:17 | |
*** Underknowledge1 has joined #openstack-nova | 14:17 | |
*** Underknowledge1 is now known as Underknowledge | 14:18 | |
*** ociuhandu has joined #openstack-nova | 14:18 | |
openstackgerrit | Ghanshyam proposed openstack/placement master: DNM: testing direct l-c https://review.opendev.org/c/openstack/placement/+/773813 | 14:37 |
lyarwood | sean-k-mooney: ^ sorry forgot to update the unit tests in that fist change | 14:44 |
* lyarwood works on docs now | 14:44 | |
lyarwood | argh git-review is still slow | 14:44 |
openstackgerrit | Lee Yarwood proposed openstack/nova master: WIP libvirt: Record the machine_type of instances in system_metadata https://review.opendev.org/c/openstack/nova/+/767533 | 14:45 |
openstackgerrit | Lee Yarwood proposed openstack/nova master: WIP nova-manage: Add commands for managing instance machine type https://review.opendev.org/c/openstack/nova/+/769548 | 14:45 |
openstackgerrit | Lee Yarwood proposed openstack/nova master: WIP nova-status: Add hw_machine_type check for libvirt instances https://review.opendev.org/c/openstack/nova/+/770643 | 14:45 |
*** ociuhandu has quit IRC | 14:45 | |
gibi | lyarwood, stephenfin: do we need both? https://review.opendev.org/c/openstack/nova/+/773727 https://review.opendev.org/c/openstack/nova/+/769920 | 14:55 |
stephenfin | Yes, I think so. The fixtures proves the stubbing isn't complete and would be useful even when it is to prevent regressions | 14:56 |
lyarwood | yeah what stephenfin said, already has shown a few things we missed AFAICT | 14:56 |
*** belmoreira has joined #openstack-nova | 15:00 | |
*** tesseract has quit IRC | 15:02 | |
gibi | thanks | 15:04 |
*** tesseract has joined #openstack-nova | 15:05 | |
*** derekh has quit IRC | 15:05 | |
*** ociuhandu has joined #openstack-nova | 15:11 | |
*** ociuhandu has quit IRC | 15:21 | |
*** pmannidi has quit IRC | 15:21 | |
*** ociuhandu has joined #openstack-nova | 15:22 | |
*** lpetrut has quit IRC | 15:23 | |
*** mkrai has joined #openstack-nova | 15:23 | |
gibi | I'm +2 on the fairly simple libvirt metadata feature https://review.opendev.org/c/openstack/nova/+/750552 | 15:23 |
*** pmannidi has joined #openstack-nova | 15:24 | |
gibi | so if some core has time then it is an easy win | 15:24 |
openstackgerrit | Sylvain Bauza proposed openstack/nova master: Add network and utils methods for getting routed networks and segments https://review.opendev.org/c/openstack/nova/+/773976 | 15:26 |
openstackgerrit | Sylvain Bauza proposed openstack/nova master: WIP: Add a routed networks scheduler pre-filter https://review.opendev.org/c/openstack/nova/+/749068 | 15:26 |
*** ociuhandu has quit IRC | 15:27 | |
bauzas | gibi: lemme look | 15:27 |
bauzas | gibi: btw. thanks for continuing to review the routed networks series | 15:28 |
gibi | bauzas: thanks | 15:28 |
bauzas | fwiw, I'm pretty done, just the last change needs to be having UTs and docs | 15:28 |
gibi | bauzas: ack, I will continue looking at it, actaully the self -1 made me stop so it is good that you stated now that it is basically ready | 15:33 |
*** LinPeiWen has quit IRC | 15:33 | |
bauzas | gibi: yeah I needed to add UTs | 15:33 |
bauzas | now it's done | 15:33 |
bauzas | those are easy peasy | 15:33 |
*** derekh has joined #openstack-nova | 15:36 | |
gibi | :) | 15:37 |
*** tbachman has joined #openstack-nova | 15:37 | |
bauzas | gibi: concerns with reliability of the guest metadata information in https://review.opendev.org/c/openstack/nova/+/750552 | 15:44 |
sean-k-mooney | bauzas: reliablity? | 15:45 |
sean-k-mooney | this is an internal debug info | 15:46 |
sean-k-mooney | so if its a little out of sync i think its ok | 15:46 |
bauzas | sean-k-mooney: well, if so, we don't need it | 15:47 |
bauzas | operators could get their infos by other means, right? | 15:47 |
sean-k-mooney | we dont need it but it does make debuging from logs simpler | 15:47 |
sean-k-mooney | they could but this would be useful for us reading sosreports | 15:48 |
bauzas | sean-k-mooney: right, but then we need it to be reliable | 15:48 |
sean-k-mooney | where we cant | 15:48 |
sean-k-mooney | ya | 15:48 |
sean-k-mooney | well | 15:48 |
sean-k-mooney | it would be preferable | 15:48 |
bauzas | sean-k-mooney: I personnally voted on the spec because I do agree with the usecase | 15:48 |
bauzas | but if we go down the road, we need this information to be correct | 15:48 |
sean-k-mooney | i have not read your concern in context in the review | 15:48 |
sean-k-mooney | you belive there is a race in the code ? | 15:49 |
bauzas | right, when detaching | 15:49 |
sean-k-mooney | i see | 15:49 |
sean-k-mooney | if that can be fixed then i agree it shoudl be. | 15:49 |
bauzas | the proposer wrote to delete the info without waiting the neutron event | 15:49 |
*** mkrai_ has joined #openstack-nova | 15:49 | |
bauzas | which could fail | 15:49 |
bauzas | and for most of the cases where operators would want to see the IPs, those would be for networking debugging | 15:50 |
sean-k-mooney | which neutron event? network-vif-unplugged? | 15:50 |
bauzas | yeah | 15:50 |
sean-k-mooney | we dont need to wait for that | 15:50 |
bauzas | sean-k-mooney: see the patch https://review.opendev.org/c/openstack/nova/+/750552 | 15:50 |
sean-k-mooney | we can but we dont need too. | 15:50 |
sean-k-mooney | once we detach it form libvirt its detacted form the vm | 15:51 |
sean-k-mooney | what could fail is removing the device owner(vm uuid) form the port | 15:51 |
*** ociuhandu has joined #openstack-nova | 15:52 | |
*** mkrai has quit IRC | 15:52 | |
bauzas | sean-k-mooney: sean-k-mooney: but then the IP would still be assigned to the instance, right? | 15:53 |
sean-k-mooney | this is the only place we use network-vif-unplugged i belvie https://opendev.org/openstack/nova/src/branch/master/nova/compute/manager.py#L10079-L10083 | 15:53 |
sean-k-mooney | bauzas: the ip is assigned to the port | 15:54 |
sean-k-mooney | if the port is not attached to the vm anymroe then even if nueton still thinks the port has teh ip packet wont get to the vm | 15:54 |
bauzas | sean-k-mooney: the comment is confusing here https://review.opendev.org/c/openstack/nova/+/750552/8/nova/virt/libvirt/driver.py#2329 | 15:54 |
bauzas | we have some internal object that awaits a neutron callback | 15:55 |
gibi | bauzas: ack, I will check | 15:55 |
*** ociuhandu has quit IRC | 15:55 | |
sean-k-mooney | bauzas: the network info cache wont be update until neutron sees the port is removed | 15:55 |
gibi | but nova meeting starts in 4 minutes on #openstack-meeting-3 | 15:55 |
sean-k-mooney | i belive that is what it is refering too | 15:55 |
sean-k-mooney | the filter however will remove it from the network info when generating the metadata | 15:56 |
sean-k-mooney | network_info = list(filter(lambda info: info['id'] != vif['id'], | 15:56 |
sean-k-mooney | instance.get_network_info())) | 15:56 |
sean-k-mooney | so regardless of if neutron has sent the event or not to cause the info cache to be refreshed the copy we pass to generate the data has it removed | 15:56 |
*** ociuhandu has joined #openstack-nova | 15:57 | |
bauzas | sean-k-mooney: my concern is not the fact it filters | 15:58 |
bauzas | he wrote the filter for a good reason | 15:58 |
bauzas | my concern is that we remove this information from the metadate while we could still need it | 15:59 |
bauzas | actually, the question is more, who is the source of truth ? nova or neutron ? | 15:59 |
bauzas | the IP address is bound to a port, which itself is attached to an instance | 16:00 |
bauzas | what if the detach event fails in the meantime ? | 16:00 |
sean-k-mooney | we remove it after libvirt has finished detaching the interface so why would we need it | 16:00 |
gibi | bauzas: if this info is in the domain xml then I would say that what matters is what the VM sees. so if the vif was removed from the VM then we can remove the metadata too | 16:01 |
bauzas | gibi: in this case, I could understand this | 16:01 |
sean-k-mooney | the sequencing is we remove the interface form the domain | 16:01 |
sean-k-mooney | then we unplug the vif form the backend | 16:02 |
sean-k-mooney | then we remove it form the metadata | 16:02 |
gibi | that sequence is OK to me | 16:02 |
sean-k-mooney | then after that i belive the compute manger update the neutron port and remvoed the device owner | 16:02 |
*** macz_ has joined #openstack-nova | 16:03 | |
*** macz_ has quit IRC | 16:03 | |
sean-k-mooney | by the way we cannot unconditionally wait for network-vif-unplugged here as not all backend will send it if im not mistaken | 16:04 |
sean-k-mooney | ml2/ovs will | 16:04 |
sean-k-mooney | after we do self.vif_driver.unplug(instance, vif) | 16:04 |
sean-k-mooney | but i doint think al backend will | 16:04 |
sean-k-mooney | yep https://github.com/openstack/nova/blob/788035add9b32fa841389d906a0e307c231456ba/nova/compute/manager.py#L7779-L7794 | 16:07 |
sean-k-mooney | we tell the dirver to detach which is what is being modified | 16:07 |
sean-k-mooney | and then if we dont raise an exceptio we do _deallocate_port_for_instance | 16:07 |
sean-k-mooney | that is what does the neutron port update/delete https://github.com/openstack/nova/blob/788035add9b32fa841389d906a0e307c231456ba/nova/network/neutron.py#L1710-L1714 | 16:09 |
*** mlavalle has quit IRC | 16:09 | |
sean-k-mooney | bauzas: hopefully ^ that makes sense | 16:11 |
bauzas | sean-k-mooney: on the nova meeting, catching up | 16:11 |
*** artom has quit IRC | 16:12 | |
bauzas | sean-k-mooney: well, gibi's point sounds reasonable to me | 16:13 |
bauzas | from a VM perspective, the nic is detached | 16:14 |
sean-k-mooney | yep before we ever touch the xml to update the metadta | 16:14 |
sean-k-mooney | so its consitent with novas/libvirt view | 16:14 |
bauzas | ok, so I'll comment but I'll leave my -1 for other nits | 16:14 |
sean-k-mooney | cool | 16:14 |
*** mlavalle has joined #openstack-nova | 16:15 | |
bauzas | humm, eavesdrop is lagging 15 mins behind, can't just provide a link yet | 16:16 |
*** hoonetorg has joined #openstack-nova | 16:20 | |
sean-k-mooney | ya it can | 16:25 |
sean-k-mooney | bauzas: its up to date now | 16:26 |
bauzas | yup, commented 5 mins before | 16:26 |
sean-k-mooney | so you did | 16:26 |
bauzas | it just updated straight while it was lagging | 16:26 |
bauzas | I guess there are crons behind eavesdrop | 16:27 |
bauzas | unless it's event-based, which would surprise me | 16:27 |
sean-k-mooney | i think its a chron/periodic sync ya | 16:27 |
sean-k-mooney | its rare that it get more then a few minutes out of date | 16:28 |
*** dklyle has joined #openstack-nova | 16:39 | |
*** pmannidi has quit IRC | 16:47 | |
*** pmannidi has joined #openstack-nova | 16:51 | |
*** mkrai_ has quit IRC | 16:59 | |
*** ociuhandu has quit IRC | 17:05 | |
lyarwood | stephenfin: mind if I address a nit in https://review.opendev.org/c/openstack/nova/+/751367/2 and rebase the series for you? | 17:06 |
*** ociuhandu has joined #openstack-nova | 17:06 | |
stephenfin | lyarwood: for sure | 17:06 |
stephenfin | go for it | 17:06 |
stephenfin | I missed the AR | 17:07 |
lyarwood | stephenfin: np and apologies for missing this series in train for so long | 17:08 |
*** ociuhandu has quit IRC | 17:12 | |
openstackgerrit | Lee Yarwood proposed openstack/nova stable/train: Only allow one scheduler service in tests https://review.opendev.org/c/openstack/nova/+/751362 | 17:14 |
openstackgerrit | Lee Yarwood proposed openstack/nova stable/train: func tests: move _run_periodics() into base class https://review.opendev.org/c/openstack/nova/+/751363 | 17:14 |
openstackgerrit | Lee Yarwood proposed openstack/nova stable/train: Helper to start computes with different HostInfos https://review.opendev.org/c/openstack/nova/+/751364 | 17:14 |
openstackgerrit | Lee Yarwood proposed openstack/nova stable/train: tests: Add reproducer for bug #1879878 https://review.opendev.org/c/openstack/nova/+/751365 | 17:14 |
openstackgerrit | Lee Yarwood proposed openstack/nova stable/train: Add generic reproducer for bug #1879878 https://review.opendev.org/c/openstack/nova/+/751366 | 17:14 |
openstackgerrit | Lee Yarwood proposed openstack/nova stable/train: Don't unset Instance.old_flavor, new_flavor until necessary https://review.opendev.org/c/openstack/nova/+/751367 | 17:14 |
openstack | bug 1879878 in OpenStack Compute (nova) train "VM become Error after confirming resize with Error info CPUUnpinningInvalid on source node " [Undecided,In progress] https://launchpad.net/bugs/1879878 - Assigned to Stephen Finucane (stephenfinucane) | 17:14 |
openstackgerrit | Lee Yarwood proposed openstack/nova stable/train: Move confirm resize under semaphore https://review.opendev.org/c/openstack/nova/+/751368 | 17:14 |
openstackgerrit | Lee Yarwood proposed openstack/nova stable/train: Move revert resize under semaphore https://review.opendev.org/c/openstack/nova/+/751369 | 17:14 |
*** tesseract has quit IRC | 17:15 | |
*** ociuhandu has joined #openstack-nova | 17:19 | |
*** openstack has joined #openstack-nova | 17:23 | |
*** ChanServ sets mode: +o openstack | 17:23 | |
*** ociuhandu has quit IRC | 17:26 | |
*** ociuhandu has joined #openstack-nova | 17:32 | |
*** vishalmanchanda has quit IRC | 17:38 | |
*** zzzeek has quit IRC | 17:42 | |
*** artom has joined #openstack-nova | 17:43 | |
*** zzzeek has joined #openstack-nova | 17:44 | |
*** ociuhandu has quit IRC | 17:45 | |
*** ociuhandu has joined #openstack-nova | 17:47 | |
*** zul has quit IRC | 17:49 | |
*** ociuhandu has quit IRC | 17:49 | |
*** ociuhandu has joined #openstack-nova | 17:49 | |
*** ociuhandu_ has joined #openstack-nova | 17:53 | |
*** ociuhandu has quit IRC | 17:57 | |
*** ociuhandu_ has quit IRC | 17:58 | |
*** pmannidi has quit IRC | 17:59 | |
*** derekh has quit IRC | 18:00 | |
*** pmannidi has joined #openstack-nova | 18:02 | |
*** rpittau is now known as rpittau|afk | 18:10 | |
sean-k-mooney | dansmith: regarding ci resouces. one thing that i have tought about from time to time was spliting check into check and fast check with check dependt on fast check | 18:11 |
sean-k-mooney | now we said we dont want to make them dependent because we want all the result at once | 18:11 |
*** hemna has quit IRC | 18:11 | |
*** songwenping_ has joined #openstack-nova | 18:12 | |
sean-k-mooney | but if we had two pipelines liek that we might be abel to run only the logner jobs if the patch did not have -w | 18:12 |
sean-k-mooney | im not sure if that would save much resouces | 18:12 |
sean-k-mooney | but sometimes im torn between pushing code to gerrit to have a backup or make shareing ti between multiple servers simpler | 18:12 |
*** dtantsur is now known as dtantsur|afk | 18:12 | |
sean-k-mooney | and wasting gate resouces | 18:12 |
sean-k-mooney | if we had a way to say dont run the jobs yet that might help with early verions | 18:13 |
*** hemna has joined #openstack-nova | 18:13 | |
*** songwenping__ has quit IRC | 18:14 | |
*** hamalq has joined #openstack-nova | 18:14 | |
sean-k-mooney | in the grand scheme of things its proably not going to be large but it might be worth exploring having a ready-for-ci lable or something | 18:14 |
sean-k-mooney | i did that for my third party ci https://github.com/SeanMooney/ci-sean-mooney/blob/main/zuul.d/pipelines.yaml#L53-L56 | 18:16 |
sean-k-mooney | and its how the intel nfv ci used to run to save capsity | 18:16 |
sean-k-mooney | althogh it was not night and day or anything | 18:16 |
*** k_mouza has quit IRC | 18:17 | |
*** ralonsoh has quit IRC | 18:18 | |
*** hoonetorg has quit IRC | 18:24 | |
*** tesseract has joined #openstack-nova | 18:25 | |
dansmith | sean-k-mooney: yeah I'm not sure if that's really doable, but I would love to get *some* results before others, that would help a lot | 18:29 |
sean-k-mooney | what i suggested before was ll the non tempest ones first then the rest | 18:29 |
sean-k-mooney | granted you can run those simpley locally | 18:29 |
sean-k-mooney | its too bad zuul cant report back as each job finishes but when i realy want that i do go to zuul.openstack.org | 18:31 |
sean-k-mooney | and just get teh results from there instead of waiting | 18:31 |
sean-k-mooney | the results are avaiable in zuul once the indivuatl job finsihes just not in gerrit | 18:32 |
*** lyarwood has quit IRC | 18:34 | |
dansmith | sean-k-mooney: yeah I'd like to have one tempest job and the easy ones in the first go I think | 18:46 |
dansmith | worker counts being lower would make that still go faster I think | 18:47 |
sean-k-mooney | yep you could do that | 18:47 |
sean-k-mooney | choose one of the faster ones | 18:47 |
dansmith | but I think zuul lacks some persistence required to split up the job and still know when it can gate | 18:47 |
dansmith | so not sure that's really an option | 18:47 |
sean-k-mooney | im not sure about that | 18:47 |
sean-k-mooney | it curerntly does it based on lables | 18:47 |
dansmith | using something like experimental and requiring a +1 experimental run before gate would be a hack around that maybe | 18:48 |
dansmith | sean-k-mooney: well, talk to the infra folks, but my understanding is it's hard | 18:48 |
sean-k-mooney | we woudl jsut need a requires claus in the gate piple to look for verifed and fast-verfied +1 from zuul | 18:48 |
dansmith | in addition to solving this by dividing up the problem or saying "zuul should have a feature" I think there is a LOT of work we all can do to make things faster, duplicate less, and be more targetd | 18:49 |
sean-k-mooney | currently its looking for just verifed and workflow https://github.com/openstack/project-config/blob/master/zuul.d/pipelines.yaml#L80-L82 | 18:49 |
sean-k-mooney | but you could add a 3rd labple to that e.g. fast-verifed and still requrie a +1 form both form zuul | 18:50 |
dansmith | okay but without a prioritization, we'd still be hours and hours before running | 18:50 |
sean-k-mooney | dansmith: it would require use to update the gerrit config and add the feature however | 18:50 |
dansmith | I have stuff that has been in the check queue for three hours and it hasn't started to run a single thing | 18:51 |
sean-k-mooney | we have precedence: normal | 18:51 |
sean-k-mooney | for prioritisation beteween pipliens | 18:51 |
dansmith | I've already talked to infra about this, | 18:51 |
dansmith | and the other precedences are used for things | 18:51 |
dansmith | check and experimental are the same even | 18:51 |
sean-k-mooney | yep they are | 18:51 |
sean-k-mooney | yep both low | 18:51 |
sean-k-mooney | althoguh experimtal will report back first | 18:52 |
dansmith | again, I think we can do a lot without making this an infra problem | 18:52 |
sean-k-mooney | they have teh same precidence but are in differnet queues | 18:52 |
dansmith | and just making it so we get fast check in two hours and slow check in 24 hours isn't really going to help | 18:52 |
sean-k-mooney | dansmith: oh ya i know | 18:52 |
sean-k-mooney | its more if we run out of room with your current effort | 18:52 |
sean-k-mooney | there are other things we can do with infra | 18:52 |
sean-k-mooney | but its more involed | 18:52 |
sean-k-mooney | im not suggesting we start with infra changes just pointing out we can do things via infra changes if its still a proablem | 18:53 |
dansmith | there's lots we could ask infra to do, but relative to the staffing of the top five projects, I mean.. :) | 18:53 |
sean-k-mooney | the other thing too is you were just looking at 1st party ci | 18:54 |
dansmith | I get the impression the things we "could do with infra" require a very wide-scope of potential considerations, more than we can just do in our job defs, and likely would need zuul changes | 18:54 |
sean-k-mooney | dansmith: yes the infra chagnes are openstack wide | 18:54 |
sean-k-mooney | requireign both gerrit and zull configuration chagnes | 18:55 |
sean-k-mooney | so very big/wide reaching hammer | 18:55 |
* dansmith nods | 18:55 | |
sean-k-mooney | not running 8 almost identicaly jobs is relitivly local in contrast | 18:55 |
dansmith | yeah | 18:56 |
dansmith | I wish I could help accelerate us not running the two grenades because those are fairly heavy and really duplicative | 18:56 |
sean-k-mooney | well we can stop that in nova | 18:56 |
sean-k-mooney | right now if we want too | 18:57 |
dansmith | I know, but I think we agreed to wait until the ceph and zuulv3 thing was resolved | 18:57 |
dansmith | I already proposed it with -W to wait on that | 18:57 |
*** spatel has quit IRC | 18:57 | |
sean-k-mooney | well mor i ment we can drop integrated-gate-compute template then contol which grenddade jobs run rom the check and gate pipelines | 18:58 |
sean-k-mooney | https://github.com/openstack/nova/blob/master/.zuul.yaml#L421 | 18:58 |
sean-k-mooney | which means we coudl jsut run - nova-grenade-multinode | 18:59 |
dansmith | https://review.opendev.org/c/openstack/tempest/+/771499 | 18:59 |
dansmith | we agreed we would wait to do that until the ceph multinode zuulv3 thing was resolved | 18:59 |
sean-k-mooney | oh i know i just was pointing out we could do it via nova if we had resovled it and not need a tempest patch | 19:00 |
*** lemko7 has quit IRC | 19:00 | |
*** alexe9191 has joined #openstack-nova | 19:00 | |
*** lemko has joined #openstack-nova | 19:00 | |
dansmith | and when we discussed, gmann wanted it changed there ^, but yes, the mechanics aren't hard, it's the agreement required, and in this case, blocking on the zuulv3 conversion | 19:01 |
alexe9191 | good day everyone:) | 19:01 |
sean-k-mooney | alexe9191: o/ | 19:01 |
alexe9191 | I have a question about the retry filter in nova | 19:01 |
alexe9191 | I am wondering where does it get it's spec_obj from ? specefically this piece of code here: | 19:01 |
alexe9191 | def host_passes(self, host_state, spec_obj): | 19:01 |
alexe9191 | """Skip nodes that have already been attempted.""" | 19:01 |
alexe9191 | retry = spec_obj.retry | 19:01 |
sean-k-mooney | alexe9191: it has not been required for quite some time | 19:01 |
sean-k-mooney | alexe9191: its passed in by the filter schduler | 19:02 |
alexe9191 | but where is it stored? memory or db? | 19:02 |
sean-k-mooney | its the request spec | 19:02 |
sean-k-mooney | its builts in the api | 19:03 |
sean-k-mooney | then passed to the conductor and scudler | 19:03 |
sean-k-mooney | i belive we might have it in the api db | 19:03 |
alexe9191 | Interesting, let me check | 19:03 |
dansmith | yes, api_db | 19:03 |
*** tesseract has quit IRC | 19:04 | |
alexe9191 | ok, so if a host fail, it will register it's state here in that table. | 19:04 |
alexe9191 | request_specs | 19:04 |
sean-k-mooney | no | 19:04 |
sean-k-mooney | we dont commit that to the db | 19:04 |
sean-k-mooney | we only track that in memroy i belive | 19:05 |
alexe9191 | that table is quite loaded though. | 19:05 |
sean-k-mooney | the request_sepc is used for other things | 19:05 |
sean-k-mooney | so it is saved in the db | 19:05 |
sean-k-mooney | but wee dont save the failed host in the request spec during schdule and commit that back | 19:06 |
alexe9191 | so that part right here: | 19:06 |
alexe9191 | return self.filter_handler.get_filtered_objects(self.enabled_filters, | 19:06 |
alexe9191 | hosts, spec_obj, index) | 19:06 |
alexe9191 | i see it's passing hosts but the retry filter is kicking out all of the hosts in the aggregate I am trying to schedule in. They are all healthy and they are hosting VMs, they probably had issues at some time. | 19:07 |
alexe9191 | The interesting thing, this happens only with one flavor, other flavors are returning a different results for the retryfilter. | 19:07 |
alexe9191 | I restarted nova-compute on all of those hosts but that did not change the result of the scheduling:) so I am wondering to be honest where does it saves the host state. | 19:08 |
sean-k-mooney | alexe9191: what releast of nova are you using by the way | 19:08 |
alexe9191 | rocky:) | 19:08 |
sean-k-mooney | so you have placment | 19:08 |
alexe9191 | yes | 19:08 |
sean-k-mooney | then you can disable the retry filter entirly | 19:09 |
sean-k-mooney | i belive rocky is the release we stopped using it | 19:09 |
sean-k-mooney | thats what im checking now | 19:09 |
sean-k-mooney | ah it was queens https://github.com/openstack/nova/blob/master/releasenotes/notes/deprecate-retry-filter-4d1dba39a2c21836.yaml | 19:10 |
sean-k-mooney | as part of https://specs.openstack.org/openstack/nova-specs/specs/queens/implemented/return-alternate-hosts.html | 19:10 |
*** spatel has joined #openstack-nova | 19:10 | |
sean-k-mooney | alexe9191: so on rocky you can and should disable the retry filter | 19:10 |
alexe9191 | Ok! that's good to know | 19:11 |
alexe9191 | Is there a way to mitigate the effect of the retry filter right now? restarting the nova scheduler right now is probably something that's gonna cause a lot of grief | 19:12 |
alexe9191 | We have about 800~ hosts | 19:12 |
alexe9191 | 9 schedulers | 19:12 |
sean-k-mooney | really why so manny? | 19:12 |
alexe9191 | we're thinking about cells but this is in the future plans | 19:12 |
alexe9191 | it's a big infrastructure | 19:13 |
alexe9191 | that's why I was wondering if I can empty that spec_obj from a cache/db table | 19:13 |
sean-k-mooney | still scduling is typeiclaly not the largets part of a but | 19:13 |
*** belmoreira has quit IRC | 19:13 | |
sean-k-mooney | infacti its typeiclly quite a small amount | 19:13 |
sean-k-mooney | 9 schdulers is quite a lot | 19:13 |
sean-k-mooney | alexe9191: no the previsouly tired host are only updated in memory | 19:14 |
alexe9191 | in the nova-scheduler I am guessing | 19:14 |
sean-k-mooney | yep | 19:15 |
openstackgerrit | Merged openstack/nova master: Fix invalid argument formatting in exception messages https://review.opendev.org/c/openstack/nova/+/763511 | 19:15 |
sean-k-mooney | they only are tracked during the singel scheduling request | 19:15 |
alexe9191 | one more question, is that spec_obj also tied to the flavor ? | 19:15 |
sean-k-mooney | kind of | 19:15 |
dansmith | sean-k-mooney: should the retry filter even be used anymore? don't we pass selected_hosts down and have the conductor just iterate over them and then declare it dead? | 19:16 |
sean-k-mooney | dansmith: we do from queens | 19:16 |
sean-k-mooney | dansmith: thats whyi said to remove it | 19:16 |
dansmith | oh I see you said that above | 19:16 |
sean-k-mooney | dansmith: also i delete it on master a few release ago | 19:16 |
sean-k-mooney | alexe9191: this is the request_spec https://github.com/openstack/nova/blob/stable/rocky/nova/objects/request_spec.py#L50-L87 | 19:17 |
sean-k-mooney | it has the flavor and image avaiabel which the filters can use | 19:17 |
alexe9191 | :) So the flavor is a part of it | 19:17 |
alexe9191 | a combination of things then | 19:17 |
sean-k-mooney | its modeling all the requirements for scduling an instnace more or less | 19:18 |
alexe9191 | that's the reason why it's giving different results for different flavors | 19:18 |
sean-k-mooney | yes | 19:18 |
alexe9191 | :) Many thanks! I am gonna go and schedule a remove of the retry filter since it's not needed | 19:19 |
sean-k-mooney | although on the first iteration thorugh the scudler it should do nothing | 19:19 |
alexe9191 | it's doing that on the retryfilter | 19:19 |
sean-k-mooney | https://github.com/openstack/nova/blob/stable/rocky/nova/scheduler/filters/retry_filter.py#L43 | 19:20 |
*** andrewbonney has quit IRC | 19:20 | |
sean-k-mooney | retry.hosts shoudl be empty | 19:20 |
*** spatel has quit IRC | 19:20 | |
alexe9191 | I am not sure I understand why? | 19:21 |
alexe9191 | https://github.com/openstack/nova/blob/stable/rocky/nova/scheduler/filters/retry_filter.py#L34 | 19:21 |
sean-k-mooney | well spec_obj.retry will not be set | 19:22 |
sean-k-mooney | so it will return true on line 36 | 19:22 |
alexe9191 | I am not quite sure how am I ending up with this number then: | 19:22 |
alexe9191 | 2021-02-04 19:16:58.981 112479 DEBUG nova.filters [req-7daf5214-19ca-48bc-9240-7db0be15c304 03169685e1924a6fa4eee2da46335331 a0b22ddc828140beaf13bc3daaba4a93 - default default] Starting with 148 host(s) get_filtered_objects /usr/lib/python2.7/site-packages/nova/filters.py:70 | 19:22 |
alexe9191 | 2021-02-04 19:16:58.982 112479 DEBUG nova.filters [req-7daf5214-19ca-48bc-9240-7db0be15c304 03169685e1924a6fa4eee2da46335331 a0b22ddc828140beaf13bc3daaba4a93 - default default] Filter RetryFilter returned 148 host(s) get_filtered_objects /usr/lib/python2.7/site-packages/nova/filters.py:104 | 19:22 |
alexe9191 | For that specefic flavor, though other flavors are returning different numbers | 19:23 |
sean-k-mooney | well in that case the RetryFilter return 148 hosts | 19:23 |
sean-k-mooney | and it was given 148 hosts | 19:23 |
sean-k-mooney | so they all were pased | 19:23 |
alexe9191 | where was it given 148 hosts from ? | 19:24 |
sean-k-mooney | from placment | 19:24 |
alexe9191 | does placement register the failed hosts? | 19:24 |
sean-k-mooney | you started with 148 hosts form placment | 19:24 |
alexe9191 | ow | 19:24 |
sean-k-mooney | no | 19:24 |
sean-k-mooney | placment basically says ( based on our request) here are the set of host that could fit the vm | 19:25 |
sean-k-mooney | then the filter refine that | 19:25 |
*** belmoreira has joined #openstack-nova | 19:25 | |
sean-k-mooney | placemtn in rocky is basically only lookign at ram disk and cpus | 19:25 |
openstackgerrit | Artom Lifshitz proposed openstack/nova master: WIP: libvirt: start tracking NUMACell.socket for hosts https://review.opendev.org/c/openstack/nova/+/766816 | 19:26 |
openstackgerrit | Artom Lifshitz proposed openstack/nova master: WIP: extra specs/image pros: add `socket` PCI NUMA affinity https://review.opendev.org/c/openstack/nova/+/772748 | 19:26 |
openstackgerrit | Artom Lifshitz proposed openstack/nova master: WIP: Add `socket` PCI NUMA affinity policy request prefilter https://review.opendev.org/c/openstack/nova/+/772749 | 19:26 |
openstackgerrit | Artom Lifshitz proposed openstack/nova master: WIP: pci: implement the `socket` NUMA affinity policy https://review.opendev.org/c/openstack/nova/+/772779 | 19:26 |
openstackgerrit | Artom Lifshitz proposed openstack/nova master: WIP: Track host NUMA topology in PCI manager https://review.opendev.org/c/openstack/nova/+/774149 | 19:26 |
alexe9191 | not the availability zone or such then ? | 19:26 |
sean-k-mooney | it does a bit more but basically of your 800 comptue nodes it said here are the 148 that could fit your vm | 19:26 |
sean-k-mooney | not in rocky by deault at least | 19:26 |
sean-k-mooney | we did add the az later | 19:26 |
alexe9191 | ok... but the interesting thing then is, when i schedule the virtual machine directly on the host it works just fine | 19:26 |
alexe9191 | so from resources point of view there are plenty | 19:27 |
sean-k-mooney | actully it can do the az in rocky https://github.com/openstack/nova/blob/stable/rocky/nova/scheduler/request_filter.py#L63 | 19:27 |
sean-k-mooney | but i think that is off by default and you use the az filter | 19:27 |
alexe9191 | I am using that | 19:28 |
alexe9191 | but I end up with hosts that are not usable for that specefic az | 19:28 |
alexe9191 | they are all from zone 2,3,4 for instance and the one i want is zone1 | 19:28 |
alexe9191 | BUt this is happening only on this flavor. | 19:29 |
sean-k-mooney | the flavor wont change teh az interaction | 19:29 |
alexe9191 | indeed | 19:29 |
alexe9191 | if I drop the az I get more hosts to start with on the retry filter though. | 19:29 |
*** artom has quit IRC | 19:30 | |
sean-k-mooney | yes so that is placment limiting the hosts | 19:30 |
sean-k-mooney | to only those in the requeted az | 19:30 |
sean-k-mooney | its likely that one of the later fiters is failing | 19:30 |
*** spatel has joined #openstack-nova | 19:30 | |
sean-k-mooney | can you paste the fule filter logs for the spwan to http://paste.openstack.org/ | 19:31 |
alexe9191 | actually what I said was just wrong.. i end up with the same number of hosts 148 if I drop the az, the scheduling happen though cause az filter is not filtering anything out | 19:31 |
alexe9191 | yes one moment let me sanitise it | 19:31 |
sean-k-mooney | ya if you have enable the plamcent version you can also disable the az filter | 19:32 |
alexe9191 | indeed that will also be done since it can be used | 19:32 |
alexe9191 | so here is the version that works: | 19:32 |
alexe9191 | http://paste.openstack.org/show/802341/ | 19:32 |
sean-k-mooney | over the release we have slowly been moving things too placment where it makes sense | 19:32 |
sean-k-mooney | yep so you went form 148 down to 104 | 19:33 |
sean-k-mooney | then those would get weighed | 19:33 |
alexe9191 | http://paste.openstack.org/show/802342/ this is the one that does not (COmpute is reporting 0 cause those are disabled) | 19:33 |
alexe9191 | Filter AvailabilityZoneFilter returned 8 because 8 are only in zone1 | 19:34 |
alexe9191 | so I am starting with less than I should | 19:34 |
*** belmoreira has quit IRC | 19:34 | |
sean-k-mooney | so the first following filters RetryFilter AvailabilityZoneFilter AggregateDiskFilter, AggregateCoreFilter and AggregateRamFilter can be removed | 19:34 |
sean-k-mooney | and NUMATopologyFilter should come last | 19:35 |
sean-k-mooney | so the one that faild had most of the host elimiated by the AvailabilityZoneFilter | 19:36 |
alexe9191 | indeed, because none of the hosts in zone1 reported it self to placement to be a good match for that flavor | 19:36 |
alexe9191 | though there are plenty of space on those hosts to cover the needed resources | 19:36 |
alexe9191 | and I have no max placement in the config | 19:36 |
sean-k-mooney | it sound like you have stale allocation in plamcent then | 19:37 |
sean-k-mooney | although what i dont understand is why the az filter removed any hosts | 19:37 |
alexe9191 | anyway to make sure that this is the case? | 19:37 |
sean-k-mooney | you said you enabeld the plamcent az filtering | 19:37 |
alexe9191 | no I meant on the api request | 19:38 |
sean-k-mooney | oh ok | 19:38 |
alexe9191 | apologies for the confusion:) | 19:38 |
sean-k-mooney | no worries | 19:38 |
sean-k-mooney | am we have a heal allcoation command | 19:38 |
sean-k-mooney | dansmith: do you know if that is in rocky | 19:38 |
dansmith | not off hand | 19:39 |
sean-k-mooney | so we have https://github.com/openstack/nova/blob/7b5ac717bd338be32414ae25f60a4bfe4c94c0f4/nova/cmd/manage.py#L2121 | 19:40 |
sean-k-mooney | ya that is on rocky | 19:40 |
*** rcernin has joined #openstack-nova | 19:40 | |
sean-k-mooney | so you can do nova-manage --heal-allocations i think but before you do that | 19:41 |
sean-k-mooney | alexe9191: you have the aggreate ram,disk,core filters enabled | 19:41 |
alexe9191 | indeed | 19:41 |
sean-k-mooney | alexe9191: do you manage allocation ratios by aggreate | 19:41 |
alexe9191 | more or less yes | 19:42 |
sean-k-mooney | ok that is proably the issue | 19:42 |
sean-k-mooney | http://lists.openstack.org/pipermail/openstack-dev/2018-January/126283.html | 19:42 |
sean-k-mooney | we deprecated those in ocata because once you use placment you can nolonger do that | 19:43 |
sean-k-mooney | alexe9191: you have to set the allocation ratios per host | 19:43 |
alexe9191 | let me check the code I think we have that in place now | 19:43 |
sean-k-mooney | ok here are the docs on that topic if you have not got them set on each compute host | 19:44 |
sean-k-mooney | https://docs.openstack.org/nova/latest/admin/configuration/schedulers.html#allocation-ratios | 19:44 |
*** rcernin has quit IRC | 19:45 | |
alexe9191 | just checked right now and we have that on rocky | 19:45 |
alexe9191 | older versions are a different story but we are getting rid of those | 19:45 |
sean-k-mooney | ok so its in the nova.conf on each of your compute nodees | 19:45 |
alexe9191 | yes | 19:45 |
alexe9191 | docker exec -it nova_compute grep cpu /etc/nova/nova.conf | 19:46 |
alexe9191 | cpu_allocation_ratio=1.0 | 19:46 |
sean-k-mooney | ok then what you basically need to do is compare the avaiable resouces in plamcnet to those reported in the hyperviors api | 19:46 |
sean-k-mooney | if there is a missmatch due to stales allcoation the usage wont be the same in nova view and placments | 19:47 |
alexe9191 | I am actually building now a json file with the resources using openstack hypervisor show | 19:47 |
sean-k-mooney | which would be why plamcent would have elminated the host before it got to the az filter | 19:47 |
sean-k-mooney | unfortunetly im more or less done for the day so i wont be able to help you contiue debuging but my best guess is placemetn and nova are out of sync | 19:49 |
sean-k-mooney | so placment is filtering out the host in that az first. before it gets to the schduler | 19:49 |
alexe9191 | You've already sat me on the good path:) I am going to check this and check the heal command if that is the case | 19:50 |
alexe9191 | one more question though, can I query the resources per node? or is that per class only? | 19:50 |
sean-k-mooney | in placment? | 19:50 |
alexe9191 | yes | 19:50 |
sean-k-mooney | yes you can | 19:50 |
sean-k-mooney | you can use the resouce providers endpoint ot list all the invetores and the usage per host | 19:50 |
sean-k-mooney | there is an osc plugin for that too | 19:51 |
sean-k-mooney | called osc-placement if you dont have it installed | 19:51 |
sean-k-mooney | it will give you some placment command for the openstack client | 19:51 |
alexe9191 | installing now:) thanks alot | 19:52 |
sean-k-mooney | here are the docs https://docs.openstack.org/osc-placement/latest/cli/index.html#resource-provider-inventory-list | 19:53 |
sean-k-mooney | openstack resource provider show [--allocations] <uuid> will also be useful | 19:53 |
sean-k-mooney | or openstack resource provider usage show <uuid> | 19:53 |
alexe9191 | Testing it now :) | 19:54 |
alexe9191 | yup, got the numbers | 19:54 |
*** artom has joined #openstack-nova | 19:57 | |
*** hoonetorg has joined #openstack-nova | 20:01 | |
*** artom has quit IRC | 20:02 | |
*** pmannidi has quit IRC | 20:06 | |
*** pmannidi has joined #openstack-nova | 20:10 | |
alexe9191 | yup, I see differences between what nova is reporting vs what placement is reporting | 20:10 |
*** stand has quit IRC | 20:12 | |
*** hoonetorg has quit IRC | 20:13 | |
*** rcernin has joined #openstack-nova | 20:17 | |
*** k_mouza has joined #openstack-nova | 20:17 | |
*** k_mouza has quit IRC | 20:21 | |
*** k_mouza has joined #openstack-nova | 20:21 | |
*** k_mouza has quit IRC | 20:26 | |
*** hoonetorg has joined #openstack-nova | 20:30 | |
*** rcernin has quit IRC | 20:31 | |
openstackgerrit | Merged openstack/nova master: functional: Add tests for mixed CPU policy https://review.opendev.org/c/openstack/nova/+/755852 | 20:44 |
*** ociuhandu has joined #openstack-nova | 20:45 | |
*** rcernin has joined #openstack-nova | 21:00 | |
openstackgerrit | Merged openstack/nova master: Remove __unicode__() from nova unit test Exception https://review.opendev.org/c/openstack/nova/+/769894 | 21:03 |
*** whoami-rajat__ has quit IRC | 21:10 | |
*** pmannidi has quit IRC | 21:15 | |
*** pmannidi has joined #openstack-nova | 21:17 | |
*** ociuhandu has quit IRC | 21:20 | |
*** ociuhandu has joined #openstack-nova | 21:21 | |
*** ociuhandu has quit IRC | 21:22 | |
*** ociuhandu has joined #openstack-nova | 21:22 | |
*** ociuhandu has quit IRC | 21:29 | |
*** alexe9191 has quit IRC | 21:30 | |
*** nweinber has quit IRC | 21:33 | |
*** ociuhandu has joined #openstack-nova | 21:38 | |
*** rcernin has quit IRC | 21:39 | |
*** rcernin has joined #openstack-nova | 21:56 | |
*** cap has quit IRC | 22:00 | |
*** xek has quit IRC | 22:00 | |
*** ociuhandu has quit IRC | 22:07 | |
*** ociuhandu has joined #openstack-nova | 22:08 | |
*** ociuhandu has quit IRC | 22:15 | |
*** lbragstad_ has joined #openstack-nova | 22:22 | |
*** pmannidi has quit IRC | 22:24 | |
*** lbragstad has quit IRC | 22:24 | |
*** rcernin has quit IRC | 22:26 | |
*** rcernin has joined #openstack-nova | 22:26 | |
*** pmannidi has joined #openstack-nova | 22:27 | |
openstackgerrit | Merged openstack/nova stable/ussuri: Warn when starting services with older than N-1 computes https://review.opendev.org/c/openstack/nova/+/770764 | 22:39 |
openstackgerrit | Merged openstack/nova stable/ussuri: Reproduce bug 1896463 in func env https://review.opendev.org/c/openstack/nova/+/770768 | 22:40 |
openstack | bug 1896463 in OpenStack Compute (nova) ussuri "evacuation failed: Port update failed : Unable to correlate PCI slot " [Low,In progress] https://launchpad.net/bugs/1896463 - Assigned to Balazs Gibizer (balazs-gibizer) | 22:40 |
openstackgerrit | Merged openstack/nova stable/ussuri: [doc]: Fix glance image_metadata link https://review.opendev.org/c/openstack/nova/+/761977 | 22:40 |
*** spatel has quit IRC | 22:45 | |
openstackgerrit | Merged openstack/nova stable/train: Fix a hacking test https://review.opendev.org/c/openstack/nova/+/767793 | 22:55 |
*** lemko has quit IRC | 22:57 | |
*** lemko has joined #openstack-nova | 22:57 | |
*** tkajinam has quit IRC | 22:59 | |
*** tkajinam has joined #openstack-nova | 22:59 | |
*** slaweq has quit IRC | 23:03 | |
*** lbragstad_ is now known as lbragstad | 23:03 | |
*** lbragstad has quit IRC | 23:21 | |
*** lbragstad has joined #openstack-nova | 23:23 | |
*** pmannidi has quit IRC | 23:33 | |
*** pmannidi has joined #openstack-nova | 23:35 | |
*** efried has quit IRC | 23:38 | |
*** efried has joined #openstack-nova | 23:38 |
Generated by irclog2html.py 2.17.2 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!