16:00:49 #startmeeting openstack_ansible_meeting 16:00:49 Meeting started Tue Dec 18 16:00:49 2018 UTC and is due to finish in 60 minutes. The chair is mnaser. Information about MeetBot at http://wiki.debian.org/MeetBot. 16:00:51 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 16:00:52 #topic rollcall 16:00:53 The meeting name has been set to 'openstack_ansible_meeting' 16:00:53 o/ 16:02:12 quiet quiet :) 16:02:12 \o/ 16:02:35 year end is going on :-) 16:03:01 not much in last week highlights, we can discuss open topics in open discussion 16:03:10 would we wait to a quorum or can we discuss placement a bit? 16:03:20 guilhermesp: we can do that in open discussion 16:03:36 #topic bug triage 16:03:40 #link https://bugs.launchpad.net/openstack-ansible/+bug/1808830 16:03:41 Launchpad bug 1808830 in openstack-ansible "container properties were not updated from env.d file" [Undecided,New] 16:03:48 mnaser: do we need a proper blueprint/spec related to os_tempest part? 16:04:05 chandankumar: can we discuss this later in open discussion? but probably not :> 16:04:13 mnaser: sure 16:04:21 hmm 16:04:24 looks like someone is using lvm 16:04:32 for containers 16:04:48 i remember cloudnull mentioning that this swallows data 16:05:31 i dont know enough about this 16:05:53 i'll leave it, hopefully next time we triage we can see what things look like 16:06:08 properties have been documented as part of o_u_config 16:06:14 not as env.d 16:06:27 iirc 16:06:33 oh wait 16:06:33 ok so what they're doing is just wrong? 16:06:40 I meant container_vars 16:06:43 mmm 16:06:52 Jesse Pretorius (odyssey4me) proposed openstack/openstack-ansible stable/queens: Add automated migration of neutron agents to bare metal https://review.openstack.org/625331 16:06:55 so their workarounds is really the right way to do it? 16:07:02 o/ 16:07:09 yeah that's how I'd configure things 16:07:22 container_vars: 16:07:36 in that case we'll put as invalid and mention that the workaround is the right way to go about it 16:08:11 mnaser evrardjp so I helped the guy out yesterday - effectively the inventory was generated before the env.d file change, and if that's done then the env.d change does not take effect 16:08:19 aaaah 16:08:26 so it works but because the inventory already exists 16:08:28 it doesnt update 16:08:33 so we're not really overriding, its just an 'initial' value 16:08:43 odyssey4me: also I notice there is a 1 space difference 16:08:55 as you can see container_fs_size is wrongly indented 16:08:56 I don't know if we consider that as a bug - but it's not obvious, and we don't really do a good job of documenting how people can do things like this after the fact. 16:09:12 evrardjp I suspect that's just launchpad being stupid. 16:09:22 It would be good to verify it independently. 16:09:43 I mean that would be a good reason why it doesn't show up on an empty inventory. 16:10:01 after that our inventory is quite a mess, so things not updating would not be a surprise to me 16:10:02 yah, when I worked through it with him - he had the spacing right 16:10:06 lemme find the eavesdrop 16:10:42 okay so if we wanna quickly triage this, how do we feel like 16:11:50 i would defer to evrardjp or odyssey4me because this isn't my strength :) 16:12:33 if someone really cares about the inventory, (s)he/it should step up, as to my knowledge, everybody is fine with the current way of working, even if it's bugged 16:12:47 the beginning of the very long conversation is here: http://eavesdrop.openstack.org/irclogs/%23openstack-ansible/%23openstack-ansible.2018-12-17.log.html#t2018-12-17T14:23:23 16:12:50 so I'd say low 16:12:53 perhaps good to add that to the bug 16:12:57 but confirmed? 16:13:14 I would not say it's confirmed unless someone else verifies it 16:13:15 i'm not asking anyone to fix it, but just if we recognize that this is an issue or not :) 16:13:28 so triaged/low? 16:13:57 assign to me - I'll try to confirm it, it shouldn't be hard 16:14:01 so leave it new 16:14:31 ok cool 16:14:39 thank you odyssey4me 16:14:58 #link https://bugs.launchpad.net/openstack-ansible/+bug/1808693 16:15:00 Launchpad bug 1808693 in openstack-ansible "haproxy galera backend config contains httpchk statement" [Undecided,New] 16:15:09 I think it might just be a simple doc fix - a known issue somewhere. 16:15:31 on that we have an http middleware 16:15:35 I'd say invalid 16:15:43 i think this is probably invalid, this deployment probably failed to start the xinetd server 16:15:49 indeed 16:15:58 disabling checks is NOT a solution :) 16:16:24 done 16:16:28 #link https://bugs.launchpad.net/openstack-ansible/+bug/1808543 16:16:29 Launchpad bug 1808543 in openstack-ansible "Keystone Federation cannot complete SP node setup on stable/rocky" [Undecided,New] 16:16:46 oh that's redkrieg :) 16:17:11 i see a keystone traceback 16:17:49 could this be a keystone thing? i personally dont know much about federation.. a little bit 16:18:01 yeah, this was piped up in channel too and I asked for a bug report 16:18:28 It's been a while since I've done any federation things - hopefully I can confirm this... I'd say leave new and assign to me 16:18:34 would it be a dependency error following the keystone traceback? 16:18:49 unless someone else wants to have a bash at it 16:19:26 it might just be a new dependency which isn't covered 16:19:55 specifically keystone.auth.saml2? 16:20:43 Maybe we can ask redkrieg to dig in a bit if he has time. He’s done a bunch of python :) 16:20:57 it seems like it might just be that whatever that task is doing, it's not using the venv 16:21:34 Let’s leave it as is and catch up with redkrieg this week 16:22:59 #link https://bugs.launchpad.net/openstack-ansible/+bug/1807400 16:22:59 Launchpad bug 1807400 in openstack-ansible "networksegments table in neutron can not be cleared automatically" [Undecided,New] 16:23:19 Sounds like a neutron bug? 16:23:58 that does seem like it's not registered to the right project 16:24:21 mark invalid, but add neutron 16:25:04 yep done 16:25:09 #link https://bugs.launchpad.net/openstack-ansible/+bug/1807346 16:25:10 Launchpad bug 1807346 in openstack-ansible "[heat] Installations fails during Update Q->R" [Undecided,New] 16:26:02 huh thats weird 16:26:08 sounds like variable changes between Q -> R? 16:26:13 that is breaking the task 16:27:29 its almost liek 16:27:45 its trying to create a user called `stack_domain_admin` but there's already one? 16:27:56 oh wait i didnt scroll down 16:28:26 seems like ansible bugs ? 16:28:34 hmm, ok - I haven't seen this yet, but I guess I could take the time to verify it unless someone else has a little time for it? 16:28:55 can be assigned to me 16:28:59 It seems that I'm falling on the upgrade sword. :p 16:29:10 we have to do another pretty big upgrade soon 16:29:10 Yay! Thanks guilhermesp :) 16:29:36 so i might run into that i guess 16:29:49 assigned to guilherme for further heck 16:29:50 #link https://bugs.launchpad.net/openstack-ansible/+bug/1807074 16:29:51 Launchpad bug 1807074 in openstack-ansible "Neutron OVS bridge_mappings not built on computes" [Undecided,New] 16:30:53 jamesdenton: did you want to grab this? 16:31:15 jamesdenton: you know i feel this relates to that fix i did a while back 16:32:03 https://review.openstack.org/#/c/614282/ 16:32:53 seems legit, and that's in rocky - perhaps note that in the bug, but leave it new to give jamesdenton some time to look again 16:33:00 I think he's out today. 16:33:18 okay cool, ill leave it as new so we can see it again next week 16:33:25 I don't think that patch is in a released OSA - or it's only in the most recent release. 16:33:36 yeah but that patch needs another follow up one 16:33:59 #link https://bugs.launchpad.net/openstack-ansible/+bug/1806938 16:34:01 Launchpad bug 1806938 in openstack-ansible "RabbitMQ Upgrade does not work during upgrade from Q->R" [Undecided,New] 16:34:22 same guy from heat one 16:34:22 ah, a patch for that merged recently 16:34:41 yeah, I think mnaser has already fixed it 16:34:48 yep. 16:34:51 https://review.openstack.org/625200 16:34:55 that was a very expensive bug ill let you know that. 16:34:56 ;[ 16:35:14 sorry mnaser :/ 16:35:21 its okay 16:35:28 thats why i wanna try and help with some of hte ugprade stuff soon 16:36:07 gah 16:36:10 launchpad wont let me comment 16:36:39 https://media.giphy.com/media/b3hWtl6x3H3Ko/giphy.gif 16:37:05 getting a wall of red text 16:37:32 aunchpad requires a REFERER header to perform this action. There is no REFERER header present. This can be caused by configuring your browser to block REFERER headers. 16:37:38 i think this happened in summit when we were registering a bug or something on a session 16:37:52 I can't seem to find my favourite cat presses button and kaboom gif - evrardjp knows the one I mean, 16:37:53 can someoen else comment that was fixed in https://review.openstack.org/#/c/625200/ 16:38:01 lemme try mnaser 16:38:15 and set status to fix released 16:38:18 odyssey4me: I don't have it here 16:38:24 but yeah. 16:39:37 thanks guilhermesp 16:39:39 #link https://bugs.launchpad.net/openstack-ansible/+bug/1806696 16:39:40 Launchpad bug 1806696 in openstack-ansible "Galera lxc shutdown results in corrupted db" [Undecided,New] 16:40:03 oh no 16:40:28 FYI Christian Zunker and his team at Codecentric have done a wonderful job of reporting decent bugs and submitting fixes. 16:40:39 +++ 16:40:42 are they on irc 16:40:42 He was at Berlin IIRC. 16:40:47 I don't think so. 16:40:58 yeah i saw some cool patches 16:41:36 heh, 5 secs to kill a container is a bit harsh for galera - that sounds legit 16:41:54 It's probably better to make that a pretty long wait time - maybe 5 mins or something. 16:41:55 but is it the case for nspawn? 16:41:56 ive always lxc-stop'd things 16:43:11 set to confirmed 16:43:30 it'd be nice to get something sorted for that asap 16:43:42 hi guys, I'm working with Justin who reported this bug, if there's any additional info needed 16:43:44 it sounds like a bad bug indeed 16:43:55 searching SHUTDOWNDELAY in codesearch shows nothing 16:43:58 so maybe this isnt something we deploy? 16:44:18 mnaser no, it's likely a default - but we can probably implement something to change it 16:44:34 yep ok 16:44:40 francois great bug, and it'd be nice to figure out a lxc_hosts role patch to increase that timeout 16:44:43 at least make it match with the value used in timeout for galera 16:44:57 evrardjp yeah, I think that's 180 secs 16:45:02 I'd go for 5 mins myself. 16:45:07 francois: do you think you have someone who can help submit a patch for this? i can help them land it 16:45:10 i just dont wanna land it myself 16:45:11 I remember that was longer than 5 minutes in old times 16:45:23 remember we searched with mancdaz something about 3600s timeout? 16:45:25 :) 16:45:45 that's too much indeed 16:45:52 yeah definitely if there's an idea and the best way to fix it 16:45:52 oh man, that's like back in the day when the dinosaurs were still roaming 16:45:53 but 5 minutes wouldn't be that bad 16:46:07 odyssey4me: hahaha 16:47:24 ok lets sync up after the meeting 16:47:30 agreed that 5 minutes sounds much saner then 5 secondes, it looks like a pretty harsh default 16:47:37 id like to go to open discussion cause we took a bit a lot of time 16:47:40 #topic open discussion 16:47:44 I can't really help on finding out where to start for this 16:47:47 anyone has any things? 16:47:55 I do about placement 16:48:04 can someone give a feedback https://review.openstack.org/#/c/618820/ ? 16:48:09 we have greenfiled soon but 16:48:15 uwsgi app is not running yet 16:48:18 I might find a little time to look into it evrardjp mnaser francois - upgrade tests take quite a bit of time. ;) 16:48:41 guilhermesp: it looks like something is not wired up properly 16:48:44 in that last patch I added two tasks to configure uwsgi but seems that the ini file is not being placed under /etc/uwsgi 16:48:58 do you have a vm with that role in there? maybe we can dig in and see it there later? 16:49:14 yep I've been doing this this morning 16:49:26 I intend to continue dig in 16:49:44 and about the tests, placement doesnt have its own tempest tests 16:49:53 that's really f'n annoying that they dont 16:49:55 guys from the channel said to reuse compute tests 16:50:12 yep that's bad 16:50:19 well thats not easy because it's going to conflict with the placement thats alreadty deployed 16:50:26 I didn't ask the reasons they have but I can try to figure out, anyways... 16:50:27 unless we add a patch to remove placement and we have this circular dependency 16:50:28 ugh 16:50:37 indeed 16:50:46 lets get it to run and validate it locally and merge it that way.. i dunno, icant think of any other way. 16:50:53 I will need to test by hand as soon as I finish my PR so the code can be merged 16:51:17 for now, don't bother with any real tests - do something simple like validate the service is running and be done with it 16:51:18 we are up to 51 versions in my PR :P 16:51:26 odyssey4me: +1 16:51:31 once we change os_nova to use it, then we can do something 16:51:42 odyssey4me: makes sense for now 16:51:57 guilhermesp yeah, for now I'm more inclined to get that patch in even if it's not fully functional - it's getting too big 16:52:00 heads up, new ansible-lint released today, if you start seeing new linting failures 16:52:07 https://docs.ansible.com/ansible-lint/rules/default_rules.html 16:52:08 I believe today or within the next days you guys can review and merge hopefully 16:52:11 should help with what is new 16:52:13 pabelanger we pin ansible-lint :) 16:52:13 chandankumar: anything at yor side? 16:52:34 on that topic, we're long overdue for an asible-lint pin update 16:52:59 mnaser: I got the answer regarding the spec/blueprint question 16:53:21 odyssey4me: nice, I might start doing that as the new linters are some what aggressive IMO 16:53:42 mnaser: we are sending os_tempest calloboration update on openstack-discuss list , anything we can improve in that report? 16:54:02 chandankumar: i *love* it 16:54:19 im going to try and do smething similar inside OSA side, i really really like those 16:54:44 is there some plan to upgrade ansible in stable/rocky, as it seems to fix issues like https://bugs.launchpad.net/openstack-ansible/+bug/1803382 16:54:45 Launchpad bug 1803382 in openstack-ansible "unbound-client.yml failure depending on TERMCAP envvar content" [High,Confirmed] - Assigned to Francois Deppierraz (francois-ctrlaltdel) 16:54:49 pabelanger I dunno if there's a default job for ansible-lint, but it'd be nice if it supported pinning - we could then use it, possibly 16:55:18 francois: nice launchpad id 16:55:27 odyssey4me: not sure either, I still manage it via tox-linters job and tox.ini 16:55:37 mnaser: glad to know that you guys liked that , last update of the year is tomorrow, then, we can do send resume in new year 16:55:43 evrardjp: :) 16:55:57 chandankumar: wonderful, its been a pleasure working together 16:56:10 mnaser: :-) 16:56:41 francois hmm, I'm not sure we could safely upgrade to ansible 2.6 for rocky... that'd be a new minor revision for sure if we did 16:56:50 is there no way to get the right fix into ansible 2.5? 16:57:20 personally, I'd like rocky to be ansible 2.6 - but it is a big change that risks destabilising more things 16:57:35 it might be better to just document that as a known issue 16:57:42 i agree that we shouldnt make a big leap like that 16:57:59 i would want to work with upstream ansible to land those fixes instead 16:58:10 odyssey4me: I wasn't actually able to find the exact fix in ansible 16:58:27 yep, if possible - otherwise just document that people shouldn't use screen and use tmux instead :p 16:59:02 odyssey4me: maybe we can have openstack-ansible clear env 16:59:07 and we can pass only the things we want down to ansible 16:59:12 that said, I think jrosser might have actually identified the fix for it - or at least a workaround 16:59:30 mnaser that sounds a bit like a minefield 16:59:59 add a little spice to your life 17:00:13 hahaha :) 17:00:14 to end it 17:00:20 happy new year in advance. this is our last meeting of the year 17:00:26 next tuesday is the 25th 17:00:38 so unless y'all wanan celebrate xmas here together, we'll cancel that one :) 17:00:58 thanks for all the awesome work, we're doing great progress in what we discussed since the PTG 17:01:06 https://media.giphy.com/media/3o7TKtsBMu4xzIV808/giphy.gif 17:01:11 and it's awesome to see people excited about recommending OSA for large deployments in the ML (if anyone saw that thread!) 17:01:12 happy holidays you all! It's been a pleasure to work with osa team. I remember my first steps were this year 17:01:20 mid-year I think 17:01:30 and the plan is to keep working hard to contribute as many as I can 17:01:50 guilhermesp thank you for joining the party :) 17:02:03 and on that fun note, i'll end it with the lame joke we all wanted to say 17:02:04 odyssey4me: o/ let's rock 17:02:06 * odyssey4me pencils a note to catch up on the ML 17:02:15 see y'all in next years meeting 17:02:16 #endmeeting