16:01:43 #startmeeting openstack_ansible_meeting 16:01:44 Meeting started Tue Apr 18 16:01:43 2017 UTC and is due to finish in 60 minutes. The chair is evrardjp. Information about MeetBot at http://wiki.debian.org/MeetBot. 16:01:46 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 16:01:48 The meeting name has been set to 'openstack_ansible_meeting' 16:01:58 lets do this! 16:01:58 #topic Last week AP 16:02:15 All last week AP are done 16:02:25 #topic this week triage 16:02:42 #link https://bugs.launchpad.net/openstack-ansible/+bug/1682481 16:02:44 Launchpad bug 1682481 in openstack-ansible "When using openstack-ansible-lxc_container_create role with an SSH jumpbox, sits waiting forever for container even though ssh is available." [Undecided,New] 16:02:59 hmm. 16:03:14 confirmed/medium? im not sure what a good fix would be though 16:03:41 delegate_to: {{ physical_host }}" ? 16:04:23 but I'd say confirmed medium yes. 16:04:35 ansible 2.3 and wait_for_connection should be a good solution 16:04:43 but were are not there yet. 16:04:44 im not sure that'd work - basically you'd need to test the port via the SSH jumpbox i assume? 16:04:54 the only way that'd work is to test SSH rather than test a port 16:05:07 yeah, that's what wait_for_connection will do 16:05:16 yeah but thats only 2.3 i guess 16:05:33 I'll mark low-hanging-fruit and ask if delegate_to: "{{ physical_host }}" would be anough 16:05:35 enough* 16:06:03 i think only performing an SSH connection test would actually work based on the bug 16:06:05 but yeah lets move on 16:06:20 #link https://bugs.launchpad.net/openstack-ansible/+bug/1682169 16:06:23 Launchpad bug 1682169 in openstack-ansible "Upgrade N->O nova-manage cell_v2 issue" [Undecided,New] 16:06:40 i'll take a look at this one 16:06:57 ok, for triaging or for fixing? 16:06:59 i need to run some tests around upgrades but it sounds plausible and repeatable so need to figure it out 16:07:04 ok 16:07:09 fixing/confirming at the very least :) 16:07:13 #action andymccr triage https://bugs.launchpad.net/openstack-ansible/+bug/1682169 16:07:20 this way it's marked for next week. 16:07:34 next 16:07:36 #link https://bugs.launchpad.net/openstack-ansible/+bug/1682161 16:07:37 Launchpad bug 1682161 in openstack-ansible "Upgrade N->O Nova services launch with newton tag and not ocata (systemd issue)" [Undecided,New] 16:08:15 o/ 16:08:27 cloudnull is volunteering:) 16:08:31 it makes sense to me. 16:08:40 I've seen other like this sadly. 16:08:49 or I seem to recall other like this. 16:08:55 * cloudnull reading bug 16:08:58 Confirmed medium? 16:09:07 hmm 16:09:20 https://github.com/openstack/openstack-ansible-os_nova/blob/master/tasks/nova_init_systemd.yml#L78 16:10:09 evrardjp: i think we need a daemon reload too 16:10:22 im wondering if the 2 are related 16:10:34 or this: https://github.com/openstack/openstack-ansible-os_nova/blob/966ea269c9acb88052aca8982c37e0ba8a94a207/tasks/nova_init_common.yml#L20 16:10:42 cloudnull: agreed. 16:11:09 I think we need to make this daemon-reload standard on all our roles 16:11:19 I thought of this: https://review.openstack.org/#/c/440463/ 16:11:28 well.. 16:11:30 https://github.com/openstack/openstack-ansible-os_nova/blob/stable/ocata/handlers/main.yml#L71 16:11:34 so it's part of the task 16:11:39 oh ok 16:11:51 but... maybe it doesn't actually reload the daemon ? 16:11:52 I misread 16:12:27 is Grelaud (fabrice-grelaud) here ? 16:13:24 so this could be an upstream ansible bug too 16:13:27 so we are at a state where we need confirmation 16:14:02 based on the diff it makes sense that something in the service / systemd task handler is not actually doing the daemon reload 16:14:28 yeah 16:14:39 oh wait 16:15:04 this spells it out fairly clearly https://launchpadlibrarian.net/315388078/nova-ocata-systemd-bug.txt 16:17:45 so i guess the next step is trying to figure out how we fix this everywhere. 16:18:51 If we had a log of the whole run that would be great 16:18:55 do we need to ask for this? 16:19:14 or will someone try to figure it out without the complete run log? 16:19:21 that would be helpful 16:19:33 I think we've enough to confirm the issue and mark it medium 16:19:39 I only think of issues in the run after the flush_handlers. 16:19:48 did this help? https://review.openstack.org/#/c/452327/ 16:19:48 but even that 16:19:52 go ahead and assign me to it and I'll try and recreate 16:20:11 jmccrory: it should have 16:20:18 wait. 16:20:28 Maybe he is not using the latest ocata. That's right 16:20:40 cloudnull: ok 16:20:46 oh. . .https://github.com/openstack/openstack-ansible-os_nova/blob/15.1.0/handlers/main.yml#L71 16:20:55 that bug was on 15.1.0 16:20:58 haha! 16:21:03 which did not have that change 16:21:09 score! 16:21:12 good catch jmccrory we were only on the branch, not on the right tag! 16:21:20 good catch jmccrory! 16:21:20 boom, already fixed! 16:21:38 hah good 16:21:41 andymccr: when can we get a tag out for 15.1.1 16:21:54 https://review.openstack.org/#/c/456640/ 16:22:05 boom! 16:22:06 team team team 16:22:09 fixed! 16:22:11 i will try follow it up cloudnull 16:22:15 cool 16:22:29 marking as invalid 16:22:31 next 16:22:42 #link https://bugs.launchpad.net/openstack-ansible/+bug/1682108 16:22:44 Launchpad bug 1682108 in openstack-ansible "[vagrant] openstack-ansible-lxc_hosts: The `lxc` module is not importable. Check the requirements."" [Undecided,New] 16:23:09 that's hwoarang 's bug 16:23:30 ahh this has a fix already too :) 16:23:35 yeah so i posted a solution which i'm not sure if it's a workaround or not 16:23:37 Related, not cloased 16:23:42 -a 16:23:56 well at least we can review this :) 16:24:08 i need some help from someone which access on a live openstack CI host to figure out why these host use pip2 from real root instead of the one from virtualenv :( 16:24:46 otherwise i can't progress much since it's going to be a guessing game 16:24:54 yeah, I guess. 16:24:57 hahaha. 16:25:11 hwoarang: we might be able to ask infra to hold a node for us 16:25:24 that's possible 16:25:37 sadly what evrardjp said is normal. its a guessing game. 16:25:39 else, cloudnull there is a way to get latest infra image for local testing? 16:25:54 you can build it with DIB 16:26:11 there is no artifact of the daily build ? 16:26:32 hwoarang: in the past we've added rapid break points to a commit so it runs, breaks and then we go look try and move things ahead 16:27:08 I guess we should avoid that as much as possible, IIRC. 16:27:38 i see 16:27:42 cloudnull: could you have a look with hwoarang? 16:28:04 if holding a node for half a hour is possible that would be very helpful 16:28:08 sure. 16:28:30 ok lets discuss after 16:28:39 the meeting 16:28:41 #action hwoarang cloudnull investigate on https://bugs.launchpad.net/openstack-ansible/+bug/1682108 with infra 16:28:42 Launchpad bug 1682108 in openstack-ansible "[vagrant] openstack-ansible-lxc_hosts: The `lxc` module is not importable. Check the requirements."" [Undecided,New] 16:28:53 something like that, we'll remember for next week :) 16:29:03 ok next 16:29:05 #link https://bugs.launchpad.net/openstack-ansible/+bug/1681714 16:29:05 Launchpad bug 1681714 in openstack-ansible " Ceph cluster’s performance are not monitored by Telegraf" [Undecided,New] - Assigned to Bertrand Lallau (bertrand-lallau) 16:29:17 confirmed wishlist? 16:29:22 sure 16:29:22 ++ 16:29:31 ok next 16:29:32 #link https://bugs.launchpad.net/openstack-ansible/+bug/1681695 16:29:34 Launchpad bug 1681695 in openstack-ansible "Incorrect keystone with multiple memcache configuration" [Undecided,New] - Assigned to Jean-Philippe Evrard (jean-philippe-evrard) 16:30:01 I'll work on this. 16:30:04 if bertrand is here, we're intending to port a mess of ceph plugins over to monitorstack soon and get that into our telegraf setup 16:30:21 for the last one basically we have to think where to put memcache in the future 16:30:35 but it's fine I have already a commit for the start of this. 16:30:54 that's all for today 16:30:55 cool :) 16:30:57 good job! 16:31:02 thanks everyone! 16:31:08 anything else to add? 16:31:13 nope! 16:31:16 #endmeeting