15:06:03 <johnthetubaguy> #startmeeting XenAPI 15:06:03 <openstack> Meeting started Wed Nov 26 15:06:03 2014 UTC and is due to finish in 60 minutes. The chair is johnthetubaguy. Information about MeetBot at http://wiki.debian.org/MeetBot. 15:06:05 <openstack> Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 15:06:07 <johnthetubaguy> hello folks 15:06:08 <openstack> The meeting name has been set to 'xenapi' 15:06:17 <johnthetubaguy> #topic CI 15:06:23 <BobBall> Howdy. 15:06:30 <matel> Hello 15:06:42 <johnthetubaguy> BobBall: matel: hows things looking in CI land? 15:06:52 <BobBall> Good... But... 15:06:58 <BobBall> https://review.openstack.org/#/c/133975/ 15:07:02 <BobBall> Don't like that -1 15:07:14 <BobBall> I've not followed up on it and I need to I know 15:07:20 <BobBall> but it means I'm not happy to remove the restrictions yet 15:07:38 <BobBall> there are still some race conditions / circumstaces where some of the excluded tests will fail 15:08:08 <matel> that's bad news 15:08:18 <BobBall> I did note that it's the same message we've seen many times... Details: {u'message': u'No valid host was found. There are not enough hosts available.', u'created': u'2014-11-19T16:57:33Z', u'code': 500} 15:08:23 <johnthetubaguy> like 10% failure I guess 15:08:24 <matel> What was the failure ratio? 15:08:32 <BobBall> No idea 15:08:36 <johnthetubaguy> BobBall: that means you have too many instances running 15:08:37 <BobBall> we'd need maybe 100 runs to get a ratio 15:08:55 <BobBall> You think it's running out of RAM then? 15:09:03 <johnthetubaguy> oh wait, maybe not... 15:09:16 <johnthetubaguy> we need the other instance faults in the DB for that instance 15:09:22 * BobBall checks screen-n-sch... 15:09:24 <matel> It's gonna be a hard nut to crack 15:09:35 <johnthetubaguy> its possible it was retrying, and then runs out of hosts 15:09:43 <johnthetubaguy> thats more likely I guess 15:10:02 <BobBall> screen-n-sch is taking _ages_ to get from the CDN! 15:10:39 <BobBall> it's only 31k but has been downloading for over 60 seconds so far... 15:10:48 * BobBall strums his fingers on the table 15:11:19 <johnthetubaguy> doh 15:11:26 <BobBall> Let's move on and see if that loads later. 15:11:29 <johnthetubaguy> yeah 15:11:36 <johnthetubaguy> any more for any more? 15:12:07 <matel> Neutron does not run in parallel for some reason. 15:12:12 <matel> But it ran in serial. 15:12:17 <matel> Let me dig out the results. 15:12:45 <matel> http://paste.openstack.org/show/137612/ 15:13:14 <BobBall> 21 failures in parallel was that? I don't remember 15:13:30 <BobBall> or was that 21 failures when we went serial only? 15:13:35 <BobBall> That must have been serial only, sorry 15:13:39 <BobBall> worker 0 only :) 15:13:50 <johnthetubaguy> yeah, I have a feeling neutron broke in parallel 15:13:52 <BobBall> We got lots more failures running in parallel didn't we 15:14:07 <BobBall> Sure - but it's run in parallel for -infra 15:14:08 <johnthetubaguy> some races still in iptables, I think I was told 15:14:12 <johnthetubaguy> hmm, odd 15:14:18 <johnthetubaguy> I guess it muss be now 15:14:18 <BobBall> Ah - that might not be using iptables 15:14:28 <johnthetubaguy> anyways 15:14:32 <BobBall> matel: Which firewall was that with? 15:14:35 <matel> We are not using iptables either 15:14:39 <BobBall> Now using noop? 15:14:44 <matel> y 15:15:03 <BobBall> OK 15:15:15 <BobBall> Do you know who might have more details on the races you're thinking of johnthetubaguy ? 15:15:39 <johnthetubaguy> BobBall: trying to remember... 15:16:00 <johnthetubaguy> BobBall: we don't have the neutron call back libvirt have implemented to avoid lots of those races, it probably related 15:16:20 <johnthetubaguy> libvirt has instance events call back to say neutron has completed, or something like that 15:16:45 <johnthetubaguy> basically there is a wait for event thing in libivrt 15:17:24 <BobBall> I see 15:17:29 <BobBall> So what does that do? 15:17:43 <BobBall> stop neutron from running things in parallel until libvirt has said that X has finished? 15:17:58 <johnthetubaguy> not really 15:18:09 <johnthetubaguy> it waits to start the VM until neutron has set everything up 15:18:23 <BobBall> So it's the nova driver _listning_ to the neutron event 15:18:46 <BobBall> so if the issue is what you're thinking of we don't need the nova driver to emit events 15:19:02 <matel> We are experiencing failures with parallel runs 15:19:25 <matel> If we were missing something like what you said, we would see failures with serial as well, don't we? 15:19:46 <johnthetubaguy> yes thats true 15:19:47 <BobBall> No - because neutron has slowness issues 15:19:51 <matel> Or maybe the higher load... 15:19:57 <BobBall> so races will become more visible in parallel 15:20:04 <johnthetubaguy> BobBall: +1 15:20:09 <johnthetubaguy> you typed it before I did 15:20:30 <matel> BobBall said the magic word: race 15:20:46 <johnthetubaguy> yeah, it is a race 15:20:55 <johnthetubaguy> it might be ready when you start the VM, it might not 15:21:05 <BobBall> Let's just add a sleep 60. 15:21:08 <johnthetubaguy> so libvirt waits till its ready 15:21:10 <BobBall> Sleep 60's always work. 15:21:14 <johnthetubaguy> right... 15:21:20 <BobBall> ... if you have enough of them 15:21:27 <matel> And they speed up the time it takes to run the tests. 15:21:36 <johnthetubaguy> what are the failures looking like? 15:21:39 <BobBall> Paradixically there matel ... :) 15:22:24 <matel> gimme a sec 15:23:03 <matel> tempest.api.compute.servers.test_attach_interfaces.AttachInterfacesTestJSON.test_create_list_show_delete_interfaces 15:23:03 <matel> tempest.api.compute.servers.test_attach_interfaces.AttachInterfacesTestXML.test_create_list_show_delete_interfaces 15:23:03 <matel> tempest.api.network.admin.test_quotas.QuotasTest.test_lbaas_quotas 15:23:03 <matel> tempest.api.network.test_service_type_management.ServiceTypeManagementTestJSON.test_service_provider_list 15:23:07 <matel> tempest.cli.simple_read_only.network.test_neutron.SimpleReadOnlyNeutronClientTest.test_neutron_lb_healthmonitor_list 15:23:10 <matel> tempest.cli.simple_read_only.network.test_neutron.SimpleReadOnlyNeutronClientTest.test_neutron_lb_member_list 15:23:13 <matel> tempest.cli.simple_read_only.network.test_neutron.SimpleReadOnlyNeutronClientTest.test_neutron_lb_pool_list 15:23:16 <matel> tempest.cli.simple_read_only.network.test_neutron.SimpleReadOnlyNeutronClientTest.test_neutron_lb_vip_list 15:23:19 <matel> tempest.cli.simple_read_only.network.test_neutron.SimpleReadOnlyNeutronClientTest.test_neutron_meter_label_list 15:23:22 <matel> tempest.cli.simple_read_only.network.test_neutron.SimpleReadOnlyNeutronClientTest.test_neutron_meter_label_rule_list 15:23:25 <matel> tempest.scenario.test_load_balancer_basic.TestLoadBalancerBasic.test_load_balancer_basic 15:23:27 <matel> tempest.scenario.test_network_basic_ops.TestNetworkBasicOps.test_hotplug_nic 15:23:29 <matel> tempest.scenario.test_security_groups_basic_ops.TestSecurityGroupsBasicOps.test_cross_tenant_traffic 15:23:32 <matel> tempest.thirdparty.boto.test_ec2_instance_run.InstanceRunTest.test_compute_with_volumes 15:23:48 <johnthetubaguy> I don't really mean that 15:23:50 <matel> johnthetubaguy: these are the failed ones with serial 15:23:53 <BobBall> test_hotplug_nic <-- we don't do that yet do we? 15:23:55 <johnthetubaguy> what sort of failures did those see? 15:24:11 <johnthetubaguy> BobBall: I have code for that, but its not finished or merged 15:24:43 <BobBall> OK, great. 15:24:46 <johnthetubaguy> are they ssh failure? 15:24:51 <BobBall> We might want to talk about that in the new year 15:24:59 <johnthetubaguy> BobBall: put it up for review a year ago or so 15:25:11 <BobBall> *nod* 15:25:29 <johnthetubaguy> coded it in some airport after some US summit, while drinking something from starbucks 15:25:57 <johnthetubaguy> anyways 15:26:07 <johnthetubaguy> any more we want to debug now? or do this offline? 15:26:10 <BobBall> Sounds like it doesn't even need reviewing - just merge it :D 15:26:20 <BobBall> That's the perfect coding environment! 15:26:38 <BobBall> No, I don't think there is more to talk about on that front 15:26:44 <matel> no 15:27:48 <johnthetubaguy> #topic Open Discussion 15:27:57 <johnthetubaguy> any more? 15:28:02 <matel> nope 15:28:36 <BobBall> yes 15:28:40 <BobBall> maybe 15:28:45 <BobBall> should we talk about DIB? 15:29:22 <matel> We'll give a try to use DIB to create the domU's root filesystem 15:30:00 <matel> So something like this: You have a qcow2 with a XenServer inside, on the SR sits a vhd containing domU's virtual disk 15:30:04 <johnthetubaguy> OK, yeah, they are having "fun" using that in rackspace mind 15:30:23 <johnthetubaguy> crazy 15:30:28 <matel> ... Use tapdisk to mount the virtual disk.. 15:30:32 <johnthetubaguy> anyways, any more? 15:30:46 <matel> No, thanks, thanks. 15:31:03 <johnthetubaguy> thanks all 15:31:07 <johnthetubaguy> #endmeeting