22:04:13 #startmeeting nova_cells 22:04:14 Meeting started Wed Feb 4 22:04:13 2015 UTC and is due to finish in 60 minutes. The chair is alaski. Information about MeetBot at http://wiki.debian.org/MeetBot. 22:04:15 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 22:04:18 The meeting name has been set to 'nova_cells' 22:04:34 #topic Test failurs 22:04:38 argh 22:04:53 use #undo 22:04:57 #undo 22:04:58 Removing item from minutes: 22:04:59 (if you want) 22:05:02 sweet 22:05:08 I thought it was pretty appropriate 22:05:11 #topic Test failures 22:05:13 "failurs" 22:05:16 heh 22:05:30 hehe 22:05:36 :) 22:06:02 last week bauzas exluded the last test we expect to fail on the check-tempest-dsvm-cells job 22:06:16 so in theory it should be passing, but isn't 22:06:37 the most common failure I've seen is an ec2 run idempotent job 22:06:45 which I can't get to fail locally 22:07:15 http://logs.openstack.org/04/153004/3/check/check-tempest-dsvm-cells/dc1ee1f/console.html has an example 22:07:54 any eyes on that are appreciated so we can get the job voting at some point 22:08:16 I'm also (still) running tests locally. since the instance object flavor field merged failures have changed a bit and I'm investigating 22:08:49 melwitt: great 22:09:06 any help on testing is much appreciated 22:09:17 #topic WIP 22:09:33 https://review.openstack.org/#/c/150381/ 22:10:03 There's a patch up for a quick PoC of multiple db support 22:10:09 unfortunately Dheeraj is not here to comment 22:10:52 that's alright 22:11:01 he continues to look how to allow DB api to connect to different DBs 22:11:02 dansmith added some good comments to the review 22:11:18 yes, just saw it. Thanks 22:11:54 he also mentioned passing info through the context for picking a db, which I think is a good idea 22:12:44 yeah, think that's how it's going to have to work, 22:12:49 else we'll need some major refactoring of db api 22:13:12 agreed 22:13:23 ok 22:13:57 we can get into more specifics as work progresses, but that seems like the right mechanism to try 22:14:26 on my end I've been busy with other things that are wrapping up now, so I'm going to be coding furiously on this now 22:14:43 so there should be some more code to look at soon 22:15:11 (it's also why this agenda is light) 22:15:21 any other work to be called out? 22:15:34 alaski: the flavor stuff merged, so I should try fixing the libvirt driver now, right? 22:15:57 dansmith: yes, that would be excellent 22:16:10 alaski: and, does that mean that the flavor stuff didn't break cells? 22:16:13 I think that ties into what melwitt has been investigating 22:16:14 because that would be ... shocking 22:16:20 ah, melwitt you wanna do that? 22:16:23 dansmith: it appears that it didn't 22:16:28 alaski: shocking 22:16:42 dansmith: heh, we had faith in you 22:16:47 glad someone did :) 22:17:31 but yes, the libvirt driver should be looked at in light of the flavors work 22:17:32 dansmith: can you point me to the flavor stuff? 22:17:54 it's likely that a previous workaround can now be removed 22:18:02 alaski, dansmith: yes, it didn't make it worse from what I see so far, there's something different wrong now which I'm working out 22:18:03 belmoreira: https://review.openstack.org/#/c/135700/ 22:18:23 dansmith: thanks 22:18:24 melwitt: boy, you really know how to compliment someone eh? 22:18:30 :P 22:19:14 does someone want an action item to look at things? 22:19:34 melwitt: I'm guessing you'll continue digging in? 22:20:06 dansmith: lol. not what I meant, sorry. from my understanding, the flavor stuff should have made 60 failures disappear but it didn't, the failures are now different. I've *almost* got a handle on what it is that I hope to confirm later today, then I will be asking you for help :P 22:20:32 melwitt: oh, really? I didn't expect it to fix any of those 22:21:13 dansmith: it should have because the problem was that cell couldn't access the flavor that existed only at the top. with flavor staying with the instance, it should have had all the info it needed to succeed 22:21:16 it should mean that flavors don't need to exist in the cells db now 22:21:39 melwitt: ah, but only if the code that is looking up the flavor is using the per-instance interface 22:21:46 melwitt: so I think probably some more work is needed to convert those 22:21:48 I also figured out why the flavor not found problem got fixed and then returned. it got fixed the first time by garyk passing flavor obj down to driver 22:22:20 melwitt: so maybe tomorrow we can sync up on some of those failures and take a look at what changes might be easy? 22:22:27 but then it broke again when I added a get flavor call to fill in extra_specs that he asked for 22:22:54 melwitt: ahh 22:22:57 flavor get fails if you're in the non-top cell. 22:23:01 this should fix that 22:23:15 but we have to change the code to use the flavor on the instance instead of looking it up 22:23:17 yeah, the work from garyk shouldn't be necessary any longer 22:23:27 nor the extra_specs thing 22:23:37 right 22:23:46 there are some lookups in ironic that can be removed as well 22:23:52 cool 22:24:39 it sounds like melwitt has this in hand, with some possible assistance from dansmith 22:25:02 s/possible/probable :) 22:25:03 cool 22:25:06 hah 22:25:09 so we'll move on 22:25:17 I can offer probable assistance 22:25:20 melwitt: keep him busy 22:25:26 #topic open discussion 22:25:51 :) 22:26:01 There's still an open question of networking 22:26:22 I spoke with anteaya at the nova midcycle and she should be getting me in touch with someone from neutron 22:26:37 meaning whether we support n-net, or how we organize cells and neutron? 22:26:41 the latter 22:26:43 so, 22:26:52 if we associate a neutron endpoint per cell, 22:26:53 alaski: that is great 22:27:04 then they can be the same or one per cell, or one per few cells right? 22:27:29 if we're switching db/mq per cell, no reason we can't also switch the neutron being used 22:27:46 dansmith: right. but that assumes no longer using a global neutron 22:27:49 and then in the migration code, we can initially just fail if the target node isn't the same neutron endpoint as us 22:27:57 or that it partitions similarly 22:28:07 alaski: ah, I guess that might break some of our API stuff that expect a single pool of network resources 22:28:53 right, so we need some discussion on where best to address that 22:29:16 we work around it at rackspace, but I need to reach out and see what limitations that causes 22:29:25 yeah 22:30:01 this is the inital spec for nova net to neutron migration: https://review.openstack.org/#/c/147723/1 22:30:16 * tonyb is massively late. Did I get volenteered for anything? 22:30:28 tonyb: yeah, you're figuring out networking 22:30:35 ideally it would be nice if neutron partitioned along cell lines so we could do the multiple endpoint thing that you suggested dansmith 22:30:45 dansmith: awesome. I can do that. 22:30:54 tonyb: yep, please report back in a week with the solution 22:30:56 alaski: yeah, even if multiple partitions share a neutron 22:31:23 alaski: in fact I was thinking in the other way arround... having a central neutron 22:31:47 that is likely going to require cells-like support in neutron I think, 22:31:53 which is something we could wait a long time for 22:32:25 I 22:33:02 I'll talk to gus on the neutron side. It's something we should consider as part of the migration planning that's going on. 22:33:03 alaski: seems interesting... but I think that should be a long term goal... 22:33:58 long term it might make sense. I'd like to get a sense of how the neutron group sees things 22:34:42 I have the feeling that we're going to influence how neutron is deployed to work with cells 22:35:28 but primarily I think we need to get everyone talking 22:35:37 alaski, dansmith (and others) do you have time after the nova meeting tomorrow to talk to some neutron folk? (if I can arrange it) 22:36:14 which meeting time is it tomorrow? 22:36:17 that's the morning one 22:36:22 and yeah, I prolly could 22:36:27 yeah, I'll be around then 22:36:49 okay I'll see what I can do. 22:36:56 tonyb: thanks 22:37:19 any other topics for discussion? 22:37:36 move to get back to work :) 22:37:57 sounds good to me 22:38:09 #endmeeting