15:04:37 #startmeeting ceilometer 15:04:37 Meeting started Thu Mar 20 15:04:37 2014 UTC and is due to finish in 60 minutes. The chair is jd__. Information about MeetBot at http://wiki.debian.org/MeetBot. 15:04:38 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 15:04:40 The meeting name has been set to 'ceilometer' 15:04:42 hey 15:04:49 o/ 15:04:52 o/ 15:04:54 ok it's just lag :) 15:04:57 hi everyone 15:05:08 o/ 15:05:08 o/ 15:05:11 o/ 15:05:12 o/ 15:06:02 o/ 15:06:07 * eglynn lurks ... 15:06:16 o/ 15:07:19 #topic Milestone status icehouse-rc1 15:07:39 #link https://launchpad.net/ceilometer/+milestone/icehouse-rc1 15:07:49 so we have a bunch of bugs to fix we probably need to focus a lot on that 15:08:02 I've started to go through the entire bug list to triage the bug and target them as needed 15:08:04 help welcome :) 15:08:11 anything to discuss on that otherwise? 15:08:54 trying to find a link. 15:08:57 yep 15:09:13 I've started critical bug about logging 15:09:17 nm. i should really read #link.lol 15:09:29 nprivalova: what are you plans for addressing that bug? 15:10:07 gordc: am I right that ceilometer pipeline is running inside swift to publish messages? 15:10:33 nprivalova: yes 15:10:35 yep. it's used in middleware. 15:10:45 I saw that bug and i'm not sure how to fix it 15:10:52 actually syslog and s-proxy are almost the same 15:11:34 I guess it's not ceilometer problem 15:11:45 is there no way to filter logs from 'external' libraries? like how we define here: https://github.com/openstack/ceilometer/blob/master/ceilometer/service.py#L117 15:11:47 looks like swift writes in 2 log files 15:12:36 I don't know from what I saw in the bug the problem is that the log are set to debug 15:12:51 I don't see how we are supposed to change that? 15:13:21 jd__: yeah, our logs should be debug ... (some of them are audit level right now) 15:13:35 the ones in the bug report are debug as far as I saw 15:13:36 the problem is not id 'debug' 15:13:59 nprivalova: do you have a link to the bug you're discussing? 15:13:59 ok, let's proceed in local-channel 15:14:12 #link https://bugs.launchpad.net/ceilometer/+bug/1294789 15:14:14 Launchpad bug 1294789 in ceilometer "ceilometer swift module is spamming rsyslog with useless logs" [Critical,Triaged] 15:14:25 nprivalova: thanks 15:14:50 nprivalova: i'm ok with discussing this after other topics. 15:14:56 seems like it'll take a while. 15:15:00 yep 15:15:04 me too 15:15:07 is there any other topic? 15:17:36 looks like no 15:17:38 I guess not :) 15:17:46 #topic Tempest integration 15:18:15 does everybody knows the whole story about tempest :)? 15:18:27 what the whole story? 15:18:48 or what do you call the whole story? :) 15:19:10 nprivalova: I'd like to hear stories :) 15:19:11 I've started a thread [Ceilometer][QA][Tempest][Infra] Ceilometer tempest testing in gate 15:19:28 ok I saw that one I think but I didn't read everything yet 15:19:53 so we cannot run tempest tests on gating as it is 15:20:06 there are at least 2 blockers 15:20:13 1. DBDeadlock 15:20:26 2. One collector 15:21:06 but Sean noticed that cpu is very high even with 1 collector 15:21:26 nprivalova: i strung together all my patches to see if my multi collector and dbdeadlock patches work well 15:21:49 cool 15:21:58 so my plan is to continue investigation and do some profiling 15:22:01 clearly we need to improve that so all efforts should be oriented on that now 15:22:16 future kudos to gordc and nprivalova then :) 15:22:17 nprivalova: i'd imagine the cpu would be high with one collector. we're essentially passing hundred/thousands of messages to one source and letting it churn away. 15:23:15 gordc: it can't be the reason for high cpu 15:23:34 hundred/thousands of messages would cause high io latency not cpu 15:24:09 it seems that we are creating/processing lot's of objects somehow 15:24:32 Alexei_9871: sure... that and the load will always be at 100%. 15:24:51 Alexei_9871: that's a good hypothesis I think 15:24:55 doing profiling would help 15:24:57 gordc: yes but multiple collectors on same server won't help us 15:25:08 jd__: me and nprivalova are now working on it 15:25:17 #action provide profiling resuls 15:25:41 doesn't work :) 15:25:55 jd__: we disscussed a little bit one colletor issue and a better solution would be to use several native threads instead of workers 15:26:06 one collector* 15:26:19 jd__: what do you think? 15:26:42 Alexei_9871: nothing is thread safe 15:27:05 I'm not against it but you won't manage to do that until 2016 ;) 15:27:07 jd__: we still have GIL :) 15:27:23 that too, so the perf would be less than having several workers anyway 15:27:29 but I think it's out of scope here 15:27:34 anything else on testing? 15:27:40 jd__: problem with several workers is huge memory consumption 15:27:56 yep, We've created tests for pollsters 15:28:16 Alexei: how huge on collector? 15:28:29 llu-laptop: same question 15:29:25 I think we will get this info after profiling 15:29:32 ok, moving on then 15:29:40 let's go back at the end or after the meeting on that if you want 15:29:43 #topic Release python-ceilometerclient? 15:29:48 eglynn-afk: around maybe? :) 15:29:54 I think we still have patches in the queue 15:30:06 jd__: you're right, we still have left 15:30:22 ok and we've still have time to release, so let's wait a bit :) 15:30:28 #topic Open discussion 15:30:39 jd__: there was a gate issue with the clients' pypy gate, it's temporary fixed, I hope it will help in the process 15:31:29 llu-laptop: around 50Mb * num_workers 15:31:46 not that terrible 15:31:55 ildikov_: ok 15:36:18 nothing else? closing in a minute then 15:36:41 thanks to everyone who helped anamalagon with her OPW application 15:36:49 We'll find out April 21 if she got into the program 15:36:59 cool 15:37:32 hi, can I ask a question regarding _ThreadPoolWithWait? 15:37:49 I asked the question earlier in the ceilometer channel, but got no response. 15:37:54 terriyu: np, she's very active, so it's easy to help her :) 15:38:19 anybody know the reason why we want the threadPool to wait? 15:38:25 tongli: maybe nobody has the answer :) 15:38:57 I noticed some rather strange behaviors. 15:39:11 ildikov_: awesome :) I'll pass on the compliment 15:39:25 when message arrives in collector, if we have like 64 threads in the pool (default), then the thread will 15:39:45 hold for a long time, until all the 64 threads get some tasks, then all the threads start execution. 15:40:03 took a long time to figure this out, wonder if anyone knows anything about this. 15:40:57 terriyu: cool, thanks :) 15:42:40 wrapping up then, let's go on #openstack-ceilometer if needed :) 15:42:52 #endmeeting