Tuesday, 2015-12-01

*** dimtruck is now known as zz_dimtruck00:02
*** zz_dimtruck is now known as dimtruck00:23
*** mriedem_away is now known as mriedem00:42
*** dimtruck is now known as zz_dimtruck01:07
*** mriedem has quit IRC01:13
*** f13o has quit IRC01:31
*** markvoelker has quit IRC01:46
*** badari has quit IRC01:54
*** badari has joined #openstack-performance02:08
*** badari has quit IRC02:14
*** arnoldje has joined #openstack-performance02:20
*** rfolco has joined #openstack-performance02:21
*** rfolco has quit IRC02:51
*** badari has joined #openstack-performance05:19
*** arnoldje has quit IRC05:25
*** badari has quit IRC05:33
*** harlowja_at_home has joined #openstack-performance05:57
*** dims has quit IRC06:09
*** dims has joined #openstack-performance06:14
*** badari has joined #openstack-performance06:14
*** badari has quit IRC06:22
*** harlowja_at_home has quit IRC06:34
*** mlgrneff has joined #openstack-performance08:54
*** xek has joined #openstack-performance09:07
*** amaretskiy has joined #openstack-performance09:22
*** itsuugo has quit IRC10:03
*** itsuugo has joined #openstack-performance10:05
*** rfolco has joined #openstack-performance10:19
*** mlgrneff has quit IRC11:37
*** mlgrneff has joined #openstack-performance11:37
*** regXboi has joined #openstack-performance12:25
*** markvoelker_ has joined #openstack-performance13:49
*** arnoldje has joined #openstack-performance14:03
*** mdorman has joined #openstack-performance14:11
*** markvoelker has joined #openstack-performance14:34
*** markvoelker_ has quit IRC14:35
*** mriedem has joined #openstack-performance14:36
*** badari has joined #openstack-performance14:41
*** arnoldje has quit IRC14:41
*** mdorman has quit IRC14:46
kun_huangIs our meeting 5 minutes later or one hour later?14:54
DinaBelovakun_huang - sadly I did not resend the email to collect feedback on that step - so in 5 mins14:55
DinaBelovaI'll send the proposal today14:55
DinaBelovakun_huang - I forgot to do that, as was suffering from fever last week :(14:55
kun_huangDinaBelova: welcome back now :)14:56
DinaBelovakun_huang - yeah!14:56
*** manand has joined #openstack-performance14:58
DinaBelova#startmeeting Performance Team15:00
openstackMeeting started Tue Dec  1 15:00:05 2015 UTC and is due to finish in 60 minutes.  The chair is DinaBelova. Information about MeetBot at http://wiki.debian.org/MeetBot.15:00
openstackUseful Commands: #action #agreed #help #info #idea #link #topic #startvote.15:00
openstackThe meeting name has been set to 'performance_team'15:00
DinaBelovahello everyone! :)15:00
DinaBelovaraise your hands who's around :)15:00
kun_huango/15:00
amaretskiyhi15:00
DinaBelovaklindgren, harlowja? :)15:00
DinaBelovahehe, people sleeping :)15:01
mlgrneffhi, it's Rob from Flex Ciii15:01
DinaBelovamlgrneff - hey!15:01
DinaBelovanice to see u here15:01
mlgrneffGlad to finally make it.15:01
DinaBelovaso let's wait for a few more moments :) to allow people appear15:02
DinaBelovamlgrneff ;)15:02
manandhi15:02
DinaBelovaok, so let's probably start. As usual with today's agenda15:03
DinaBelova#link https://wiki.openstack.org/wiki/Meetings/Performance#Agenda_for_next_meeting15:03
DinaBelova#topic Action Items15:03
DinaBelovaandreykurilin are you around?15:03
DinaBelovawe have an action item on you15:03
andreykurilinsure15:03
DinaBelovafrom last meeting15:03
*** mlgrneff is now known as RobNeff15:03
andreykurilinI'm here:)15:03
andreykurilinhi15:03
DinaBelovaabout when load pattern tests runner will be available in Rally15:04
DinaBelovaI know there is already some change for it on review - any feeling when it'll be merged?15:04
*** rvasilets___ has joined #openstack-performance15:04
DinaBelova#link https://review.openstack.org/#/c/234195/15:04
*** AugieMena has joined #openstack-performance15:04
DinaBelovakun_huang or I may asl you as well - as a reviewer :)15:05
rvasilets___o/15:05
DinaBelovaany bad/good feeling on this patch?15:05
DinaBelovarvasilets___ o/15:05
andreykurilinrvasilets is an owner of this patch )15:05
andreykurilinhe can give a better estimates15:05
DinaBelovawell, his progress depends much on the reviewers :)15:05
DinaBelovarvasilets___?15:05
DinaBelovaI remember this change was hot enough last time - any estimates on merging?15:06
rvasilets___Patch is ready. Just waiting for review. Its already work15:06
*** Jeff__ has joined #openstack-performance15:06
DinaBelovaok, so I'm right and reviewing effort needs to be spend :)15:07
andreykurilinGeneral news are: patch by rvasilets adds a base ability to "stress runner", but it doesn't decrease load on SLA failure15:07
rvasilets___pboldin and kun are reviewed it. That is all15:07
kun_huangDinaBelova: I will update my review on new patch set ;)15:07
DinaBelovaandreykurilin - well, we need to do first steps first :)15:07
DinaBelovakun_huang, ok, thanks sir :)15:07
andreykurilinDinaBelova: I don't know anyone who works on the second step15:07
DinaBelovaso waiting for it with great wish to see it merged as a first step - we can add work item for second one15:08
DinaBelovaandreykurilin ack15:08
DinaBelovacool, so let's go further15:08
*** dims has quit IRC15:08
DinaBelovaso about devstack-gate n-api-meta  job15:08
DinaBelovathere is an email to openstack-operators15:08
DinaBelovalemme check it15:08
DinaBelovamriedem proposed one job to run metadata as well15:09
DinaBelovahm, cannot find the email in the arcives15:10
DinaBelovaI'll check it later15:10
DinaBelovaso anyway, action item is done15:10
mriedemneutron large ops15:10
DinaBelovamriedem - yep, thank you sir15:10
DinaBelovaand about my action items - done both15:10
DinaBelovareally encouraging you guys to review https://review.openstack.org/#/c/251343/15:11
DinaBelovaand all chain related15:11
*** pglass has joined #openstack-performance15:11
DinaBelovaandreykurilin, harlowja, kun_huang - please feel free to leave comments :)15:11
DinaBelovaok, cool, so that's it for action items15:11
kun_huanggot it15:12
DinaBelovaany questions regarding this topic?15:12
andreykurilinmy todo list is extended :) will review your patches a little bit later15:12
DinaBelovaandreykurilin ;)15:12
DinaBelovacool, thanks andreykurilin kun_huang15:12
DinaBelovalet's move forward15:12
DinaBelova#topic Nova-conductor performance issues15:12
DinaBelovaok, so we have some news on that field :)15:12
DinaBelovaklindgren has sent me the results of the following patch15:13
DinaBelova#link http://paste.openstack.org/show/480426/15:13
DinaBelovait's dumping the conductor workers info once a minute15:13
DinaBelovavisual results may be found here15:13
DinaBelova#link https://drive.google.com/a/mirantis.com/file/d/0ByRtVrZu5ifzUkZTQVZzMERYQWc/view15:13
DinaBelovaand today dims and I had a chance to take a look15:14
kun_huangopening it...15:14
DinaBelovafrankly speaking it's not super obvious...15:14
DinaBelovathere are dumps of 2 workers there15:14
*** dansmith has joined #openstack-performance15:14
DinaBelovai suspect that some of the reds are related to the bug we have observed in MOS as well - https://bugs.launchpad.net/mos/+bug/1380220 - this is related to the eventlet red svg boxes (that can be seen on first picture from conductor worker with 8401 pid) - we did not observe this bug before in community, so it looks like we need to file it on the upstream as well (related to the heartbits probably)15:14
openstackLaunchpad bug 1380220 in Mirantis OpenStack "OpenStack services excessively poll socket events when oslo.messaging is used" [Medium,Triaged] - Assigned to MOS QA Team (mos-qa)15:14
DinaBelovawe have seen the same behavior as 8401 worker on some of the MOS installations15:15
johnthetubaguydo we know what DB driver they are using for the nova-conductor?15:15
*** rpodolyaka has joined #openstack-performance15:15
*** ctrath has joined #openstack-performance15:15
DinaBelovajohnthetubaguy - it's not listed in the https://etherpad.openstack.org/p/remote-conductor-performance15:15
DinaBelovahm...15:15
DinaBelovaI hope klindgren will appear :)15:16
mriedemmysql-python i think15:16
DinaBelovaI know it's a bit early for him now15:16
johnthetubaguywould be good to check if its pymysql15:16
mriedemb/c it's older oslo.db pre-pymysql15:16
mriedemjohnthetubaguy: i don't think it is15:16
johnthetubaguymriedem: ah, that is what I was wondinerg15:16
dansmithmriedem: they could still be running it with the older one, right?15:16
dansmithit's just the default changed..15:16
mriedemoslo.db==1.7.1, MySQL-python==1.2.3 (kilo reqs are: oslo.db<1.8.0,>=1.7.0)15:16
DinaBelovaah, stop-stop15:16
DinaBelovaoslo.db==1.7.1, MySQL-python==1.2.3 (kilo reqs are: oslo.db<1.8.0,>=1.7.0)15:16
DinaBelovajohnthetubaguy ^^15:16
kun_huangjohnthetubaguy: has anyone deployed pymysql yet?15:16
dansmithjohnthetubaguy: either way, does that impact the messaging performance?15:16
DinaBelovamriedem, yep, thanks15:16
mriedemdansmith: it might impact eventlet, is probably why people are asking15:17
mriedemsince mysql-python doesn't support eventlet right?15:17
johnthetubaguydansmith: not sure what you mean by messaging performance15:17
johnthetubaguymriedem: yeah, each DB call locks up the whole thread15:17
dansmithmriedem: it will affect eventlet, but the hotspots are banging hard on rabbit sockets, which doesn't seem like it would be related to the db driver15:17
DinaBelovabut still 8401 worker is having the https://bugs.launchpad.net/mos/+bug/1380220 - issue - but in fact it sould not influence %CPU used15:17
openstackLaunchpad bug 1380220 in Mirantis OpenStack "OpenStack services excessively poll socket events when oslo.messaging is used" [Medium,Triaged] - Assigned to MOS QA Team (mos-qa)15:17
johnthetubaguyand eventlet lets each one in turn do its DB call, before letting it process the response, I am told15:18
DinaBelovainteresting moment is with 840215:18
DinaBelovalots of RabbitMQ-related timeouts15:18
johnthetubaguydansmith: ah, so my head is mush today, can I can't even open the files somehow15:18
dansmithjohnthetubaguy: right so what DinaBelova is talking about right now ... doesn't seem db driver related to me15:18
DinaBelovadansmith - yep..15:19
dansmithjohnthetubaguy: and that's what I observed when looking at their profile traces a couple weeks ago.. something seems to be banging really hard on rabbit15:19
DinaBelovaand the only feeling I have now - to check that their CPU load is really related to the RabbitMQ15:19
dansmithjohnthetubaguy: I thought maybe it was heartbeats or something, but they say that's disabled15:19
DinaBelovaso we need more dumps + top screens15:19
dansmithDinaBelova: ++15:19
DinaBelovaif yes15:19
*** bauzas has joined #openstack-performance15:19
DinaBelovaone possible variant to fix it is rabbitmq upgrade 0_o15:19
DinaBelovaor some tcp / whatever tuning...15:20
DinaBelovaas we simply see  RabbitMQ waiting for reading things on the wire - and that’s it15:20
DinaBelovaif all their CPU issues are about this - well, it'll be other (one more) RabbitMQ story15:20
DinaBelovathey have RabbitMQ 3.3.x15:20
johnthetubaguydansmith: ah, interesting, that does seem separate, unless eventlet is making it lie15:20
DinaBelovajohnthetubaguy - yep, sir15:21
DinaBelovaso I'll ask klindgren to make more dumps for more workers and for longer time + include tops for the same moments15:21
johnthetubaguydansmith: I know belliott hit some issues with DB locking sending the elapsed times crazy, but I don't remember the details now15:21
dansmithjohnthetubaguy: yeah, eventlet could be making it lie for sure, I'm just not sure that the db driver could be making the numbers look like they're elsewhere15:21
mriedemdansmith: didn't they see a gain when turning off ssl too?15:21
dansmithmriedem: a small gain15:21
dansmithmriedem: the first big gain was because they mistyped the config :/15:22
DinaBelova#action DinaBelova klindgren more dumps for more workers and for longer time + include tops for the same moments to ensure we see the same RabbitMQ related reds at the same time conductors CPU is going crazy15:22
DinaBelovamriedem - without ssl it was just a little drop15:22
johnthetubaguydansmith: oh, wait, this does sound like what brian found, you want to reduce the number of eventlet works, and it stops thrashing the hub15:22
*** rohanion has joined #openstack-performance15:22
dansmithjohnthetubaguy: s/works/workers/ ?15:22
*** harlowja_at_home has joined #openstack-performance15:23
johnthetubaguyyeah, sorry15:23
johnthetubaguyworkers15:23
dansmithjohnthetubaguy: can you explain more?15:23
johnthetubaguywell, greenlet threads I mean15:23
harlowja_at_home\o15:23
johnthetubaguyso I think eventlet got very busy scheduling between lots of active threads15:23
*** f13o has joined #openstack-performance15:23
DinaBelovaharlowja_at_home o/15:23
johnthetubaguyso we turned down the number of workers (this was on the scheduler, rather than conductor)15:23
johnthetubaguyand we found it was better at pushing through DB queries, when using mysql-python15:24
*** arnoldje has joined #openstack-performance15:24
dansmithhmm15:24
mriedemscheduler workers?15:24
dansmithjohnthetubaguy: that means more queuing in rabbit instead of queuing in memory on a conductor itself?15:24
dansmithmriedem: greenlet workers15:24
mriedemoh15:24
DinaBelovajohnthetubaguy - interesting, may you please fill your proposal in the https://etherpad.openstack.org/p/remote-conductor-performance somewhere15:24
johnthetubaguywell, more waiting to be restored, while the thread is busy doing DB stuff15:24
dansmithjohnthetubaguy: well, what I mean is,15:25
johnthetubaguyas it lets all the DB stuff happen before letting the python code process the response, or something like that?15:25
DinaBelovajohnthetubaguy - in fact due to the cprofile data DB operations were not so busy15:25
dansmithjohnthetubaguy: we don't dequeue a thousand things from rabbit and then try to balance them even though we can only do one at a time15:25
DinaBelovabut who knows15:25
dansmithDinaBelova: right, that's fine15:25
dansmithDinaBelova: they wouldn't in this case johnthetubaguy is talking about15:25
DinaBelovadansmith - yeah, i just understood it15:26
DinaBelovadansmith thanks15:26
johnthetubaguyits more a starvation issue, as I understood it15:26
dansmithyeah, I guess I can see that15:26
dansmiththe thing is,15:26
dansmiththey don't have much if any real db traffic needing to be services15:26
dansmither, serviced15:26
dansmithso I'm not sure why there would be a pile of requests needing a pile of threads15:27
dansmithbasically, just periodics and service checkins15:27
dansmiththeir cloud is otherwise mostly idle15:27
DinaBelovathey have lots of nova metadata requests15:27
mriedemisn't cells always syncing up too?15:27
DinaBelovadue to periodical puppet scripts running15:27
johnthetubaguymriedem: nova via conductor though15:27
dansmithDinaBelova: that's true, forgot about those15:27
mriedemjohnthetubaguy: ok15:28
DinaBelovaand they sure have the cache as well, but metadata is periodically knocking the conductor15:28
mriedemwhich is why we talked about turning on n-api-meta in one of the large ops jobs15:28
dansmithso anyway, seems like worth a try15:28
johnthetubaguyhonestly, those service updates are all blocking DB calls, but they should be quick-ish though15:28
DinaBelovamriedem precisely15:28
*** zz_dimtruck is now known as dimtruck15:28
*** markvoelker_ has joined #openstack-performance15:28
DinaBelovajohnthetubaguy - may you please add your idea to the https://etherpad.openstack.org/p/remote-conductor-performance just to have it written up?15:29
johnthetubaguyah... n-api-meta uses conductor... I never quite realised that15:29
johnthetubaguyDinaBelova: will do15:29
dansmithjohnthetubaguy: yeah, so you can have a db-less compute node15:29
DinaBelovajohnthetubaguy - thank you sir15:29
DinaBelovaso yeah15:29
DinaBelovawe need more data!15:29
DinaBelovawill ping klindgren after the meeting :)15:29
DinaBelovaand I guess we may leave this topic for a while15:29
DinaBelovalet's move forward15:29
johnthetubaguydansmith: yeah, only just made that connection, oops15:29
DinaBelova#topic Some hardware to reproduce the issues15:30
DinaBelovakun_huang - your topic, sir15:30
kun_huangoh15:30
kun_huangsince we are talking about issue on performance everyday15:30
DinaBelovamay you please explain what do you mean here? do you have the HW or dod you want to have some?15:30
DinaBelova:)15:30
kun_huangI have some15:31
kun_huangand want to make good use of them15:31
DinaBelovakun_huang wow, that will be simply perfect15:31
DinaBelovaand that will make easier possible issues debug, etc15:31
*** markvoelker has quit IRC15:32
kun_huangso my first question, who need those first15:32
kun_huangI know intel&rackspace have public lab in U.S15:32
kun_huangHas everyone used their resources?15:32
*** mdorman has joined #openstack-performance15:33
*** markvoelker_ has quit IRC15:33
DinaBelovakun_huang - they have pretty big env, yes. but this lab is having competiting schedule between people who want to use it afaik15:33
DinaBelovawe don't use it for now15:34
DinaBelovait or any other big labs15:34
kun_huangDinaBelova: I'll apply some resource from my company15:34
kun_huangat least, my boss support this idea15:34
DinaBelovakun_huang - that is very promising, thanks!15:35
kun_huangand I need write some materials...15:35
kun_huangsome paper work15:35
DinaBelovain case of success - may you please write up some instructions15:35
DinaBelovaoh yeah :)15:35
DinaBelovakun_huang - nobody loves it :)15:35
DinaBelovakun_huang thanks in advance!15:35
*** mdorman has quit IRC15:35
DinaBelovakun_huang - the resources we're using inside mirantis sadly are for mirantis usage only... but we can extend the test plans regarding peoples opinion15:36
DinaBelovaI hope to solve issue with these documents placing with TC15:36
DinaBelovaand then we'll start feedback collection15:36
DinaBelovafrom you and others15:36
DinaBelovareally hope to make this stuff clearer this week15:37
DinaBelovakun_huang - once more time - thanks for your effort15:37
* regXboi wanders in late15:37
DinaBelovaregXboi o/15:37
DinaBelovaregXboi I PROMISE to send an email with +1 hour to the meeting start time suggestion15:37
DinaBelovaI feel people are sufferng15:37
harlowja_at_home:)15:37
kun_huangokay, I'll keep this channel posted with any update15:37
DinaBelovakun_huang thanks!15:38
kun_huangor I need any help15:38
DinaBelovasure, feel free to ping me15:38
* harlowja_at_home is suffering from not enough coffee 15:38
DinaBelovaharlowja_at_home :d15:38
DinaBelovalol15:38
DinaBelova#topic OSProfiler weekly update15:38
kun_huanggood morning guys harlowja_at_home regXboi15:38
DinaBelovak, so let's go to the profiler15:38
* harlowja_at_home coffeeeeeee15:38
regXboiDinaBelova: my problem is that I've got too many meetings stacking up on each other :(15:38
* regXboi skims scrollback15:39
DinaBelovaregXboi, yes, sir :(15:39
DinaBelovathat's the issue15:39
DinaBelovatimeframes comfortable for both US and Europeans are overcrouded15:39
DinaBelovacrowded*15:39
DinaBelova:(15:39
regXboiack15:39
DinaBelovaso going back to the osprofiler - harlowja_at_home - chain https://review.openstack.org/#/c/251343/ is pretty done15:39
DinaBelovaso I need reviews!15:40
DinaBelovalol15:40
harlowja_at_homecool beans, i'll check it out15:40
DinaBelovaand I need you to finish https://review.openstack.org/#/c/246116/ :)15:40
harlowja_at_homeit will be my today mission :)15:40
DinaBelovaso I can play with ELK for 100% here :)15:40
harlowja_at_homeyes ma'm15:40
DinaBelovaharlowja_at_home ack15:40
DinaBelova:)15:40
harlowja_at_homeu are playing with elk?15:40
harlowja_at_homeis that a thing people do in europe?15:40
DinaBelova#action harlowja_at_home review https://review.openstack.org/#/c/251343/15:40
DinaBelovaharlowja_at_home :D15:40
DinaBelovathat's what tough Russian wifes are doing in the meanwhile15:41
DinaBelovalol15:41
harlowja_at_home;)15:41
DinaBelovaso speaking seriously - I want to continue working on this direction15:41
harlowja_at_homedancing with elk (the new movie, based off dancing with wolves)15:41
DinaBelovaof adding more logging and analysing opportunities15:42
DinaBelova:)15:42
harlowja_at_homedef15:42
DinaBelovaso I'm kindly asking you to polish your change15:42
harlowja_at_homesure thing15:42
DinaBelovaand I'll be able to go with ELK here and make some experiments15:42
harlowja_at_home(only if i get to dance with elk to)15:42
DinaBelovaharlowja_at_home thank you sir15:42
DinaBelova:D15:42
DinaBelovaabout spec news15:43
regXboiDinaBelova: is there a plan to extend osprofiler deeper into what it tracks?:)15:43
harlowja_at_homehttps://review.openstack.org/#/c/103825/ also btw, but dims had some questions there...15:43
harlowja_at_homemaybe boris can followup on 103825 (or other person?)15:43
regXboier s/plan/patch/15:43
DinaBelovaregXboi - not patch but plans :)15:43
harlowja_at_home103825 is also somewhat ambiguous about if it wants oslo to adopt osprofiler15:43
harlowja_at_homeit'd be nice if that was like stated (is that a goal?)15:44
DinaBelovaharlowja_at_home - indeed15:44
DinaBelovaharlowja_at_home yes, it is15:44
DinaBelovaBoris is currently communicating with dims about his conserns15:44
harlowja_at_homek15:44
DinaBelovaas right now developing approach is a bit different15:44
DinaBelovafrom what dims is asking about15:44
DinaBelovathe issue is that 2 years ago Boris got -2 on his patch to oslo.messaging and oslo.db15:45
harlowja_at_homeya15:45
* regXboi wonders about decoration15:45
DinaBelovawith that profiling thing15:45
harlowja_at_homeDinaBelova, its been boris life goal to get that merged15:45
harlowja_at_homebefore boris retires he might get it merged15:45
DinaBelovaregXboi - you can use decoration now everywhere already15:45
DinaBelovaharlowja_at_home :D15:45
harlowja_at_homeboris the old, lol15:45
harlowja_at_homehopefully before then it will merge15:46
regXboiDinaBelova: yes, but it's not mentioned that I could see in 10382515:46
DinaBelovaregXboi, hm, lemme check15:46
DinaBelovait was there I believe15:46
DinaBelovaoh15:46
* harlowja_at_home remembers boris trying decoration, people still complain about random crap (like oh decoration will add code... blah blah)15:46
regXboidecoration can be an dependent add on patch15:47
harlowja_at_homeanyways, let's work through these weird issues that people have, and finally get it in (i hope)15:47
DinaBelovaregXboi - it has disappeared15:47
regXboibut folks are proposing decoration profiling in other projects15:47
regXboiso it's silly not to have it here15:47
DinaBelovaregXboi - agreed15:47
regXboibut - in order to get this merged15:47
regXboilet's save that for a follow on :)15:47
DinaBelova#action boris-42 add information about ways of profiling to to 10382515:47
DinaBelovaharlowja_at_home agreed15:47
DinaBelova:D15:48
regXboikeep it short and simple ( the OpenStack KISS :) )15:48
DinaBelova:)15:48
harlowja_at_homemerging before boris retires would be superb to, lol15:48
DinaBelovaok, so anything else here about osprofiler for now?15:48
harlowja_at_homeit needs a dancing with elk logo15:48
DinaBelova:D15:48
regXboiI like that15:48
* harlowja_at_home can't draw though15:48
regXboiand I'd want *that* patch :)15:48
DinaBelovaok, so open discussion15:49
DinaBelova#topic Open Discussion15:49
DinaBelovaand u can joke here :D15:49
* harlowja_at_home i never joke15:49
harlowja_at_homei'm always serious15:49
DinaBelovaharlowja_at_home I suspected that15:49
DinaBelova:d15:49
DinaBelovaok, so any topics to cover?15:50
DinaBelovaideas to share?15:50
harlowja_at_homehttps://en.wikipedia.org/wiki/Dances_with_Wolves (the dancing with wolves movie) btw15:50
DinaBelovapossible work items to add here? https://etherpad.openstack.org/p/perf-zoom-zoom15:50
harlowja_at_homeso about that rally upload results, public website thing15:50
DinaBelovaRobNeff - probably yuo can share something?15:51
manandwe discussed about figuring our ratio of controllers to compute nodes few weeks ago, is that a topic worth discussing?15:51
DinaBelovaharlowja_at_home, yep, sir?15:51
harlowja_at_homedo people think we should try to do that, or save it for later...15:51
DinaBelovamanand - yes, it is15:51
DinaBelovamanand - we just did not have much response yet collected15:51
DinaBelovaprobably I'll need to refresh the discussion pinging some of the operators directly15:51
* harlowja_at_home would really like a way for people to get involved, uploading there useful results with some metadata, and periodically do this, so that we as a community can gather data about each other, and use it for trending, analysis...15:52
DinaBelovaharlowja_at_home - I think it's useful for sure15:52
DinaBelovathe only thing is that it's not only about rally15:52
DinaBelovaanything in fact may land there15:52
RobNeffDo you have a how-to on the Rally Upload yet?15:52
andreykurilin`so about that rally upload results, public website thing` I like this idea15:52
harlowja_at_homeDinaBelova, sure, although kitty-kat pictures hopefully aren't uploaded15:52
DinaBelovaharlowja_at_home :D15:52
DinaBelovaRobNeff - we do not have this web site yet15:53
DinaBelovabut we think it's a good idea15:53
DinaBelovaimportant moment here15:53
DinaBelovarally results mean nothing without the cloud topology shared15:53
DinaBelova:(15:53
DinaBelovaso people need to be open enough to share some of the details15:53
DinaBelovaharlowja_at_home - do you think it'll be possible?15:54
harlowja_at_homeagreed the critical part is open-enough15:54
harlowja_at_homei think we have to start by letting people upload what they can, and we can improve on uploading what is better15:54
harlowja_at_home*with more metadata about there topology...15:54
regXboiDinaBelova: did I forget to point you at https://etherpad.openstack.org/p/hyper-scale ?15:54
DinaBelovaregXboi - yep :)15:54
regXboiand if I did - I'm sorry15:54
DinaBelovawill go through it15:54
regXboithat's open to all to go look at read15:54
harlowja_at_homebut initially i think we need to just get people to upload the basics, and as they get less 'scared' or whatever the can upload more info15:54
DinaBelova#action DinaBelova go through the https://etherpad.openstack.org/p/hyper-scale15:55
*** dims has joined #openstack-performance15:55
regXboithe back half is neutron specific15:55
DinaBelovaregXboi thanks!15:55
regXboibut I'm wondering if the front half would make sense as a devref documentation *somewhere*15:55
DinaBelovaregXboi will take a look, if yes, it'll be really good to do that15:55
DinaBelovaharlowja_at_home - so about the web site - are you the volunteer here? :)15:55
regXboiDinaBelova: if you can suggest a *where*, I'm all ears15:56
harlowja_at_homeDinaBelova, ummm, errr15:56
harlowja_at_homelet me get back to u on that, ha15:56
DinaBelovaregXboi - well, we can grad docs team and shake it a bit to find that out15:56
DinaBelovagrab*15:56
regXboiack15:56
DinaBelova:)15:56
DinaBelovaharlowja_at_home - ok15:56
DinaBelovaharlowja_at_home - simply I'm a bit busy with profiler now and conductor investigations15:57
harlowja_at_home(and elk dancing)15:57
DinaBelovatherefore right now personally I cannot work on that15:57
DinaBelovayeah15:57
harlowja_at_homenp15:57
DinaBelovaso help is super appreciated15:57
DinaBelovaharlowja_at_home :)15:57
harlowja_at_homeunderstood15:57
DinaBelovak, cool15:58
DinaBelovaanything else here?15:58
DinaBelovathanks everyone for hot and productive discussion!15:58
johnthetubaguyklindgren dansmith mriedem DinaBelova: I have attempted to write up my ideas around executor_thread_pool_size in the etherpad: https://etherpad.openstack.org/p/remote-conductor-performance let me know if any of that is unclear.15:58
DinaBelovajohnthetubaguy thank you sir!15:58
dansmithjohnthetubaguy: cool15:58
regXboiDinaBelova: I'm pinging somebody I know in the docs project right now15:58
DinaBelovaok, cool15:58
DinaBelovabuy!15:59
DinaBelova#endmeeting15:59
openstackMeeting ended Tue Dec  1 15:59:08 2015 UTC.  Information about MeetBot at http://wiki.debian.org/MeetBot . (v 0.1.4)15:59
openstackMinutes:        http://eavesdrop.openstack.org/meetings/performance_team/2015/performance_team.2015-12-01-15.00.html15:59
openstackMinutes (text): http://eavesdrop.openstack.org/meetings/performance_team/2015/performance_team.2015-12-01-15.00.txt15:59
openstackLog:            http://eavesdrop.openstack.org/meetings/performance_team/2015/performance_team.2015-12-01-15.00.log.html15:59
*** Jeff__ has quit IRC15:59
*** harlowja_at_home has quit IRC15:59
mriedemdansmith: johnthetubaguy: i see a few things in the n-api-meta flow that could be optimized, mostly around lazy-loading instance object fields, not sure if it would help much though16:00
mriedembut we are loading up security groups, metadata and system_metadata after getting the instance object16:00
*** markvoelker has joined #openstack-performance16:00
dansmithmriedem:  we checked for lazy loads in his logs I think16:00
dansmithand found only a few16:00
mriedemseems we should just join up front when getting hte instance from the db16:00
klindgrenand once again /me meeting fail16:00
mriedemdansmith: the SecurityGroupList is retrieved separately, rather than via the instance16:01
mriedemso that's 3 fields that could be joined up front on the instance get query16:01
dansmithmriedem: you mean not lazy loaded via objects, just separately-loaded?16:01
johnthetubaguymriedem: seems worth a try16:01
mriedemright16:01
mriedemi can write up a change to see how it looks16:01
*** rohanion has quit IRC16:02
johnthetubaguybut it does sounds like the nova-conductor is saturated (not in a CPU resources sense, in a possible through put sense) which is slowing everything down16:02
mriedemthe metadata service also proxies back to neutron and we could be smarter in the neutron api code wrt filtering out fields that we don't need in the response, but that would probably not even be noticeable, just cuts down on the amount of content passed around16:02
mriedemi was trying to figure out if the neutron-metadata-agent was polling the n-api-meta service,16:03
mriedembut not getting any feedback on that in the -neutron channel16:03
mriedemthe q-meta agent is doing something every 30 seconds http://logs.openstack.org/43/251543/1/check/gate-tempest-dsvm-neutron-large-ops/eb34603/logs/screen-q-meta.txt.gz#_2015-11-30_22_06_48_07516:03
mriedembut that could just be health checking with the server, idk16:04
DinaBelovaklindgren o/16:05
DinaBelovamriedem - personally I was not able to reproduce enough load via nova-metadata calls, but that leads simple to the API service saturaiton16:06
DinaBelovanot the metadata16:06
DinaBelova:(16:07
DinaBelovathat's why I ended up with asking klindgren to make some dumps16:07
DinaBelovaklindgren - please got through todays nova-conductor section in the logs16:08
*** dims_ has joined #openstack-performance16:08
klindgrenyep - will do16:08
johnthetubaguyklindgren: sorry, if folks already asked, but do you know what db driver you are using here? mysql-python or pymysql?16:09
mriedemmysql-python16:09
DinaBelovaMySQL-python==1.2.316:09
johnthetubaguymriedem: sorry, I thought that was an open question still, my bad16:10
*** dims has quit IRC16:10
DinaBelovajohnthetubaguy - nope, that's listed in the etherpad you added info to :)16:10
DinaBelovaklindgren - so please take a look on the lines johnthetubaguy has added to the https://etherpad.openstack.org/p/remote-conductor-performance too16:12
DinaBelovathanks :)16:12
johnthetubaguyDinaBelova: I see it now, thanks16:12
johnthetubaguymriedem: added a note on the rackspace comments, I don't think we have tried using pymysql, I think thats actually a different thing they were talking about switching back to16:14
mriedemjohnthetubaguy: yeah, i talked to alaski about that a few weeks ago,16:14
mriedemi don't think rax moved to that yet b/c the direct to sql code depends on mysql-python16:15
klindgrenso re: adding more workers are you talkign about adding more physical boxes?  Or are you talking about setting workers=xx?16:16
alaskimriedem: that's part of it.  we don't use the mysql db api everywhere though, so we could try it in some places16:16
alaskithough where we're hitting the most performance issues we are running mysqldb, so that's really where we would want to try it16:17
mriedemklindgren: do you see anything in the neutron-metadata-agent polling the n-api-meta service?16:17
klindgrenmriedem, we don't run the neutron-metadata-agent16:17
klindgrenwe run metadata on the compute nodes directly16:17
johnthetubaguyklindgren: there were to things, greenlet threads and workers, covered it more in the etherpad16:17
klindgrensince we run flat/dhcp networks16:18
klindgrenwe just bind 169.254.169.254 to loopback and the normal iptables rules route metadata requests locally, occasionaly it will fall back to another server for metadata - but 99% of the time it using its local metadata instance - when I checked.16:19
johnthetubaguyklindgren: have you tried executor_thread_pool_size=2 and rpc_conn_pool_size=2 in an attempt to reduce the CPU usage for each worker process?16:24
klindgrenwe haven't I will do that and see what it does16:24
johnthetubaguyklindgren: awesome, that will be an interesting test, certainly seen that help with nova-scheduler in the past (although the caching scheduler sort of removes the need for that)16:25
klindgrenunder juno we use to run workers=40 on two boxes and never had cpu alarms, in kilo we started getting 100% cpu alarms.  I adjusted the workers down a bit to releive pressure on the boxes, and we added a third physical server.16:26
klindgrenadding the third box jsut chewed up 100% cpu on that box and did nothing to relieve the cpu usage on the other two servers16:26
johnthetubaguyklindgren: what did the rabbit queue lengths look like during this?16:27
johnthetubaguyfor conductor16:27
klindgrenI can try starting up multiple conductor workers in differnt complete processes too and see if that makes everything better.  IE have 2 worker threads but 10 conductor services16:28
klindgrenI can check generally we run with pretty much no queue lengths - except for notifications.info where we have to do work in external systems that are not super fast16:28
johnthetubaguyklindgren: I guess I assumed these were separate processes, good point16:28
klindgrenwe do see ~500-1k messages/second in this cell16:29
klindgrentypically in the 500 range16:30
*** dims has joined #openstack-performance16:30
*** harlowja_at_home has joined #openstack-performance16:31
*** dims_ has quit IRC16:33
johnthetubaguyklindgren: thats for the extra context, it does sound like a regression, beyond the stuff I am suggesting, although interested to see if they help16:40
klindgrenyea - I am starting some testing/cahgnes now - will keep the channel/ehterpad updated16:40
*** RobNeff has quit IRC16:47
*** arnoldje has quit IRC16:53
*** mriedem is now known as mriedem_meeting16:58
*** arnoldje has joined #openstack-performance17:17
*** amaretskiy has quit IRC17:37
*** mriedem_meeting is now known as mriedem17:43
*** RobNeff has joined #openstack-performance17:45
*** harlowja_at_home has quit IRC17:51
*** RobNeff has quit IRC18:04
*** ctrath has quit IRC18:06
*** ctrath has joined #openstack-performance18:13
*** pglass has quit IRC18:25
*** arnoldje has quit IRC18:28
*** arnoldje has joined #openstack-performance18:49
*** ctrath has quit IRC18:55
*** mriedem has quit IRC18:56
*** pglass has joined #openstack-performance18:58
*** mriedem has joined #openstack-performance19:01
*** ctrath1 has joined #openstack-performance19:03
*** AugieMena has quit IRC19:38
*** ctrath1 has quit IRC19:56
*** ctrath has joined #openstack-performance19:58
*** pglass has quit IRC20:01
*** ctrath has quit IRC20:03
*** arnoldje has quit IRC20:03
*** ctrath has joined #openstack-performance20:11
*** arnoldje has joined #openstack-performance20:24
harlowjaDinaBelova ok, i added some hopefully useful feedback on https://review.openstack.org/#/c/247005/20:30
*** rfolco has quit IRC20:31
harlowjaDinaBelova what tooz does20:31
harlowjahttps://github.com/openstack/tooz/blob/master/tooz/coordination.py#L40720:31
harlowja(taskflow is a little more complicated than that, but similar)20:31
harlowjaso if we are going to do connection_url for things, imho we might as well do something similar...20:32
*** SpamapS is now known as TheKettle21:04
*** TheKettle is now known as SpamapS21:04
*** ctrath1 has joined #openstack-performance21:12
*** ctrath has quit IRC21:15
*** nihilifer has quit IRC21:25
*** nihilifer has joined #openstack-performance21:27
*** harlowja has quit IRC21:27
*** harlowja has joined #openstack-performance21:28
klindgrenDinaBelova, so just for refrence we see ~130-150 messages/s on the conductor queue but a backlog of 022:09
klindgrenjohnthetubaguy, I set executor_thread_pool_size = 5 and rpc_request_woekres to 222:09
klindgrenno real change22:09
*** arnoldje has quit IRC22:12
klindgrenthe biggest changes in cpu consumption was setting metadata cache timeout to 3 minutes, turning off ssl to rabbitmq and adding in that change from dansmith to stop doing flavor migrations22:13
harlowjaklindgren did u try pymysql ?22:17
klindgrennewp22:17
klindgrennot yet22:17
harlowjak22:17
harlowjait may help (or not)22:17
klindgrenstarting with the easier win's/attempts22:17
klindgrenand letthign those bake fore a while to see what we see22:17
klindgrenletting*22:18
harlowjak22:19
harlowjafair nuff22:19
*** harlowja has quit IRC22:19
*** harlowja has joined #openstack-performance22:21
*** arnoldje has joined #openstack-performance22:32
klindgrentrying the running multiple separate nova-conductor processes - currently running 4 separate with 5 workers22:39
*** ctrath1 has quit IRC22:52
*** harlowja has quit IRC22:57
*** mriedem has quit IRC22:58
*** ctrath has joined #openstack-performance22:59
*** harlowja has joined #openstack-performance23:01
*** regXboi has quit IRC23:02
*** dimtruck is now known as zz_dimtruck23:10
*** arnoldje has quit IRC23:24
*** manand has quit IRC23:37
klindgrenalso no real change in cpu usage23:38

Generated by irclog2html.py 2.14.0 by Marius Gedminas - find it at mg.pov.lt!