11:00:30 #startmeeting scientific-wg 11:00:31 Meeting started Wed Jul 5 11:00:30 2017 UTC and is due to finish in 60 minutes. The chair is oneswig. Information about MeetBot at http://wiki.debian.org/MeetBot. 11:00:32 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 11:00:34 The meeting name has been set to 'scientific_wg' 11:00:46 #link Agenda for today is here https://wiki.openstack.org/wiki/Scientific_working_group#IRC_Meeting_July_5th_2017 11:01:03 ... Who noticed the time change? 11:01:24 evening 11:01:27 #chair b1airo 11:01:28 Current chairs: b1airo oneswig 11:01:34 Hi b1airo 11:01:42 Hi, Tim joining 11:01:51 Hello! 11:01:56 Hi noggin143 priteau 11:02:01 well remembered :-) 11:02:25 We've got a shortish agenda for today 11:02:37 #topic Supercomputing BoF 11:02:51 Good afternoon. 11:03:00 First item was that the submission deadline is approaching for BoFs at SC 11:03:13 Which is in Denver the week after the Sydney summit 11:03:24 b1airo: do you have the deadline to hand? 11:04:06 At SC2016 we did a panel session which pretty much filled a big room 11:05:28 The question to discuss is, is there appetite for putting something on for SC again? 11:05:32 end of July off the top of my head... 11:05:50 oneswig: was there something at ISC? 11:06:01 verdurin: nothing official AFAIK 11:06:10 My colleague John attended, I did no. 11:06:25 yep 31st July 11:06:44 i would like to do something at ISC next year 11:07:11 b1airo: me too 11:07:38 One action on SC - I think we should check with hogepodge to see if he's returning. 11:08:06 There's also the possibility of a second edition of the book we wrote, which was very popular at the conference 11:08:10 yes good idea 11:08:50 #action oneswig to follow up on book and foundation attendance for SC 11:08:59 so who, if anyone, is planning to attend SC17 ? 11:09:23 I might - its on the way home from Sydney for me 11:09:30 hahaha 11:09:37 Less so for you b1airo :-) 11:09:58 at least the Summit will be short, so i can go home in between 11:10:36 re: openstack summit - I had a look at submitting papers, are the categories all changed from last time? 11:11:11 b1airo: I think most WG members at SC16 were from the US time zone, we should follow up with this next week 11:11:26 i haven't looked yet to be honest, don't really think i've got enough new material to warrant speaking 11:11:43 Can't say yet whether I'll be at SC. 11:11:50 wondered if the new format meant fewer categories 11:12:08 the TPC was interesting 11:12:16 TPC? 11:12:27 technical programme committee 11:12:33 for SC17 I mean 11:13:00 submissions already selected for presentations? 11:13:44 mostly 11:13:44 I won't be at SC. I will ask Kate if she'll be there, she generally attends. 11:14:05 priteau: she's involved in running it isn't she? 11:14:12 there was not a huge amount of competition from great papers to be frank 11:14:34 b1airo: what was the category? 11:14:37 only a handful accepted for the clouds and distributed computing stream 11:15:11 i'm still working on shepherding one of them, but that's turning out ok now 11:15:15 oneswig: she's chair of the Clouds & Distributed Computing committee 11:16:01 So it's very likely she'll be there 11:16:10 yeah i expect kate would be at the conference itself given she is chair, but i suppose you never know - she did say her travel diary was pretty crazy 11:16:27 b1airo: after seeing a few talks last time I assumed the bar was pretty high, perhaps it's worth having a go next time round 11:16:47 oneswig, yes i think so 11:17:07 noted for 2018 :-) 11:17:18 probably a practice and experience paper 11:17:51 might be just the thing to get that Maldives meeting lined up ;-) 11:18:25 Oh, if it'll get us to the Maldives I'll do it!! 11:19:06 noggin143, what is happening at CERN? i see luminosity is surpassing expectations since the restart 11:19:21 yup, running great 11:19:37 we’re getting some more tape in stock 11:19:59 also just got ironic working 11:20:15 OK so to sum up on this, I think there's scope for some activity (including possibly a BoF) but we should follow up with the Americas session next week. 11:20:31 noggin143: in a cell or a separate deploy? 11:20:44 cool! i have been wanting to ask you guys about that as i thought i'd heard you had it going with cellsV1 and was interested to know if there were any gremlins there 11:20:47 in a cell, in the production cloud 11:20:59 we’re also just starting testing cells v2 with upstream 11:22:01 is there a blog or some such to follow, or maybe we can get belmiro along to talk about it..? 11:22:39 I’m sure belmiro will share the details, although he’s a bit busy with the new arrival at home 11:22:41 oh, belmoreira, are you here already? 11:22:44 We recently saw some oddity relating to the timing of Ironic registering hypervisors and populating a cell 11:23:31 did he get a new gaming rig...? :-P 11:23:58 b1airo: too good! :-) 11:24:15 OK - want to talk THP? 11:24:28 #topic b1airo's issues with THP 11:25:03 i don't have a great deal of detail on this yet sorry but maybe someone will just know what our problem is 11:25:06 here's hoping! 11:25:26 What are the symptoms? 11:25:45 basically we have noticed on some of our virtualised HPC nodes that certain jobs sometimes run very slowly 11:26:17 the guests are CentOS7 backed by hugepages (either static or THP) on Ubuntu hypervisors (4.4 kernel) 11:26:34 the THP issues are inside the guest itself 11:27:05 the host THP usage stays steady after the guest is started 11:27:12 b1airo: only with CentOS 7 guests, so a wide disparity in guest-host kernels? 11:28:12 so far we only have data for CentOS7 guests as we don't yet have a repeatable trigger and so have not been able to narrow things down or look at kernel bisections etc 11:28:39 in the guests, when performance goes bad we see very high system time 11:29:32 and when khugepaged runs periodically it uses noticeably more CPU than normal 11:29:43 Is there some kind of steady-state activity that can provoke this? What I'm getting at is stochastic kernel profiling using perf - http://www.brendangregg.com/FlameGraphs/cpuflamegraphs.html 11:30:18 this seems to be related to THP defrag / memory compaction, as disabled defrag makes the high system time disappear 11:30:52 oneswig, perf is a good suggestion 11:31:12 only problem is, you need to provoke the problem continuously for (say) a minute 11:31:17 to get a sample 11:31:37 as we're suspecting some kind of memory fragmentation issue we've started collecting into influx (with grafana bolted on) vmstat and buddyinfo data 11:32:00 for all compute guests 11:32:35 we also see a high rate of compact_fail in the THP counters 11:33:18 What are the causes of failure - any thoughts? 11:33:39 it seems like it could be some interaction with pagecache as when we drop_caches the compact_fail stops the dramatic increase and hugepage count goes up again 11:34:16 so far we haven't seen any obvious job/application patterns in the slurm logs 11:34:54 but i note that THP only applies to anonymous memory, so currently linux does not hugepage-ify pagecache 11:36:22 the one thing we do have is an application that runs dramatically faster with hugepage backing in the guest (as well as the hypervisor), so we do have a test-case to detect problems that should finish in 1min if working properly 11:36:37 Does it count as a failure if there's insufficient contiguous physical memory to create a hugepage? 11:37:18 oneswig, that's a great question - i have not had a lot of luck finding murky details of the compaction algorithm 11:37:59 there is an LWN article from about 2010 that talks about it generally but it is not in relation to THP 11:38:11 Hopefully, failure is nothing more than that - no room in the inn 11:38:33 and might explain why evicting caches indirectly helps 11:38:52 in the default CentOS7 (and maybe RHEL7) guest setup /sys/kernel/mm/transparent_hugepage/defrag is enabled 11:39:53 which means that the vm (as in linux virtual memory management, not virtual machine) will perform defrag on-demand to allocate a hugepage in response to a page fault 11:40:54 i think that is why we see the system time drop off when 'echo never > /sys/kernel/mm/transparent_hugepage/defrag', as we stop slowing down page faults and just fall back to using 4k pages 11:41:04 I wonder if a paravirtualised guest sees some data passed through on memory mapping, similar to numa awareness 11:41:32 hmm no idea 11:42:05 I'm trying to picture these interactions between the page mapping of host and guest and not doing very well... 11:42:38 i'm still wondering if there is not a much simpler way to do guest to physical memory mappings than the 2-dimensional method used today 11:42:51 I think what you've got here is a top-drawer presentation subject for Sydney - if you get to the bottom of it 11:43:46 Keep us updated? 11:44:08 if you can assume a guest's memory is mlocked (reasonable for high-performance use-cases) then it seems like you can reduce the guest page lockup to an offset 11:44:54 but i'm sure that's a stupidly naive idea and if i find the right mailing list someone will suitably shoot me down :-) 11:45:14 oneswig, fingers crossed! 11:45:58 Should we move on? 11:46:04 we'll be trying to build some nice graphs to visualise it more and hopefully i can share something pictorial next week 11:46:10 yep 11:46:16 Thanks b1airo, be really interesting to see what you find 11:46:31 #topic scientific application catalogue 11:46:52 OK, I had an action to investigate Murano on tripleO 11:46:59 verdurin: you did too - any luch? 11:47:05 I mean luck? 11:47:24 I got my TripleO deploy to the point where I can add Murano, that's all I managed alas 11:48:11 I did however have discussions with several others who were also interested in this kind of service 11:49:19 verdurin: you there Adam? 11:49:32 we’re looking at Murano this summer. Given that the OpenStack application catalog at openstack.org is being retired, it would be interesting to share some scientific apps 11:49:57 is it, totally missed that - how come? 11:50:55 it was competing with other repositories like bitnami and dockerhub 11:51:15 noggin143: is the CERN investigation being done in a way that we can participate in? 11:51:52 it’s a summer student project for the next few weeks and they will do a writeup once it is done. 11:52:25 Back now oneswig 11:52:28 Sounds useful, let's keep it on the agenda. 11:52:39 Hi verdurin - any luck with you? 11:53:10 oneswig: not yet. Hopefully very soon. 11:53:26 me too... it's a three horse race :-) 11:53:59 noggin143: is there an intern assigned for this work? 11:54:08 we could start a therad 11:54:09 thread 11:54:16 noggin143, so was the reason to do with those other places having larger contributor communities and so the openstack catalog looking bad because it is so sparse ? 11:54:58 and because it was viewed as a competitive solution. Better to integrate with other communities than have our own 11:55:40 The mailing list thread about the app catalog is http://lists.openstack.org/pipermail/openstack-dev/2017-March/113362.html 11:55:42 is there an alternative that would allow us to share murano apps for example? 11:55:50 thanks priteau 11:56:29 I think we’re probably best off just setting up a github repo and the syncing to it 11:56:30 it predates my recent decision to try and read more of openstack-dev again 11:56:50 b1airo: the NSF guys had a Murano catalog in draft form 11:57:12 and were sharing packagings on that. Don't think it went mainstream but we should pursue that. 11:57:35 NSF work sounds interesting. 11:58:10 #link presentation here https://www.openstack.org/videos/barcelona-2016/image-is-everything-dynamic-hpc-vm-repositories-using-murano 11:59:09 Murano on Nectar is getting a steady flow of contributions now, i haven't actually looked at our index in a while but i see the "new murano package" tickets come in for review, maybe one every fortnight at the moment 11:59:19 Anything more - AOB? 11:59:47 there is a relevant ML thread on rebranding WGs to SIGs 11:59:47 b1airo: sounds like it has enough critical mass on Nectar to be self-sustaining 12:00:14 we would certainly love to share with other science clouds 12:00:30 love to hear more about it, b1airo 12:00:52 out of time... thanks everyone 12:01:13 cheers all, goodnight! 12:01:20 Bye 12:01:24 #endmeeting