21:00:21 #startmeeting scientific-sig 21:00:22 Meeting started Tue Jul 9 21:00:21 2019 UTC and is due to finish in 60 minutes. The chair is oneswig. Information about MeetBot at http://wiki.debian.org/MeetBot. 21:00:24 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 21:00:26 The meeting name has been set to 'scientific_sig' 21:00:44 hello world? 21:00:56 #link agenda for today https://wiki.openstack.org/wiki/Scientific_SIG#IRC_Meeting_July_9th_2019 21:00:59 hello 21:01:14 Hey Bob, how's the Cinder work going? 21:02:08 Going good 21:02:20 As this meeting is running, I'm paring back OSDs from a Ceph cluster, one by one... 21:02:22 Migrated everything off GPFS onto our new NetApp filers 21:02:23 :) 21:02:35 NetApp, fancy... 21:02:47 I’ll be very happy to get off the broken GPFS drivers for Cinder 21:03:03 How are they breaking for you? 21:03:09 Showing signs of age? 21:03:20 bugs that haven’t been patched for quite some time 21:03:30 cinder retype is has been broken since Liberty 21:03:39 so we couldn’t live migrate out backend storage 21:04:19 That somewhat negates the advantage of network filesystems... 21:04:44 the GPFS driver doesn’t properly update all the UUIDs so you get orphaned and broken UUID entries in you database (among other problems) 21:05:39 That sounds unfortunate. Is there nobody outside of IBM maintaining it (and nobody inside IBM either)? 21:05:50 not sure 21:05:57 i’ve been patching things myself as needed 21:06:09 Live migration was broken for awhile as well 21:06:41 So today's a good day overall... 21:06:48 oh yes 21:06:51 manually patched Nova for that until patched in libvirt 5.1 (which isn’t out for CentOS yet) 21:07:05 it’ll be nice to have things work smoothly :) 21:07:35 already seeing the benefits of dedup as we speak 21:07:45 is it just us today? 21:08:13 so far... trandles sent his apologies 21:08:31 martial also 21:08:35 gotcha 21:08:57 i have to jet about halfway to pick up my son 21:09:05 No problem 21:09:11 I've been working on an interesting problem with CephFS which might amuse you 21:09:56 I've got a Slurm deployment with compute partitions in Cambridge, Oxfordshire and Edinburgh with Manchester coming up. 21:10:17 nice 21:10:29 how are you liking CephFS? 21:10:39 I don't have a filer in any of those places so I made a scalable home dir using CephFS converged on the compute nodes 21:10:44 CephFS has potential. 21:11:10 It's deficient in some ways but does some remarkable things occasionally 21:11:24 did you guys start playing with ceph on that surplus hardware? 21:11:37 For this system, I can make directories with local IO, eg /cambridge, /edinburgh etc. 21:11:53 a copy from one to the other sees the data jetting off up the country... 21:11:58 Hi jmlowe 21:12:01 hi 21:12:10 surplus hardware? 21:12:27 Was that to rbudden? 21:12:45 I think Goddard had some extra hardware rbudden was going to try some ceph experiments on 21:12:57 This project was more about making do with a shortage of hardware... 21:13:05 jmlowe: we haven’t yet, but it’s on the radar for possibly our TDS 21:13:38 jmlowe: today was Cinder backend from GPFS -> NetApp (mentioned earlier, before you joined) 21:13:53 right, I figured 21:15:51 Are you involved in OSC's OpenOnDemand? I've heard some good things about it. 21:17:10 I know PSC helped test OOD while they were developing it 21:17:26 we have some potential interest at Goddard 21:17:35 for Jupyter notebooks if I recall 21:17:56 i don’t have much hands on with it personally 21:19:52 Be interesting to hear if there's a broader pulse in the project. 21:20:41 jmlowe: rbudden: I forget, is either of you planning to go to Shanghai? 21:21:12 no way I can swing that 21:21:42 no 21:21:56 Ah, too bad. 21:22:26 any word on the 2020 North American Summit? 21:22:28 eh, i wouldn’t want to deal with the paperwork 21:22:31 ;) 21:22:40 yeah, any news? i was curious if it was announced 21:23:11 The rate things are going, not sure if Brits will be allowed in the States by then :-) 21:23:30 I haven't heard anything, to answer your question 21:24:00 lol 21:24:10 China is just off limits completely… 21:24:23 other internation is possible, just more involved 21:24:34 hoping for somewhere nice for 2020! 21:25:07 I can't use NSF funds to travel internationally so I have to use IU funds, comes with a different set of strings attached 21:26:12 The freedom of the private company is a double-edged sword :-) 21:26:19 hi gang 21:26:45 apologies for being tardy - school holiday head-f*&kery here :-/ 21:26:54 and i' 21:26:57 Morning b1airo, glad you made it 21:27:02 #chair b1airo 21:27:03 Current chairs: b1airo oneswig 21:27:10 i'm about to join a face to face, so will follow along from sidelines 21:27:38 Actually we didn't have much to cover - rbudden was talking Cinder migration and I was talking CephFS 21:27:55 How's your Bright system? 21:27:56 how'd you go with that CephFS metadata conundrum? 21:28:03 urgh 21:28:17 Aha, CephFS MDS - had a breakthrough from Dan van der Ster today. 21:28:45 So, it turns out you can pin metadata service to specific directories and MDS instances 21:28:48 nice. i assume you were already sharding the dirs? 21:29:00 It's not documented anywhere apart from a blog post 21:29:18 yes, was going to ask whether you'd tried that last week, but was too slow before you left 21:29:51 directories have affinity to local OSDs in the same region. 21:30:10 metadata was really painful but Dan's given me some good ideas on that. 21:30:22 oh, so you've got multiple data pools under this too? 21:30:52 yup, it's across multiple sites 21:31:03 #link CephFS subtree pinning https://ceph.com/community/new-luminous-cephfs-subtree-pinning/ 21:31:19 yeah i'd say pinning is a good start, i think Dan found the auto-balancing is still a bit iffy 21:32:05 Are you using Ceph at nesi? 21:32:22 no not at the moment 21:32:26 sorry to interrupt, gotta jet to pick up my son. have a good day/evening everyone! 21:32:30 a couple of people keen on trying it though 21:32:34 same to you rbudden 21:33:17 b1airo: so what's the OpenStack takeover looking like now? 21:35:18 BTW did you see the news on the Zoom vulnerability? 21:35:43 a slow burn i'm afraid. we're taking some slow steps, but won't have much latitude to make major changes/inroads until we setup NeSI.3, which has been pushed out two years by the Gov 21:36:11 yes, only impacts people using Apple Mail client though i believe? 21:36:34 I believe so. Which is unfortunate for me... 21:37:03 i figured you'd be using an emacs plugin... ;-) 21:37:32 ha - emacs... great OS, shame about the editor 21:38:05 careful with that trash talk around here o_0 21:38:28 M-x undo! 21:38:50 dad joke! 21:41:46 anyway - we are pretty quiet on content today. I'm thinking of other OS-related things that have happened here this week. I've been down the rabbithole with some weird fluentd behaviour. Not very exciting 21:44:04 Any other business? 21:46:03 OK, let's close for the day - thanks guys 21:46:07 #endmeeting