21:00:03 #startmeeting swift 21:00:04 Meeting started Wed Mar 24 21:00:03 2021 UTC and is due to finish in 60 minutes. The chair is timburke__. Information about MeetBot at http://wiki.debian.org/MeetBot. 21:00:05 Useful Commands: #action #agreed #help #info #idea #link #topic #startvote. 21:00:07 The meeting name has been set to 'swift' 21:00:10 who's here for the swift meeting? 21:00:20 o/ 21:00:27 o/ 21:01:03 o/ 21:01:26 o/ 21:03:42 o/ 21:03:47 as usual, the agenda's at https://wiki.openstack.org/wiki/Meetings/Swift 21:03:51 first up 21:03:55 #topic PTG 21:04:27 if you haven't yet, please register and respond on the doodle poll for meeting times 21:04:37 #link https://april2021-ptg.eventbrite.com/ 21:04:42 #link https://doodle.com/poll/cnr3qb7y2afiad9v 21:05:39 and we've got an etherpad for gathering topics 21:05:41 #link https://etherpad.opendev.org/p/swift-ptg-xena 21:06:05 i really ought to book (virtual) rooms and such today 21:06:41 then doodle poll needs to be a priority.. so please do it if you haven't yet! 21:08:08 any questions or comments on it? less than a month away now! 21:11:01 all right then 21:11:08 on to some updates! 21:11:12 #topic relinker 21:11:41 so we did another part power at nvidia! just finished up this morning 21:12:02 went better than the last time -- we're definitely getting better at it 21:12:16 but there are still some rough edges 21:14:00 notably, we saw a bunch of extra (empty) handoff partitions getting created because we ran the relinker more than once, and it'd want to do a rehash on a partition that never got any links put in it 21:14:59 i think https://review.opendev.org/c/openstack/swift/+/781908 should deal with it -- basically, if we call get_hashes for a partition that doesn't exist, we should be able to return early 21:16:55 since last week, we had a bunch of great stuff land -- operators can now target specific policies and even partitions, and we've merged acoles's great refactoring to make relink and cleanup reuse a lot of the same code 21:18:21 still to come, mattoliverau has some more improvements to progress visibility, and i rebased my patch to parallelize the relinker 21:18:32 nice 21:19:06 if you're interested in part power increases, those are likely the best places to jump in! 21:19:14 cool. the relinker is getting better 21:19:23 So how high are you now? 21:19:32 And I don't mean marijuana 21:19:49 only 18. we've done it twice, and intend to do at least two more 21:19:50 we also learnt some stuff doing this increase after a previous one. For example, we discovered we had some tombstones that had not been cleaned up in their original partitions, so during this relink we had *two* locations trying to create a link in the same new location 21:20:04 Hmm. So even 18 is not enough for you? 21:20:13 we just love doing it :D 21:20:17 (speaking of, https://review.opendev.org/c/openstack/swift/+/781591 is a first crack at being able to step by more than one) 21:20:21 That's like 262000 partitions 21:20:58 if you assume 10k disks, that's still only 26/disk ;-) 21:21:24 oh cool, jump more then 1, interesting 21:22:21 one question for relink. in my case, I forgot to increase reclaim age before the relink, so there are a lots of cleanup job is failed.. 21:22:59 I fi 21:23:42 as long as the failures are for tombstones (.ts files) i expect it should be mostly fine 21:24:30 I finished the part power job and the object replicator is running again. The failed diskfiles will be delete? Most of diskfiles are .data and it was deleted in other node during relink 21:27:11 the .datas -- are the objects still in listings? 21:27:36 nop. It is not in list 21:28:23 i think those will need to be manually cleaned up :-( 21:28:58 though zaitcev's dark data audit watcher could come in handy when you upgrade 21:30:02 It does not reattach them though 21:30:06 Can only delete 21:30:37 But what the heck happened? Relinker misfired, or they were orphans even before partition power increase? 21:30:39 fwiw, newer versions also make it so the relinker can respect the cluster's reclaim_age by pointing it at object-server.conf 21:31:04 Oh, n/m I see the comment about reclaim age. 21:31:09 sounds like they were deletes that weren't propergated fully so delete makes sense. 21:32:23 i'm guessing there was a delete after the increased ring went out but before the cleanup processed the partition. then a reclaim age passed. cleanup finally got to it, went to open the diskfile in the new location -- and cleaned up the tombstone before erroring 21:33:14 ...makes me think the relinker (when not given a config file) should use a reclaim_age that's intentionally longer than anyone would likely have it configured... 21:33:17 That's going to be the first time ever watcher got useful to anyone then. 21:33:28 If seongsoocho uses it. 21:33:34 yes. during the relink , the object was deleted in other 2 nodes but not deleted in one node (because relinker use that disk too heavy , so object server return error) 21:33:41 peace of mind is still a valid use-case ;-) 21:33:56 zaitcev: cool. 21:34:48 all right 21:34:57 #topic sharding and shrinking 21:35:17 i know mattoliverau and acoles have continued working on this 21:35:44 I've been playing with al's repair and anaylze chain, and it's pretty awesome. 21:36:08 and mattoliverau, you put together an etherpad about how to tell that a shard is old enough to consider shrinking, yeah? 21:36:10 #link https://etherpad.opendev.org/p/swift-shrinking-active-age 21:36:26 oh yeah. The current shrinking blocker is ^ 21:36:53 zaitcev pulled me up on some __hash__ kludge in that for which I am grateful - made it better 21:37:46 I finally got some ideas about tombstone reporting up https://review.opendev.org/c/openstack/swift/+/782832 - very much WIP though 21:38:50 there is an edge case that one primary is small and shards before the others, the new shards ranges become active and then being pretty empty will be considered candidates for shrinking. The etherpad explains current progress, problems, pain points. 21:39:34 turns out what we thought should be an easy fix seems to get more and more complicated :P 21:39:35 mattoliverau: thanks for starting the etherpad 21:40:08 So if your interested, read, ask questions and comment in the etherpad. 21:40:11 mattoliverau: that's ditributed systems for you! ;) 21:40:19 lol yeah :) 21:40:55 in the meantime I wanted to give Als' repair stuff some love, because they'd also be great to have and have been up a long time. 21:41:38 I plugged them into my WIP shard graph visualiser tool and now you can get a green path for the best path. Which is nice when testing als changes. 21:41:39 sounds good. we'll be able to fix up busted sharding yet! 21:41:50 #topic open discussion 21:41:58 what else should we bring up this week? 21:42:19 ah 21:42:33 one thing, it is not actually related to Swift 21:43:10 before a week from PTG, NVIDIA GTC will be held at online and I have a session to introduce my recent work. 21:43:13 just FYI. 21:43:24 kota_: great! 21:43:29 cool! congrats! 21:43:40 kota_: cool!!!! 21:43:43 do you have a link / date/ time ? 21:43:48 nice! 21:43:50 it's free so you can watch the pre-recorded video 21:44:19 the registration is at https://gtc21.event.nvidia.com/ 21:45:12 it looks like the direct link to the video is not work unless someone registered to see. 21:45:14 kota_: thanks for sharing that 21:45:22 :D 21:45:26 kota_: thanks 21:45:29 +1 21:45:55 Although I'm a little sick of everything remote. 21:46:25 exactly, I like the face to face meeting we did in the past. 21:46:32 yeah, me too. with a little luck, maybe we'll be able to actually see each other in person the next time 21:46:37 yeah. but at least it makes it easier to attend conferences you might not be able to convince someone to send you too.. esp if your on the otherside of the world :) 21:47:20 but +100 for missing going to f2f conferences 21:49:14 all right, i think i'll call it then 21:49:25 thank you all for coming, and thank you for working on swift! 21:49:29 #endmeeting