Wednesday, 2021-01-13

*** tosky has quit IRC00:01
*** ianychoi has joined #openstack-meeting00:14
*** macz_ has quit IRC00:15
*** zbr has quit IRC00:22
*** zbr has joined #openstack-meeting00:22
*** bbowen has quit IRC00:36
*** bbowen has joined #openstack-meeting00:37
*** mlavalle has quit IRC00:47
*** jmasud has joined #openstack-meeting00:52
*** kevinz has joined #openstack-meeting01:03
*** jmasud has quit IRC01:08
*** jamesmcarthur has quit IRC01:13
*** jamesmcarthur has joined #openstack-meeting01:15
*** icey has quit IRC01:15
*** icey has joined #openstack-meeting01:16
*** jmasud has joined #openstack-meeting01:19
*** jamesmcarthur has quit IRC01:36
*** jamesmcarthur has joined #openstack-meeting01:38
*** dklyle has quit IRC01:40
*** jamesmcarthur has quit IRC01:43
*** armax has quit IRC01:47
*** jamesmcarthur has joined #openstack-meeting01:48
*** dklyle has joined #openstack-meeting01:48
*** baojg has joined #openstack-meeting01:52
*** armax has joined #openstack-meeting02:08
*** tinwood has quit IRC02:08
*** armax has quit IRC02:10
*** tinwood has joined #openstack-meeting02:11
*** macz_ has joined #openstack-meeting02:16
*** macz_ has quit IRC02:21
*** manpreet has quit IRC02:44
*** jmasud has quit IRC02:44
*** dklyle has quit IRC02:45
*** armax has joined #openstack-meeting02:48
*** armax has quit IRC02:54
*** jamesmcarthur has quit IRC02:57
*** jamesmcarthur has joined #openstack-meeting02:57
*** rcernin has quit IRC02:57
*** jamesmcarthur has quit IRC03:03
*** rcernin has joined #openstack-meeting03:18
*** rcernin has quit IRC03:21
*** rcernin has joined #openstack-meeting03:21
*** jamesmcarthur has joined #openstack-meeting03:29
*** psachin has joined #openstack-meeting03:33
*** jmasud has joined #openstack-meeting03:35
*** jmasud has quit IRC03:37
*** ricolin_ has joined #openstack-meeting03:39
*** manpreet has joined #openstack-meeting03:46
*** ricolin_ has quit IRC03:54
*** jamesmcarthur has quit IRC04:07
*** armstrong has joined #openstack-meeting04:21
*** rcernin has quit IRC04:35
*** rcernin has joined #openstack-meeting04:35
*** jmasud has joined #openstack-meeting04:41
*** gyee has quit IRC05:08
*** vishalmanchanda has joined #openstack-meeting05:11
*** evrardjp has quit IRC05:20
*** evrardjp has joined #openstack-meeting05:24
*** evrardjp has quit IRC05:33
*** evrardjp has joined #openstack-meeting05:35
*** jmasud has quit IRC05:37
*** jmasud has joined #openstack-meeting05:40
*** rcernin_ has joined #openstack-meeting05:42
*** rcernin has quit IRC05:42
*** jmasud has quit IRC05:49
*** jmasud_ has joined #openstack-meeting05:49
*** jamesmcarthur has joined #openstack-meeting06:04
*** jamesmcarthur has quit IRC06:09
*** e0ne has joined #openstack-meeting06:31
*** e0ne has quit IRC06:49
*** timburke__ has quit IRC06:52
*** dsariel has joined #openstack-meeting06:55
*** jmasud_ has quit IRC07:18
*** ralonsoh has joined #openstack-meeting07:19
*** rcernin_ has quit IRC07:28
*** armstrong has quit IRC07:40
*** e0ne has joined #openstack-meeting07:54
*** e0ne has quit IRC07:54
*** e0ne has joined #openstack-meeting07:55
*** e0ne has quit IRC07:55
*** slaweq has joined #openstack-meeting07:59
*** diablo_rojo__ has quit IRC08:01
*** slaweq has quit IRC08:04
*** rcernin_ has joined #openstack-meeting08:06
*** slaweq has joined #openstack-meeting08:10
*** jmasud has joined #openstack-meeting08:19
*** e0ne has joined #openstack-meeting08:21
*** e0ne has quit IRC08:22
*** rpittau|afk is now known as rpittau08:25
*** rcernin_ has quit IRC08:26
*** tosky has joined #openstack-meeting08:39
*** e0ne has joined #openstack-meeting09:00
*** rfolco has joined #openstack-meeting09:05
*** jamesmcarthur has joined #openstack-meeting09:06
*** dasp_ has quit IRC09:16
*** dasp has joined #openstack-meeting09:17
*** e0ne has quit IRC09:20
*** e0ne has joined #openstack-meeting09:21
*** e0ne has quit IRC09:25
*** e0ne has joined #openstack-meeting09:26
*** e0ne has quit IRC09:26
*** jamesmcarthur has quit IRC09:28
*** dmacpher has joined #openstack-meeting10:17
*** ociuhandu has joined #openstack-meeting10:40
*** jmasud has quit IRC10:47
*** jmasud has joined #openstack-meeting10:49
*** oneswig has joined #openstack-meeting10:59
oneswig#startmeeting scientific-sig11:00
openstackMeeting started Wed Jan 13 11:00:08 2021 UTC and is due to finish in 60 minutes.  The chair is oneswig. Information about MeetBot at http://wiki.debian.org/MeetBot.11:00
openstackUseful Commands: #action #agreed #help #info #idea #link #topic #startvote.11:00
*** openstack changes topic to " (Meeting topic: scientific-sig)"11:00
openstackThe meeting name has been set to 'scientific_sig'11:00
oneswigApologies, no set agenda for today (been busy!)11:00
*** dh3 has joined #openstack-meeting11:03
oneswigWrangling with diskimage-builder in the other terminal11:04
dh3o/ my other meeting was cancelled so I thought I'd stick my head in11:06
oneswigHello dh3, nice to see you.11:07
oneswigI don't think there's anything on my agenda for this week11:07
oneswigAny news to report from you?11:07
dh3things are ticking over in our "new normal"11:08
oneswigGood to hear.11:08
oneswigThe very phrase "ticking over" is evoking all kinds of envy :-)11:09
oneswig(been a busy return to work)11:09
*** PeteC61 has joined #openstack-meeting11:10
dh3There is plenty going on but not much firefighting which is a good place to be :)11:10
oneswigHappy to hear it, for many reasons11:11
PeteC61It is only Jan11:11
oneswigHello PeteC61, nice to see you11:12
oneswigJan indeed.  2021 is just getting started.11:12
oneswigNext week we have our team "design summit" to plan our R&D activities11:13
oneswigAn unusually high number of exciting things to work on...11:14
dh3are you thinking more infra/deployment tools, or user-facing features?11:14
oneswigWell it covers everything but infrastructure is going to be "the big enchilada" this time round11:15
*** b1airo has joined #openstack-meeting11:17
oneswigOften the priorities we set are akin to new years resolutions in their enduring effect, but we've adopted the 10% time approach to R&D to change that11:18
oneswigDo you have that approach - setting aside some time in the week to focus on R&D and improvement?11:19
b1airoo/ hello and happy new year. i just remembered I'd opened irccloud but then not actually logged in :-)11:19
dh3That's an interesting one, I'd expect to see some overlap between "I want to work on this" and "I have to work on this to keep my customers happy"11:19
dh3No official 10% time here but recently we gained an official "no meetings" afternoon each week which is heading in that direction, I think.11:20
b1airosounds blissful!11:20
PeteC61That is my plan. In the interim, if something is of potential interest and is proposed, we can potentially make space.11:20
oneswigHi b1airo, happy new year11:21
oneswig#chair b1airo11:21
openstackCurrent chairs: b1airo oneswig11:21
PeteC61always looking to hear how others are progressing though. This can helkp direct us to what else is out there that is working for others.11:21
b1airointeresting culture conversation i've stumbled into by the sounds of it11:22
b1airoi've got some tech/openstack investigations underway if anyone wants to talk shop...11:24
oneswigIn some ways it's the challenge of getting beyond chapter 1 in "the phoenix project"11:25
oneswigb1airo: go ahead!11:25
b1airofirst, let's start with multi-tenant options for Spectrum Scale. we're designing an OpenStack cloud/hpc hybrid thing at the moment (as you know oneswig), and i'm trying to ensure i've got an up-to-date picture of parallel filesystem options that could fit in and service multiple tenants from shared storage infrastructure11:27
b1airoI have a pretty good idea of what's possible with Lustre, but I'm still trying to figure out if Spectrum Scale even has sufficient isolation controls11:29
b1airoCephFS is obviously a possibility, but I'm a bit wary about it from several perspectives - maturity, supportability, and write/iops performance for scratch storage - we'll probably do other things with it at more modest scale and learn from there11:32
oneswigHave you evaluated Weka b1airo?  They have a multi-tenancy story as well.11:35
b1airomy next one is wondering whether anyone has current experience architecting (and running!) Ceph all-flash pools. thinking about choice of data devices and WAL/DB devices etc, also CPUs. wanting to make sure i've learned from someone else's mistakes :-)11:35
*** ociuhandu has quit IRC11:36
oneswigOn that subject, this looks like interesting kit https://h20195.www2.hpe.com/v2/GetPDF.aspx/a50000084enw.pdf11:36
dh3I know Spectrum Scale/GPFS has Cinder drivers etc but not clear if it can present the shared filesystem as a volume (assuming you want to access the same data from both OpenStack and HPC)11:37
dh3Our Ceph is predominantly HDD (NVME for journals) so the only advice I have is to make sure you use devices with high endurance (DWPD)11:39
b1airoand finally (for now), we've got a partner who will be deploying a cluster (and bunch of related side machines/services) into this environment and they need pretty tight integration with other on-prem infrastructure like the AD domain controllers, so they their own private networking space routed between their on-prem networks and our cloud env. there are plenty of ways of skinning that cat outside of11:39
b1airoOpenStack, but I'm wondering if there is already some relevant routing functionality in Neutron that could make the whole thing more software defined and repeatable for other future tenants11:39
b1airoand finally (for now), we've got a partner who will be deploying a cluster (and bunch of related side machines/services) into this environment and they need pretty tight integration with other on-prem infrastructure like the AD domain controllers, so they want own private networking space routed between their on-prem networks (multiple sites) and our cloud env. there are plenty of ways of skinning that cat11:40
b1airooutside of OpenStack, but I'm wondering if there is already some relevant routing functionality in Neutron that could make the whole thing more software defined and repeatable for other future tenants11:40
*** jmasud has quit IRC11:42
b1airoyep oneswig, that could be a reasonable choice of box. I believe the DL325 is the platform on which HPE offer their Spectrum Scale (ECE based) product/solution too11:44
b1airothough as we will have other general purpose "bulk storage" HDD based pools i was leaning towards scattering NVMe around the cluster. but perhaps it would make more sense to try and get at least 4 dedicated nodes, even if they start out only 1/4 or 1/2 populated from a storage and memory perspective11:46
b1airodh3: on endurance, that's an interesting point... are you referring to the data devices or WAL/DB? i suspect i've previously been too conservative on this front. some of the all-flash ceph product marketing i've read recently is using QLC based drives with what seem to be very read focused performance11:49
b1airo(data drives that is, often optane for WAL/DB though)11:50
oneswigWhat's the write endurance quoted for optane?  It's certainly a lot quicker and consistent for writes11:52
dh3I was imagining for the WAL, the DB too assuming your data churn is high enough to justify it. I wonder if the speed differential between Optane and NVME would make it worth it.11:53
b1airodh3: there is also a Manila driver for Scale/GPFS, however it requires the filesystem to be exported via NFS from GPFS protocol nodes, so it's not a real parallel filesystem option - we do require something high performance11:53
b1airobelieve it also requires a shared flat provider network over which all tenants mount their shares, so that immediately limits the practical usage of it11:54
dh3oneswig we had Intel in to talk about Optane and they didn't put a number on it, just to say "lots" or "plenty" (shades of Rolls-Royce and "adequate horsepower")11:55
b1airooptane is insanely high IOPs both R & W, and very high endurance, but yes i do agree that it could be overly expensive depending on how much you need for DB duties (which in itself seems to be something that changes whereever you look - 4% of data capacity in one place, 6% elsewhere, some GBs per TB number elsewhere, or just 30GB or 300GB per OSD if you dig into mailing list and bugs that detail how things11:59
b1airowork (currently) with RocksDB11:59
oneswigb1airo: the integration with multi-tenant networking could present a significant challenge that HPC-class storage simply doesn't face in it's home turf12:00
oneswigah, we are out of time12:00
dh3b1airo I'm interested to know what you end up with, we haven't (yet) found the performance motivation to have anything all-flash12:00
oneswigWe did this - but it's out of date now: https://www.stackhpc.com/ceph-on-the-brain-a-year-with-the-human-brain-project.html12:01
oneswigI think this image is still relevant for NVME though - https://www.stackhpc.com/images/julia-nvme-sustained-writes.png12:01
oneswigBetter close the meeting - thanks all12:03
oneswig#endmeeting12:03
b1airoto make matters more complicated i've just discovered there's a hybrid Optane memory option now too12:03
*** openstack changes topic to "OpenStack Meetings || https://wiki.openstack.org/wiki/Meetings/"12:03
openstackMeeting ended Wed Jan 13 12:03:46 2021 UTC.  Information about MeetBot at http://wiki.debian.org/MeetBot . (v 0.1.4)12:03
openstackMinutes:        http://eavesdrop.openstack.org/meetings/scientific_sig/2021/scientific_sig.2021-01-13-11.00.html12:03
openstackMinutes (text): http://eavesdrop.openstack.org/meetings/scientific_sig/2021/scientific_sig.2021-01-13-11.00.txt12:03
openstackLog:            http://eavesdrop.openstack.org/meetings/scientific_sig/2021/scientific_sig.2021-01-13-11.00.log.html12:03
oneswigoops, sorry b1airo, messages crossed12:03
b1airo(the H10 if you're interested)12:04
b1airono worries12:04
oneswiggot a link?12:04
b1airoand re endurance for Optane: "The first-generation Optane DC P4800X launched with a 30 DWPD write endurance rating, later increased to 60 DWPD. The new P5800X further increases write endurance to 100 DWPD. Capacities will range from 400GB to 3.2TB."12:05
b1airohttps://www.storagereview.com/review/intel-optane-memory-h10-review12:05
b1airoguess i should find the Ceph IRC too12:06
oneswigGood information, thanks b1airo12:06
dh3interesting reading, thanks, see you soon12:06
oneswigUntil next time!12:06
b1airocya12:06
*** oneswig has quit IRC12:06
*** dh3 has quit IRC12:10
*** raildo has joined #openstack-meeting12:16
*** armax has joined #openstack-meeting12:24
*** armstrong has joined #openstack-meeting12:38
*** ociuhandu has joined #openstack-meeting12:45
*** rh-jelabarre has joined #openstack-meeting12:53
*** bbowen has quit IRC12:53
*** ociuhandu has quit IRC12:55
*** ociuhandu has joined #openstack-meeting12:56
*** bbowen has joined #openstack-meeting12:56
*** ociuhandu has quit IRC12:56
*** vishalmanchanda has quit IRC13:01
*** ociuhandu has joined #openstack-meeting13:01
*** jamesmcarthur has joined #openstack-meeting13:26
*** jamesmcarthur has quit IRC13:30
*** e0ne has joined #openstack-meeting13:55
*** lajoskatona has joined #openstack-meeting14:00
*** liuyulong has joined #openstack-meeting14:01
liuyulong#startmeeting neutron_l314:02
openstackMeeting started Wed Jan 13 14:02:17 2021 UTC and is due to finish in 60 minutes.  The chair is liuyulong. Information about MeetBot at http://wiki.debian.org/MeetBot.14:02
openstackUseful Commands: #action #agreed #help #info #idea #link #topic #startvote.14:02
*** openstack changes topic to " (Meeting topic: neutron_l3)"14:02
openstackThe meeting name has been set to 'neutron_l3'14:02
*** liuyulong has quit IRC14:03
slaweqhi14:03
*** liuyulong has joined #openstack-meeting14:04
liuyulong#startmeeting neutron_l314:04
openstackliuyulong: Error: Can't start another meeting, one is in progress.  Use #endmeeting first.14:04
haleybhi14:04
liuyulongSorry, I lost the internet connection.14:04
liuyulonghi14:04
liuyulongThe meeting is started?14:04
haleyblooks like it14:04
lajoskatonaHi14:05
* haleyb runs to get a snack quickly14:05
*** liuyulong has quit IRC14:05
*** liuyulong has joined #openstack-meeting14:06
liuyulongLost again...14:06
liuyulong#topic Bugs14:07
*** openstack changes topic to "Bugs (Meeting topic: neutron_l3)"14:07
liuyulongWe should have at least 3 list from our bug deputy.14:07
*** liuyulong has quit IRC14:08
*** liuyulong has joined #openstack-meeting14:09
liuyulong#link http://lists.openstack.org/pipermail/openstack-discuss/2021-January/019602.html14:09
liuyulong#chair haleyb14:09
openstackCurrent chairs: haleyb liuyulong14:09
liuyulong#chair slaweq14:09
openstackCurrent chairs: haleyb liuyulong slaweq14:09
liuyulongSorry, I've lost the network connection 3 times...14:10
liuyulong#link http://lists.openstack.org/pipermail/openstack-discuss/2021-January/019781.html14:10
liuyulongI have one RFE need the team to pay attention.14:10
liuyulong#link https://bugs.launchpad.net/neutron/+bug/191112614:10
openstackLaunchpad bug 1911126 in neutron "[RFE][L3] add ability to control router SNAT more granularly" [Wishlist,New] - Assigned to LIU Yulong (dragon889)14:10
liuyulong#link https://review.opendev.org/c/openstack/neutron-specs/+/77054014:10
liuyulongI've uploaded the spec for it.14:11
slaweqliuyulong: I have this on my todo list to read through it14:11
slaweqI will probably add it to this week's drivers meeting14:11
*** dsariel has quit IRC14:11
liuyulonghaleyb, would you mind take over the meeting chair. I'm afraid I will lose the connection again...14:11
*** dsariel has joined #openstack-meeting14:12
liuyulongslaweq, thank you. I will catch the meeting.14:12
liuyulong#link http://lists.openstack.org/pipermail/openstack-discuss/2020-December/019572.html14:12
haleybi'm back14:14
haleybSo i don't see any new l3 bugs from last weeks report14:16
haleybany other bugs to discuss?14:17
haleyb#topic On demand agenda14:18
*** openstack changes topic to "On demand agenda (Meeting topic: neutron_l3)"14:18
*** liuyulong has quit IRC14:18
haleybAny other topics to discuss?14:19
slaweqI don't have anything14:19
haleybOr reviews that need attention?14:19
*** liuyulong has joined #openstack-meeting14:19
*** liuyulong has quit IRC14:20
haleybok, i guess short meeting, liu can just ping anyone in neutron channel if we missed it14:20
haleyb#endmeeting14:20
*** openstack changes topic to "OpenStack Meetings || https://wiki.openstack.org/wiki/Meetings/"14:20
openstackMeeting ended Wed Jan 13 14:20:49 2021 UTC.  Information about MeetBot at http://wiki.debian.org/MeetBot . (v 0.1.4)14:20
openstackMinutes:        http://eavesdrop.openstack.org/meetings/neutron_l3/2021/neutron_l3.2021-01-13-14.02.html14:20
openstackMinutes (text): http://eavesdrop.openstack.org/meetings/neutron_l3/2021/neutron_l3.2021-01-13-14.02.txt14:20
openstackLog:            http://eavesdrop.openstack.org/meetings/neutron_l3/2021/neutron_l3.2021-01-13-14.02.log.html14:20
slaweqo/14:20
slaweqthx14:20
lajoskatonao/14:22
*** vishalmanchanda has joined #openstack-meeting14:35
*** TrevorV has joined #openstack-meeting14:41
*** jamesmcarthur has joined #openstack-meeting14:44
*** armax has quit IRC14:48
*** armax has joined #openstack-meeting14:50
*** belmoreira has joined #openstack-meeting14:57
*** ociuhandu_ has joined #openstack-meeting14:58
*** ociuhandu has quit IRC15:02
*** macz_ has joined #openstack-meeting15:13
*** macz_ has quit IRC15:17
*** psachin has quit IRC15:28
*** ociuhandu_ has quit IRC15:35
*** ociuhandu has joined #openstack-meeting15:35
*** ricolin has quit IRC15:38
*** jamesmcarthur has quit IRC15:45
*** dklyle has joined #openstack-meeting15:46
*** jamesmcarthur has joined #openstack-meeting15:46
*** ricolin has joined #openstack-meeting15:47
*** jamesmcarthur has quit IRC15:51
*** jamesmcarthur has joined #openstack-meeting15:54
*** macz_ has joined #openstack-meeting16:02
*** ociuhandu_ has joined #openstack-meeting16:11
*** ociuhandu has quit IRC16:11
*** e0ne has quit IRC16:15
*** armstrong has quit IRC16:27
*** jamesmcarthur has quit IRC16:32
*** jamesmcarthur has joined #openstack-meeting16:34
*** dsariel has quit IRC16:34
*** slaweq has quit IRC16:34
*** dsariel has joined #openstack-meeting16:34
*** slaweq has joined #openstack-meeting16:36
*** jamesmcarthur has quit IRC16:38
*** adrianc has quit IRC16:41
*** tosky has quit IRC16:41
*** adrianc has joined #openstack-meeting16:42
*** tosky has joined #openstack-meeting16:42
*** jmasud has joined #openstack-meeting16:46
*** jamesmcarthur has joined #openstack-meeting16:50
*** jmasud has quit IRC16:57
*** jmasud has joined #openstack-meeting16:57
*** gyee has joined #openstack-meeting17:00
*** ociuhandu_ has quit IRC17:06
*** ociuhandu has joined #openstack-meeting17:09
*** e0ne has joined #openstack-meeting17:11
*** ociuhandu_ has joined #openstack-meeting17:13
*** ociuhandu has quit IRC17:16
*** ociuhandu_ has quit IRC17:17
*** ociuhandu has joined #openstack-meeting17:22
*** ociuhandu has quit IRC17:27
*** jamesmcarthur has quit IRC17:29
*** jamesmcarthur has joined #openstack-meeting17:30
*** baojg has quit IRC17:41
*** baojg has joined #openstack-meeting17:41
*** mlavalle has joined #openstack-meeting17:57
*** timburke has joined #openstack-meeting17:59
*** rpittau is now known as rpittau|afk18:05
*** baojg has quit IRC18:11
*** baojg has joined #openstack-meeting18:11
*** timburke_ has joined #openstack-meeting18:13
*** PeteC61 has quit IRC18:15
*** timburke has quit IRC18:15
*** baojg has quit IRC18:18
*** baojg has joined #openstack-meeting18:19
*** ralonsoh has quit IRC18:20
*** e0ne has quit IRC18:39
*** jmasud has quit IRC18:46
*** lajoskatona has quit IRC18:48
*** belmoreira has quit IRC19:03
*** manpreet has quit IRC19:06
*** jamesmcarthur has quit IRC19:06
*** jamesmcarthur_ has joined #openstack-meeting19:11
*** jmasud has joined #openstack-meeting19:13
*** jmasud has quit IRC19:36
*** baojg has quit IRC19:39
*** baojg has joined #openstack-meeting19:39
*** jamesmcarthur_ has quit IRC19:54
*** jamesmcarthur has joined #openstack-meeting19:54
*** jamesmcarthur has quit IRC20:14
*** armstrong has joined #openstack-meeting20:20
*** slaweq has quit IRC20:41
*** acoles has joined #openstack-meeting20:58
timburke_#startmeeting swift21:00
openstackMeeting started Wed Jan 13 21:00:11 2021 UTC and is due to finish in 60 minutes.  The chair is timburke_. Information about MeetBot at http://wiki.debian.org/MeetBot.21:00
openstackUseful Commands: #action #agreed #help #info #idea #link #topic #startvote.21:00
*** openstack changes topic to " (Meeting topic: swift)"21:00
openstackThe meeting name has been set to 'swift'21:00
timburke_who's here for the swift meeting?21:00
mattoliverauo/21:00
seongsoochoo/21:00
*** e0ne has joined #openstack-meeting21:00
rledisezo/21:00
acoleso/21:01
dsarielo/21:01
*** e0ne has quit IRC21:01
kota_o/21:01
timburke_as usual, the agenda's at https://wiki.openstack.org/wiki/Meetings/Swift21:01
timburke_first up21:02
timburke_#topic reconciler/ec/encryption21:02
*** openstack changes topic to "reconciler/ec/encryption (Meeting topic: swift)"21:02
timburke_#link https://bugs.launchpad.net/swift/+bug/191080421:02
openstackLaunchpad bug 1910804 in OpenStack Object Storage (swift) "Encryption doesn't play well with processes that copy cleartext data while preserving timestamps" [Undecided,New]21:02
timburke_so i had a customer report an issue with an object that would consistently 50321:03
claygohai21:03
timburke_digging in more, we found that they had 11 frags of it for an 8+4 policy... but those had 3 separate sets of crypto meta between them21:04
timburke_...and no set of crypto meta had more than 7 frags21:04
clayglawl21:04
acoles(I had to think about this at first) meaning frags have been encrypted with three different body keys...for same object!!!21:05
timburke_root cause was traced back to a couple issues: (1) we deploy with encryption in the reconciler pipeline and (2) we have every (container?) node running a reconciler21:06
timburke_(well, that and the fact that it was moved to an EC policy. if it were going to a replicated policy, any replica regardless of crypto meta would be capable of generating a client response)21:07
timburke_i've got a fix up to pull encryption out of the reconciler pipeline if it was misconfigured -- https://review.opendev.org/c/openstack/swift/+/77052221:08
timburke_but i wanted to raise awareness of the issue so no one else finds themselves in this situation21:09
timburke_also worth noting: i think you could run into a similar issue *without encryption* if your EC backend is non-deterministic21:10
*** zaitcev has joined #openstack-meeting21:12
timburke_the open source backends are deterministic as i recall (that is, the frag outputs only depend on the EC params from swift.conf and the input data), but i don't know the details of shss, for example21:12
timburke_does anyone have any questions about the bug or its impact?21:13
timburke_all right21:14
mattoliverauNice investigation!21:14
timburke_#topic SSYNC and non-durable frags21:15
*** openstack changes topic to "SSYNC and non-durable frags (Meeting topic: swift)"21:15
timburke_#link https://bugs.launchpad.net/swift/+bug/177800221:15
openstackLaunchpad bug 1778002 in OpenStack Object Storage (swift) "EC non-durable fragment won't be deleted by reconstructor. " [High,Confirmed]21:15
timburke_i know acoles (and clayg?) has been working on this problem a bit lately, though i'm not sure where things stand21:15
kota_shss might be impacted. i'll check it.21:15
acolesI just got my probe test working!21:16
timburke_\o/21:16
acolesbackground: we noticed some partitions were never cleaned up on handoffs21:16
acolesturned out they had non-durable data frags on them , so the dir would not be deleted21:16
acolesbut reconstructor/ssync does not sync non-durable frags21:17
acoles:(21:17
acolesso https://review.opendev.org/c/openstack/swift/+/770047 should fix that21:17
acolesby (a) sync'ing non-durables (they could still be useful data) and (b) then removing non-durables on the handoff21:18
clayghttps://bugs.launchpad.net/swift/+bug/1778002 has been around for awhile - anyone doing EC rebalances has probably noticed it21:18
openstackLaunchpad bug 1778002 in OpenStack Object Storage (swift) "EC non-durable fragment won't be deleted by reconstructor. " [High,Confirmed]21:18
zaitcevHrm. I never noticed because I have excess space.21:20
timburke_i think we mainly noticed because we monitor handoffs as part of our rebalances21:21
acolesthe commit message on the patch details the various changes needed to get the non-durables yielded to ssync and then have ssync sync them21:22
timburke_acoles, are there any questions that might need answering, or is this something that everyone should just anticipate getting better Real Soon Now?21:22
acolesreview always welcome, but there's no specific issue I  have in mind for feedback21:23
timburke_excellent21:23
acolesI'm about to push a new patchset - and I have one more test to write21:23
timburke_#topic cleaning up shards when root DB is deleted and reclaimed21:25
*** openstack changes topic to "cleaning up shards when root DB is deleted and reclaimed (Meeting topic: swift)"21:25
timburke_meanwhile, mattoliverau has picked up21:25
timburke_#link https://bugs.launchpad.net/swift/+bug/191123221:25
openstackLaunchpad bug 1911232 in OpenStack Object Storage (swift) "empty shards fail audit with reclaimed root db " [Undecided,Confirmed] - Assigned to Matthew Oliver (matt-0)21:25
timburke_how's that going?21:26
mattoliverauYeah things are moving along. I have https://review.opendev.org/c/openstack/swift/+/77052921:26
mattoliverauit's not fixed yet, just worked on a probe test that shows he problem.21:26
acolesa very good place to start :)21:27
mattoliverauIn an ideal world we'd have shrinking and autosharding so shards with nothing in them was suppose to collapse into the root before reclaim_age21:27
mattoliveraubut we don't have that, and there is still an edge case where their not getting cleaned up.21:28
mattoliverauI'll have another pathset up today that should have an initial version of a fix. Currently still on my laptop as it needs some debugging and tests21:28
mattoliveraukeep an eye out for that and then please review and we can make sure we don't leave any pesky shards around :)21:29
timburke_sounds good21:29
timburke_#topic s3api and allowable clock skew21:30
*** openstack changes topic to "s3api and allowable clock skew (Meeting topic: swift)"21:30
timburke_i've had some clients getting back RequestTimeTooSkewed errors for a while -- not real common, but it's a fairly persistent problem21:31
timburke_i'm fairly certain it's that they retry a failed request verbatim, rather than re-signing with the new request time21:31
timburke_eventually, given the right retry/backoff options, the retry goes longer than 5mins and they get back a 40321:32
zaitcevso, there's nothing we can do, right?21:33
timburke_i *think* AWS has an allowable skew of more like 15mins (though can't remember whether i read it somewhere or determined it experimentally)21:33
zaitcevThat's what I remember, too.21:34
timburke_so i proposed a patch to make it configurable, with a default of (what i recall as being) AWS's limit21:34
timburke_#link https://review.opendev.org/c/openstack/swift/+/77000521:34
zaitcevIt was mentioned in the old Developer's Guide. But that document is gone, replaced with API Reference.21:34
timburke_i wanted to check if anyone had concerns about increasing this default value (it would of course be called out in release notes later)21:35
kota_should we extend the default value too?21:35
* kota_ said same thing :P21:36
timburke_kota_, yeah, the patch as written increases the timeout from 5mins to 15mins (if you don't explicitly set a value)21:36
timburke_ok, seems like we're generally ok with it :-)21:37
timburke_#topic relinker21:38
*** openstack changes topic to "relinker (Meeting topic: swift)"21:38
timburke_i found a couple issues recently that might be good to know about if anyone's planning a part-power increase (or two) soon21:39
timburke_#link https://bugs.launchpad.net/swift/+bug/191058921:39
openstackLaunchpad bug 1910589 in OpenStack Object Storage (swift) "Multiple part power increases leads to misplaced data" [Undecided,New]21:39
timburke_^^^ characterizes something i think i mentioned last week, but hadn't gotten a clean repro for21:39
timburke_rledisez, do you think you might have time to review https://review.opendev.org/c/openstack/swift/+/769855 (which should address it)?21:40
zaitcevChristian is no longer around essentially, we have to do without.21:40
clayg😢 I hope he's doing well tho!  😁21:41
*** vishalmanchanda has quit IRC21:41
rlediseztimburke_: absolutely, I'll do that this week21:41
timburke_thanks! only thing worth calling out (i think) is that the state file format changed in such a way that any old state files will just be discarded21:42
rlediseznot a big deal. don't upgrade if you're relinking, and worst case sceniario, it restart from zero21:43
timburke_but that should only really be a concern if someone is doing a swift upgrade mid-part-power-increase, which doesn't seem like a great plan anyway21:43
*** baojg has quit IRC21:43
clayghahaha21:43
timburke_the other one i noticed is a little thornier21:43
timburke_#link https://bugs.launchpad.net/swift/+bug/191047021:44
openstackLaunchpad bug 1910470 in OpenStack Object Storage (swift) "swift-object-relinker does not handle unmounted disks well" [Undecided,New]21:44
*** baojg has joined #openstack-meeting21:44
timburke_essentially, on master, if the relinker hits an unmounted disk, you get no feedback about it at all21:44
timburke_i've got a patch that at least has us log the fact that the disk is getting skipped -- https://review.opendev.org/c/openstack/swift/+/76963221:45
timburke_but it doesn't exit with a non-zero status code or anything21:45
seongsoochoSo now, is it safe to increase the partition power only once until the patch is applied?21:46
rledisezseongsoocho: from production experience, it is. we did it on multiple clusters with the current status of the relinker21:47
timburke_seongsoocho, yes, increasing it once will definitely be fine. once it's been increased, you could go manually clear the state files -- then it would be safe to do it again21:47
rledisezbut you should care about the last bug mentioned by timburke_, ensure your rights are ok (root:root) on unmounted disk to avoid bad surprises21:48
timburke_they'd be named something like /srv/node/*/.relink.*.json21:48
seongsoochoaha. ok thanks :)21:48
rledisezat some point, it would be useful to have a recon option that return the values of the relink.json and tells when one is missing (eg: because unmounted)21:49
timburke_good thought!21:49
*** TrevorV has quit IRC21:50
timburke_all right, i mostly wanted to raise awareness on those -- i'll let you know if i get a good idea on a better resolution for that second one21:50
timburke_#topic open discussion21:50
*** openstack changes topic to "open discussion (Meeting topic: swift)"21:50
timburke_what else should we talk about this week?21:51
acolesOMM I'm seeing this test fail in virtualenvs (e.g. tox -e py36) but not outside virtualenv: 'nosetests ./test/unit/common/test_manager.py:TestManagerModule.test_verify_server' - anyone else noticed that? I'm baffled21:52
acolesAFAICT the test is asserting that swift-Object-server is not on my PATH21:53
acolesnote the capital 'O'21:53
claygdoes it always fail?21:53
acolesinside virtualenv yes - I mean, I just noticed in last 20mins21:54
claygare you on a case insenstive file system 🤣21:54
acolesvsaio and macos both the same21:54
acolesapart from it failing, I don't like that a unit test is making assertions about what I might have on my PATH21:55
claygoh, i just tried venv in my vsaio and it worked 🤷‍♂️21:55
acolesif no-one else has noticed I'll dig some more21:55
claygpy3.8 tho21:55
zaitcevacoles: so, if you comment verify_server, does it fail?21:56
zaitcevor, well21:56
zaitcevit's a test, so it's a little artificial21:56
acolespy2.7 fails21:57
zaitcevse21:58
zaitceva second21:58
timburke_maybe related to https://review.opendev.org/c/openstack/swift/+/769848 ?21:58
zaitcevyes, that one21:58
timburke_i should go review that... or maybe acoles should ;-)21:59
*** rcernin has joined #openstack-meeting21:59
zaitcevYeah21:59
timburke_all right21:59
acolesdon't think so, in my virtualenvs, '$ which swift-Object-server' actually finds a match21:59
acoles:/22:00
timburke_O.o22:00
zaitcevMaybe just back out the whole thing. It's an option. But I hoped that just backing out the effects in the decorator, and _only_ screw with the exit code, would let use preserve it.22:00
zaitcevOh22:00
timburke_zaitcev, seems likely to be reasonable22:00
timburke_acoles, well where did it come from!?22:01
timburke_anyway, we're at time22:01
acolesI think there's a case-insensitivity thing going on in my virtualenvs ?!? v weird22:01
timburke_thank you all for coming, and thank you for working on swift!22:01
timburke_there's a lot going on, and i'm excited to see it all happening22:01
timburke_#endmeeting22:01
*** openstack changes topic to "OpenStack Meetings || https://wiki.openstack.org/wiki/Meetings/"22:01
openstackMeeting ended Wed Jan 13 22:01:50 2021 UTC.  Information about MeetBot at http://wiki.debian.org/MeetBot . (v 0.1.4)22:01
openstackMinutes:        http://eavesdrop.openstack.org/meetings/swift/2021/swift.2021-01-13-21.00.html22:01
openstackMinutes (text): http://eavesdrop.openstack.org/meetings/swift/2021/swift.2021-01-13-21.00.txt22:01
openstackLog:            http://eavesdrop.openstack.org/meetings/swift/2021/swift.2021-01-13-21.00.log.html22:01
*** dsariel has left #openstack-meeting22:02
*** zaitcev has left #openstack-meeting22:03
*** acoles has left #openstack-meeting22:14
*** jmasud has joined #openstack-meeting22:18
*** jmasud has quit IRC22:20
*** baojg has quit IRC22:52
*** baojg has joined #openstack-meeting22:53
*** timburke_ has quit IRC23:11
*** timburke_ has joined #openstack-meeting23:11
*** rh-jelabarre has quit IRC23:17
*** yamamoto has joined #openstack-meeting23:32
*** jmasud has joined #openstack-meeting23:33
*** baojg has quit IRC23:33
*** baojg has joined #openstack-meeting23:34
*** jmasud has quit IRC23:37

Generated by irclog2html.py 2.17.2 by Marius Gedminas - find it at https://mg.pov.lt/irclog2html/!