Opened at 2010-08-27T00:18:25Z
Last modified at 2016-07-11T16:47:31Z
#1187 new defect
mitigate the performance bottleneck of slow servers in download
Reported by: | davidsarah | Owned by: | |
---|---|---|---|
Priority: | major | Milestone: | undecided |
Component: | code-network | Version: | 1.8β |
Keywords: | download performance | Cc: | tahoe-lafs.org@… |
Launchpad Bug: |
Description (last modified by lpirl)
Ticket #1170 showed that server selection can have a big impact on download performance (presumably the same is true of upload performance) when the chosen servers have different effective bandwidths. For example, ticket:1170#comment:94 showed an overall download bandwidth of ~90 Kbps vs ~190 Kbps depending on how many shares were taken from each of two servers.
My hypothesis is that this is mainly due to the time to download each segment being bottlenecked by the slowest server. Once we've downloaded the share(s) for a given segment from the faster servers and are waiting for the slower ones, the faster servers are left idle until the next segment.
In principle, we can use the erasure coding to mitigate this bottleneck. If k were larger than it typically is now, we could download as many shares per segment from each server as its current effective bandwidth is able to support. Since the individual shares would be a smaller fraction of the segment size, so would be the wasted time per segment.
A further optimization would be to pipeline requests across consecutive segments. That is, when we're getting near to finishing the download of shares for a segment, we can also be downloading shares for the next segment (but only the next segment) from servers that would otherwise be idle. Note that this depends on downloading a variable number of shares from each server for each segment (otherwise, the servers that finish first on one segment would also typically finish first on the next segment, so we would either have unbounded memory usage or still have to leave the faster servers servers that finish first idle some of the time).
It isn't entirely clear how large k would need to be for this approach to work. If it would need to be larger than the number of servers, then we would have to rethink the servers-of-happiness criterion. The current definition basically only credits each server for having one share, and it would have to be changed to credit each server for having multiple shares. I think this can still be given a fairly simple definition in the bipartite matching model.
Change History (4)
comment:1 Changed at 2010-08-27T00:22:16Z by davidsarah
comment:2 Changed at 2010-08-27T00:25:04Z by davidsarah
- Description modified (diff)
comment:3 Changed at 2010-09-07T00:48:40Z by davidsarah
See also #1110, which proposes pipelining without downloading a variable number of shares from each server per segment.
comment:4 Changed at 2016-07-11T16:47:31Z by lpirl
- Cc tahoe-lafs.org@… added
- Description modified (diff)
In the ticket description, when I say "idle", I mean not working on this download. Of course servers that are idle in that sense might be doing other useful work, but probably most Tahoe grids have very bursty download traffic, and so they often (usually?) won't be.