[tahoe-lafs-trac-stream] [tahoe-lafs] #1816: add a lease renewal method that only renews some shares in a shareset, to be used by repair/rebalancing

tahoe-lafs trac at tahoe-lafs.org
Thu Dec 6 22:32:26 UTC 2012


#1816: add a lease renewal method that only renews some shares in a shareset, to
be used by repair/rebalancing
-------------------------+-------------------------------------------------
     Reporter:           |      Owner:  warner
  davidsarah             |     Status:  new
         Type:  defect   |  Milestone:  undecided
     Priority:  normal   |    Version:  1.9.2
    Component:  code-    |   Keywords:  forward-compatibility rebalancing
  storage                |  repair RIStorageServer design-review-needed
   Resolution:           |
Launchpad Bug:           |
-------------------------+-------------------------------------------------

Comment (by davidsarah):

 Merged from #881 (cancel leases on extra shares in repairer, check-and-
 add-lease, upload, and publish):

 > The ideal state of a file is to have exactly N distinct shares on N
 distinct servers. Anything beyond that is "extra": they might improve
 reliability but also consume extra storage space. We'd like to remove
 these extra shares to bring the total consumed storage space back down to
 the target implied by the user's choice of the N/k "expansion ratio".
 >
 > For mutable files, anyone with a writecap can simply delete the extra
 shares. We should modify the "publish" operation to identify and delete
 the extra shares (after successfully updating the non-extra shares).
 >
 > But there is no appropriate way to explicitly delete an immutable share:
 we intentionally do not provide a "destroycap". So the way to get rid of
 these shares is through garbage collection.
 >
 > The operations that add leases (check --add-lease, and the repairer)
 should pay attention to how many shares have been seen, and identify the
 extra shares, and then cancel any leases that we can on them.
 >
 > Check-and-add-lease pipelines both operations: it sends a DYHB and an
 add-lease-to-anything-you-have message together, ignoring the response
 from the add-lease message, and counting the DYHB responses to form the
 checker results. This speeds up the operation: if we allowed the code to
 have an unbounded number of outstanding messages in flight, the entire
 operation could be finished in one RTT.
 >
 > Instead, this code should watch the DYHB responses and identify the
 extra shares, then send out cancel-lease messages for the extra shares.
 This increases the required time to two RTT (since we can't send out any
 cancel-lease messages until we've seen enough DYHB responses to correctly
 identify shares as being extra), but only in the (hopefully rare) case
 where there are extra shares. In the common case, check-and-add-lease
 should proceed at full speed and never need to send out additional
 messages.
 >
 > Sending out cancel-lease messages is also easier than carefully
 refraining from sending out add-lease messages on the extra shares. To
 accomplish that, we'd have to do a full check run (i.e. DYHB messages to
 everyone), and only after most of those came back could we do the
 selective add-lease messages. By sending out cancel-messages instead,
 we're sending more messages (DYHB, add-lease, cancel-lease), but we can
 pipeline them more efficiently.
 >
 > Extra shares can arise in a variety of ways. The most common is when a
 mutable file is modified while some of the servers are offline: new shares
 (to replace the unavailable ones) will be created and sent to new servers,
 and then on a subsequent publish, all shares will be updated. This
 typically results in e.g. sh1 being present on both servers A and B.
 >
 > Another cause is the immutable repairer, which (because immutable upload
 is still pretty simplistic) will place a share on a server before checking
 to see if that same share is on a different server, or before seeing if
 there are any other shares on that server already. This typically results
 in e.g. sh1 and sh2 being present on server A, while sh2 is also present
 on server B.
 >
 > The storage server's add/cancel lease operations need to be enhanced to
 allow clients to selectively manipulate leases on each share, not just the
 bucket as a whole. This is needed to allow the sh2 on server A to expire,
 while preserving the sh1 on server A. This also argues against some of the
 storage-server changes that I've recommended elsewhere (#600), in which
 the lease information would be pulled out of the per-share files and into
 a per-bucket structure, since that would make it impossible to cancel a
 lease on one share but not the other.

 The above was written before [https://github.com/davidsarah/tahoe-
 lafs/blob/1818-leasedb/docs/proposed/leasedb.rst leasedb] (#1818) and also
 before the fix to #1528 which removed the remote methods to explicitly
 cancel leases. Now, it would have to be implemented in terms of a
 selective lease renewal operation as proposed in comment:2.

-- 
Ticket URL: <https://tahoe-lafs.org/trac/tahoe-lafs/ticket/1816#comment:5>
tahoe-lafs <https://tahoe-lafs.org>
secure decentralized storage


More information about the tahoe-lafs-trac-stream mailing list