Opened at 2009-12-20T23:26:52Z
Last modified at 2014-09-11T22:18:51Z
#869 new enhancement
Allow Tahoe filesystem to be run over a different key-value-store / DHT implementation — at Version 3
Reported by: | davidsarah | Owned by: | nobody |
---|---|---|---|
Priority: | major | Milestone: | undecided |
Component: | code-network | Version: | 1.5.0 |
Keywords: | scalability performance forward-compatibility backward-compatibility availability newcaps docs anti-censorship | Cc: | |
Launchpad Bug: |
Description (last modified by davidsarah)
source:docs/architecture.txt describes Tahoe as comprising three layers: key-value store, filesystem, and application.
Most of what makes Tahoe different from other systems is in the filesystem layer -- the layer that implements a cryptographic capability filesystem. The key-value store layer implements (a little bit more than) a Distributed Hash Table, which is a fairly well-understood primitive with many implementations. The Tahoe filesystem and applications could in principle run on a different DHT, and it would still behave like Tahoe -- with different (perhaps better, depending on the DHT) scalability, performance, and availability properties, but with confidentiality and integrity ensured by Tahoe without relying on the DHT severs.
However, there are some obstacles to running the Tahoe filesystem layer on another DHT:
- the code isn't strictly factored into layers (even though most code files belong mainly to one layer), so there isn't a narrow API between the grid and filesystem-related abstractions.
- the communication with servers currently needs to be encrypted (independently of the share encryption), and other DHTs probably wouldn't support that.
- because the filesystem has only been used with one key-value store layer up to now, it may make assumptions about that layer that haven't been clearly documented.
Note that even if the Tahoe code was strictly layered, we should still expect there to be some significant effort to port Tahoe to a particular DHT. The DHT servers would probably have to run some Tahoe code in order to verify shares, for example.
Change History (3)
comment:1 Changed at 2009-12-22T05:28:49Z by warner
comment:2 Changed at 2010-01-20T07:10:09Z by davidsarah
- Summary changed from Allow Tahoe filesystem to be run over a different grid/DHT implementation to Allow Tahoe filesystem to be run over a different key-value-store / DHT implementation
The "grid layer" is now called the "key-value store layer".
comment:3 Changed at 2010-03-25T00:10:17Z by davidsarah
- Description modified (diff)
Hmm, good points. This ties in closely to the docs outline that we wrote up (but which we haven't finished by writing the actual documentation it calls for): docs/specifications/outline.rst .
As you note, there are several abstraction-layer leaks which would need to be plugged or accomodated to switch to a general-purpose DHT for the bottom-most layer. Here are a few thoughts.
There might also be some better ways of describing Tahoe's nominal layers, in a sense refactoring the description or shuffling around the dotted lines. I've been trying to write up a presentation using the following arrangement:
One way to look at Tahoe is in terms of that top-most API: you don't care what it does, you just need to know about filecaps and dircaps. Another view is about some client code, the API, the gateway node, and the servers that the gateway connects to: this diagram would show different sorts of message traversing the different connections. A third view would abstract the servers and the DHT/erasure-coding stuff into a lookup table, and focus on the crypto-and-above layers.