http://pypi.python.org/pypi/dupfilefind usage: dupfilefind [-h] [-v] [-I IGNORE_DIRS] [-H] [-D] [-m M] [-p] [dir [dir ...]] Find files with identical contents. positional arguments: dir a directory to recursively examine optional arguments: -h, --help show this help message and exit -v, --verbose Emit more information. -I IGNORE_DIRS, --ignore-dirs IGNORE_DIRS comma-separated list of directories to skip (if you need to name a directory which has a comma in its name then escape that command twice) (default _darcs,.svn -H, --hard-link-them Whenever a file is found with identical contents to a previously discovered file, replace the new one with a hard link to the old one. This option is very dangerous because hard links are confusing and dangerous things to have around. -D, --delete-them Whenever a file is found with identical contents to a previously discovered file, delete the new one. This option is dangerous. -m M, --min-size M Ignore files smaller than this (default 1024). -p, --profiles Generate a profile of your files which doesn't contain very much potentially-sensitive information (it contains only the file sizes rounded up to 4096 bytes, how many files were duplicates, and an 16-bit checksum of each file). This could be useful for a p2p storage project to measure how valuable convergent encryption is. == Starting Points == * TracGuide -- Built-in Documentation * [http://trac.edgewall.org/ The Trac project] -- Trac Open Source Project * [http://trac.edgewall.org/wiki/TracFaq Trac FAQ] -- Frequently Asked Questions * TracSupport -- Trac Support For a complete list of local wiki pages, see TitleIndex.