Original article: http://www.codeproject.com/KB/files/DuplicateFinder.aspx
Download source code: http://files.cnblogs.com/zhangjidong/DuplicateFinder_src.zip
Introduction
Once a year, I do a terrible job, clean up the files I have created or downloaded, and last time I tried to do this, I think it is complicated to do this in semi-automated mode. I need some free and easy-to-use tools to find duplicate files, but I didn't find anything that fits my needs.
I decided to write one.
Background
CRC algorithm methods, see http://www.fesersoft.com/dotNet can be found. I use the MD5 , I add an event to MD5 In the calculation method, this process is used to calculate the hash. It is used to read the stream. When MD5 when the computing method reads the same stream.
Code of use
The utility uses two main classes, directorycrawler, and hashing. the use is so obvious
Two classes are used:ClassesAndDirectorycrawlerAnd their usage is obvious.
Point of interest
# Optimize file movement. When moving large files, the interface is slow.
# Add an optionalCRC32OrMD5Hash option.
# UseXMLFor the configuration file, move the duplicate fileD: "duplicatefiles(This is hardcoded. Long live Microsoft !)
It is sufficient for me to skip the scan.
# Do not forget to outline your post.
# (Already done): Write an event and activate it.MD5Class used to report the hashed process.10 GB.
History
V0.2:
Add List fileinfo > and List string [] > replaces the array list.
MD5Hash ChannelCRC32
Skip source directory Addition
Add and remove sub-Directories
Some Optimizations
License
This Article describes how to connect any source code or file that follows cpol .