Wednesday, 25 April 2012
Fancy file systems are all the rage. "ZFS":http://hub.opensolaris.org/bin/view/Community+Group+zfs/whatis, "btrfs":https://btrfs.wiki.kernel.org, and even Microsoft's new "ReFS":http://blogs.msdn.com/b/b8/archive/2012/01/16/building-the-next-generation-file-system-for-windows-refs.aspx include data deduplication features. However, these techniques can use a lot of memory, and new file systems are often not nearly as stable as tried and true file systems, such as "ext3/ext4":https://ext4.wiki.kernel.org/ or "XFS":http://oss.sgi.com/projects/xfs/. Experimenting with file systems is fun; however, in this case, I am not about to trust all of my data to a new file system just to remove duplicate files. Instead, I decided to deduplicate my files using a few scripts. The end result: over 12GB of saved space after just a few minutes of scanning my hard drive. With hard drive prices sky high (but slowly coming down), this will help me last another 6 months before I have to upgrade my hard drives again.
Continue reading "Finding and Eliminating Duplicated Files"