I’m looking for a deduplication software probably using hash sums to tidy up my pc…
Let me explain, I downloaded all sorts of photos from my smartphone to the pc, using warpinator among others. In short, I guess my photos have been downloaded several times, in different folders. In short, it’s a mess.
I’d like to sort it out once and for all. Do you know of any software that can do the job?
One gets interactively to decide which file to keep and delete all others.
-r --recurse for every directory given follow subdirectories
encountered within
-d --delete prompt user for files to preserve and delete all
others; important: under particular circumstances,
data may be lost when using this option together
with -s or --symlinks, or when specifying a
particular directory more than once; refer to the
fdupes documentation for additional information
Duplicate: A file that matches the original. Note that depending on rmlint settings, “match” may mean an exact match or just that the files have matching hash values.
For different locations a working example is this:
This will find files in path1 which have duplicates (same data content) in path2. It will create a shell script which, if run, will remove the duplicates under path1, leaving only the unique files.
indeed rmlink does a great job
I accidentally launched it without doing a dry run first.
The result is excellent work on the photos, but scratching in my personal documents for modifications that I should have validated one by one…
Fortunately, a good backup of my /home and /data by Borg-backup enabled me to pinpoint the differences. And yes, Borg allows you to make a diff between two successive backups…
Now it’s up to me to restore a number of things.
Too bad rmlint doesn’t keep a log for each launch…