Registered Member
|
I'm not a programmer, so my understanding may be lacking. I think that when trying to sort for duplicate files, programs build a database, then query it for multiples of values of attributes like date, size, crc/hash sum etc.
This sounds like something for which Nepomuk could be very useful, since it keeps tracks through Strigi of the file system and stores data in the database. Since most dups occur for photos, music files and documents, which Strigi probably indexes anyway, finding duplicates would mean just running the correct queries and presenting them to users. It helps to trim down collections and free space on the system. And I think that the infrastructure is already present. Even if there are no crc sums in the database. Again, I'm not a programmer, so please correct me if I'm wrong. I have 2 similar ideas, which would be presented at later posts. |
Registered users: Baidu [Spider], Bing [Bot], Google [Bot], Yahoo [Bot]