If digital image stocks have been built up over years of work with different people, it is inevitable that duplicates will occur. These redundant data records are of course a thorn in the side of every data curator and he or she will try to merge or delete the corresponding data records with every finding. As soon as a certain mass of images has been reached, the finding of these duplicates is often left to chance or results from information provided by users. Thanks to the major developments in the field of image similarity search, however, the topic can now also be dealt with more systematically. For example, with the similARiTy tool, which compares two convolutes of digital images on the basis of percepual hashing and BK-tree. First, a json file is created for each directory to be compared, which is then compared with each other. Prerequisites for using this command-line tool are python and imagemagick.

Date de début
Date de fin

Site internet de la manifestation

Infrastructures de recherche

Nom du responsable

Thorsten Wübbena

Thorsten Wübbena

Directeur de recherche / jusqu'à 2019
Screenshot (Detail)