/etc

Following up on my previous post on finding visually-similar images in the Rijksmuseum with Pastec, I thought it would be interesting to apply the same techniques to the New York Public Library’s recent release of 180,000+ public domain images. After tweaking some tools to download the images associated with the data, I again built a Pastec index of the images and searched for matches within the collection. The results should be downloadable here:

One facet made apparent by this process is the differences in the within-collection match results between the Rijksmuseum and the NYPL. Using the same process I’m using to automatically post GIFs of the Rijksmuseum matches to Tumblr let me do the same to generate GIFs for the NYPL’s “unique” matches to get a sense for what was being matched. The first thing that struck me from this was the size of the results: despite having only 2,166 more “unique” matches, the NYPL GIF output was a massive 134GB compared to the Rijksmuseum’s 40GB. Of these, 10,230 (out of 35,195) exceeded Tumblr’s 2MB upload limit, compared to 2,532 (out of 33,029) for the Rijksmuseum. So we have a lot more matches where the number of matches to other items within the collection is greater.

By looking at the resulting images, it also quickly becomes apparent that the types of matches found are also very different. While the Rijksmuseum matches are mostly prints, objects, and photographs, the majority of the NYPL matches fall into a few different categories, where the visual similarity (according to Pastec, at least) of items within these categories results in the very different distribution of matches found:

Subtracting some of these types of objects might winnow things down to some of the more interesting matches, depending on what you’re interested in.

Another thing I did with this data was to try matching the NYPL images against the Rijksmuseum index, to try to find matches between the two collections. That resulted in 1,329 matches, which I’ve made available as a clickable Google Docs spreadsheet (which you can also download as CSV).

Ultimately, both of these experiments suggest to me that it might be useful to use a process like this to surface related items within digital collections webpages when images are available. I’ve come across many matches turned up by this process which would be helpful to know about when looking at any of the objects contained in the match set, but would be nearly impossible to find manually from the existing metadata. It might also be interesting to merge multiple indexes together to achieve something like a “Cultural Heritage TinEye” (imagine also adding images from Europeana, WikiMedia Commons, etc.).