Content Based Image Retrieval with Siamese Networks

Content Based Image Retrieval with Siamese Networks

Background

This project was created at University as part of my final year project in order to learn more about the use of Convolutional Neural Networks and their efficacy within the branch of computer vision known as Content Based Image Retrieval.

Network Structure

Abstract

Content Based Image Retrieval, otherwise known as ‘CBIR’ is a method used to extract visually similar images from a large database, based on the features of a given query image. This project explores and presents a modified approach to CBIR by using artificial neural networks, as opposed to classical computer vision techniques. The method presented in the report utilises a Siamese Neural Network paired with a variation of One Shot Learning, which is trained on pairs of dissimilar images in order to extract image feature vectors. Visual similarity is then deduced by utilising a distance function that compares the distance between the two feature vectors returned by the network and this will therefore provide a measure of similarity based on the image content learned by the network.

Code Repository and Datasets

You can access the full PDF write-up here and the code associated with the project here.

The datasets used within this project were the Omniglot dataset and a modifed version of the Stanford Dogs dataset.

Credits

The implemented network architecture is based on that of Gregory Koch et al. with the original paper is listed here.


Michele Pascale

Michele Pascale

PhD Student in Mathematics @ Queen Mary University of London

rss facebook twitter github gitlab youtube mail spotify lastfm instagram linkedin google google-plus pinterest medium vimeo stackoverflow reddit quora quora