Image Caption Generation: CNN-RNN Architecture

Using computer vision and natural language processing techniques to generate detailed textual captions that describe the content of images. The model is trained on the Flickr8k dataset, which consists of 8000 unique images, each described by five different sentences.

(c) Mark Hodierne

Get in touch