Implemented the style transfer technique from "Image Style Transfer Using Convolutional Neural Networks" (Gatys et al., CVPR 2015). The general idea is to take two images, and produce a new image that reflects the content of one but the artistic "style" of the other.
Vanilla recurrent neural networks to train a model that can generate novel captions for images. Used the 2014 release of the Microsoft COCO dataset for image captioning, which consists of 80,000 training images and 40,000 validation images, each annotated with 5 captions.
Vanilla RNNs can be tough to train on long sequences due to vanishing and exploding gradiants caused by repeated matrix multiplication. LSTMs solve this problem by replacing the simple update rule of the vanilla RNN with a gating mechanism as follows.