Thursday, June 14, 2018

"Dank Learning: Generating Memes Using Deep Neural Networks"

This is not the future I was promised.

From Stanford University's Computer Science mavens via
Abel L Peirson V, E Meltem Tolunay
We introduce a novel meme generation system, which given any image can produce a humorous and relevant caption. Furthermore, the system can be conditioned on not only an image but also a user-defined label relating to the meme template, giving a handle to the user on meme content. The system uses a pretrained Inception-v3 network to return an image embedding which is passed to an attention-based deep-layer LSTM model producing the caption - inspired by the widely recognised Show and Tell Model. We implement a modified beam search to encourage diversity in the captions. We evaluate the quality of our model using perplexity and human assessment on both the quality of memes generated and whether they can be differentiated from real ones. Our model produces original memes that cannot on the whole be differentiated from real ones.
Comments: Stanford CS 224n Project
Subjects: Computation and Language (cs.CL); Learning (cs.LG)
Cite as: arXiv:1806.04510 [cs.CL]
(or arXiv:1806.04510v1 [cs.CL] for this version)

arXiv download page (9 page PDF)