Day 11 In-class Notes
most RNN’s belong to the general category of “generative” models, in that these models provide a departure from strict classification or prediction tasks. it’s important to remember, though, that even generative models still tell us alot about how we project rational thought onto the world, and depend greatly on how our design, contextualization, objectives, and choice of training data!
for a loose narrative of using RNN’s for artistic purposes see ross goodwin ”adventures in narrated reality” - see also “Sunspring” the movie he made with RNN-generated script
allison parrish: “When we teach computers to write, the computers don’t replace us any more than pianos replace pianists—in a certain way, they become our pens, and we become more than writers. We become writers of writers.”
some thoughts on exploring semantic space with experimental writing from Allison Parrish, and these two posts from katie rose pipkin (see also their moth generator twitter bot)
for technical background on RNN’s:
- intro to RNN’s from andrej karpathy
- intro to RNN’s from “machine learning is fun”
- intro to RNN’s from wildML
- intro to RNN’s from chris olah (chinese translation here)
- collection of many resources related to learning about RNN’s
- attention and memory
- augmented RNN’s (more on attention and memory)
compared to maximum-likelihood language models (e.g. Markov chains), RNN’s manage to balance structure at different scales (context-awareness) while generating completely new material (no copying).
compared to using other natural langauge processing tools on their own (like “bag-of-words”), RNN’s in general, and LTSM’s in particular, help keep track of overall sentiment or longer arcs.
some projects using RNN’s:
- generating click-bait news stories
- bot based on a collection of one human’s communications
- musical compositions (with interesting repetitive features!)
- generating new choreography from existing dances suggests the possibility of using RNN’s to parse the relationship between points in space aka structured feature learning
- memo atken’s realtime sequence generator / style shifter
- ”recurrent net dreams up fake chinese characters” from otoro blog
- a neural representation of sketch drawings
some code repositories of RNN-based projects: