Generative Modeling with Sparse Transformers
CRANK

We've developed the Sparse Transformer, a deep neural network which sets new records at predicting what comes next in a sequence—whether text, images, or sound. It uses an algorithmic improvement of the attention mechanism to extract patterns from sequences 30x longer than possible previously. Read Paper [ View Code [ existing challenge in AI research is modeling long-r

openai.com
Related Topics: Deep Learning