Generating Sequences With Recurrent Neural Networks

504 views
Skip to first unread message

James Bowery

unread,
Oct 10, 2013, 7:50:03 PM10/10/13
to hutter...@googlegroups.com

Alex Graves
Department of Computer Science
University of Toronto

Abstract

This paper shows how Long Short-term Memory recurrent neural networks can be used to generate complex sequences with long-range structure, simply by predicting one data point at a time. The approach is
demonstrated for text (where the data are discrete) and online handwriting (where the data are real-valued). It is then extended to handwriting
synthesis by allowing the network to condition its predictions on a text
sequence. The resulting system is able to generate highly realistic cursive
handwriting in a wide variety of styles.
...
Section 3 applies the prediction network to text from the Penn Treebank and Hutter Prize Wikipedia datasets.
...
Table 2: Wikipedia Results (bits-per-character)
Train Validation (static) Validation (dynamic)
1.42 1.67 1.33

To put the results in context, the current winner of the Hutter Prize (a variant of the PAQ-8 compression algorithm [20]) achieves 1.28 BPC on the same data (including the code required to implement the algorithm), mainstream compressors such as zip generally get more than 2, and a character level RNN applied to a text-only version of the data (i.e. with all the XML, markup tags etc. removed) achieved 1.54 on held-out data, which improved to 1.47 when the RNN was combined with a maximum entropy model [24].


Matt Mahoney

unread,
Nov 13, 2013, 10:11:04 PM11/13/13
to Hutter-Prize
I finally got to read the paper. I did some other tests on enwik8 by
dividing it into a 96 MB training set and 4 MB test set and computed
the test ratio by subtracting the compressed training file from the
compressed complete file. It beats most of the top compressors. I
posted results at
http://encode.ru/threads/1825-Compression-with-recurrent-neural-networks
> --
> You received this message because you are subscribed to the Google Groups
> "Hutter Prize" group.
> To unsubscribe from this group and stop receiving emails from it, send an
> email to hutter-prize...@googlegroups.com.
> To post to this group, send email to hutter...@googlegroups.com.
> Visit this group at http://groups.google.com/group/hutter-prize.
> For more options, visit https://groups.google.com/groups/opt_out.



--
-- Matt Mahoney, mattma...@gmail.com

James Bowery

unread,
Nov 14, 2013, 3:44:20 AM11/14/13
to Hutter Prize

James Bowery

unread,
Nov 14, 2013, 3:47:30 AM11/14/13
to Hutter Prize
The attached (unpublished) paper by Robert Johnson derives control probabilities in program flow.  Loops produce complex valued probabilities.  

I have to wonder whether recurrence in neural networks would be more tractable if treated in terms of normed division algebras or Clifford algebras.




On Wed, Nov 13, 2013 at 9:11 PM, Matt Mahoney <mattma...@gmail.com> wrote:
Reply all
Reply to author
Forward
0 new messages