Hi folks -
Sorry about the late notice for this meeting. We'll meet for the first
time this semester on Friday the 13th, in GDC 3.516 at 2:30pm. This
week we'll discuss the following paper (it's relatively short):
Do Deep Nets Really Need to be Deep?
Lei Jimmy Ba, Rich Caruana
http://arxiv.org/pdf/1312.6184v5.pdf
Currently, deep neural networks are the state of the art on problems
such as speech recognition and computer vision. In this extended
abstract, we show that shallow feed-forward networks can learn the
complex functions previously learned by deep nets and achieve
accuracies previously only achievable with deep models. Moreover, in
some cases the shallow neural nets can learn these deep functions
using a total number of parameters similar to the original deep model.
We evaluate our method on the TIMIT phoneme recognition task and are
able to train shallow fully-connected nets that perform similarly to
complex, well-engineered, deep convolutional architectures. Our
success in training shallow neural nets to mimic deeper models
suggests that there probably exist better algorithms for training
shallow feed-forward nets than those currently available.
--
http://www.cs.utexas.edu/~leif