Hi all,
I'm trying to train an RNN based on Karpathy's code at
https://github.com/karpathy/char-rnn.
My RNN has ~200 steps, 128 hidden unit and the vocab is ~20k. In clone_many_times (cloning RNN units with shared params and grads), even if collectgarbage('count') shows that about 5MB is in used, `top` shows that the memory is used pretty quickly and it eventually got a out-of-memory problem...
I don't think 200 clones could eat up all memory on the machine which has >100g memory..
Does anyone have any idea? Thanks!!
He