Below are results and comments from Urvashi using transformer_prepend
and transformer_decoder as model. A PR with a ROUGE script will be coming,
but the email can give some hints for now.
Lukasz
----------
The rouge scores from the last generated model are:
INFO:tensorflow:rouge_1_f_score: 0.2557
INFO:tensorflow:rouge_1_precision: 0.2579
INFO:tensorflow:rouge_1_recall: 0.2558
INFO:tensorflow:rouge_2_f_score: 0.1292
INFO:tensorflow:rouge_2_precision: 0.1305
INFO:tensorflow:rouge_2_recall: 0.1290
INFO:tensorflow:rouge_l_f_score: 0.2438
INFO:tensorflow:rouge_l_precision: 0.2459
INFO:tensorflow:rouge_l_recall: 0.2438
The decoder does output some empty predictions. The rouge script
identifies this as empty input and prints a warning to the console,
but note that the examples are simply treated as those with rouge
score 0.0 which is the expected behavior.
In this paper, the results of a baseline seq2seq with attention and a
50k vocabulary on the test set are:
Rouge 1=31.33
Rouge 2=11.81
Rouge L=28.83
Qualitatively, the outputs seem to be rather short, in fact, cut off
mid sentence, example: “Sam Fuller Jr. says he decided to leave the
Church of ND in 2007 after almost taking his own life. He spent 28
days in a mental institution following the incident, and that is when
he finally left the church. The 47-year-old”
The average length of a target is 56 tokens and max length 1722
tokens, while the average length of a decoded sequence is 39 tokens
and max length is 97. I noticed the hparam "max length" is set to 256.
Does this apply to both source and target sequences? If so, we might
need to make it longer. I ran a quick experiment with placing some dev
example sources in a file and decoding with the decode_from_file flag
turned on. The outputs for these were extremely bizarre. I’ve attached
a log with the outputs, but they are identical for all the examples,
they are extremely repetitive and most importantly they are completely
different from when I decode the dev set directly. Any ideas on what
might be causing this? Looks like a definite bug, but not sure if its
cropped up before.
> --
> You received this message because you are subscribed to the Google Groups
> "tensor2tensor" group.
> To unsubscribe from this group and stop receiving emails from it, send an
> email to
tensor2tenso...@googlegroups.com.
> To post to this group, send email to
tensor...@googlegroups.com.
> To view this discussion on the web visit
>
https://groups.google.com/d/msgid/tensor2tensor/58bc05e8-291b-43b3-b28c-4f1b3acbaeab%40googlegroups.com.