Better Summarization

(David Gutman) #1

An interesting addition to standard seq2seq + attention models for summarization.

Main additions are a probability of copying a word from the original (allows out of vocabulary words) and a coverage penalty (minimize overlap of current attention vector with sum of previous attention vectors) to avoid repetition.

Looks very promising and has a nice write up ( in addition to the paper:

(Brendan Fortuner) #2

Really cool, thanks for sharing.

(Jeremy Howard) #3

I really like the approach, and I think the writeup is one of the clearest I’ve read :slight_smile:

(Arvind Nagaraj) #4

Thanks for sharing!!