Content deleted Content added
Neonrights (talk | contribs) m added new section |
Neonrights (talk | contribs) |
||
Line 25:
=== Long short-term memory ===
There has been success in using [[Long short-term memory|LSTM]] models to generate paraphrases.<ref name=Prakash>{{Citation|last1=Prakash|first1=Aaditya|last2=Hasan|first2=Sadid A.|last3=Lee|first3=Kathy|last4=Datla|first4=Vivek|last5=Qadir|first5=Ashequl|last6=Liu|first6=Joey|last7=Farri|first7=Oladimeji|title=Neural Paraphrase Generation with Staked Residual LSTM Networks|year=2016|url=https://arxiv.org/abs/1610.03098}}</ref> In short, the model consists of an encoder and decoder component, both implemented using variations of an LSTM, i.e. bi-directional, stacked, etc. First, the encoding LSTM takes the [[word embedding|word embeddings]] of a sentence as input and produces a final hidden vector, which can be viewed as a representation of the input sentence. The decoding LSTM then takes the hidden vector as input and generates new sentence, terminating in an end-of-sentence token.
== Paraphrase recognition ==
|