Extraction: SpaCy TextRank: Captured 4 longer sequences long input sentences high bleu score the complete input sequence vector the input sequence information long dependencies a candidate translation simpler words the entire sequence fixed length valuable parts parallel processing one or more reference translations Seq2Seq Models some input words encodes the input sequence hidden state Vaswani et al a context vector this context vector this fixed-length context vector design the complete sequence an output word image captioning various problems Attention attention ------------------------------ Gensim TextRank: Captured 4 learning learn attention words word encoder encodes translation translated translations translate decoder paper input sequence sequences vector bleu sentence sentences unit ideas idea parts models model like image length designed design long context works work processes processed processing hidden evaluation information ------------------------------ Rake: Captured 1 2014 paper “ neural machine translation neural machine translation using seq2seq models various problems like image captioning “ thought vector ”) every encoder hidden state vaswani et al ., decoder unit works well high bleu score ). length context vector design retain longer sequences bilingual evaluation understudy decoder unit fails whole long sentence famous paper attention deep learning community deep learning arena long input sentences complete input sequence candidate translation et al seq2seq model context vector paper laid complete sequence long dependencies jointly learning shorter sentences fixed length input sequence decoder architecture ------------------------------ Rakun: 01-Mar-21 20:38:03 - Initiated a keyword detector instance. 01-Mar-21 20:38:03 - Number of nodes reduced from 128 to 119 Captured 5 attention sequence input sequence attention model paper using encoder seq2seq encoder-decoder vector information architecture learning decoder ideas prominent translation hence processes previous extension natural reads translate align candidate comparing score instead concentrated ------------------------------ Yake: Captured 5 neural machine translation input sequence deep learning community context vector sequence input neural machine machine translation attention model complete input sequence learning community deep learning context vector bilingual evaluation understudy encoder-decoder translation learning prominent ideas fixed-length context vector context vector design long input sentences machine encoder-decoder unit neural encoder-decoder model words encoder complete sequence ------------------------------ KeyBERT: 01-Mar-21 20:38:03 - Load pretrained SentenceTransformer: distilbert-base-nli-mean-tokens 01-Mar-21 20:38:03 - Did not find folder distilbert-base-nli-mean-tokens 01-Mar-21 20:38:03 - Try to download model from server: https://sbert.net/models/distilbert-base-nli-mean-tokens.zip 01-Mar-21 20:38:03 - Load SentenceTransformer from folder: /Users/irene/.cache/torch/sentence_transformers/sbert.net_models_distilbert-base-nli-mean-tokens 01-Mar-21 20:38:04 - Use pytorch device: cpu Batches: 100%|██████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 1/1 [00:00<00:00, 9.71it/s] Batches: 100%|████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 11/11 [00:01<00:00, 6.87it/s] Captured 1 deep learning neural machine encoder decoder learning arena ideas deep context neural decoder architecture machine translation transformers revolutionized graph encoder memorize long architecture encoder learning community decoder initialized prominent ideas reflected graph revolutionized deep translations graph valuable parts connecting encoder famous paper focus valuable candidate translation paper neural composed encoder arena concept helps model output critical memorize foundation famous