Extraction:
SpaCy TextRank:
Captured 4
longer sequences
long input sentences
high bleu score
the complete input sequence
vector
the input sequence
information
long dependencies
a candidate translation
simpler words
the entire sequence
fixed length
valuable parts
parallel processing
one or more
reference translations
Seq2Seq Models
some input words
encodes
the input
sequence
hidden state
Vaswani et al
a context vector
this context vector
this fixed-length context vector design
the
complete sequence
an output word
image captioning
various problems
Attention
attention
------------------------------
Gensim TextRank:
Captured 4
learning
learn
attention
words
word
encoder
encodes
translation
translated
translations
translate
decoder
paper
input sequence
sequences
vector
bleu
sentence
sentences
unit
ideas
idea
parts
models
model
like image
length
designed
design
long
context
works
work
processes
processed
processing
hidden
evaluation
information
------------------------------
Rake:
Captured 1
2014 paper “ neural machine translation
neural machine translation using seq2seq models
various problems like image captioning
“ thought vector ”)
every encoder hidden state
vaswani et al .,
decoder unit works well
high bleu score ).
length context vector design
retain longer sequences
bilingual evaluation understudy
decoder unit fails
whole long sentence
famous paper attention
deep learning community
deep learning arena
long input sentences
complete input sequence
candidate translation
et al
seq2seq model
context vector
paper laid
complete sequence
long dependencies
jointly learning
shorter sentences
fixed length
input sequence
decoder architecture
------------------------------
Rakun:
01-Mar-21 20:38:03 - Initiated a keyword detector instance.
01-Mar-21 20:38:03 - Number of nodes reduced from 128 to 119
Captured 5
attention
sequence
input sequence attention
model
paper
using
encoder
seq2seq
encoder-decoder
vector
information
architecture
learning
decoder
ideas
prominent
translation
hence
processes
previous
extension
natural
reads
translate
align
candidate
comparing
score
instead
concentrated
------------------------------
Yake:
Captured 5
neural machine translation
input sequence
deep learning community
context vector
sequence
input
neural machine
machine translation
attention
model
complete input sequence
learning community
deep learning
context
vector
bilingual evaluation understudy
encoder-decoder
translation
learning
prominent ideas
fixed-length context vector
context vector design
long input sentences
machine
encoder-decoder unit
neural
encoder-decoder model
words
encoder
complete sequence
------------------------------
KeyBERT:
01-Mar-21 20:38:03 - Load pretrained SentenceTransformer: distilbert-base-nli-mean-tokens
01-Mar-21 20:38:03 - Did not find folder distilbert-base-nli-mean-tokens
01-Mar-21 20:38:03 - Try to download model from server: https://sbert.net/models/distilbert-base-nli-mean-tokens.zip
01-Mar-21 20:38:03 - Load SentenceTransformer from folder: /Users/irene/.cache/torch/sentence_transformers/sbert.net_models_distilbert-base-nli-mean-tokens
01-Mar-21 20:38:04 - Use pytorch device: cpu
Batches: 100%|██████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 1/1 [00:00<00:00, 9.71it/s]
Batches: 100%|████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 11/11 [00:01<00:00, 6.87it/s]
Captured 1
deep learning
neural machine
encoder decoder
learning arena
ideas deep
context neural
decoder architecture
machine translation
transformers revolutionized
graph encoder
memorize long
architecture encoder
learning community
decoder initialized
prominent ideas
reflected graph
revolutionized deep
translations graph
valuable parts
connecting encoder
famous paper
focus valuable
candidate translation
paper neural
composed encoder
arena concept
helps model
output critical
memorize
foundation famous