4 minute read

Few years past with introduction of word vectors, NLP research has again rapidly escalated again with idea of Pretrained Language modeling, which in turn gave significant boost in field of generating unsupervised contextual embedding of sentences.

We will look into series of notable papers and compare their performance in SentEval task to calculate sentence similarity.

Wait .. Why need Unsupervised Embeddings ?

Apart from lack of supervised data-set for most of real world problems, contextual embedding of text is one decent leap towards make computer close to humans in language Understanding. But don’t worry, we are nowhere there yet. The ideas discussed below may be state of the art , but just good in a restricted settings.

In my own experience as ML Practitioner, when working on real life problems related to Natural Language Understanding like, Question Answering with FAQ, Fact checking systems e.t.c, chief problem is lack of supervised dataset. Say a Garage company approaches to you and tells, they need automation of emails through FAQs they have prepared to generic queries. Before that, a human had to manually search the FAQ and reply to clients mail.

We will discuss the following sentence embedding techniques and their performance in senteval task.

by michilov brought along significant boost in NLP research with notion of deep learning based transfer learning. Further research extended to using word vectors to generate sentence embedding, most significant to mention are these techniuqes below

1. Simple Average of Each words of sentences

It’s mathmatically proved that, simple average of words gives empirically correct sentence embeddings.

2. A SIMPLE BUT TOUGH-TO-BEAT BASELINE FOR SENTENCE EMBEDDINGS

3. Skip Thought Vectors

3.1 Basic Architecture

Trained in similar way as skip-gram word2vec model was trained. We have a encoder decoder architecture, closely resembling same one as in landmark paper neural machine translation||PAPER_LINK .

As with any encoder decoder architecture, encoder will generate a thought vector, which is processed by the decoder to generate the words

The training architecture is , (s_(i-1) , s_i , s_(i+1) ) , the encoder takes in s_i , which itself is composed of t words.

##

LOOK DETAILS HERE ::

https://towardsdatascience.com/paper-summary-evaluation-of-sentence-embeddings-in-downstream-and-linguistic-probing-tasks-5e6a8c63aab1

https://mlexplained.com/2017/12/28/an-overview-of-sentence-embedding-methods/

https://paperswithcode.com/paper/concatenated-power-mean-word-embeddings-as

https://paperswithcode.com/paper/evaluation-of-sentence-embeddings-in

Explaine history in youtube video

https://www.youtube.com/watch?v=nFCxTtBqF5U

Sentence embeddings for automated factchecking - Lev Konstantinovskiy

https://www.youtube.com/watch?v=ddf0lgPCoSo

as sentences and paragraphs. Surprisingly, Wieting et al (ICLR’16) showed that such complicated methods are outperformed, especially in out-of-domain (transfer learning) settings, by simpler methods involving mild retraining of word embeddings and basic linear regression. The method of Wieting et al. requires retraining with a substantial labeled dataset such as Paraphrase Database (Ganitkevitch et al., 2013). The current paper goes further, showing that the following completely unsupervised sentence embedding is a formidable baseline: Use word embeddings computed using one of the popular methods on unlabeled corpus like Wikipedia, represent the sentence by a weighted average of the word vectors, and then modify them a bit using PCA/SVD. This weighting improves performance by about 10% to 30% in textual similarity tasks, and beats sophisticated supervised methods including RNN’s and LSTM’s. It even improves Wieting et al.’s embeddings. This simple method should be used as the baseline to beat in future, especially when labeled training data is scarce or nonexistent. The paper also gives a theoretical explanation of the success of the above unsupervised method using a latent variable generative model for sentences, which is a simple extension of the model in Arora et al. (TACL’16) with new “smoothing” terms that allow for words occurring out o

Word2vec   PAPER:LINK by michilov brought along significant boost in NLP research with notion of deep learning based transfer learning. Further research extended to using word vectors to sentence embedding, most significant to mention are these techniuqes below
  1. Simple Average of Each words of sentences

It’s mathmatically proved that, simple average of words

  1. A SIMPLE BUT TOUGH-TO-BEAT BASELINE FOR SENTENCE EMBEDDINGS
  2. Skip Thought Vectors

along with it’s modifications variants

Few years past with introduction of word vectors, NLP research has again rapidly escalated again with idea of Pretrained Language modeling, which in turn gave significant boost in field of generating unsupervised contextual embedding of sentences.

We will look into series of notable papers and their working mechanism.

Wait .. Why need unsupervised embeddings ?

Apart from lack of supervised data-set for most of real world problems, contextual embedding of text is one decent leap towards make computer close to humans in language Understanding. But don’t worry, we are nowhere there yet. The ideas discussed below may be state of the art , but just good in a restricted settings.

In my own experience as ML Practitioner, when working on real life problems related to Natural Language Understanding like, Question Answering with FAQ, Fact checking systems e.t.c, chief problem is lack of supervised dataset. Say a Garage company approaches to you and tells, they need automation of emails through FAQs they have prepared to generic queries. Before that, a human had to manually search the FAQ and reply to clients mail.

STATIC NLP FOR TEXT EMBEDDING ## HISTORY LEKH VAI

We will discuss the following sentence embedding techniques and their variants in details.

  1. Word Vectors
    • Average of words vectors
    • A simple but tough to beat baseline sentence embedding
    • Concatenated Power Mean Word Embeddings as Universal Cross-Lingual Sentence Representations
  2. Infersent
  3. Universal Sentence Encoder
  4. BERT and it’s variations

use here and see which outperformes , visualize the vector space of sentence embedding

https://github.com/brmson/dataset-sts/tree/master/data/sts/semeval-sts

Use simple text classification dataset, and check how

nearest neighbour on classification performs with these unsupervised approaches. i.e without training

def print_hi(name)
  puts "Hi, #{name}"
end
print_hi('Tom')
#=> prints 'Hi, Tom' to STDOUT.