Introduction
The gloVe is an efficient word vector learning algorithm discussed in this post. This article will explain why GloVe is superior and the motivation behind GloVe's cost function, which is the most critical aspect of the algorithm.
The gloVe is a word vector technology that, after a brief pause, rode the wave of word vectors. To refresh our memory, word vectors organize words into a pleasant vector space where related words cluster together, and different words repel each other. The gloVe benefits that, unlike Word2vec, it does not rely solely on local statistics (local context information of terms) to generate word vectors but also incorporates worldwide statistics (word co-occurrence).
Reason Behind Using GloVe
On the premise that "you shall know a word by the company it keeps," word vectors were created. You take a large corpus and turn it into a dataset of tuples, with each tuple containing a single value (some word x, a word in the context of x). Then, given the word x, you'd utilize your old pal, a neural network, to learn to predict the context word of x. Why not stay with Word2vec after its impressive performance? The cause is the fundamentals of the solution formulation, not performance. Keep in mind that Word2vec only uses language information from the local area. The semantics learned for a specific word are only influenced by the words that surround it.
Take the line "The cat sat on the carpet," for example. If we use Word2vec, we won't be able to find out whether "the" is a particular context for the terms "cat" and "carpet." Is "the" merely a stopword?
This can be considered suboptimal, especially among theoreticians.
This is when GloVe enters the picture. "Global Vectors" is the abbreviation for "Global Vectors." GloVe, as previously indicated, captures both global and local statistics of a corpus to generate word vectors. Do we, however, require both international and local statistics? It turns out that each form of statistic has its benefit. Word2vec, which records local statistics, performs exceptionally well in analogy tasks. However, a method like LSA that uses global statistics performs poorly in analogy tasks. However, because the Word2vec technique relies solely on local statistics, it has several drawbacks (as stated above).
Read more:- Machine Learning






