Listen "Word2Vec"
Episode Synopsis
The sources explore word embeddings, representing words as numerical vectors to capture meaning. The Skip-gram model is a key method for learning these high-quality, distributed vector representations from large text datasets. This model predicts surrounding words in a sentence, resulting in word vectors that encode linguistic patterns. To enhance the Skip-gram model, the sources introduce techniques like subsampling frequent words and negative sampling for faster, more accurate training. These word vectors can be combined using mathematical operations, enabling analogical reasoning, and the approach is extended to phrase representations.
More episodes of the podcast Large Language Model (LLM) Talk
Kimi K2
22/07/2025
Mixture-of-Recursions (MoR)
18/07/2025
MeanFlow
10/07/2025
Mamba
10/07/2025
LLM Alignment
14/06/2025
Why We Think
20/05/2025
Deep Research
12/05/2025
vLLM
04/05/2025
Qwen3: Thinking Deeper, Acting Faster
04/05/2025
ZARZA We are Zarza, the prestigious firm behind major projects in information technology.