# Word2Vec The skip-gram model is simply taking a word and predicting the surrounding words from that word. ![](Screen%20Shot%202021-05-18%20at%2010.47.11%20AM.png) ### Continuous Bag of Words (CBOW) A question I had was why is it called continuous bag of words. The answer is touched on here: [natural language - Why CBOW model is called "continuous"? - Cross Validated](https://stats.stackexchange.com/questions/444270/why-cbow-model-is-called-continuous#:~:text=%22We%20denote%20this%20model%20further,representations%20previously%20used%20in%20NLP.) Basically, we can reason about this as follows: > "We denote this model further as CBOW, as unlike standard bag-of-words model, it uses continuous distributed representation of the context". > Since word vectors are elements in $\mathbb{R}^n$, they are inherently continuous, as opposed to the discrete one-hot representations previously used in NLP. ### Distributed Bag of Words (DBOW) --- Date: 20210518 Links to: [Projection-vs-Embedding](/posts//posts//posts/Projection-vs-Embedding.md) References: * [Word2Vec: Distributed Representations of Words and Phrases and their Compositionality](https://www.youtube.com/watch?v=yexR53My2O4) * [Word2Vec tutorial](https://medium.com/@zafaralibagh6/a-simple-word2vec-tutorial-61e64e38a6a1) * [My blog post](https://www.nathanieldake.com/Deep_Learning/03-Recurrent_Neural_Networks-02-Word-Embeddings-and-Poetry.html) * [Learn word embedding](https://lilianweng.github.io/lil-log/2017/10/15/learning-word-embedding.html) * [Word to vec, whiteboard](https://photos.google.com/photo/AF1QipOxnUmAuyA5ZhGLV2aF3bVtxjoY_0ia3dnwhLyH) * [Word2Vec : Natural Language Processing - YouTube](https://www.youtube.com/watch?v=f7o8aDNxf7k) * [TFIDF : Data Science Concepts - YouTube](https://www.youtube.com/watch?v=OymqCnh-APA) * [A simple Word2vec tutorial. In this tutorial we are going to… | by Zafar Ali | Medium](https://medium.com/@zafaralibagh6/a-simple-word2vec-tutorial-61e64e38a6a1) * Notability, word2vec