- 快召唤伙伴们来围观吧
- 微博 QQ QQ空间 贴吧
- 文档嵌入链接
- 复制
- 微信扫一扫分享
- 已成功复制到剪贴板
Introduction to NLP
展开查看详情
1 .Introduction to nlp Natural language processing Gene Olafsen
2 .NLP unleashed Translation Summarization Parsing Sentiment Analysis
3 .NLP unleashed Translation Summarization Parsing Sentiment Analysis
4 .Word encoding One-hot Vector 10,000 word vocabulary represented as an incredibly sparse 10,000 dimensional vector Embedding Operates in a small and dense vector The neural network training phase allows the model to learn the embedding position of words in a vector. Backpropagation moves the word embeddings around. "Similar" words typically cluster close to each other.
5 .Word relationships Source: https://www.tensorflow.org/tutorials/word2vec
6 .Word relationships Source: https://www.tensorflow.org/tutorials/word2vec
7 .Walk through the embedding process Create the variable representing the embeddings for every word in the vocabulary. Perform random initialization. Vocabulary_size = 50000 Embedding_size = 150 init_embedding = tf.random_uniform ([ vocabulary_size , embedding_size ], -1.0, 1.0) embeddings = tf.Variable ( init_embedding )
8 .Trained Embedding It is common to reuse your trained embeddings in other NLP applications. There are pretrained word embeddings that are available for download. Similar to transfer learning described in earlier slide decks, which froze specified layers, pretrained embedding layers can likewise be frozen. http://ahogrammer.com/2017/01/20/the-list-of-pretrained-word-embeddings/ https://github.com/Hironsan/awesome-embedding-models https://github.com/bheinzerling/bpemb
9 .