CS224N Lecture3 筆記

主要內容:

Why do we need approximations of the original skip-gram formulation?

How is the problem mitigated by using negative sampling?

How is CBOW different from skip-gram?

What are the limitations of using SVD on the co-occurrence matrix to

get word vectors?

How does GloVe combine the advantages of count-based models and

predictive models?

How to evaluate word vectors? What are the most commonly used tasks

for intrinsic evaluation?

Which factors could affect the quality of learned vectors?

推薦閱讀:

【Neural Networks and Deep Learning】3.改進神經網路的學習方法(過擬合)
使用py-faster-rcnn進行目標檢測(object detect)
【Neural Networks and Deep Learning】4.神經網路可以計算任何函數的可視化證明(1)
Inception-v2/v3結構解析(原創)

TAG:計算機科學 | 神經網路 | 自然語言處理 |