Differences
This shows you the differences between two versions of the page.
Both sides previous revision Previous revision Next revision | Previous revision | ||
courses:rg:2014:wr [2014/11/04 01:24] hoangt |
courses:rg:2014:wr [2014/11/04 14:41] (current) popel |
||
---|---|---|---|
Line 8: | Line 8: | ||
dog 11010111010 | dog 11010111010 | ||
cat 11010111010 | cat 11010111010 | ||
- | 2. Section 4.1 defines a corrupted (or noise) n-gram, but there is a tiny error/typo in the definition. Try to be nitpicking and point it out. | + | 2. Section 4.1 defines a corrupted (or noise) n-gram, but there is a tiny error/typo in the definition. Try nitpicking and point it out. |
3. Section 7.4 states that "word representations in NER brought larger gains on the out-of-domain data than on the in-domain data." Try to guess what is the reason. | 3. Section 7.4 states that "word representations in NER brought larger gains on the out-of-domain data than on the in-domain data." Try to guess what is the reason. | ||
Line 16: | Line 16: | ||
b) Does it contain any compound feature with word representations? | b) Does it contain any compound feature with word representations? | ||
c) Give an example of a possible compound feature with word representations for the NER task. | c) Give an example of a possible compound feature with word representations for the NER task. | ||
+ | |||
+ | 5. | ||
+ | Consider the C&W embedding vectors with 50 dimensions. Guess which word has the embedding vector most similar (by Euclidean distance) to the following vector: | ||
+ | a) vector(king) - vector(man) + vector(woman) | ||
+ | b) vector(dollars) - vector(dollar) + vector(mouse) | ||
**Hint** : The paper is 11-page long. You can skip section 2 and section 3.2 which are the literature review. | **Hint** : The paper is 11-page long. You can skip section 2 and section 3.2 which are the literature review. | ||