[ Skip to the content ]

Institute of Formal and Applied Linguistics Wiki


[ Back to the navigation ]

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revision Previous revision
Next revision
Previous revision
Last revision Both sides next revision
user:zeman:parser-adaptation [2008/03/14 13:55]
zeman UMIACS NLP Wiki.
user:zeman:parser-adaptation [2008/03/14 14:25]
zeman make all
Line 4: Line 4:
  
 Yet older notes exist at [[https://wiki.cs.umd.edu/nlpwiki/index.php?title=Parser_Adaptation|UMIACS NLP Wiki]]. Yet older notes exist at [[https://wiki.cs.umd.edu/nlpwiki/index.php?title=Parser_Adaptation|UMIACS NLP Wiki]].
 +
 +<code>setenv PADAPT /net/work/people/zeman/padapt
 +cd $PADAPT</code>
 +
 +This thing is not (yet) under version control.
 +
 +There is a ''Makefile'' in ''$PADAPT''. ''make all'' should take care of all numbers in all experiments. However, some of the procedures (especially reranker training) are better run separately on the LRC cluster.
 +
 +===== To do =====
 +
 +  * Classifier combination. Merge Charniak n-best lists from gloss and delex (and acquis-gloss and acquis-delex). Either let them vote, or let the reranker select from the merged list.
 +  * Find a way to estimate trustworthiness of parses in self-training. Charniak’s rank in n-best list? Voting of three or more parsers? Some sort of sentence classifier that would estimate how easy it is for Charniak to parse? Maybe start with short sentences only? If we can reliably say what parses by Charniak can be trusted, we can restrict self training to those and see what happens.
 +  * Immediately follows the bootstrapping experiment we discussed earlier. It failed but maybe it would not, had we been able to distinguish good from bad examples.
 +  * Reverse the gloss experiment. “Translate” (gloss) Danish training data to Swedish, train a parser on it, test the parser on Swedish.
 +  * Refine glossing. Translate N (100?) most frequent words. Of the rest, translate only M (4? 5?)-letter suffixes. Suffixation is another way of delexicalization and it would also be interesting how it affects accuracy of monolingual parsing.
 +  * Explore the learning curve of the adapted parsers, as one of the reviewers suggested. Gradually increase the amount of Danish training data, monitor the changes in Swedish parsing accuracy.
 +  * Test other languages with different degrees of relatedness. No new research, thus low priority; on the other hand, this could give answer to the doubts one of the reviewers had.
 +  * Test a dependency parser in similar setting. After all, the treebanks we work with are of dependency origin.
 +  * As reviewers suggested more error analysis. What is the nature of the most frequent errors that the adapted parser does and the Danish parser does not? Are they caused by lexical divergences? Morpho-syntactic? Domain mismatch?
 +  * Use Swedish training data instead of Acquis. Strip the structure from them but keep the gold-standard POS tags. This experiment could show the impact of tagging errors. It provides a different domain, too.
  

[ Back to the navigation ] [ Back to the content ]