[ Skip to the content ]

Institute of Formal and Applied Linguistics Wiki


[ Back to the navigation ]

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Next revision
Previous revision
courses:rg:2012:sigtest-mt-zilka [2012/11/14 17:19]
zilka vytvořeno
courses:rg:2012:sigtest-mt-zilka [2013/12/02 22:18] (current)
popel
Line 13: Line 13:
 Does the bootstrap resampling (Section 5) assume normal (Gaussian) distribution of the scores of samples? Does the bootstrap resampling (Section 5) assume normal (Gaussian) distribution of the scores of samples?
 ===== Question 4 ===== ===== Question 4 =====
-We bootstrapped 1000 test sets, computed scoreA-scoreB on each, and we got -1000,-950,-900,-850 ... -5,0,0,0,0,0,0,0,0,0,0,1,2,3 ... 970. +We bootstrapped 1000 test sets, computed scoreA-scoreB on each, and we got -1000,-950,-900,-850 ... -50,0,0,0,0,0,0,0,0,0,0,1,2,3 ... 970. 
  
 Based on Section 6, which system is better - A or B? Based on Section 6, which system is better - A or B?
Line 32: Line 32:
 Can you reformulate Section 4 using this view? What is the observed test statistic and what is the null hypothesis? Can you reformulate Section 4 using this view? What is the observed test statistic and what is the null hypothesis?
  
 +====== Presentation ======
 +  * We answered:
 +    * Question 1 - BLEU scores are: 1 - 1.0, 2 - not defined (0.0 or some smoothed value in practice), 3 - 0.2 (based on the incorrect formula in the paper which is missing 1/4)
 +    * Question 2 - broad sampling, samples far apart distributed -> {data_1, data_101, data_201, ...}
 +
 +===== Section 3 =====
 +  * motivation: we don't usually have 30k sentences for testing, so we need an approximate method to obtain reliable scores
 +  * method: divide test set into 100 smaller test sets (300 sentences each)
 +    * consecutive samples - for each of the sets BLEU score varies in range +-8 %
 +    * non-consecutive samples (broad apart) - for each of the sets BLEU varies much less - +-1.5 %
 +  * they make an assumption and claim that there is no difference between comparing output of 2 different MT systems and output of 1 MT systems that is trained just with different data
 +    * Lukas Zilka complained about this assumption - they should have conducted some experiments to support their claim, as there is nothing that suggests we can generalize like that
 +
 +===== Section 4, 5 =====
 +  * we cannot use Student's T distribution to estimate confidence interval for BLEU, because it cannot be constructed in the form of sum of terms to give us mean and variance
 +  * so for estimating the confidence intervals we will use randomized test set generation (= bootstrap resampling) - e.g. we build 1000 new test sets of size 300 sentences out of our small test set of 300 sentences (i.e. we draw (with replacement) samples from the small test set; so we should get 1000 different test sets)
 +  * answer to Question3 - they do not assume there is any particular distribution in the set of BLEU scores of the 1000 test sets (i.e. their method would work regardless of whether the distribution is normal, uniform or any other), but it is perhaps normally distributed
 +
 +===== Section 6 =====
 +  * they use bootstrap resampling to compare 2 systems; we want to determine whether system 1 is better than system 2; we want to determine that from a set of differences of system's performances (i.e. difference of score of system 1 and score of system 2)
 +    * so we determine in what percent of cases system 1 beats system 2, and that's our final confidence that system 1 is better than system 2 (e.g. 45 times out of 50 -> 90% confidence)
 +  * the rest of the paper just proves that the assumption is correct
 +
 +===== Martin's explanation of p-values =====
 +  * two philosophical views of p-value - Fisher's and Pearson's - unfortunately their are mixed in modern textbooks which only confuses us
 +  * we usually set a null hypothesis H0 as: systems are the same, and alternative hypothesis HA: there is difference in the systems; P(H0) + P(HA) = 1
 +  * p-value =
 +    * P(T(X)>=T(x_orig)|H0) = P(x|H0) = //if the compared systems are the same, what's the probability that we see this data//
 +    * unfortunately we tend to view the p-value as P(H0|x) which it is not and we need to apply the Bayes' theorem to get it
 +  * bootstrap resampling can be viewed as p-value = P(d(x) < 0|H0) = P(d(x) > 2*d(x_orig)|H0), and is approximated by S/B; where S is number of system 2 beating system 1 and B is number of measurements
  

[ Back to the navigation ] [ Back to the content ]