You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
For all systems, we report results from parameter tuning to optimize average precision (AP) at rank 1000 on the newswire collections, WT10g, and Gov2, and NDCG@20 for the ClueWeb collections.
There was no separation of training and test data, so these results should be interpreted as oracle settings.
If you're going to refer to these effectiveness results, please be aware of what you're comparing!
Additional note: The values produced by these scripts are slightly different than those reported in the article.
The reason for these differences stems from the fact that Anserini evolved throughout the peer review process; the values reported in the article were those generated when the manuscript was submitted.
By the time the article was published, the implementation of Anserini has progressed.
As Anserini continues to improve we will update these scripts, which will lead to further divergences between the published values.
Unfortunately, this is an unavoidable aspect of empirical research on software artifacts.
Update (12/18/2018):
Regression effectiveness values further changed at commit e71df7aee42c7776a63b9845600a4075632fa11c Tue Dec 18 07:45:30 2018 -0500 with upgrade to Lucene 7.6.
Parameter Tuning
Invoke the tuning script on various collections as follows, on tuna: