This experiment is based on the ClueWeb09 dataset, the Category B subset (the first 50 million English webpages).
Our scripts assume the following resources:
index
: symlink to the ClueWeb09B indri indexDrQA
: symlink to the DrQA repowiki.en.bin
: from the pretrained fasttext embeddings
To build the feature files for all experimental runs (may take a while):
make
To run the experiment:
- Use python script
make_kfold_split.py
to partition data into 10 folds. - Train a model for each fold with RankLib options
-ranker 4 -metric2t NDCG@20 -norm zscore
- Make predictions and compile all results into a TREC run file.
- Evaluate the run using trec_eval and the
qrels
file.
To use the data:
- A set of TREC run files can be found in the directory
runs
. - The CQA answers for TREC Web queries 1-200 are stored in the file
ya.json.gz
.