Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Prune test data to reduce inference time #14

Open
zeynepakkalyoncu opened this issue Jun 10, 2019 · 3 comments
Open

Prune test data to reduce inference time #14

zeynepakkalyoncu opened this issue Jun 10, 2019 · 3 comments
Assignees

Comments

@zeynepakkalyoncu
Copy link
Member

Throw away all sentence that don't at least have a term that matches the sentence? Other pruning scenarios?

@zeynepakkalyoncu zeynepakkalyoncu self-assigned this Jun 10, 2019
@zeynepakkalyoncu
Copy link
Member Author

Removing sentences that do not contain any of the terms in the query reduces the size from 10.6m to 1.4m. However, the scores also take a significant hit:

Experiment: mb_5cv_pruned
1S:
map                   	all	0.3029
P_20                  	all	0.4157
2S:
map                   	all	0.3045
P_20                  	all	0.4163
SS:
map                   	all	0.3034
P_20                  	all	0.4175

I will explore other pruning methods, but it doesn't look too promising.

@lintool
Copy link
Member

lintool commented Nov 22, 2019

Let's try reranking only the top 100 docs, instead of top 1000.
NDCG is the right metric here; AP will likely be bad.

@zeynepakkalyoncu
Copy link
Member Author

NDCG@20 for BERT(MSMARCO, MB) on sentences of top 1000/100 Robust04 docs:

Top 1000 Top 100 (optimized wrt NDCG) Top 100 (optimized wrt MAP)
1S 0.5239 0.5131 0.5117
2S 0.5324 0.5206 0.5200
3S 0.5325 0.5228 0.5196

Note that hyperparameters for the first top 100 column are tuned to maximize NDCG@20 (not MAP, which is the default shown in the last column, as I wanted to see the difference). AP is pretty bad for top 100 as expected, but NDCG@20 is reasonable considering it gives us a ~10x speedup.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants