Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Why is the effect obtained after running the evaluated code less than 0.05 instead of 0.8 in the table (such as Thai BM25)? #10

Open
talk2much opened this issue Oct 25, 2023 · 3 comments

Comments

@talk2much
Copy link

Why is the effect obtained after running the evaluated code less than 0.05 instead of 0.8 in the table (such as Thai BM25)?

@lintool
Copy link
Member

lintool commented Nov 1, 2023

Can you provide more details please?

@talk2much
Copy link
Author

talk2much commented Nov 1, 2023 via email

@crystina-z
Copy link
Member

Hi @talk2much! Thanks for your interest in our work!

I would suggest on verifying the content on your dataset/qrels.test.txt? I'm running the following comment on my end:

First:

$ python -m pyserini.search --bm25 \
> --language th \
> --topics mrtydi-v1.1-thai-test \
> --index mrtydi-v1.1-thai \
> --output runs/run.bm25.mrtydi-v1.1-thai.test.txt

which is the same as you shared; and then evaluate using:

$ python -m pyserini.eval.trec_eval -c -m recip_rank -m recall.100 mrtydi-v1.1-thai-test runs/run.bm25.mrtydi-v1.1-thai.test.txt

This gives results that match to the paper:

Results:
recip_rank              all     0.4016
recall_100              all     0.8529

In addition, I have another question, if we can directly calculate the results of recall from topic.test.csv. So what does topic.train.csv do here? Is my thinking or procedure wrong?

I'm not sure if I fully understand the question, but the topic.train.csv is only needed if you want to fine-tune the model using the training set, and just for evaluation we'd only need topic.test.csv.

Let us know if you have more questions!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants