-
Notifications
You must be signed in to change notification settings - Fork 303
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
fix: Add datasets in CodeRAG-Bench #1595
base: main
Are you sure you want to change the base?
Conversation
@KennethEnevoldsen @isaac-chung Hi, can I get the approval for the workflow. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks! Good first try. Added a few suggestions. And since the dataset_transform and split_by_first_newline are mostly repeated, let's put these classes in the same file. That way the functions can be written once and reused.
mteb/tasks/Retrieval/code/CodeRAGLibraryDocumentationRetrieval.py
Outdated
Show resolved
Hide resolved
mteb/tasks/Retrieval/code/CodeRAGLibraryDocumentationRetrieval.py
Outdated
Show resolved
Hide resolved
mteb/tasks/Retrieval/code/CodeRAGLibraryDocumentationRetrieval.py
Outdated
Show resolved
Hide resolved
mteb/tasks/Retrieval/code/CodeRAGLibraryDocumentationRetrieval.py
Outdated
Show resolved
Hide resolved
mteb/tasks/Retrieval/code/CodeRAGLibraryDocumentationRetrieval.py
Outdated
Show resolved
Hide resolved
I noticed that |
That can be one option. I'd prefer to limit this PR to CodeRAG-bench and avoid any refactoring other files. |
I also noticed that the evaluation score for |
…rivate by adding a underscore prefix
This is common on many tasks. E.g. voyage results on |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Nice. I see two main items from docs before we merge:
- Add dataset metrics: "Add metadata to the task (run task.calculate_metadata_metrics())"
- Add a benchmark entry in https://github.com/embeddings-benchmark/mteb/blob/main/mteb/benchmarks/benchmarks.py to reference these datasets.
That will complete the PR. Let us know if you have any questions. Thanks again for iterating!
Any way to run |
@hepengfe hmm I don't have a good suggestion. In light of that, I'd say adding a benchmark entry (more important) and whatever dataset metrics you already have should complete the PR. That way we can look at the descriptive stats issue separately. |
This PR address #1151
It has a blocker as the dataset cannot be downloaded as reported here code-rag-bench/code-rag-bench#5
Update on 1/3/2024: the dataset server has recovered.
Checklist
make test
.make lint
.Adding datasets checklist
Reason for dataset addition: ...
mteb -m {model_name} -t {task_name}
command.sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2
intfloat/multilingual-e5-small
self.stratified_subsampling() under dataset_transform()
make test
.make lint
.