Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Run r1 evals for algorithmic datasets #221

Open
joesharratt1229 opened this issue Feb 26, 2025 · 0 comments
Open

Run r1 evals for algorithmic datasets #221

joesharratt1229 opened this issue Feb 26, 2025 · 0 comments
Assignees

Comments

@joesharratt1229
Copy link
Collaborator

joesharratt1229 commented Feb 26, 2025

Given the latency of evaluating R1 comparative to other datasets, it makes sense to segment this into running evaluations by category type. This sub-issue relating to running R1 evals for datasets of the algorithmic category

@joesharratt1229 joesharratt1229 self-assigned this Feb 26, 2025
@joesharratt1229 joesharratt1229 changed the title Run r1 evals for algorithmic dataset Run r1 evals for algorithmic datasets Feb 26, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

1 participant