Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

the source of the training data #6

Open
darkpromise98 opened this issue Dec 23, 2024 · 2 comments
Open

the source of the training data #6

darkpromise98 opened this issue Dec 23, 2024 · 2 comments

Comments

@darkpromise98
Copy link

Thanks for your open-source data of public_long_form_thought_data_5k.jsonl

I want to know the source of the data. These data come from the "Distill" (generated by QwQ/Deepseek-R1) or "Explore" (generated by STILL-2) ?

@EliverQ
Copy link
Contributor

EliverQ commented Dec 23, 2024

Thanks for your attention. These data come from the "Distill" (generated by QwQ/Deepseek-R1).

@why11699
Copy link

why11699 commented Jan 2, 2025

@EliverQ @Timothy023
Thank you for the inspiring work and the open-source data.
Just wanna ask the difference between the data you guys released and the one in your paper, as there are only 3.9k distilled data at most in your experiments.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants