Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

What prompt is used to generate the GPT2 datasets? #50

Open
veenapaddy opened this issue Apr 20, 2023 · 1 comment
Open

What prompt is used to generate the GPT2 datasets? #50

veenapaddy opened this issue Apr 20, 2023 · 1 comment

Comments

@veenapaddy
Copy link

veenapaddy commented Apr 20, 2023

I see that GPT2 is trained on webtext, but not sure how the datasets here are generated? Specifically what prompt was used with GPT2 to generate the "fake" datasets?

@MilerCt
Copy link

MilerCt commented Apr 24, 2023

datasets are generated by asking the gpt-3 or gpt-4 or gpt-2 gpt-1 which count as the data sets but you have to ask the ai 600 times to prefect one question from asking the ai which than you could use next time to see if that question might be close to the other one which is kind of nice not spending 6 hours per question almost like the video when someone was trying to make a ai that plays n64 mario racing cart perfectly but it took a big amount of training even with the gpu

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants