-
Notifications
You must be signed in to change notification settings - Fork 6.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Is this possible to produce text with few shot learning? #252
Comments
At the scale of nanoGPT basically the answer is no. ICL (in context learning) emerges a few B parameters down the road. |
Then may i ask if i would fine tune the gpt model i trained on a prompt-answer dataset, can i get a kind of ChatGPT like model? The reason i want is to have a model in my language answering questions on some of the domains i want. Thanks for the reply. |
Hi! Try loading gpt-XL weights and fine tune to your prompt-answer dataset, It should be able to produce your desired output |
@C080 |
It could pick up Turkish if it has been trained on a multi-lingual dataset with turkish inside! Anyway try using two layers of Google Translates after & before so all the reasoning happens in english! |
This is totally possible if you scale this a lot, but there are much better models for this like bert finetuned or sentiment analysis, my repo uses a similar style, but for chat messages, like so:
|
Add MLP Expansion factor control and sweep
I trained a gpt model using this repo. I tried to produce text using few shot learning like the one below:
The result i get isn't something related. Does this repo enables that feature or is my model bad?
The text was updated successfully, but these errors were encountered: