Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Questions about the use of the model #1

Closed
ScottishFold007 opened this issue Dec 21, 2021 · 1 comment
Closed

Questions about the use of the model #1

ScottishFold007 opened this issue Dec 21, 2021 · 1 comment

Comments

@ScottishFold007
Copy link

image

Is the above code a direct run out of a model that can be used directly? Or does it need tuning or further training?Like

from transformers import pipeline
from transformers import AutoModelForMaskedLM, AutoTokenizer

tokenizer = AutoTokenizer.from_pretrained("new")
model = AutoModelForMaskedLM.from_pretrained("new")


unmasker = pipeline("fill-mask",model=model,tokenizer=tokenizer,device = 0)
from pprint import pprint

pprint(unmasker(f"I come for {unmasker.tokenizer.mask_token} in last time!"))

Will the above application code come out with a good result? Based on your experience?

@bminixhofer
Copy link
Member

Hi, thanks for your interest!

In general, WECHSEL just acts as good initialization, you will need further training to get good results.

You can try the zero-shot case, and it should be better than random, but I don't expect results will be usable for any real-world application.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants