diff --git a/transformers_doc/en/masked_language_modeling.ipynb b/transformers_doc/en/masked_language_modeling.ipynb index 52ac325f..d7c4a196 100644 --- a/transformers_doc/en/masked_language_modeling.ipynb +++ b/transformers_doc/en/masked_language_modeling.ipynb @@ -742,7 +742,7 @@ "source": [ "from transformers import AutoTokenizer\n", "\n", - "tokenizer = AutoTokenizer.from_pretrained(\"my_awesome_eli5_mlm_model\")\n", + "tokenizer = AutoTokenizer.from_pretrained(\"stevhliu/my_awesome_eli5_mlm_model\")\n", "inputs = tokenizer(text, return_tensors=\"pt\")\n", "mask_token_index = torch.where(inputs[\"input_ids\"] == tokenizer.mask_token_id)[1]" ] @@ -814,7 +814,7 @@ "source": [ "from transformers import AutoTokenizer\n", "\n", - "tokenizer = AutoTokenizer.from_pretrained(\"my_awesome_eli5_mlm_model\")\n", + "tokenizer = AutoTokenizer.from_pretrained(\"stevhliu/my_awesome_eli5_mlm_model\")\n", "inputs = tokenizer(text, return_tensors=\"tf\")\n", "mask_token_index = tf.where(inputs[\"input_ids\"] == tokenizer.mask_token_id)[0, 1]" ] diff --git a/transformers_doc/en/pytorch/masked_language_modeling.ipynb b/transformers_doc/en/pytorch/masked_language_modeling.ipynb index 58177150..b75b09e7 100644 --- a/transformers_doc/en/pytorch/masked_language_modeling.ipynb +++ b/transformers_doc/en/pytorch/masked_language_modeling.ipynb @@ -598,7 +598,7 @@ "source": [ "from transformers import AutoTokenizer\n", "\n", - "tokenizer = AutoTokenizer.from_pretrained(\"my_awesome_eli5_mlm_model\")\n", + "tokenizer = AutoTokenizer.from_pretrained(\"stevhliu/my_awesome_eli5_mlm_model\")\n", "inputs = tokenizer(text, return_tensors=\"pt\")\n", "mask_token_index = torch.where(inputs[\"input_ids\"] == tokenizer.mask_token_id)[1]" ] diff --git a/transformers_doc/en/tensorflow/masked_language_modeling.ipynb b/transformers_doc/en/tensorflow/masked_language_modeling.ipynb index 21821e53..86e17553 100644 --- a/transformers_doc/en/tensorflow/masked_language_modeling.ipynb +++ b/transformers_doc/en/tensorflow/masked_language_modeling.ipynb @@ -616,7 +616,7 @@ "source": [ "from transformers import AutoTokenizer\n", "\n", - "tokenizer = AutoTokenizer.from_pretrained(\"my_awesome_eli5_mlm_model\")\n", + "tokenizer = AutoTokenizer.from_pretrained(\"stevhliu/my_awesome_eli5_mlm_model\")\n", "inputs = tokenizer(text, return_tensors=\"tf\")\n", "mask_token_index = tf.where(inputs[\"input_ids\"] == tokenizer.mask_token_id)[0, 1]" ]