diff --git a/paddlenlp/transformers/bart/tokenizer.py b/paddlenlp/transformers/bart/tokenizer.py index ad93d0a0355b..591e56e0abac 100644 --- a/paddlenlp/transformers/bart/tokenizer.py +++ b/paddlenlp/transformers/bart/tokenizer.py @@ -115,6 +115,9 @@ def __init__(self, mask_token="", **kwargs): + super(BartTokenizer, self).__init__(vocab_file, merges_file, errors, + max_len, pad_token, eos_token) + bos_token = AddedToken(bos_token, lstrip=False, rstrip=False) if isinstance( bos_token, str) else bos_token @@ -147,9 +150,6 @@ def __init__(self, pad_token=pad_token, mask_token=mask_token) - super(BartTokenizer, self).__init__(vocab_file, merges_file, errors, - max_len, pad_token, eos_token) - def _bpe_encode(self, text): bpe_tokens = [] re = try_import("regex")