# Throw an error if we can pad because there is no padding token
ifpad_to_max_lengthandself.pad_token_idisNone:
raiseValueError(
"Unable to set proper padding strategy as the tokenizer does not have a padding token. In this case please set the `pad_token` `(tokenizer.pad_token = tokenizer.eos_token e.g.)` or add a new pad token via the function add_special_tokens if you want to use a padding strategy"
@@ -1115,6 +1121,12 @@ class PreTrainedTokenizer(object):
...
@@ -1115,6 +1121,12 @@ class PreTrainedTokenizer(object):
"Input is not valid. Should be a string, a list/tuple of strings or a list/tuple of integers."
"Input is not valid. Should be a string, a list/tuple of strings or a list/tuple of integers."
)
)
# Throw an error if we can pad because there is no padding token
ifpad_to_max_lengthandself.pad_token_idisNone:
raiseValueError(
"Unable to set proper padding strategy as the tokenizer does not have a padding token. In this case please set the `pad_token` `(tokenizer.pad_token = tokenizer.eos_token e.g.)` or add a new pad token via the function add_special_tokens if you want to use a padding strategy"
)
ifreturn_offsets_mapping:
ifreturn_offsets_mapping:
raiseNotImplementedError(
raiseNotImplementedError(
"return_offset_mapping is not available when using Python tokenizers."
"return_offset_mapping is not available when using Python tokenizers."
...
@@ -1126,8 +1138,7 @@ class PreTrainedTokenizer(object):
...
@@ -1126,8 +1138,7 @@ class PreTrainedTokenizer(object):