"git@developer.sourcefind.cn:chenpangpang/ComfyUI.git" did not exist on "5ce4b9eca3c601965f93852274b2e428b9a49179"
Commit ada24def authored by Julien Chaumond's avatar Julien Chaumond
Browse files

[run_lm_finetuning] Tweak fix for non-long tensor, close #2728

see 1ebfeb79

 and #2728
Co-Authored-By: default avatarLysandre Debut <lysandre.debut@reseau.eseo.fr>
parent 2184f870
......@@ -118,7 +118,7 @@ class TextDataset(Dataset):
return len(self.examples)
def __getitem__(self, item):
return torch.tensor(self.examples[item])
return torch.tensor(self.examples[item], dtype=torch.long)
class LineByLineTextDataset(Dataset):
......@@ -138,7 +138,7 @@ class LineByLineTextDataset(Dataset):
return len(self.examples)
def __getitem__(self, i):
return torch.tensor(self.examples[i])
return torch.tensor(self.examples[i], dtype=torch.long)
def load_and_cache_examples(args, tokenizer, evaluate=False):
......@@ -195,7 +195,6 @@ def _rotate_checkpoints(args, checkpoint_prefix="checkpoint", use_mtime=False) -
def mask_tokens(inputs: torch.Tensor, tokenizer: PreTrainedTokenizer, args) -> Tuple[torch.Tensor, torch.Tensor]:
""" Prepare masked tokens inputs/labels for masked language modeling: 80% MASK, 10% random, 10% original. """
inputs = inputs.clone().type(dtype=torch.long)
labels = inputs.clone()
# We sample a few tokens in each sequence for masked-LM training (with probability args.mlm_probability defaults to 0.15 in Bert/RoBERTa)
probability_matrix = torch.full(labels.shape, args.mlm_probability)
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment