"test/vscode:/vscode.git/clone" did not exist on "9f3bd2ad39e63a8738a7bc66c98361f62ee53637"
Commit 651cc107 authored by Mostofa Patwary's avatar Mostofa Patwary
Browse files

Merging megatron with ICT

parents e919dd8e c601d751
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
...@@ -149,6 +149,7 @@ def train_valid_test_datasets_provider(train_val_test_num_samples): ...@@ -149,6 +149,7 @@ def train_valid_test_datasets_provider(train_val_test_num_samples):
short_seq_prob=args.short_seq_prob, short_seq_prob=args.short_seq_prob,
seed=args.seed, seed=args.seed,
skip_warmup=(not args.mmap_warmup), skip_warmup=(not args.mmap_warmup),
binary_head=False,
dataset_type='ict') dataset_type='ict')
print_rank_0("> finished creating BERT ICT datasets ...") print_rank_0("> finished creating BERT ICT datasets ...")
......
This diff is collapsed.
This diff is collapsed.
...@@ -61,7 +61,7 @@ if __name__ == '__main__': ...@@ -61,7 +61,7 @@ if __name__ == '__main__':
elif args.task in ['MNLI', 'QQP']: elif args.task in ['MNLI', 'QQP']:
from glue.finetune import main from glue.finetune import main
elif args.task in ['LAMBADA', 'WIKITEXT103']: elif args.task in ['LAMBADA', 'WIKITEXT103']:
from zeroshot_gpt2.evaluate import main from zeroshot_gpt.evaluate import main
else: else:
raise NotImplementedError('Task {} is not implemented.'.format( raise NotImplementedError('Task {} is not implemented.'.format(
args.task)) args.task))
......
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
def test_import():
import megatron
This diff is collapsed.
...@@ -85,7 +85,7 @@ class Encoder(object): ...@@ -85,7 +85,7 @@ class Encoder(object):
sentence_ids = Encoder.tokenizer.tokenize(sentence) sentence_ids = Encoder.tokenizer.tokenize(sentence)
if len(sentence_ids) > 0: if len(sentence_ids) > 0:
doc_ids.append(sentence_ids) doc_ids.append(sentence_ids)
if self.args.append_eod: if len(doc_ids) > 0 and self.args.append_eod:
doc_ids[-1].append(Encoder.tokenizer.eod) doc_ids[-1].append(Encoder.tokenizer.eod)
ids[key] = doc_ids ids[key] = doc_ids
return ids, len(json_line) return ids, len(json_line)
...@@ -182,6 +182,8 @@ def main(): ...@@ -182,6 +182,8 @@ def main():
for i, (doc, bytes_processed) in enumerate(encoded_docs, start=1): for i, (doc, bytes_processed) in enumerate(encoded_docs, start=1):
total_bytes_processed += bytes_processed total_bytes_processed += bytes_processed
for key, sentences in doc.items(): for key, sentences in doc.items():
if len(sentences) == 0:
continue
for sentence in sentences: for sentence in sentences:
builders[key].add_item(torch.IntTensor(sentence)) builders[key].add_item(torch.IntTensor(sentence))
builders[key].end_document() builders[key].end_document()
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment