Commit 423c51b0 authored by Neel Kant's avatar Neel Kant
Browse files

Bugfix and remove unneeded script

parent e949a5c5
......@@ -966,7 +966,7 @@ class InverseClozeDataset(data.Dataset):
padless_max_len = self.max_seq_len - 2
# select a random sentence from the document as input
input_sentence_idx = rng.randint(num_sentences)
input_sentence_idx = rng.randint(0, num_sentences - 1)
tokens, token_types = self.sentence_tokenize(doc[input_sentence_idx], 0)
input_tokens, input_token_types = tokens[:target_seq_length], token_types[:target_seq_length]
if not len(input_tokens) > 0:
......
#!/bin/bash
LENGTH=512
CHKPT="chkpts/debug"
COMMAND="/home/scratch.gcf/adlr-utils/release/cluster-interface/latest/mp_launch python pretrain_bert_ict.py \
--num-layers 6 \
--hidden-size 768\
--num-attention-heads 12 \
--batch-size 1 \
--checkpoint-activations \
--seq-length $LENGTH \
--max-position-embeddings $LENGTH \
--train-iters 1000 \
--no-save-optim --no-save-rng \
--save $CHKPT \
--resume-dataloader \
--train-data /home/universal-lm-data.cosmos549/datasets/wikipedia/wikidump_lines.json \
--presplit-sentences \
--loose-json \
--text-key text \
--data-loader lazy \
--tokenizer-type BertWordPieceTokenizer \
--cache-dir cache \
--split 58,1,1 \
--distributed-backend nccl \
--lr 0.00015 \
--num-workers 0 \
--no-load-optim --finetune \
--lr-decay-style cosine \
--weight-decay 1e-2 \
--clip-grad 1.0 \
--warmup .01 \
--save-interval 1000 \
--fp16 --adlr-autoresume --adlr-autoresume-interval 5000"
submit_job --image 'http://gitlab-master.nvidia.com/adlr/megatron-lm/megatron:rouge_score' --mounts /home/universal-lm-data.cosmos549,/home/raulp -c "${COMMAND}" --name ict_test --partition interactive --gpu 8 --nodes 2 --autoresume_timer 300 -i
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment