Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
Megatron-LM
Commits
492fdf83
Commit
492fdf83
authored
Nov 25, 2021
by
zihanl
Browse files
update knowledgeable dialogue running script
parent
4676493c
Changes
2
Hide whitespace changes
Inline
Side-by-side
Showing
2 changed files
with
75 additions
and
0 deletions
+75
-0
examples/prompt_knowledge_generation.sh
examples/prompt_knowledge_generation.sh
+38
-0
examples/prompt_response_generation.sh
examples/prompt_response_generation.sh
+37
-0
No files found.
examples/prompt_knowledge_generation.sh
0 → 100644
View file @
492fdf83
#!/bin/bash
WORLD_SIZE
=
8
DISTRIBUTED_ARGS
=
"--nproc_per_node
$WORLD_SIZE
\
--nnodes 1
\
--node_rank 0
\
--master_addr localhost
\
--master_port 6000"
CHECKPOINT_PATH
=
<Specify path
for
the language model>
INPUT_PATH
=
<Specific path
for
the input
test
dataset>
OUTPUT_PATH
=
<Speicifc path
for
the output>
PROMPT_PATH
=
<Specific path
for
the prompts>
python
-m
torch.distributed.launch
$DISTRIBUTED_ARGS
./tasks/main.py
\
--tensor-model-parallel-size
1
\
--pipeline-model-parallel-size
1
\
--num-layers
24
\
--hidden-size
1024
\
--num-attention-heads
16
\
--seq-length
2048
\
--max-position-embeddings
2048
\
--micro-batch-size
1
\
--vocab-file
/gpfs/fs1/projects/gpu_adlr/datasets/nlp/gpt2_indexed_dataset/bpe/gpt2-vocab.json
\
--merge-file
/gpfs/fs1/projects/gpu_adlr/datasets/nlp/gpt2_indexed_dataset/bpe/gpt2-merges.txt
\
--load
${
CHECKPOINT_PATH
}
\
--fp16
\
--DDP-impl
torch
\
--tokenizer-type
GPT2BPETokenizer
\
--out-seq-length
100
\
--sample-input-file
${
INPUT_PATH
}
\
--sample-output-file
${
OUTPUT_PATH
}
\
--prompt-file
${
PROMPT_PATH
}
\
--prompt-type
knowledge
\
--num-prompt-examples
10
\
--dynamic-prompt
\
--task
knwl-dialo-prompt
examples/prompt_response_generation.sh
0 → 100644
View file @
492fdf83
#!/bin/bash
WORLD_SIZE
=
8
DISTRIBUTED_ARGS
=
"--nproc_per_node
$WORLD_SIZE
\
--nnodes 1
\
--node_rank 0
\
--master_addr localhost
\
--master_port 6000"
CHECKPOINT_PATH
=
<Specify path
for
the language model>
INPUT_PATH
=
<Specific path
for
the input
test
dataset>
OUTPUT_PATH
=
<Speicifc path
for
the output>
PROMPT_PATH
=
<Specific path
for
the prompts>
python
-m
torch.distributed.launch
$DISTRIBUTED_ARGS
./tasks/main.py
\
--tensor-model-parallel-size
1
\
--pipeline-model-parallel-size
1
\
--num-layers
24
\
--hidden-size
1024
\
--num-attention-heads
16
\
--seq-length
2048
\
--max-position-embeddings
2048
\
--micro-batch-size
1
\
--vocab-file
/gpfs/fs1/projects/gpu_adlr/datasets/nlp/gpt2_indexed_dataset/bpe/gpt2-vocab.json
\
--merge-file
/gpfs/fs1/projects/gpu_adlr/datasets/nlp/gpt2_indexed_dataset/bpe/gpt2-merges.txt
\
--load
${
CHECKPOINT_PATH
}
\
--fp16
\
--DDP-impl
torch
\
--tokenizer-type
GPT2BPETokenizer
\
--out-seq-length
100
\
--sample-input-file
${
INPUT_PATH
}
\
--sample-output-file
${
OUTPUT_PATH
}
\
--prompt-file
${
PROMPT_PATH
}
\
--prompt-type
response
\
--num-prompt-examples
20
\
--task
knwl-dialo-prompt
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment