Commit 8126d777 authored by mshoeybi's avatar mshoeybi
Browse files

addressed Jareds comments

parent b10ca211
......@@ -38,7 +38,7 @@ def generate_and_post_process(model,
temperature=1.0,
add_BOS=False,
use_eod_token_for_early_termination=True):
"""Run inferecne and post-process outputs, i.e., detokenize,
"""Run inference and post-process outputs, i.e., detokenize,
move to cpu and convert to list."""
# Main inference.
......
......@@ -61,7 +61,8 @@ class ForwardStep:
max_sequence_len)
# Pipelining arguments.
args = get_args()
self.pipeline_size_larger_than_one = args.pipeline_model_parallel_size
self.pipeline_size_larger_than_one = (
args.pipeline_model_parallel_size > 1)
# Threshold of pipelining.
self.pipelining_batch_x_seqlen = \
args.inference_batch_times_seqlen_threshold
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment