Commit 8126d777 authored by mshoeybi's avatar mshoeybi
Browse files

addressed Jareds comments

parent b10ca211
...@@ -38,7 +38,7 @@ def generate_and_post_process(model, ...@@ -38,7 +38,7 @@ def generate_and_post_process(model,
temperature=1.0, temperature=1.0,
add_BOS=False, add_BOS=False,
use_eod_token_for_early_termination=True): use_eod_token_for_early_termination=True):
"""Run inferecne and post-process outputs, i.e., detokenize, """Run inference and post-process outputs, i.e., detokenize,
move to cpu and convert to list.""" move to cpu and convert to list."""
# Main inference. # Main inference.
......
...@@ -61,7 +61,8 @@ class ForwardStep: ...@@ -61,7 +61,8 @@ class ForwardStep:
max_sequence_len) max_sequence_len)
# Pipelining arguments. # Pipelining arguments.
args = get_args() args = get_args()
self.pipeline_size_larger_than_one = args.pipeline_model_parallel_size self.pipeline_size_larger_than_one = (
args.pipeline_model_parallel_size > 1)
# Threshold of pipelining. # Threshold of pipelining.
self.pipelining_batch_x_seqlen = \ self.pipelining_batch_x_seqlen = \
args.inference_batch_times_seqlen_threshold args.inference_batch_times_seqlen_threshold
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment