Unverified Commit 916b3cdd authored by Jani Monoses's avatar Jani Monoses Committed by GitHub
Browse files

Allow passing dtype and max_new_tokens to HF reference script (#1903)

parent 838dcda1
...@@ -36,7 +36,7 @@ def normal_text(args): ...@@ -36,7 +36,7 @@ def normal_text(args):
t = get_tokenizer(args.model_path, trust_remote_code=True) t = get_tokenizer(args.model_path, trust_remote_code=True)
m = AutoModelForCausalLM.from_pretrained( m = AutoModelForCausalLM.from_pretrained(
args.model_path, args.model_path,
torch_dtype=torch.float16, torch_dtype=args.dtype,
low_cpu_mem_usage=True, low_cpu_mem_usage=True,
device_map="auto", device_map="auto",
trust_remote_code=True, trust_remote_code=True,
...@@ -47,7 +47,7 @@ def normal_text(args): ...@@ -47,7 +47,7 @@ def normal_text(args):
"The capital of the United Kindom is", "The capital of the United Kindom is",
"Today is a sunny day and I like", "Today is a sunny day and I like",
] ]
max_new_tokens = 16 max_new_tokens = args.max_new_tokens
torch.cuda.set_device(0) torch.cuda.set_device(0)
...@@ -104,6 +104,16 @@ if __name__ == "__main__": ...@@ -104,6 +104,16 @@ if __name__ == "__main__":
default="TinyLlama/TinyLlama-1.1B-Chat-v0.4", default="TinyLlama/TinyLlama-1.1B-Chat-v0.4",
# default="meta-llama/Llama-2-7b-chat-hf", # default="meta-llama/Llama-2-7b-chat-hf",
) )
parser.add_argument(
"--max-new-tokens",
type=int,
default=16)
parser.add_argument(
"--dtype",
type=str,
default="float16")
args = parser.parse_args() args = parser.parse_args()
normal_text(args) normal_text(args)
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment