"_backup/profiling/tf_bench.py" did not exist on "0d6504434befdf609d34709891eecf85f27e0934"
Unverified Commit 8b6a4486 authored by giorgiopiatti-dfinity's avatar giorgiopiatti-dfinity Committed by GitHub
Browse files

fix missing revision arg when loading tokenizer (#2982)

parent a69cb5cf
...@@ -71,6 +71,7 @@ class DetokenizerManager: ...@@ -71,6 +71,7 @@ class DetokenizerManager:
server_args.tokenizer_path, server_args.tokenizer_path,
tokenizer_mode=server_args.tokenizer_mode, tokenizer_mode=server_args.tokenizer_mode,
trust_remote_code=server_args.trust_remote_code, trust_remote_code=server_args.trust_remote_code,
revision=server_args.revision,
) )
self.decode_status = LimitedCapacityDict() self.decode_status = LimitedCapacityDict()
......
...@@ -206,6 +206,7 @@ class Scheduler: ...@@ -206,6 +206,7 @@ class Scheduler:
server_args.tokenizer_path, server_args.tokenizer_path,
tokenizer_mode=server_args.tokenizer_mode, tokenizer_mode=server_args.tokenizer_mode,
trust_remote_code=server_args.trust_remote_code, trust_remote_code=server_args.trust_remote_code,
revision=server_args.revision,
) )
self.tokenizer = self.processor.tokenizer self.tokenizer = self.processor.tokenizer
else: else:
...@@ -213,6 +214,7 @@ class Scheduler: ...@@ -213,6 +214,7 @@ class Scheduler:
server_args.tokenizer_path, server_args.tokenizer_path,
tokenizer_mode=server_args.tokenizer_mode, tokenizer_mode=server_args.tokenizer_mode,
trust_remote_code=server_args.trust_remote_code, trust_remote_code=server_args.trust_remote_code,
revision=server_args.revision,
) )
# Check whether overlap can be enabled # Check whether overlap can be enabled
......
...@@ -158,6 +158,7 @@ class TokenizerManager: ...@@ -158,6 +158,7 @@ class TokenizerManager:
server_args.tokenizer_path, server_args.tokenizer_path,
tokenizer_mode=server_args.tokenizer_mode, tokenizer_mode=server_args.tokenizer_mode,
trust_remote_code=server_args.trust_remote_code, trust_remote_code=server_args.trust_remote_code,
revision=server_args.revision,
) )
self.tokenizer = self.processor.tokenizer self.tokenizer = self.processor.tokenizer
os.environ["TOKENIZERS_PARALLELISM"] = "false" os.environ["TOKENIZERS_PARALLELISM"] = "false"
...@@ -171,6 +172,7 @@ class TokenizerManager: ...@@ -171,6 +172,7 @@ class TokenizerManager:
server_args.tokenizer_path, server_args.tokenizer_path,
tokenizer_mode=server_args.tokenizer_mode, tokenizer_mode=server_args.tokenizer_mode,
trust_remote_code=server_args.trust_remote_code, trust_remote_code=server_args.trust_remote_code,
revision=server_args.revision,
) )
# Store states # Store states
......
...@@ -83,6 +83,7 @@ class TpModelWorker: ...@@ -83,6 +83,7 @@ class TpModelWorker:
server_args.tokenizer_path, server_args.tokenizer_path,
tokenizer_mode=server_args.tokenizer_mode, tokenizer_mode=server_args.tokenizer_mode,
trust_remote_code=server_args.trust_remote_code, trust_remote_code=server_args.trust_remote_code,
revision=server_args.revision,
) )
self.tokenizer = self.processor.tokenizer self.tokenizer = self.processor.tokenizer
else: else:
...@@ -90,6 +91,7 @@ class TpModelWorker: ...@@ -90,6 +91,7 @@ class TpModelWorker:
server_args.tokenizer_path, server_args.tokenizer_path,
tokenizer_mode=server_args.tokenizer_mode, tokenizer_mode=server_args.tokenizer_mode,
trust_remote_code=server_args.trust_remote_code, trust_remote_code=server_args.trust_remote_code,
revision=server_args.revision,
) )
self.device = self.model_runner.device self.device = self.model_runner.device
......
...@@ -1027,6 +1027,7 @@ class Runtime: ...@@ -1027,6 +1027,7 @@ class Runtime:
self.server_args.tokenizer_path, self.server_args.tokenizer_path,
tokenizer_mode=self.server_args.tokenizer_mode, tokenizer_mode=self.server_args.tokenizer_mode,
trust_remote_code=self.server_args.trust_remote_code, trust_remote_code=self.server_args.trust_remote_code,
revision=self.server_args.revision,
) )
async def async_generate( async def async_generate(
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment