"vscode:/vscode.git/clone" did not exist on "c48c1d7c4672a2483ddd938825bd5e5bac5dc6f1"
Commit 0371f45c authored by bzantium's avatar bzantium
Browse files

fix for merge from master

parent baa8b0d3
...@@ -52,28 +52,6 @@ class HFLM(BaseLM): ...@@ -52,28 +52,6 @@ class HFLM(BaseLM):
self.tokenizer = transformers.AutoTokenizer.from_pretrained( self.tokenizer = transformers.AutoTokenizer.from_pretrained(
pretrained if tokenizer is None else tokenizer, pretrained if tokenizer is None else tokenizer,
<<<<<<< HEAD
revision=revision + ("/" + subfolder if subfolder is not None else ""))
# assert isinstance(self.tokenizer, (
# transformers.GPT2Tokenizer, transformers.GPT2TokenizerFast,
# transformers.T5Tokenizer, transformers.T5TokenizerFast,
# )), "this tokenizer has not been checked for compatibility yet!"
self.vocab_size = self.tokenizer.vocab_size
# if isinstance(self.tokenizer, (transformers.GPT2Tokenizer, transformers.GPT2TokenizerFast)):
# assert self.tokenizer.encode('hello\n\nhello') == [31373, 198, 198, 31373], \
# self.tokenizer.encode('hello\n\nhello')
# multithreading and batching
self.batch_size_per_gpu = batch_size # todo: adaptive batch size
# TODO: fix multi-gpu
# gpus = torch.cuda.device_count()
# if gpus > 1:
# self.gpt2 = nn.DataParallel(self.gpt2)
=======
revision=revision, revision=revision,
trust_remote_code=trust_remote_code, trust_remote_code=trust_remote_code,
) )
...@@ -95,7 +73,6 @@ class HFLM(BaseLM): ...@@ -95,7 +73,6 @@ class HFLM(BaseLM):
self.batch_size_per_gpu = batch_size self.batch_size_per_gpu = batch_size
else: else:
self.batch_size_per_gpu = int(batch_size) self.batch_size_per_gpu = int(batch_size)
>>>>>>> d145167959c2b1826d900524912cb99c44d5fb30
@property @property
def eot_token_id(self): def eot_token_id(self):
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment