warnings.warn(f'Input type into Linear4bit is torch.float16, but bnb_4bit_compute_type=torch.float32 (default). This will lead to slow inference or training speed.')
warnings.warn(f'Input type into Linear4bit is torch.float16, but bnb_4bit_compute_type=torch.float32 (default). This will lead to slow inference or training speed.')
warnings.filterwarnings('ignore',message='.*inference or training')
warnings.filterwarnings('ignore',message='.*inference or training')
def_update_buffers(self):
defstring_to_tensor(s):
"""stores string as ints for serialization. assumes codes fit int16"""