import torch from transformers import AutoModelForCausalLM, AutoTokenizer from transformers.generation.utils import GenerationConfig from peft import PeftModel, PeftConfig model_path = "/home/wanglch/projects/DISC-FinLLM/FinLLM" model = AutoModelForCausalLM.from_pretrained( model_path, torch_dtype=torch.float16, device_map="auto", trust_remote_code=True ) model.generation_config = GenerationConfig.from_pretrained(model_path) tokenizer = AutoTokenizer.from_pretrained( model_path, use_fast=False, trust_remote_code=True, ) messages = [ {"role": "user", "content": "请解释一下什么是银行不良资产?"}, ] response = model.chat(tokenizer, messages) print(response)