Commit 5a50b346 authored by Azure's avatar Azure
Browse files

fix hard coding caused by rope dim calculation, load from config now

parent 476b1d8d
...@@ -78,7 +78,7 @@ def local_chat(): ...@@ -78,7 +78,7 @@ def local_chat():
else: else:
content += line + "\n" content += line + "\n"
if content == "": if content == "":
if True: # config.prompt_file == None or config.prompt_file == "": if config.prompt_file == None or config.prompt_file == "":
content = "hi" content = "hi"
else: else:
content = open(config.prompt_file, "r").read() content = open(config.prompt_file, "r").read()
......
...@@ -19,7 +19,7 @@ from transformers.generation import GenerationMixin ...@@ -19,7 +19,7 @@ from transformers.generation import GenerationMixin
from transformers.modeling_attn_mask_utils import AttentionMaskConverter from transformers.modeling_attn_mask_utils import AttentionMaskConverter
# from transformers.modeling_flash_attention_utils import FlashAttentionKwargs # from transformers.modeling_flash_attention_utils import FlashAttentionKwargs
from transformers.modeling_outputs import BaseModelOutputWithPast, CausalLMOutputWithPast, SequenceClassifierOutputWithPast from transformers.modeling_outputs import BaseModelOutputWithPast, CausalLMOutputWithPast, SequenceClassifierOutputWithPast
from transformers.modeling_rope_utils import ROPE_INIT_FUNCTIONS from ktransformers.util.modeling_rope_utils import ROPE_INIT_FUNCTIONS
from transformers.modeling_utils import PreTrainedModel # ALL_ATTENTION_FUNCTIONS, PreTrainedModel from transformers.modeling_utils import PreTrainedModel # ALL_ATTENTION_FUNCTIONS, PreTrainedModel
# from transformers.processing_utils import Unpack # from transformers.processing_utils import Unpack
from transformers.utils import ( from transformers.utils import (
......
This diff is collapsed.
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment