Commit 7e1d5e53 authored by zhuwenwen's avatar zhuwenwen
Browse files

merge v0.3.1

parents e3378b20 5f08050d
This diff is collapsed.
This diff is collapsed.
...@@ -111,13 +111,13 @@ class LLM: ...@@ -111,13 +111,13 @@ class LLM:
def get_tokenizer( def get_tokenizer(
self) -> Union[PreTrainedTokenizer, PreTrainedTokenizerFast]: self) -> Union[PreTrainedTokenizer, PreTrainedTokenizerFast]:
return self.llm_engine.tokenizer return self.llm_engine.tokenizer.tokenizer
def set_tokenizer( def set_tokenizer(
self, self,
tokenizer: Union[PreTrainedTokenizer, PreTrainedTokenizerFast], tokenizer: Union[PreTrainedTokenizer, PreTrainedTokenizerFast],
) -> None: ) -> None:
self.llm_engine.tokenizer = tokenizer self.llm_engine.tokenizer.tokenizer = tokenizer
def generate( def generate(
self, self,
......
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
...@@ -89,9 +89,7 @@ class ScaledActivation(nn.Module): ...@@ -89,9 +89,7 @@ class ScaledActivation(nn.Module):
if params_dtype is None: if params_dtype is None:
params_dtype = torch.get_default_dtype() params_dtype = torch.get_default_dtype()
self.scales = nn.Parameter( self.scales = nn.Parameter(
torch.empty(intermediate_size_per_partition, torch.empty(intermediate_size_per_partition, dtype=params_dtype))
dtype=params_dtype,
device="cuda"))
set_weight_attrs(self.scales, {"weight_loader": self.weight_loader}) set_weight_attrs(self.scales, {"weight_loader": self.weight_loader})
def forward(self, x: torch.Tensor) -> torch.Tensor: def forward(self, x: torch.Tensor) -> torch.Tensor:
......
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
This diff is collapsed.
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment