Commit 956b561b authored by Hongxin Liu's avatar Hongxin Liu Committed by ver217
Browse files

[moe] fix mixtral forward default value (#5329)

parent b60be18d
...@@ -437,7 +437,7 @@ class MixtralPipelineForwards: ...@@ -437,7 +437,7 @@ class MixtralPipelineForwards:
use_cache: Optional[bool] = None, use_cache: Optional[bool] = None,
output_attentions: Optional[bool] = None, output_attentions: Optional[bool] = None,
output_hidden_states: Optional[bool] = None, output_hidden_states: Optional[bool] = None,
output_router_logits: Optional[bool] = True, output_router_logits: Optional[bool] = None,
return_dict: Optional[bool] = None, return_dict: Optional[bool] = None,
stage_manager: Optional[PipelineStageManager] = None, stage_manager: Optional[PipelineStageManager] = None,
hidden_states: Optional[torch.FloatTensor] = None, hidden_states: Optional[torch.FloatTensor] = None,
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment