Commit fc413707 authored by wangshankun's avatar wangshankun
Browse files

del script private info

parent 9e120289
...@@ -39,31 +39,18 @@ def main(): ...@@ -39,31 +39,18 @@ def main():
type=str, type=str,
required=True, required=True,
choices=[ choices=[
"wan2.1", "wan2.1",
"hunyuan", "hunyuan",
"wan2.1_distill", "wan2.1_distill",
"wan2.1_causvid", "wan2.1_causvid",
"wan2.1_skyreels_v2_df", "wan2.1_skyreels_v2_df",
"cogvideox", "cogvideox",
"wan2.1_audio", "wan2.1_audio",
"wan2.2_moe", "wan2.2_moe",
"wan2.2", "wan2.2",
"wan2.2_moe_audio", "wan2.2_moe_audio",
"wan2.2_audio", "wan2.2_audio",
"wan2.2_moe_distill", "wan2.2_moe_distill",
,
], ],
default="wan2.1", default="wan2.1",
) )
......
...@@ -343,7 +343,7 @@ class VideoGenerator: ...@@ -343,7 +343,7 @@ class VideoGenerator:
self.model.scheduler.reset() self.model.scheduler.reset()
inputs["previmg_encoder_output"] = self.prepare_prev_latents(prev_video, prev_frame_length) inputs["previmg_encoder_output"] = self.prepare_prev_latents(prev_video, prev_frame_length)
# Run inference loop # Run inference loop
if total_steps is None: if total_steps is None:
total_steps = self.model.scheduler.infer_steps total_steps = self.model.scheduler.infer_steps
...@@ -385,11 +385,6 @@ class WanAudioRunner(WanRunner): # type:ignore ...@@ -385,11 +385,6 @@ class WanAudioRunner(WanRunner): # type:ignore
self._video_generator = None self._video_generator = None
self._audio_preprocess = None self._audio_preprocess = None
if self.seq_p_group is None:
self.sp_size = 1
else:
self.sp_size = dist.get_world_size(self.seq_p_group)
def initialize(self): def initialize(self):
"""Initialize all models once for multiple runs""" """Initialize all models once for multiple runs"""
...@@ -613,7 +608,7 @@ class WanAudioRunner(WanRunner): # type:ignore ...@@ -613,7 +608,7 @@ class WanAudioRunner(WanRunner): # type:ignore
def load_transformer(self): def load_transformer(self):
"""Load transformer with LoRA support""" """Load transformer with LoRA support"""
base_model = WanAudioModel(self.config.model_path, self.config, self.init_device, self.seq_p_group) base_model = WanAudioModel(self.config.model_path, self.config, self.init_device)
if self.config.get("lora_configs") and self.config.lora_configs: if self.config.get("lora_configs") and self.config.lora_configs:
assert not self.config.get("dit_quantized", False) or self.config.mm_config.get("weight_auto_quant", False) assert not self.config.get("dit_quantized", False) or self.config.mm_config.get("weight_auto_quant", False)
lora_wrapper = WanLoraWrapper(base_model) lora_wrapper = WanLoraWrapper(base_model)
......
#!/bin/bash #!/bin/bash
# set path and first # set path and first
lightx2v_path="/home/wangshankun/code/LightX2V" lightx2v_path=
model_path="/data/nvme0/gushiqiao/models/Wan2.2-R2V812-Audio-5B" model_path=
export CUDA_VISIBLE_DEVICES=0,1,2,3 export CUDA_VISIBLE_DEVICES=0,1,2,3
......
#!/bin/bash #!/bin/bash
# set path and first # set path and first
lightx2v_path="/home/wangshankun/code/LightX2V" lightx2v_path=
model_path="/data/nvme0/models/Wan2.1-R2V721-Audio-14B-720P" model_path=
export CUDA_VISIBLE_DEVICES=0 export CUDA_VISIBLE_DEVICES=0
...@@ -11,7 +11,6 @@ source ${lightx2v_path}/scripts/base/base.sh ...@@ -11,7 +11,6 @@ source ${lightx2v_path}/scripts/base/base.sh
export TORCH_CUDA_ARCH_LIST="9.0" export TORCH_CUDA_ARCH_LIST="9.0"
export PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True export PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True
export ENABLE_GRAPH_MODE=false
export ENABLE_GRAPH_MODE=false export ENABLE_GRAPH_MODE=false
export SENSITIVE_LAYER_DTYPE=None export SENSITIVE_LAYER_DTYPE=None
......
...@@ -2,8 +2,8 @@ ...@@ -2,8 +2,8 @@
# set path and first # set path and first
lightx2v_path="/home/wangshankun/code/LightX2V" lightx2v_path=
model_path="/data/nvme0/models/Wan2.1-R2V721-Audio-14B-720P" model_path=
# set environment variables # set environment variables
source ${lightx2v_path}/scripts/base/base.sh source ${lightx2v_path}/scripts/base/base.sh
......
#!/bin/bash #!/bin/bash
# set path and first # set path and first
lightx2v_path="/home/wangshankun/code/LightX2V" lightx2v_path=
model_path="/data/nvme0/gushiqiao/models/official_models/wan2.2/Wan2.2-TI2V-5B" model_path=
export CUDA_VISIBLE_DEVICES=0 export CUDA_VISIBLE_DEVICES=0
# set environment variables # set environment variables
source ${lightx2v_path}/scripts/base/base.sh source ${lightx2v_path}/scripts/base/base.sh
export ENABLE_GRAPH_MODE=false
python -m lightx2v.infer \ python -m lightx2v.infer \
--model_cls wan2.2 \ --model_cls wan2.2 \
--task i2v \ --task i2v \
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment