pipeline_output.py 1.38 KB
Newer Older
Aryan's avatar
Aryan committed
1
from dataclasses import dataclass
Aryan's avatar
Aryan committed
2
from typing import List, Union
Aryan's avatar
Aryan committed
3

Aryan's avatar
Aryan committed
4
5
import numpy as np
import PIL.Image
Aryan's avatar
Aryan committed
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
import torch

from diffusers.utils import BaseOutput


@dataclass
class HunyuanVideoPipelineOutput(BaseOutput):
    r"""
    Output class for HunyuanVideo pipelines.

    Args:
        frames (`torch.Tensor`, `np.ndarray`, or List[List[PIL.Image.Image]]):
            List of video outputs - It can be a nested list of length `batch_size,` with each sub-list containing
            denoised PIL image sequences of length `num_frames.` It can also be a NumPy array or Torch tensor of shape
            `(batch_size, num_frames, channels, height, width)`.
    """

    frames: torch.Tensor
Aryan's avatar
Aryan committed
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39


@dataclass
class HunyuanVideoFramepackPipelineOutput(BaseOutput):
    r"""
    Output class for HunyuanVideo pipelines.

    Args:
        frames (`torch.Tensor`, `np.ndarray`, or List[List[PIL.Image.Image]]):
            List of video outputs - It can be a nested list of length `batch_size,` with each sub-list containing
            denoised PIL image sequences of length `num_frames.` It can also be a NumPy array or Torch tensor of shape
            `(batch_size, num_frames, channels, height, width)`. Or, a list of torch tensors where each tensor
            corresponds to a latent that decodes to multiple frames.
    """

    frames: Union[torch.Tensor, np.ndarray, List[List[PIL.Image.Image]], List[torch.Tensor]]