Skip to content

Commit e439012

Browse files
committed
Fix ruff lint and format issues
1 parent bd45ef6 commit e439012

4 files changed

Lines changed: 26 additions & 15 deletions

File tree

src/diffusers/modular_pipelines/hunyuan_video1_5/before_denoise.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -245,9 +245,11 @@ def description(self) -> str:
245245

246246
@property
247247
def expected_components(self) -> list[ComponentSpec]:
248+
from transformers import SiglipImageProcessor, SiglipVisionModel
249+
248250
from ...models import AutoencoderKLHunyuanVideo15
249251
from ...pipelines.hunyuan_video1_5.image_processor import HunyuanVideo15ImageProcessor
250-
from transformers import SiglipVisionModel, SiglipImageProcessor
252+
251253
return [
252254
ComponentSpec("vae", AutoencoderKLHunyuanVideo15),
253255
ComponentSpec(

src/diffusers/modular_pipelines/hunyuan_video1_5/decoders.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -75,7 +75,9 @@ def __call__(self, components, state: PipelineState) -> PipelineState:
7575
else:
7676
latents = block_state.latents.to(components.vae.dtype) / components.vae.config.scaling_factor
7777
video = components.vae.decode(latents, return_dict=False)[0]
78-
block_state.videos = components.video_processor.postprocess_video(video, output_type=block_state.output_type)
78+
block_state.videos = components.video_processor.postprocess_video(
79+
video, output_type=block_state.output_type
80+
)
7981

8082
self.set_block_state(state, block_state)
8183
return components, state

src/diffusers/modular_pipelines/hunyuan_video1_5/denoise.py

Lines changed: 8 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,6 @@
1212
# See the License for the specific language governing permissions and
1313
# limitations under the License.
1414

15-
from typing import Any
1615

1716
import torch
1817

@@ -115,7 +114,9 @@ def __call__(
115114

116115
# Step 1: Collect model inputs
117116
guider_inputs = {
118-
input_name: tuple(getattr(block_state, v) for v in value) if isinstance(value, tuple) else getattr(block_state, value)
117+
input_name: tuple(getattr(block_state, v) for v in value)
118+
if isinstance(value, tuple)
119+
else getattr(block_state, value)
119120
for input_name, value in self._guider_input_fields.items()
120121
}
121122

@@ -129,9 +130,7 @@ def __call__(
129130
for guider_state_batch in guider_state:
130131
components.guider.prepare_models(components.transformer)
131132

132-
cond_kwargs = {
133-
input_name: getattr(guider_state_batch, input_name) for input_name in guider_inputs.keys()
134-
}
133+
cond_kwargs = {input_name: getattr(guider_state_batch, input_name) for input_name in guider_inputs.keys()}
135134

136135
context_name = getattr(guider_state_batch, components.guider._identifier_key)
137136
with components.transformer.cache_context(context_name):
@@ -306,7 +305,9 @@ def __call__(
306305
timestep_r = None
307306

308307
guider_inputs = {
309-
input_name: tuple(getattr(block_state, v) for v in value) if isinstance(value, tuple) else getattr(block_state, value)
308+
input_name: tuple(getattr(block_state, v) for v in value)
309+
if isinstance(value, tuple)
310+
else getattr(block_state, value)
310311
for input_name, value in self._guider_input_fields.items()
311312
}
312313

@@ -316,9 +317,7 @@ def __call__(
316317
for guider_state_batch in guider_state:
317318
components.guider.prepare_models(components.transformer)
318319

319-
cond_kwargs = {
320-
input_name: getattr(guider_state_batch, input_name) for input_name in guider_inputs.keys()
321-
}
320+
cond_kwargs = {input_name: getattr(guider_state_batch, input_name) for input_name in guider_inputs.keys()}
322321

323322
context_name = getattr(guider_state_batch, components.guider._identifier_key)
324323
with components.transformer.cache_context(context_name):

src/diffusers/modular_pipelines/hunyuan_video1_5/encoders.py

Lines changed: 12 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -226,12 +226,20 @@ def encode_prompt(
226226
)
227227

228228
_, seq_len, _ = prompt_embeds.shape
229-
prompt_embeds = prompt_embeds.repeat(1, num_videos_per_prompt, 1).view(batch_size * num_videos_per_prompt, seq_len, -1)
230-
prompt_embeds_mask = prompt_embeds_mask.repeat(1, num_videos_per_prompt, 1).view(batch_size * num_videos_per_prompt, seq_len)
229+
prompt_embeds = prompt_embeds.repeat(1, num_videos_per_prompt, 1).view(
230+
batch_size * num_videos_per_prompt, seq_len, -1
231+
)
232+
prompt_embeds_mask = prompt_embeds_mask.repeat(1, num_videos_per_prompt, 1).view(
233+
batch_size * num_videos_per_prompt, seq_len
234+
)
231235

232236
_, seq_len_2, _ = prompt_embeds_2.shape
233-
prompt_embeds_2 = prompt_embeds_2.repeat(1, num_videos_per_prompt, 1).view(batch_size * num_videos_per_prompt, seq_len_2, -1)
234-
prompt_embeds_mask_2 = prompt_embeds_mask_2.repeat(1, num_videos_per_prompt, 1).view(batch_size * num_videos_per_prompt, seq_len_2)
237+
prompt_embeds_2 = prompt_embeds_2.repeat(1, num_videos_per_prompt, 1).view(
238+
batch_size * num_videos_per_prompt, seq_len_2, -1
239+
)
240+
prompt_embeds_mask_2 = prompt_embeds_mask_2.repeat(1, num_videos_per_prompt, 1).view(
241+
batch_size * num_videos_per_prompt, seq_len_2
242+
)
235243

236244
prompt_embeds = prompt_embeds.to(dtype=dtype, device=device)
237245
prompt_embeds_mask = prompt_embeds_mask.to(dtype=dtype, device=device)

0 commit comments

Comments
 (0)