diff --git a/modelscope/models/cv/image_super_resolution_pasd/unet_2d_blocks.py b/modelscope/models/cv/image_super_resolution_pasd/unet_2d_blocks.py index 67fec789..a36316e4 100644 --- a/modelscope/models/cv/image_super_resolution_pasd/unet_2d_blocks.py +++ b/modelscope/models/cv/image_super_resolution_pasd/unet_2d_blocks.py @@ -8,10 +8,11 @@ import torch.nn.functional as F from diffusers.models.attention_processor import (Attention, AttnAddedKVProcessor, AttnAddedKVProcessor2_0) -from diffusers.models.transformers.dual_transformer_2d import DualTransformer2DModel from diffusers.models.resnet import (Downsample2D, FirDownsample2D, FirUpsample2D, KDownsample2D, KUpsample2D, ResnetBlock2D, Upsample2D) +from diffusers.models.transformers.dual_transformer_2d import \ + DualTransformer2DModel from torch import nn from .attention import AdaGroupNorm diff --git a/modelscope/models/cv/image_super_resolution_pasd_v2/unet_2d_blocks.py b/modelscope/models/cv/image_super_resolution_pasd_v2/unet_2d_blocks.py index b76467b2..8704cceb 100644 --- a/modelscope/models/cv/image_super_resolution_pasd_v2/unet_2d_blocks.py +++ b/modelscope/models/cv/image_super_resolution_pasd_v2/unet_2d_blocks.py @@ -20,11 +20,12 @@ from diffusers.models.activations import get_activation from diffusers.models.attention_processor import (Attention, AttnAddedKVProcessor, AttnAddedKVProcessor2_0) -from diffusers.models.transformers.dual_transformer_2d import DualTransformer2DModel from diffusers.models.normalization import AdaLayerNorm from diffusers.models.resnet import (Downsample2D, FirDownsample2D, FirUpsample2D, KDownsample2D, KUpsample2D, ResnetBlock2D, Upsample2D) +from diffusers.models.transformers.dual_transformer_2d import \ + DualTransformer2DModel from diffusers.models.transformers.transformer_2d import Transformer2DModel from diffusers.utils import is_torch_version, logging from einops import rearrange