diff --git a/modelscope/models/cv/image_super_resolution_pasd/unet_2d_blocks.py b/modelscope/models/cv/image_super_resolution_pasd/unet_2d_blocks.py index c95639e6..a36316e4 100644 --- a/modelscope/models/cv/image_super_resolution_pasd/unet_2d_blocks.py +++ b/modelscope/models/cv/image_super_resolution_pasd/unet_2d_blocks.py @@ -8,10 +8,11 @@ import torch.nn.functional as F from diffusers.models.attention_processor import (Attention, AttnAddedKVProcessor, AttnAddedKVProcessor2_0) -from diffusers.models.dual_transformer_2d import DualTransformer2DModel from diffusers.models.resnet import (Downsample2D, FirDownsample2D, FirUpsample2D, KDownsample2D, KUpsample2D, ResnetBlock2D, Upsample2D) +from diffusers.models.transformers.dual_transformer_2d import \ + DualTransformer2DModel from torch import nn from .attention import AdaGroupNorm diff --git a/modelscope/models/cv/image_super_resolution_pasd_v2/unet_2d_blocks.py b/modelscope/models/cv/image_super_resolution_pasd_v2/unet_2d_blocks.py index 414eae89..8704cceb 100644 --- a/modelscope/models/cv/image_super_resolution_pasd_v2/unet_2d_blocks.py +++ b/modelscope/models/cv/image_super_resolution_pasd_v2/unet_2d_blocks.py @@ -20,12 +20,13 @@ from diffusers.models.activations import get_activation from diffusers.models.attention_processor import (Attention, AttnAddedKVProcessor, AttnAddedKVProcessor2_0) -from diffusers.models.dual_transformer_2d import DualTransformer2DModel from diffusers.models.normalization import AdaLayerNorm from diffusers.models.resnet import (Downsample2D, FirDownsample2D, FirUpsample2D, KDownsample2D, KUpsample2D, ResnetBlock2D, Upsample2D) -from diffusers.models.transformer_2d import Transformer2DModel +from diffusers.models.transformers.dual_transformer_2d import \ + DualTransformer2DModel +from diffusers.models.transformers.transformer_2d import Transformer2DModel from diffusers.utils import is_torch_version, logging from einops import rearrange from torch import nn diff --git a/modelscope/models/cv/image_view_transform/ldm/autoencoder.py b/modelscope/models/cv/image_view_transform/ldm/autoencoder.py index de702b35..5b0d16d2 100755 --- a/modelscope/models/cv/image_view_transform/ldm/autoencoder.py +++ b/modelscope/models/cv/image_view_transform/ldm/autoencoder.py @@ -3,7 +3,7 @@ from contextlib import contextmanager import pytorch_lightning as pl import torch import torch.nn.functional as F -from taming.modules.vqvae.quantize import VectorQuantizer2 as VectorQuantizer +from taming.modules.vqvae.quantize import VectorQuantizer from ..util import instantiate_from_config from .distributions import DiagonalGaussianDistribution diff --git a/modelscope/models/cv/image_view_transform/ldm/ddpm.py b/modelscope/models/cv/image_view_transform/ldm/ddpm.py index 4f57d456..cc665df4 100755 --- a/modelscope/models/cv/image_view_transform/ldm/ddpm.py +++ b/modelscope/models/cv/image_view_transform/ldm/ddpm.py @@ -16,7 +16,7 @@ import torch import torch.nn as nn from einops import rearrange, repeat from omegaconf import ListConfig -from pytorch_lightning.utilities.distributed import rank_zero_only +from pytorch_lightning.utilities.rank_zero import rank_zero_only from torch.optim.lr_scheduler import LambdaLR from torchvision.utils import make_grid from tqdm import tqdm