From 1231490022a76dd0c60f6cd1c002075ed0e2bef7 Mon Sep 17 00:00:00 2001 From: Jintao Huang Date: Thu, 22 May 2025 00:10:15 +0800 Subject: [PATCH] merge release/1.26 --- docker/Dockerfile.ubuntu | 13 ++++++++++--- modelscope/version.py | 4 ++-- 2 files changed, 12 insertions(+), 5 deletions(-) diff --git a/docker/Dockerfile.ubuntu b/docker/Dockerfile.ubuntu index 3db3f857..b3a613fe 100644 --- a/docker/Dockerfile.ubuntu +++ b/docker/Dockerfile.ubuntu @@ -59,7 +59,7 @@ RUN echo $CUR_TIME RUN sh /tmp/install.sh {version_args} && \ curl -fsSL https://ollama.com/install.sh | sh && \ pip install --no-cache-dir -U funasr scikit-learn && \ - pip install --no-cache-dir -U qwen_vl_utils qwen_omni_utils pyav librosa timm transformers accelerate peft trl safetensors && \ + pip install --no-cache-dir -U qwen_vl_utils qwen_omni_utils librosa timm transformers accelerate peft trl safetensors && \ cd /tmp && GIT_LFS_SKIP_SMUDGE=1 git clone -b {modelscope_branch} --single-branch https://github.com/modelscope/modelscope.git && \ cd modelscope && pip install . -f https://modelscope.oss-cn-beijing.aliyuncs.com/releases/repo.html && \ cd / && rm -fr /tmp/modelscope && pip cache purge; \ @@ -68,7 +68,7 @@ RUN sh /tmp/install.sh {version_args} && \ pip install .[eval] && pip install evalscope -U --no-dependencies && pip install xtuner --no-dependencies && \ cd / && rm -fr /tmp/ms-swift && pip cache purge; \ pip install --no-cache-dir torch=={torch_version} torchvision=={torchvision_version} torchaudio=={torchaudio_version} {index_url} && \ - pip install --no-cache-dir transformers timm>=0.9.0 && pip cache purge; \ + pip install --no-cache-dir transformers diffusers timm>=0.9.0 && pip cache purge; \ pip install --no-cache-dir omegaconf==2.3.0 && pip cache purge; \ pip config set global.index-url https://mirrors.aliyun.com/pypi/simple && \ pip config set install.trusted-host mirrors.aliyun.com && \ @@ -76,7 +76,7 @@ RUN sh /tmp/install.sh {version_args} && \ RUN if [ "$INSTALL_MS_DEPS" = "True" ]; then \ - pip install --no-cache-dir huggingface-hub==0.25.* -U; \ + pip install --no-cache-dir huggingface-hub==0.25.* transformers==4.48.3 -U; \ fi; \ if [ "$INSTALL_MEGATRON_DEPS" = "True" ]; then \ pip install liger_kernel nvitop pre-commit "transformers==4.51.3" huggingface-hub -U && \ @@ -89,6 +89,13 @@ if [ "$INSTALL_MEGATRON_DEPS" = "True" ]; then \ pip install git+https://github.com/NVIDIA/Megatron-LM.git@core_r0.12.0; \ fi +# install nvm and set node version to 18 +ENV NVM_DIR=/root/.nvm +RUN curl -o- https://raw.githubusercontent.com/nvm-sh/nvm/v0.39.5/install.sh | bash && \ + . $NVM_DIR/nvm.sh && \ + nvm install 18 && \ + nvm use 18 + ENV SETUPTOOLS_USE_DISTUTILS=stdlib ENV VLLM_USE_MODELSCOPE=True ENV LMDEPLOY_USE_MODELSCOPE=True diff --git a/modelscope/version.py b/modelscope/version.py index 031a86b4..274d323d 100644 --- a/modelscope/version.py +++ b/modelscope/version.py @@ -1,5 +1,5 @@ # Make sure to modify __release_datetime__ to release time when making official release. -__version__ = '2.0.0' +__version__ = '1.26.0' # default release datetime for branches under active development is set # to be a time far-far-away-into-the-future -__release_datetime__ = '2099-09-06 00:00:00' +__release_datetime__ = '2025-05-08 12:00:00'