mirror of
https://github.com/modelscope/modelscope.git
synced 2025-12-16 16:27:45 +01:00
revert cuda12.8
This commit is contained in:
@@ -348,9 +348,9 @@ class SwiftImageBuilder(LLMImageBuilder):
|
|||||||
|
|
||||||
def init_args(self, args) -> Any:
|
def init_args(self, args) -> Any:
|
||||||
if not args.base_image:
|
if not args.base_image:
|
||||||
args.base_image = 'nvidia/cuda:12.9.1-cudnn-devel-ubuntu22.04'
|
args.base_image = 'nvidia/cuda:12.8.1-cudnn-devel-ubuntu22.04'
|
||||||
if not args.cuda_version:
|
if not args.cuda_version:
|
||||||
args.cuda_version = '12.9.1'
|
args.cuda_version = '12.8.1'
|
||||||
if not args.torch_version:
|
if not args.torch_version:
|
||||||
args.torch_version = '2.8.0'
|
args.torch_version = '2.8.0'
|
||||||
args.torchaudio_version = '2.8.0'
|
args.torchaudio_version = '2.8.0'
|
||||||
@@ -360,7 +360,7 @@ class SwiftImageBuilder(LLMImageBuilder):
|
|||||||
if not args.lmdeploy_version:
|
if not args.lmdeploy_version:
|
||||||
args.lmdeploy_version = '0.10.2'
|
args.lmdeploy_version = '0.10.2'
|
||||||
if not args.flashattn_version:
|
if not args.flashattn_version:
|
||||||
args.flashattn_version = '2.8.1'
|
args.flashattn_version = '2.7.4.post1'
|
||||||
return super().init_args(args)
|
return super().init_args(args)
|
||||||
|
|
||||||
def generate_dockerfile(self) -> str:
|
def generate_dockerfile(self) -> str:
|
||||||
|
|||||||
@@ -10,18 +10,18 @@ flashattn_version=${7:-2.7.1.post4}
|
|||||||
|
|
||||||
pip uninstall -y torch torchvision torchaudio
|
pip uninstall -y torch torchvision torchaudio
|
||||||
|
|
||||||
pip install --no-cache-dir torch==$torch_version torchvision==$torchvision_version torchaudio==$torchaudio_version --index-url https://download.pytorch.org/whl/cu129
|
pip install --no-cache-dir torch==$torch_version torchvision==$torchvision_version torchaudio==$torchaudio_version
|
||||||
|
|
||||||
pip install --no-cache-dir -U autoawq lmdeploy==$lmdeploy_version
|
pip install --no-cache-dir -U autoawq lmdeploy==$lmdeploy_version
|
||||||
|
|
||||||
pip install --no-cache-dir torch==$torch_version torchvision==$torchvision_version torchaudio==$torchaudio_version --index-url https://download.pytorch.org/whl/cu129
|
pip install --no-cache-dir torch==$torch_version torchvision==$torchvision_version torchaudio==$torchaudio_version
|
||||||
|
|
||||||
pip install --no-cache-dir tiktoken transformers_stream_generator bitsandbytes deepspeed torchmetrics decord optimum openai-whisper
|
pip install --no-cache-dir tiktoken transformers_stream_generator bitsandbytes deepspeed torchmetrics decord optimum openai-whisper
|
||||||
|
|
||||||
# pip install https://github.com/Dao-AILab/flash-attention/releases/download/v2.6.3/flash_attn-2.6.3+cu123torch2.4cxx11abiTRUE-cp310-cp310-linux_x86_64.whl
|
# pip install https://github.com/Dao-AILab/flash-attention/releases/download/v2.6.3/flash_attn-2.6.3+cu123torch2.4cxx11abiTRUE-cp310-cp310-linux_x86_64.whl
|
||||||
# find on: https://github.com/Dao-AILab/flash-attention/releases
|
# find on: https://github.com/Dao-AILab/flash-attention/releases
|
||||||
# cd /tmp && git clone https://github.com/Dao-AILab/flash-attention.git && cd flash-attention && python setup.py install && cd / && rm -fr /tmp/flash-attention && pip cache purge;
|
# cd /tmp && git clone https://github.com/Dao-AILab/flash-attention.git && cd flash-attention && python setup.py install && cd / && rm -fr /tmp/flash-attention && pip cache purge;
|
||||||
MAX_JOBS=4 pip install --no-cache-dir flash_attn==$flashattn_version --no-build-isolation
|
MAX_JOBS=16 pip install --no-cache-dir flash_attn==$flashattn_version --no-build-isolation
|
||||||
|
|
||||||
pip install --no-cache-dir triton auto-gptq==$autogptq_version -U && pip cache purge
|
pip install --no-cache-dir triton auto-gptq==$autogptq_version -U && pip cache purge
|
||||||
|
|
||||||
|
|||||||
Reference in New Issue
Block a user