mirror of
https://github.com/modelscope/modelscope.git
synced 2026-02-24 20:19:51 +01:00
add printout prompt
This commit is contained in:
@@ -145,6 +145,8 @@ class LlamafileCMD(CLICommand):
|
||||
# if torch is not available, we will just assume gpu cannot be used
|
||||
pass
|
||||
if has_gpu:
|
||||
print(
|
||||
'GPU detected, launching model with llamafile GPU option >>>')
|
||||
execute_cmd = f'{execute_cmd} -ngl 999'
|
||||
os.system(execute_cmd)
|
||||
|
||||
|
||||
Reference in New Issue
Block a user