diff --git a/clone_voice.ipynb b/clone_voice.ipynb index a5e3391..52284ea 100644 --- a/clone_voice.ipynb +++ b/clone_voice.ipynb @@ -119,11 +119,30 @@ "outputs": [], "source": [ "from bark.api import generate_audio\n", - "from bark.generation import SAMPLE_RATE\n", + "from bark.generation import SAMPLE_RATE, preload_models, codec_decode, generate_coarse, generate_fine, generate_text_semantic\n", "text_prompt = \"Hello, my name is Suno. And, uh — and I like pizza. [laughs]\"\n", "voice_name = \"speaker_0\" # use your custom voice name here if you have one" ] }, + { + "cell_type": "code", + "execution_count": null, + "metadata": {}, + "outputs": [], + "source": [ + "# download and load all models\n", + "preload_models(\n", + " text_use_gpu=True,\n", + " text_use_small=False,\n", + " coarse_use_gpu=True,\n", + " coarse_use_small=False,\n", + " fine_use_gpu=True,\n", + " fine_use_small=False,\n", + " codec_use_gpu=True,\n", + " force_reload=False\n", + ")" + ] + }, { "cell_type": "code", "execution_count": null, @@ -141,7 +160,6 @@ "outputs": [], "source": [ "# generation with more control\n", - "from bark.generation import codec_decode, generate_coarse, generate_fine, generate_text_semantic\n", "x_semantic = generate_text_semantic(\n", " text_prompt,\n", " history_prompt=voice_name,\n", diff --git a/generate.ipynb b/generate.ipynb index 4f2bb89..f60e165 100644 --- a/generate.ipynb +++ b/generate.ipynb @@ -7,11 +7,30 @@ "outputs": [], "source": [ "from bark.api import generate_audio\n", - "from bark.generation import SAMPLE_RATE\n", + "from bark.generation import SAMPLE_RATE, preload_models, codec_decode, generate_coarse, generate_fine, generate_text_semantic\n", "text_prompt = \"Hello, my name is Suno. And, uh — and I like pizza. [laughs]\"\n", "voice_name = \"speaker_0\" # use your custom voice name here if you have one" ] }, + { + "cell_type": "code", + "execution_count": null, + "metadata": {}, + "outputs": [], + "source": [ + "# download and load all models\n", + "preload_models(\n", + " text_use_gpu=True,\n", + " text_use_small=False,\n", + " coarse_use_gpu=True,\n", + " coarse_use_small=False,\n", + " fine_use_gpu=True,\n", + " fine_use_small=False,\n", + " codec_use_gpu=True,\n", + " force_reload=False\n", + ")" + ] + }, { "cell_type": "code", "execution_count": null, @@ -29,7 +48,6 @@ "outputs": [], "source": [ "# generation with more control\n", - "from bark.generation import codec_decode, generate_coarse, generate_fine, generate_text_semantic\n", "x_semantic = generate_text_semantic(\n", " text_prompt,\n", " history_prompt=voice_name,\n",