From 72f0afe683cefba6af0d61239446229b72f71a76 Mon Sep 17 00:00:00 2001 From: Francis LaBounty <73464335+francislabountyjr@users.noreply.github.com> Date: Sat, 22 Apr 2023 16:43:14 -0600 Subject: [PATCH] QoL updates --- clone_voice.ipynb | 22 ++++++++++++++++++++-- generate.ipynb | 22 ++++++++++++++++++++-- 2 files changed, 40 insertions(+), 4 deletions(-) diff --git a/clone_voice.ipynb b/clone_voice.ipynb index a5e3391..52284ea 100644 --- a/clone_voice.ipynb +++ b/clone_voice.ipynb @@ -119,11 +119,30 @@ "outputs": [], "source": [ "from bark.api import generate_audio\n", - "from bark.generation import SAMPLE_RATE\n", + "from bark.generation import SAMPLE_RATE, preload_models, codec_decode, generate_coarse, generate_fine, generate_text_semantic\n", "text_prompt = \"Hello, my name is Suno. And, uh — and I like pizza. [laughs]\"\n", "voice_name = \"speaker_0\" # use your custom voice name here if you have one" ] }, + { + "cell_type": "code", + "execution_count": null, + "metadata": {}, + "outputs": [], + "source": [ + "# download and load all models\n", + "preload_models(\n", + " text_use_gpu=True,\n", + " text_use_small=False,\n", + " coarse_use_gpu=True,\n", + " coarse_use_small=False,\n", + " fine_use_gpu=True,\n", + " fine_use_small=False,\n", + " codec_use_gpu=True,\n", + " force_reload=False\n", + ")" + ] + }, { "cell_type": "code", "execution_count": null, @@ -141,7 +160,6 @@ "outputs": [], "source": [ "# generation with more control\n", - "from bark.generation import codec_decode, generate_coarse, generate_fine, generate_text_semantic\n", "x_semantic = generate_text_semantic(\n", " text_prompt,\n", " history_prompt=voice_name,\n", diff --git a/generate.ipynb b/generate.ipynb index 4f2bb89..f60e165 100644 --- a/generate.ipynb +++ b/generate.ipynb @@ -7,11 +7,30 @@ "outputs": [], "source": [ "from bark.api import generate_audio\n", - "from bark.generation import SAMPLE_RATE\n", + "from bark.generation import SAMPLE_RATE, preload_models, codec_decode, generate_coarse, generate_fine, generate_text_semantic\n", "text_prompt = \"Hello, my name is Suno. And, uh — and I like pizza. [laughs]\"\n", "voice_name = \"speaker_0\" # use your custom voice name here if you have one" ] }, + { + "cell_type": "code", + "execution_count": null, + "metadata": {}, + "outputs": [], + "source": [ + "# download and load all models\n", + "preload_models(\n", + " text_use_gpu=True,\n", + " text_use_small=False,\n", + " coarse_use_gpu=True,\n", + " coarse_use_small=False,\n", + " fine_use_gpu=True,\n", + " fine_use_small=False,\n", + " codec_use_gpu=True,\n", + " force_reload=False\n", + ")" + ] + }, { "cell_type": "code", "execution_count": null, @@ -29,7 +48,6 @@ "outputs": [], "source": [ "# generation with more control\n", - "from bark.generation import codec_decode, generate_coarse, generate_fine, generate_text_semantic\n", "x_semantic = generate_text_semantic(\n", " text_prompt,\n", " history_prompt=voice_name,\n",