From 6a852e8f0f90ac380e0b4086acd295c01b80a608 Mon Sep 17 00:00:00 2001 From: w-okada <48346627+w-okada@users.noreply.github.com> Date: Sun, 30 Jun 2024 16:12:37 +0900 Subject: [PATCH] =?UTF-8?q?Colab=20=E3=82=92=E4=BD=BF=E7=94=A8=E3=81=97?= =?UTF-8?q?=E3=81=A6=E4=BD=9C=E6=88=90=E3=81=95=E3=82=8C=E3=81=BE=E3=81=97?= =?UTF-8?q?=E3=81=9F?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- w_okada's_Voice_Changer_version_2_x.ipynb | 24 +++++++++++++++++++---- 1 file changed, 20 insertions(+), 4 deletions(-) diff --git a/w_okada's_Voice_Changer_version_2_x.ipynb b/w_okada's_Voice_Changer_version_2_x.ipynb index 572cf64a..b9f6abed 100644 --- a/w_okada's_Voice_Changer_version_2_x.ipynb +++ b/w_okada's_Voice_Changer_version_2_x.ipynb @@ -216,8 +216,7 @@ "colab": { "base_uri": "https://localhost:8080/", "height": 148 - }, - "cellView": "form" + } }, "outputs": [ { @@ -252,7 +251,8 @@ "# @title **[2]** Start server\n", "# @markdown This cell will start the server, the first time that you run it will download the models, so it can take a while (2~4 minutes)\n", "\n", - "#@markdown - Options:\n", + "\n", + "# @markdown ### Options:\n", "ClearConsole = True # @param {type:\"boolean\"}\n", "Play_Notification = True # @param {type:\"boolean\"}\n", "\n", @@ -312,13 +312,29 @@ " play_notification_sound('https://huggingface.co/wok000/voices/resolve/main/vcclient001_vctk229_gpt-sovits_vcclient-ready.wav')\n", "wait_for_server()\n" ] + }, + { + "cell_type": "code", + "source": [ + "# @title **[FAQ]**\n", + "\n", + "# @markdown #### Q. Delay is too large, why?\n", + "# @markdown #### A. In the Colab edition, the audio output buffer is set larger to account for network latency. Depending on the environment, reducing this buffer may reduce delay. However, in some environments, this may cause the sound to become choppy due to network latency. The following diagram shows how to change the buffer size.\n", + "# @markdown \n" + ], + "metadata": { + "cellView": "form", + "id": "1uN-kGd5-IYx" + }, + "execution_count": null, + "outputs": [] } ], "metadata": { "colab": { "provenance": [], "gpuType": "T4", - "authorship_tag": "ABX9TyO4xBCjHQ5g0B28Cfbnr1eo", + "authorship_tag": "ABX9TyPEPBkGLPAGJGofiLPcnQhG", "include_colab_link": true }, "kernelspec": {