From bb6cdc8e72a3948b685be845a0b0ea9640492e32 Mon Sep 17 00:00:00 2001 From: w-okada <48346627+w-okada@users.noreply.github.com> Date: Wed, 7 Aug 2024 02:18:44 +0900 Subject: [PATCH] =?UTF-8?q?Colab=20=E3=82=92=E4=BD=BF=E7=94=A8=E3=81=97?= =?UTF-8?q?=E3=81=A6=E4=BD=9C=E6=88=90=E3=81=95=E3=82=8C=E3=81=BE=E3=81=97?= =?UTF-8?q?=E3=81=9F?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- w_okada's_Voice_Changer_version_2_x.ipynb | 74 +++-------------------- 1 file changed, 10 insertions(+), 64 deletions(-) diff --git a/w_okada's_Voice_Changer_version_2_x.ipynb b/w_okada's_Voice_Changer_version_2_x.ipynb index 98db70cc..e6bfe8d3 100644 --- a/w_okada's_Voice_Changer_version_2_x.ipynb +++ b/w_okada's_Voice_Changer_version_2_x.ipynb @@ -52,33 +52,12 @@ }, { "cell_type": "code", - "execution_count": 1, + "execution_count": null, "metadata": { "id": "W2GYWTHWmRIY", - "outputId": "88177d0e-d7a2-4a0b-fb09-578a5c27c430", - "colab": { - "base_uri": "https://localhost:8080/" - } + "cellView": "form" }, - "outputs": [ - { - "output_type": "stream", - "name": "stdout", - "text": [ - "GPU is available\n", - "GPU Name: Tesla T4\n", - "Welcome to ColabMod\n", - "Downloading the latest vcclient... \n", - " % Total % Received % Xferd Average Speed Time Time Time Current\n", - " Dload Upload Total Spent Left Speed\n", - "100 1149 100 1149 0 0 3004 0 --:--:-- --:--:-- --:--:-- 3007\n", - "100 2967M 100 2967M 0 0 154M 0 0:00:19 0:00:19 --:--:-- 219M\n", - "Download is done.\n", - "/content\n", - "Installing modules... Install is done.\n" - ] - } - ], + "outputs": [], "source": [ "#=================Updated=================\n", "# @title **[1]** Clone repository and install dependencies\n", @@ -160,8 +139,8 @@ " latest_hash_path=f'{work_dir}/latest_hash.txt'\n", "\n", " if mode == \"elf\":\n", - " # !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_latest_for_colab -o {work_dir}/vcclient_latest_for_colab\n", - " !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_2.0.44-alpha_for_colab -o {work_dir}/vcclient_latest_for_colab\n", + " !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_latest_for_colab -o {work_dir}/vcclient_latest_for_colab\n", + " # !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_2.0.44-alpha_for_colab -o {work_dir}/vcclient_latest_for_colab\n", "\n", " elif mode == \"zip\":\n", " !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_latest_for_colab.zip -o {work_dir}/vcclient_latest_for_colab.zip\n", @@ -178,8 +157,8 @@ " latest_version_hash = open('latest_hash.txt').read().strip()\n", "\n", " if mode == \"elf\":\n", - " # !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_latest_for_colab -o {work_dir}/vcclient_latest_for_colab\n", - " !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_2.0.44-alpha_for_colab -o {work_dir}/vcclient_latest_for_colab\n", + " !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_latest_for_colab -o {work_dir}/vcclient_latest_for_colab\n", + " # !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_2.0.44-alpha_for_colab -o {work_dir}/vcclient_latest_for_colab\n", " elif mode == \"zip\":\n", " !curl -L https://huggingface.co/wok000/vcclient000_colab/resolve/main/vcclient_latest_for_colab.zip -o {work_dir}/vcclient_latest_for_colab.zip\n", "\n", @@ -207,45 +186,12 @@ }, { "cell_type": "code", - "execution_count": 2, + "execution_count": null, "metadata": { "id": "s7mYqKtW6VOI", - "outputId": "3c88d3c3-5888-479f-9256-c6719d15b1f2", - "colab": { - "base_uri": "https://localhost:8080/", - "height": 148 - }, "cellView": "form" }, - "outputs": [ - { - "output_type": "stream", - "name": "stdout", - "text": [ - "--------- SERVER READY! ---------\n", - "Your server is available. elapsed: 98sec\n", - "https://wdc8fs91tuo-496ff2e9c6d22116-8003-colab.googleusercontent.com/\n", - "---------------------------------\n" - ] - }, - { - "output_type": "display_data", - "data": { - "text/plain": [ - "" - ], - "text/html": [ - "\n", - " \n", - " " - ] - }, - "metadata": {} - } - ], + "outputs": [], "source": [ "# @title **[2]** Start server\n", "# @markdown This cell will start the server, the first time that you run it will download the models, so it can take a while (2~4 minutes)\n", @@ -333,7 +279,7 @@ "colab": { "provenance": [], "gpuType": "T4", - "authorship_tag": "ABX9TyPL8HlIEcd7fZuat/q9Ir1k", + "authorship_tag": "ABX9TyPpbFBssircyl/qT8uDX2zy", "include_colab_link": true }, "kernelspec": {