diff --git a/main.py b/main.py
new file mode 100755
index 0000000..8ce0146
--- /dev/null
+++ b/main.py
@@ -0,0 +1,27 @@
+import webui as mrq
+
+if __name__ == "__main__":
+    mrq.args = mrq.setup_args()
+
+    if mrq.args.listen_path is not None and mrq.args.listen_path != "/":
+        import uvicorn
+        uvicorn.run("main:app", host=mrq.args.listen_host, port=mrq.args.listen_port if not None else 8000)
+    else:
+        mrq.webui = mrq.setup_gradio()
+        mrq.webui.launch(share=mrq.args.share, prevent_thread_lock=True, server_name=mrq.args.listen_host, server_port=mrq.args.listen_port)
+        mrq.tts = mrq.setup_tortoise()
+
+        mrq.webui.block_thread()
+elif __name__ == "main":
+    from fastapi import FastAPI
+    import gradio as gr
+
+    import sys
+    sys.argv = [sys.argv[0]]
+
+    app = FastAPI()
+    mrq.args = mrq.setup_args()
+    mrq.webui = mrq.setup_gradio()
+    app = gr.mount_gradio_app(app, mrq.webui, path=mrq.args.listen_path)
+
+    mrq.tts = mrq.setup_tortoise()
diff --git a/start.bat b/start.bat
index f0d9dfa..4cbd131 100755
--- a/start.bat
+++ b/start.bat
@@ -1,4 +1,4 @@
 call .\tortoise-venv\Scripts\activate.bat
-python app.py
+python main.py
 deactivate
 pause
\ No newline at end of file
diff --git a/start.sh b/start.sh
index a63f2c9..b67f402 100755
--- a/start.sh
+++ b/start.sh
@@ -1,3 +1,3 @@
 source ./tortoise-venv/bin/activate
-python ./app.py
+python3 ./main.py
 deactivate
diff --git a/tortoise_tts.ipynb b/tortoise_tts.ipynb
old mode 100644
new mode 100755
index b0230e3..47f4488
--- a/tortoise_tts.ipynb
+++ b/tortoise_tts.ipynb
@@ -1,185 +1 @@
-{
-  "nbformat": 4,
-  "nbformat_minor": 0,
-  "metadata": {
-    "colab": {
-      "name": "tortoise-tts.ipynb",
-      "provenance": [],
-      "collapsed_sections": []
-    },
-    "kernelspec": {
-      "name": "python3",
-      "display_name": "Python 3"
-    },
-    "language_info": {
-      "name": "python"
-    },
-    "accelerator": "GPU"
-  },
-  "cells": [
-    {
-      "cell_type": "markdown",
-      "source": [
-        "Welcome to Tortoise! 🐒🐒🐒🐒\n",
-        "\n",
-        "Before you begin, I **strongly** recommend you turn on a GPU runtime.\n",
-        "\n",
-        "There's a reason this is called \"Tortoise\" - this model takes up to a minute to perform inference for a single sentence on a GPU. Expect waits on the order of hours on a CPU."
-      ],
-      "metadata": {
-        "id": "_pIZ3ZXNp7cf"
-      }
-    },
-    {
-      "cell_type": "code",
-      "execution_count": null,
-      "metadata": {
-        "id": "JrK20I32grP6"
-      },
-      "outputs": [],
-      "source": [
-        "!git clone https://github.com/neonbjb/tortoise-tts.git\n",
-        "%cd tortoise-tts\n",
-        "!pip3 install -r requirements.txt\n",
-        "!python3 setup.py install"
-      ]
-    },
-    {
-      "cell_type": "code",
-      "source": [
-        "# Imports used through the rest of the notebook.\n",
-        "import torch\n",
-        "import torchaudio\n",
-        "import torch.nn as nn\n",
-        "import torch.nn.functional as F\n",
-        "\n",
-        "import IPython\n",
-        "\n",
-        "from tortoise.api import TextToSpeech\n",
-        "from tortoise.utils.audio import load_audio, load_voice, load_voices\n",
-        "\n",
-        "# This will download all the models used by Tortoise from the HF hub.\n",
-        "tts = TextToSpeech()"
-      ],
-      "metadata": {
-        "id": "Gen09NM4hONQ"
-      },
-      "execution_count": null,
-      "outputs": []
-    },
-    {
-      "cell_type": "code",
-      "source": [
-        "# This is the text that will be spoken.\n",
-        "text = \"Joining two modalities results in a surprising increase in generalization! What would happen if we combined them all?\"\n",
-        "\n",
-        "# Here's something for the poetically inclined.. (set text=)\n",
-        "\"\"\"\n",
-        "Then took the other, as just as fair,\n",
-        "And having perhaps the better claim,\n",
-        "Because it was grassy and wanted wear;\n",
-        "Though as for that the passing there\n",
-        "Had worn them really about the same,\"\"\"\n",
-        "\n",
-        "# Pick a \"preset mode\" to determine quality. Options: {\"ultra_fast\", \"fast\" (default), \"standard\", \"high_quality\"}. See docs in api.py\n",
-        "preset = \"fast\""
-      ],
-      "metadata": {
-        "id": "bt_aoxONjfL2"
-      },
-      "execution_count": null,
-      "outputs": []
-    },
-    {
-      "cell_type": "code",
-      "source": [
-        "# Tortoise will attempt to mimic voices you provide. It comes pre-packaged\n",
-        "# with some voices you might recognize.\n",
-        "\n",
-        "# Let's list all the voices available. These are just some random clips I've gathered\n",
-        "# from the internet as well as a few voices from the training dataset.\n",
-        "# Feel free to add your own clips to the voices/ folder.\n",
-        "%ls tortoise/voices\n",
-        "\n",
-        "IPython.display.Audio('tortoise/voices/tom/1.wav')"
-      ],
-      "metadata": {
-        "id": "SSleVnRAiEE2"
-      },
-      "execution_count": null,
-      "outputs": []
-    },
-    {
-      "cell_type": "code",
-      "source": [
-        "# Pick one of the voices from the output above\n",
-        "voice = 'tom'\n",
-        "\n",
-        "# Load it and send it through Tortoise.\n",
-        "voice_samples, conditioning_latents = load_voice(voice)\n",
-        "gen = tts.tts_with_preset(text, voice_samples=voice_samples, conditioning_latents=conditioning_latents, \n",
-        "                          preset=preset)\n",
-        "torchaudio.save('generated.wav', gen.squeeze(0).cpu(), 24000)\n",
-        "IPython.display.Audio('generated.wav')"
-      ],
-      "metadata": {
-        "id": "KEXOKjIvn6NW"
-      },
-      "execution_count": null,
-      "outputs": []
-    },
-    {
-      "cell_type": "code",
-      "source": [
-        "# Tortoise can also generate speech using a random voice. The voice changes each time you execute this!\n",
-        "# (Note: random voices can be prone to strange utterances)\n",
-        "gen = tts.tts_with_preset(text, voice_samples=None, conditioning_latents=None, preset=preset)\n",
-        "torchaudio.save('generated.wav', gen.squeeze(0).cpu(), 24000)\n",
-        "IPython.display.Audio('generated.wav')"
-      ],
-      "metadata": {
-        "id": "16Xs2SSC3BXa"
-      },
-      "execution_count": null,
-      "outputs": []
-    },
-    {
-      "cell_type": "code",
-      "source": [
-        "# You can also combine conditioning voices. Combining voices produces a new voice\n",
-        "# with traits from all the parents.\n",
-        "#\n",
-        "# Lets see what it would sound like if Picard and Kirk had a kid with a penchant for philosophy:\n",
-        "voice_samples, conditioning_latents = load_voices(['pat', 'william'])\n",
-        "\n",
-        "gen = tts.tts_with_preset(\"They used to say that if man was meant to fly, he’d have wings. But he did fly. He discovered he had to.\", \n",
-        "                          voice_samples=None, conditioning_latents=None, preset=preset)\n",
-        "torchaudio.save('captain_kirkard.wav', gen.squeeze(0).cpu(), 24000)\n",
-        "IPython.display.Audio('captain_kirkard.wav')"
-      ],
-      "metadata": {
-        "id": "fYTk8KUezUr5"
-      },
-      "execution_count": null,
-      "outputs": []
-    },
-    {
-      "cell_type": "code",
-      "source": [
-        "del tts  # Will break other cells, but necessary to conserve RAM if you want to run this cell.\n",
-        "\n",
-        "# Tortoise comes with some scripts that does a lot of the lifting for you. For example,\n",
-        "# read.py will read a text file for you.\n",
-        "!python3 tortoise/read.py --voice=train_atkins --textfile=tortoise/data/riding_hood.txt --preset=ultra_fast --output_path=.\n",
-        "\n",
-        "IPython.display.Audio('train_atkins/combined.wav')\n",
-        "# This will take awhile.."
-      ],
-      "metadata": {
-        "id": "t66yqWgu68KL"
-      },
-      "execution_count": null,
-      "outputs": []
-    }
-  ]
-}
\ No newline at end of file
+{"nbformat":4,"nbformat_minor":0,"metadata":{"colab":{"private_outputs":true,"provenance":[]},"kernelspec":{"name":"python3","display_name":"Python 3"},"language_info":{"name":"python"},"accelerator":"GPU","gpuClass":"standard"},"cells":[{"cell_type":"markdown","source":["## Initialization"],"metadata":{"id":"ni41hmE03DL6"}},{"cell_type":"code","execution_count":null,"metadata":{"id":"FtsMKKfH18iM"},"outputs":[],"source":["!git clone https://git.ecker.tech/mrq/tortoise-tts/\n","%cd tortoise-tts\n","!python -m pip install --upgrade pip\n","!pip install torch torchvision torchaudio --extra-index-url https://download.pytorch.org/whl/cu116\n","!python -m pip install -r ./requirements.txt\n","!pip install Pillow==9.0.0 # errors out only when importing\n","!python setup.py install"]},{"cell_type":"markdown","source":["## Running"],"metadata":{"id":"o1gkfw3B3JSk"}},{"cell_type":"code","source":["import webui as mrq\n","\n","mrq.args = mrq.setup_args()\n","mrq.webui = mrq.setup_gradio()\n","mrq.webui.launch(share=True, prevent_thread_lock=True)\n","mrq.tts = mrq.setup_tortoise()\n","mrq.webui.block_thread()"],"metadata":{"id":"c_EQZLTA19c7"},"execution_count":null,"outputs":[]}]}
\ No newline at end of file
diff --git a/app.py b/webui.py
similarity index 94%
rename from app.py
rename to webui.py
index 631333e..e868528 100755
--- a/app.py
+++ b/webui.py
@@ -20,6 +20,10 @@ from tortoise.api import TextToSpeech
 from tortoise.utils.audio import load_audio, load_voice, load_voices
 from tortoise.utils.text import split_and_recombine_text
 
+args = None
+webui = None
+tts = None
+
 def generate(text, delimiter, emotion, prompt, voice, mic_audio, seed, candidates, num_autoregressive_samples, diffusion_iterations, temperature, diffusion_sampler, breathing_room, cvvp_weight, experimentals, progress=gr.Progress(track_tqdm=True)):
     try:
         tts
@@ -424,7 +428,7 @@ def setup_args():
     args.listen_host = None
     args.listen_port = None
     args.listen_path = None
-    if args.listen is not None:
+    if args.listen:
         match = re.findall(r"^(?:(.+?):(\d+))?(\/.+?)?$", args.listen)[0]
 
         args.listen_host = match[0] if match[0] != "" else "127.0.0.1"
@@ -623,29 +627,4 @@ def setup_gradio():
 
     webui.queue(concurrency_count=args.concurrency_count)
 
-    return webui
-
-if __name__ == "__main__":
-    args = setup_args()
-
-    if args.listen_path is not None and args.listen_path != "/":
-        import uvicorn
-        uvicorn.run("app:app", host=args.listen_host, port=args.listen_port if not None else 8000)
-    else:
-        webui = setup_gradio()
-        webui.launch(share=args.share, prevent_thread_lock=True, server_name=args.listen_host, server_port=args.listen_port)
-        tts = setup_tortoise()
-
-        webui.block_thread()
-elif __name__ == "app":
-    import sys
-    from fastapi import FastAPI
-
-    sys.argv = [sys.argv[0]]
-
-    app = FastAPI()
-    args = setup_args()
-    webui = setup_gradio()
-    app = gr.mount_gradio_app(app, webui, path=args.listen_path)
-
-    tts = setup_tortoise()
+    return webui
\ No newline at end of file