2023-04-18 07:23:09 +02:00
import argparse
import glob
import os
import shutil
import site
import subprocess
import sys
script_dir = os . getcwd ( )
2023-05-31 19:41:03 +02:00
conda_env_path = os . path . join ( script_dir , " installer_files " , " env " )
2023-04-18 07:23:09 +02:00
2023-05-10 22:54:12 +02:00
# Use this to set your command-line flags. For the full list, see:
# https://github.com/oobabooga/text-generation-webui/#starting-the-web-ui
2023-06-01 06:20:56 +02:00
CMD_FLAGS = ' --chat '
2023-04-18 07:23:09 +02:00
2023-05-10 23:03:12 +02:00
2023-05-31 19:43:22 +02:00
# Allows users to set flags in "OOBABOOGA_FLAGS" environment variable
if " OOBABOOGA_FLAGS " in os . environ :
CMD_FLAGS = os . environ [ " OOBABOOGA_FLAGS " ]
2023-06-01 06:20:56 +02:00
print ( " The following flags have been taken from the environment variable ' OOBABOOGA_FLAGS ' : " )
2023-05-31 19:43:22 +02:00
print ( CMD_FLAGS )
2023-06-01 06:20:56 +02:00
print ( " To use the CMD_FLAGS Inside webui.py, unset ' OOBABOOGA_FLAGS ' . \n " )
def print_big_message ( message ) :
message = message . strip ( )
lines = message . split ( ' \n ' )
print ( " \n \n ******************************************************************* " )
for line in lines :
if line . strip ( ) != ' ' :
print ( " * " , line )
print ( " ******************************************************************* \n \n " )
2023-05-31 19:43:22 +02:00
2023-05-02 17:28:20 +02:00
def run_cmd ( cmd , assert_success = False , environment = False , capture_output = False , env = None ) :
# Use the conda environment
if environment :
if sys . platform . startswith ( " win " ) :
conda_bat_path = os . path . join ( script_dir , " installer_files " , " conda " , " condabin " , " conda.bat " )
cmd = " \" " + conda_bat_path + " \" activate \" " + conda_env_path + " \" >nul && " + cmd
else :
conda_sh_path = os . path . join ( script_dir , " installer_files " , " conda " , " etc " , " profile.d " , " conda.sh " )
cmd = " . \" " + conda_sh_path + " \" && conda activate \" " + conda_env_path + " \" && " + cmd
2023-06-01 06:20:56 +02:00
2023-04-18 07:23:09 +02:00
# Run shell commands
2023-05-02 17:28:20 +02:00
result = subprocess . run ( cmd , shell = True , capture_output = capture_output , env = env )
2023-06-01 06:20:56 +02:00
2023-05-02 17:28:20 +02:00
# Assert the command ran successfully
if assert_success and result . returncode != 0 :
print ( " Command ' " + cmd + " ' failed with exit status code ' " + str ( result . returncode ) + " ' . Exiting... " )
sys . exit ( )
2023-06-01 06:20:56 +02:00
2023-05-02 17:28:20 +02:00
return result
2023-04-18 07:23:09 +02:00
def check_env ( ) :
# If we have access to conda, we are probably in an environment
2023-05-02 17:28:20 +02:00
conda_exist = run_cmd ( " conda " , environment = True , capture_output = True ) . returncode == 0
if not conda_exist :
2023-04-18 07:23:09 +02:00
print ( " Conda is not installed. Exiting... " )
sys . exit ( )
2023-06-01 06:20:56 +02:00
2023-04-18 07:23:09 +02:00
# Ensure this is a new environment and not the base environment
if os . environ [ " CONDA_DEFAULT_ENV " ] == " base " :
print ( " Create an environment for this project and activate it. Exiting... " )
sys . exit ( )
def install_dependencies ( ) :
2023-06-18 00:09:22 +02:00
# Check for special characters in installation path on Windows
if sys . platform . startswith ( " win " ) :
# punctuation contains: !"#$%&'()*+,-./:;<=>?@[\]^_`{|}~
from string import punctuation
# Allow some characters: _-:\/.'"
special_characters = punctuation . translate ( { ord ( char ) : None for char in ' _-: \\ /. \' " ' } )
if any ( char in script_dir for char in special_characters ) :
print_big_message ( " WARNING: Special characters were detected in the installation path! \n This can cause the installation to fail! " )
2023-04-18 07:23:09 +02:00
# Select your GPU or, choose to run in CPU mode
print ( " What is your GPU " )
print ( )
print ( " A) NVIDIA " )
print ( " B) AMD " )
print ( " C) Apple M Series " )
print ( " D) None (I want to run in CPU mode) " )
print ( )
gpuchoice = input ( " Input> " ) . lower ( )
2023-05-25 15:57:52 +02:00
if gpuchoice == " d " :
2023-06-01 06:38:48 +02:00
print_big_message ( " Once the installation ends, make sure to open webui.py with a text editor \n and add the --cpu flag to CMD_FLAGS. " )
2023-05-25 15:57:52 +02:00
2023-04-18 07:23:09 +02:00
# Install the version of PyTorch needed
if gpuchoice == " a " :
2023-05-02 17:28:20 +02:00
run_cmd ( " conda install -y -k pytorch[version=2,build=py3.10_cuda11.7*] torchvision torchaudio pytorch-cuda=11.7 cuda-toolkit ninja git -c pytorch -c nvidia/label/cuda-11.7.0 -c nvidia " , assert_success = True , environment = True )
2023-04-18 07:23:09 +02:00
elif gpuchoice == " b " :
print ( " AMD GPUs are not supported. Exiting... " )
sys . exit ( )
elif gpuchoice == " c " or gpuchoice == " d " :
2023-05-02 17:28:20 +02:00
run_cmd ( " conda install -y -k pytorch torchvision torchaudio cpuonly git -c pytorch " , assert_success = True , environment = True )
2023-04-18 07:23:09 +02:00
else :
print ( " Invalid choice. Exiting... " )
sys . exit ( )
# Clone webui to our computer
2023-05-02 17:28:20 +02:00
run_cmd ( " git clone https://github.com/oobabooga/text-generation-webui.git " , assert_success = True , environment = True )
2023-05-25 15:50:20 +02:00
# if sys.platform.startswith("win"):
# # Fix a bitsandbytes compatibility issue with Windows
# run_cmd("python -m pip install https://github.com/jllllll/bitsandbytes-windows-webui/raw/main/bitsandbytes-0.38.1-py3-none-any.whl", assert_success=True, environment=True)
2023-06-01 06:20:56 +02:00
2023-04-18 07:23:09 +02:00
# Install the webui dependencies
update_dependencies ( )
def update_dependencies ( ) :
os . chdir ( " text-generation-webui " )
2023-05-02 17:28:20 +02:00
run_cmd ( " git pull " , assert_success = True , environment = True )
2023-04-18 07:23:09 +02:00
# Installs/Updates dependencies from all requirements.txt
2023-05-02 17:28:20 +02:00
run_cmd ( " python -m pip install -r requirements.txt --upgrade " , assert_success = True , environment = True )
2023-04-18 07:23:09 +02:00
extensions = next ( os . walk ( " extensions " ) ) [ 1 ]
for extension in extensions :
2023-05-09 19:17:08 +02:00
if extension in [ ' superbooga ' ] : # No wheels available for dependencies
continue
2023-06-01 06:20:56 +02:00
2023-04-18 07:23:09 +02:00
extension_req_path = os . path . join ( " extensions " , extension , " requirements.txt " )
if os . path . exists ( extension_req_path ) :
2023-05-02 17:28:20 +02:00
run_cmd ( " python -m pip install -r " + extension_req_path + " --upgrade " , assert_success = True , environment = True )
2023-04-18 07:23:09 +02:00
2023-05-31 19:41:03 +02:00
# Latest bitsandbytes requires minimum compute 7.0
nvcc_device_query = " __nvcc_device_query " if not sys . platform . startswith ( " win " ) else " __nvcc_device_query.exe "
min_compute = 70
compute_array = run_cmd ( os . path . join ( conda_env_path , " bin " , nvcc_device_query ) , environment = True , capture_output = True )
old_bnb = " bitsandbytes==0.38.1 " if not sys . platform . startswith ( " win " ) else " https://github.com/jllllll/bitsandbytes-windows-webui/raw/main/bitsandbytes-0.38.1-py3-none-any.whl "
if compute_array . returncode == 0 and not any ( int ( compute ) > = min_compute for compute in compute_array . stdout . decode ( ' utf-8 ' ) . split ( ' , ' ) ) :
old_bnb_install = run_cmd ( f " python -m pip install { old_bnb } --force-reinstall --no-deps " , environment = True ) . returncode == 0
2023-06-01 06:20:56 +02:00
message = " \n \n WARNING: GPU with compute < 7.0 detected! \n "
2023-05-31 19:41:03 +02:00
if old_bnb_install :
2023-06-01 06:20:56 +02:00
message + = " Older version of bitsandbytes has been installed to maintain compatibility. \n "
message + = " You will be unable to use --load-in-4bit! \n "
2023-05-31 19:41:03 +02:00
else :
2023-06-01 06:20:56 +02:00
message + = " You will be unable to use --load-in-8bit until you install bitsandbytes 0.38.1! \n "
print_big_message ( message )
2023-05-31 19:41:03 +02:00
2023-04-18 07:23:09 +02:00
# The following dependencies are for CUDA, not CPU
# Check if the package cpuonly exists to determine if torch uses CUDA or not
2023-05-02 17:28:20 +02:00
cpuonly_exist = run_cmd ( " conda list cpuonly | grep cpuonly " , environment = True , capture_output = True ) . returncode == 0
2023-04-18 07:23:09 +02:00
if cpuonly_exist :
return
# Finds the path to your dependencies
for sitedir in site . getsitepackages ( ) :
if " site-packages " in sitedir :
site_packages_path = sitedir
break
# This path is critical to installing the following dependencies
if site_packages_path is None :
print ( " Could not find the path to your Python packages. Exiting... " )
sys . exit ( )
# Fix a bitsandbytes compatibility issue with Linux
if sys . platform . startswith ( " linux " ) :
shutil . copy ( os . path . join ( site_packages_path , " bitsandbytes " , " libbitsandbytes_cuda117.so " ) , os . path . join ( site_packages_path , " bitsandbytes " , " libbitsandbytes_cpu.so " ) )
if not os . path . exists ( " repositories/ " ) :
os . mkdir ( " repositories " )
2023-06-01 06:20:56 +02:00
2023-04-18 07:23:09 +02:00
# Install GPTQ-for-LLaMa which enables 4bit CUDA quantization
os . chdir ( " repositories " )
if not os . path . exists ( " GPTQ-for-LLaMa/ " ) :
2023-05-02 17:28:20 +02:00
run_cmd ( " git clone https://github.com/oobabooga/GPTQ-for-LLaMa.git -b cuda " , assert_success = True , environment = True )
2023-06-01 06:20:56 +02:00
2023-04-18 07:23:09 +02:00
# Install GPTQ-for-LLaMa dependencies
os . chdir ( " GPTQ-for-LLaMa " )
2023-05-02 17:28:20 +02:00
run_cmd ( " git pull " , assert_success = True , environment = True )
2023-06-01 06:20:56 +02:00
2023-04-18 07:23:09 +02:00
# On some Linux distributions, g++ may not exist or be the wrong version to compile GPTQ-for-LLaMa
if sys . platform . startswith ( " linux " ) :
2023-05-31 19:44:36 +02:00
gxx_output = run_cmd ( " g++ -dumpfullversion -dumpversion " , environment = True , capture_output = True )
if gxx_output . returncode != 0 or int ( gxx_output . stdout . strip ( ) . split ( b " . " ) [ 0 ] ) > 11 :
2023-04-18 07:23:09 +02:00
# Install the correct version of g++
2023-05-02 17:28:20 +02:00
run_cmd ( " conda install -y -k gxx_linux-64=11.2.0 " , environment = True )
2023-04-18 07:23:09 +02:00
2023-05-02 17:28:20 +02:00
# Compile and install GPTQ-for-LLaMa
2023-05-06 06:14:09 +02:00
if os . path . exists ( ' setup_cuda.py ' ) :
os . rename ( " setup_cuda.py " , " setup.py " )
2023-06-01 06:20:56 +02:00
2023-05-02 17:28:20 +02:00
run_cmd ( " python -m pip install . " , environment = True )
2023-06-01 06:20:56 +02:00
2023-05-02 17:28:20 +02:00
# Wheel installation can fail while in the build directory of a package with the same name
os . chdir ( " .. " )
2023-06-01 06:20:56 +02:00
2023-04-18 07:23:09 +02:00
# If the path does not exist, then the install failed
quant_cuda_path_regex = os . path . join ( site_packages_path , " quant_cuda*/ " )
if not glob . glob ( quant_cuda_path_regex ) :
2023-05-31 19:41:54 +02:00
# Attempt installation via alternative, Windows/Linux-specific method
if sys . platform . startswith ( " win " ) or sys . platform . startswith ( " linux " ) :
2023-06-01 06:20:56 +02:00
print_big_message ( " WARNING: GPTQ-for-LLaMa compilation failed, but this is FINE and can be ignored! \n The installer will proceed to install a pre-compiled wheel. " )
2023-05-31 19:41:54 +02:00
url = " https://github.com/jllllll/GPTQ-for-LLaMa-Wheels/raw/main/quant_cuda-0.0.0-cp310-cp310-win_amd64.whl "
if sys . platform . startswith ( " linux " ) :
url = " https://github.com/jllllll/GPTQ-for-LLaMa-Wheels/raw/Linux-x64/quant_cuda-0.0.0-cp310-cp310-linux_x86_64.whl "
result = run_cmd ( " python -m pip install " + url , environment = True )
2023-05-02 17:28:20 +02:00
if result . returncode == 0 :
print ( " Wheel installation success! " )
else :
print ( " ERROR: GPTQ wheel installation failed. You will not be able to use GPTQ-based models. " )
else :
2023-05-18 15:56:49 +02:00
print ( " ERROR: GPTQ CUDA kernel compilation failed. " )
2023-05-02 17:28:20 +02:00
print ( " You will not be able to use GPTQ-based models. " )
2023-06-01 06:20:56 +02:00
2023-05-02 17:28:20 +02:00
print ( " Continuing with install.. " )
2023-04-18 07:23:09 +02:00
def download_model ( ) :
os . chdir ( " text-generation-webui " )
2023-05-02 17:28:20 +02:00
run_cmd ( " python download-model.py " , environment = True )
2023-04-18 07:23:09 +02:00
2023-06-01 06:20:56 +02:00
def launch_webui ( ) :
2023-04-18 07:23:09 +02:00
os . chdir ( " text-generation-webui " )
2023-05-10 23:01:04 +02:00
run_cmd ( f " python server.py { CMD_FLAGS } " , environment = True )
2023-04-18 07:23:09 +02:00
if __name__ == " __main__ " :
# Verifies we are in a conda environment
check_env ( )
parser = argparse . ArgumentParser ( )
parser . add_argument ( ' --update ' , action = ' store_true ' , help = ' Update the web UI. ' )
args = parser . parse_args ( )
if args . update :
update_dependencies ( )
else :
# If webui has already been installed, skip and run
if not os . path . exists ( " text-generation-webui/ " ) :
install_dependencies ( )
os . chdir ( script_dir )
# Check if a model has been downloaded yet
2023-06-01 19:01:19 +02:00
if len ( [ item for item in glob . glob ( ' text-generation-webui/models/* ' ) if not item . endswith ( ( ' .txt ' , ' .yaml ' ) ) ] ) == 0 :
2023-06-01 06:20:56 +02:00
print_big_message ( " WARNING: You haven ' t downloaded any model yet. \n Once the web UI launches, head over to the bottom of the \" Model \" tab and download one. " )
2023-04-18 07:23:09 +02:00
2023-05-31 19:41:03 +02:00
# Workaround for llama-cpp-python loading paths in CUDA env vars even if they do not exist
conda_path_bin = os . path . join ( conda_env_path , " bin " )
if not os . path . exists ( conda_path_bin ) :
os . mkdir ( conda_path_bin )
2023-06-01 06:20:56 +02:00
# Launch the webui
launch_webui ( )