Change label to avoid confusion - rocm hipblas users should obtain binaries from yellowrosecx fork. The rocm support in this repo requires self-compilation
This commit is contained in:
parent
8b8eb18567
commit
63fcbbb3f1
1 changed files with 8 additions and 8 deletions
16
koboldcpp.py
16
koboldcpp.py
|
@ -765,7 +765,7 @@ def show_new_gui():
|
||||||
lib_option_pairs = [
|
lib_option_pairs = [
|
||||||
(lib_openblas, "Use OpenBLAS"),
|
(lib_openblas, "Use OpenBLAS"),
|
||||||
(lib_clblast, "Use CLBlast"),
|
(lib_clblast, "Use CLBlast"),
|
||||||
(lib_cublas, "Use CuBLAS/hipBLAS"),
|
(lib_cublas, "Use CuBLAS"),
|
||||||
(lib_default, "Use No BLAS"),
|
(lib_default, "Use No BLAS"),
|
||||||
(lib_noavx2, "NoAVX2 Mode (Old CPU)"),
|
(lib_noavx2, "NoAVX2 Mode (Old CPU)"),
|
||||||
(lib_failsafe, "Failsafe Mode (Old CPU)")]
|
(lib_failsafe, "Failsafe Mode (Old CPU)")]
|
||||||
|
@ -922,7 +922,7 @@ def show_new_gui():
|
||||||
|
|
||||||
def changerunmode(a,b,c):
|
def changerunmode(a,b,c):
|
||||||
index = runopts_var.get()
|
index = runopts_var.get()
|
||||||
if index == "Use CLBlast" or index == "Use CuBLAS/hipBLAS":
|
if index == "Use CLBlast" or index == "Use CuBLAS":
|
||||||
gpu_selector_label.grid(row=3, column=0, padx = 8, pady=1, stick="nw")
|
gpu_selector_label.grid(row=3, column=0, padx = 8, pady=1, stick="nw")
|
||||||
quick_gpu_selector_label.grid(row=3, column=0, padx = 8, pady=1, stick="nw")
|
quick_gpu_selector_label.grid(row=3, column=0, padx = 8, pady=1, stick="nw")
|
||||||
if index == "Use CLBlast":
|
if index == "Use CLBlast":
|
||||||
|
@ -930,7 +930,7 @@ def show_new_gui():
|
||||||
quick_gpu_selector_box.grid(row=3, column=1, padx=8, pady=1, stick="nw")
|
quick_gpu_selector_box.grid(row=3, column=1, padx=8, pady=1, stick="nw")
|
||||||
if gpu_choice_var.get()=="All":
|
if gpu_choice_var.get()=="All":
|
||||||
gpu_choice_var.set("1")
|
gpu_choice_var.set("1")
|
||||||
elif index == "Use CuBLAS/hipBLAS":
|
elif index == "Use CuBLAS":
|
||||||
CUDA_gpu_selector_box.grid(row=3, column=1, padx=8, pady=1, stick="nw")
|
CUDA_gpu_selector_box.grid(row=3, column=1, padx=8, pady=1, stick="nw")
|
||||||
CUDA_quick_gpu_selector_box.grid(row=3, column=1, padx=8, pady=1, stick="nw")
|
CUDA_quick_gpu_selector_box.grid(row=3, column=1, padx=8, pady=1, stick="nw")
|
||||||
else:
|
else:
|
||||||
|
@ -941,7 +941,7 @@ def show_new_gui():
|
||||||
quick_gpu_selector_box.grid_forget()
|
quick_gpu_selector_box.grid_forget()
|
||||||
CUDA_quick_gpu_selector_box.grid_forget()
|
CUDA_quick_gpu_selector_box.grid_forget()
|
||||||
|
|
||||||
if index == "Use CuBLAS/hipBLAS":
|
if index == "Use CuBLAS":
|
||||||
lowvram_box.grid(row=4, column=0, padx=8, pady=1, stick="nw")
|
lowvram_box.grid(row=4, column=0, padx=8, pady=1, stick="nw")
|
||||||
quick_lowvram_box.grid(row=4, column=0, padx=8, pady=1, stick="nw")
|
quick_lowvram_box.grid(row=4, column=0, padx=8, pady=1, stick="nw")
|
||||||
mmq_box.grid(row=4, column=1, padx=8, pady=1, stick="nw")
|
mmq_box.grid(row=4, column=1, padx=8, pady=1, stick="nw")
|
||||||
|
@ -952,7 +952,7 @@ def show_new_gui():
|
||||||
mmq_box.grid_forget()
|
mmq_box.grid_forget()
|
||||||
quick_mmq_box.grid_forget()
|
quick_mmq_box.grid_forget()
|
||||||
|
|
||||||
if index == "Use CLBlast" or index == "Use CuBLAS/hipBLAS":
|
if index == "Use CLBlast" or index == "Use CuBLAS":
|
||||||
gpu_layers_label.grid(row=5, column=0, padx = 8, pady=1, stick="nw")
|
gpu_layers_label.grid(row=5, column=0, padx = 8, pady=1, stick="nw")
|
||||||
gpu_layers_entry.grid(row=5, column=1, padx=8, pady=1, stick="nw")
|
gpu_layers_entry.grid(row=5, column=1, padx=8, pady=1, stick="nw")
|
||||||
quick_gpu_layers_label.grid(row=5, column=0, padx = 8, pady=1, stick="nw")
|
quick_gpu_layers_label.grid(row=5, column=0, padx = 8, pady=1, stick="nw")
|
||||||
|
@ -1147,7 +1147,7 @@ def show_new_gui():
|
||||||
gpuchoiceidx = int(gpu_choice_var.get())-1
|
gpuchoiceidx = int(gpu_choice_var.get())-1
|
||||||
if runopts_var.get() == "Use CLBlast":
|
if runopts_var.get() == "Use CLBlast":
|
||||||
args.useclblast = [[0,0], [1,0], [0,1], [1,1]][gpuchoiceidx]
|
args.useclblast = [[0,0], [1,0], [0,1], [1,1]][gpuchoiceidx]
|
||||||
if runopts_var.get() == "Use CuBLAS/hipBLAS":
|
if runopts_var.get() == "Use CuBLAS":
|
||||||
if gpu_choice_var.get()=="All":
|
if gpu_choice_var.get()=="All":
|
||||||
args.usecublas = ["lowvram"] if lowvram_var.get() == 1 else ["normal"]
|
args.usecublas = ["lowvram"] if lowvram_var.get() == 1 else ["normal"]
|
||||||
else:
|
else:
|
||||||
|
@ -1373,7 +1373,7 @@ def show_old_gui():
|
||||||
blaschoice = tk.StringVar()
|
blaschoice = tk.StringVar()
|
||||||
blaschoice.set("BLAS = 512")
|
blaschoice.set("BLAS = 512")
|
||||||
|
|
||||||
runopts = ["Use OpenBLAS","Use CLBLast GPU #1","Use CLBLast GPU #2","Use CLBLast GPU #3","Use CuBLAS/hipBLAS GPU","Use No BLAS","NoAVX2 Mode (Old CPU)","Failsafe Mode (Old CPU)"]
|
runopts = ["Use OpenBLAS","Use CLBLast GPU #1","Use CLBLast GPU #2","Use CLBLast GPU #3","Use CuBLAS GPU","Use No BLAS","NoAVX2 Mode (Old CPU)","Failsafe Mode (Old CPU)"]
|
||||||
runchoice = tk.StringVar()
|
runchoice = tk.StringVar()
|
||||||
runchoice.set("Use OpenBLAS")
|
runchoice.set("Use OpenBLAS")
|
||||||
|
|
||||||
|
@ -1824,7 +1824,7 @@ if __name__ == '__main__':
|
||||||
compatgroup = parser.add_mutually_exclusive_group()
|
compatgroup = parser.add_mutually_exclusive_group()
|
||||||
compatgroup.add_argument("--noblas", help="Do not use OpenBLAS for accelerated prompt ingestion", action='store_true')
|
compatgroup.add_argument("--noblas", help="Do not use OpenBLAS for accelerated prompt ingestion", action='store_true')
|
||||||
compatgroup.add_argument("--useclblast", help="Use CLBlast for GPU Acceleration. Must specify exactly 2 arguments, platform ID and device ID (e.g. --useclblast 1 0).", type=int, choices=range(0,9), nargs=2)
|
compatgroup.add_argument("--useclblast", help="Use CLBlast for GPU Acceleration. Must specify exactly 2 arguments, platform ID and device ID (e.g. --useclblast 1 0).", type=int, choices=range(0,9), nargs=2)
|
||||||
compatgroup.add_argument("--usecublas", help="Use CuBLAS/hipBLAS for GPU Acceleration. Requires CUDA. Select lowvram to not allocate VRAM scratch buffer. Enter a number afterwards to select and use 1 GPU. Leaving no number will use all GPUs.", nargs='*',metavar=('[lowvram|normal] [main GPU ID] [mmq]'), choices=['normal', 'lowvram', '0', '1', '2', '3', 'mmq'])
|
compatgroup.add_argument("--usecublas", help="Use CuBLAS for GPU Acceleration. Requires CUDA. Select lowvram to not allocate VRAM scratch buffer. Enter a number afterwards to select and use 1 GPU. Leaving no number will use all GPUs. For hipBLAS binaries, please check YellowRoseCx rocm fork.", nargs='*',metavar=('[lowvram|normal] [main GPU ID] [mmq]'), choices=['normal', 'lowvram', '0', '1', '2', '3', 'mmq'])
|
||||||
parser.add_argument("--gpulayers", help="Set number of layers to offload to GPU when using GPU. Requires GPU.",metavar=('[GPU layers]'), type=int, default=0)
|
parser.add_argument("--gpulayers", help="Set number of layers to offload to GPU when using GPU. Requires GPU.",metavar=('[GPU layers]'), type=int, default=0)
|
||||||
parser.add_argument("--tensor_split", help="For CUDA with ALL GPU set only, ratio to split tensors across multiple GPUs, space-separated list of proportions, e.g. 7 3", metavar=('[Ratios]'), type=float, nargs='+')
|
parser.add_argument("--tensor_split", help="For CUDA with ALL GPU set only, ratio to split tensors across multiple GPUs, space-separated list of proportions, e.g. 7 3", metavar=('[Ratios]'), type=float, nargs='+')
|
||||||
parser.add_argument("--onready", help="An optional shell command to execute after the model has been loaded.", type=str, default="",nargs=1)
|
parser.add_argument("--onready", help="An optional shell command to execute after the model has been loaded.", type=str, default="",nargs=1)
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue