wip cleanup before big merge
This commit is contained in:
parent
010b2753d9
commit
4e86a07e57
5 changed files with 2976 additions and 2989 deletions
|
@ -1080,7 +1080,6 @@ static void llama_model_load_internal(
|
||||||
const int n_gpu = std::min(n_gpu_layers, int(hparams.n_layer));
|
const int n_gpu = std::min(n_gpu_layers, int(hparams.n_layer));
|
||||||
if(GetQuantsUnshuffled())
|
if(GetQuantsUnshuffled())
|
||||||
{
|
{
|
||||||
SetGPULayers(n_gpu);
|
|
||||||
|
|
||||||
fprintf(stderr, "%s: [opencl] offloading %d layers to GPU\n", __func__, n_gpu);
|
fprintf(stderr, "%s: [opencl] offloading %d layers to GPU\n", __func__, n_gpu);
|
||||||
|
|
||||||
|
|
|
@ -333,7 +333,6 @@ ModelLoadResult gpt2_model_load(const std::string & fname, gpt2_model & model, g
|
||||||
// const int n_gpu = std::min(gpulayers, int(hparams.n_layer));
|
// const int n_gpu = std::min(gpulayers, int(hparams.n_layer));
|
||||||
// if(GetQuantsUnshuffled())
|
// if(GetQuantsUnshuffled())
|
||||||
// {
|
// {
|
||||||
// SetGPULayers(n_gpu);
|
|
||||||
|
|
||||||
// fprintf(stderr, "%s: [opencl] offloading %d layers to GPU\n", __func__, n_gpu);
|
// fprintf(stderr, "%s: [opencl] offloading %d layers to GPU\n", __func__, n_gpu);
|
||||||
|
|
||||||
|
|
|
@ -336,7 +336,6 @@ ModelLoadResult gptj_model_load(const std::string & fname, gptj_model & model, g
|
||||||
// const int n_gpu = std::min(gpulayers, int(hparams.n_layer));
|
// const int n_gpu = std::min(gpulayers, int(hparams.n_layer));
|
||||||
// if(GetQuantsUnshuffled())
|
// if(GetQuantsUnshuffled())
|
||||||
// {
|
// {
|
||||||
// SetGPULayers(n_gpu);
|
|
||||||
|
|
||||||
// fprintf(stderr, "%s: [opencl] offloading %d layers to GPU\n", __func__, n_gpu);
|
// fprintf(stderr, "%s: [opencl] offloading %d layers to GPU\n", __func__, n_gpu);
|
||||||
|
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue