diff --git a/examples/server/README.md b/examples/server/README.md index 364291ee3..4ff4f5f6d 100644 --- a/examples/server/README.md +++ b/examples/server/README.md @@ -110,7 +110,7 @@ node . `prompt`: Provide a prompt. Internally, the prompt is compared, and it detects if a part has already been evaluated, and the remaining part will be evaluate. - `stop`: Specify the strings that indicate a stop. + `stop`: Specify a JSON array of stopping strings. These words will not be included in the completion, so make sure to add them to the prompt for the next iteration (default: []). `tfs_z`: Enable tail free sampling with parameter z (default: 1.0, 1.0 = disabled). diff --git a/examples/server/server.cpp b/examples/server/server.cpp index afe0cc03c..ee645c143 100644 --- a/examples/server/server.cpp +++ b/examples/server/server.cpp @@ -774,7 +774,6 @@ int main(int argc, char ** argv) // struct that contains llama context and inference llama_server_context llama; - params.model = "ggml-model.bin"; server_params_parse(argc, argv, sparams, params);