revert test addr change
This commit is contained in:
parent
a5603ded45
commit
624214aedf
4 changed files with 4 additions and 4 deletions
|
@ -2,7 +2,7 @@
|
|||
Feature: Parallel
|
||||
|
||||
Background: Server startup
|
||||
Given a server listening on 0.0.0.0:8080
|
||||
Given a server listening on localhost:8080
|
||||
And a model file stories260K.gguf
|
||||
And a model alias tinyllama-2
|
||||
And 42 as server seed
|
||||
|
|
|
@ -2,7 +2,7 @@
|
|||
Feature: Security
|
||||
|
||||
Background: Server startup with an api key defined
|
||||
Given a server listening on 0.0.0.0:8080
|
||||
Given a server listening on localhost:8080
|
||||
And a model file stories260K.gguf
|
||||
And a server api key llama.cpp
|
||||
Then the server is starting
|
||||
|
|
|
@ -2,7 +2,7 @@
|
|||
Feature: llama.cpp server
|
||||
|
||||
Background: Server startup
|
||||
Given a server listening on 0.0.0.0:8080
|
||||
Given a server listening on localhost:8080
|
||||
And a model file stories260K.gguf
|
||||
And a model alias tinyllama-2
|
||||
And 42 as server seed
|
||||
|
|
|
@ -6,7 +6,7 @@ Feature: Wrong usage of llama.cpp server
|
|||
# to cap the number of tokens any completion request can generate
|
||||
# or pass n_predict/max_tokens in the request.
|
||||
Scenario: Infinite loop
|
||||
Given a server listening on 0.0.0.0:8080
|
||||
Given a server listening on localhost:8080
|
||||
And a model file stories260K.gguf
|
||||
# Uncomment below to fix the issue
|
||||
#And 64 server max tokens to predict
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue