revert test addr change
This commit is contained in:
parent
a5603ded45
commit
624214aedf
4 changed files with 4 additions and 4 deletions
|
@ -2,7 +2,7 @@
|
||||||
Feature: Parallel
|
Feature: Parallel
|
||||||
|
|
||||||
Background: Server startup
|
Background: Server startup
|
||||||
Given a server listening on 0.0.0.0:8080
|
Given a server listening on localhost:8080
|
||||||
And a model file stories260K.gguf
|
And a model file stories260K.gguf
|
||||||
And a model alias tinyllama-2
|
And a model alias tinyllama-2
|
||||||
And 42 as server seed
|
And 42 as server seed
|
||||||
|
|
|
@ -2,7 +2,7 @@
|
||||||
Feature: Security
|
Feature: Security
|
||||||
|
|
||||||
Background: Server startup with an api key defined
|
Background: Server startup with an api key defined
|
||||||
Given a server listening on 0.0.0.0:8080
|
Given a server listening on localhost:8080
|
||||||
And a model file stories260K.gguf
|
And a model file stories260K.gguf
|
||||||
And a server api key llama.cpp
|
And a server api key llama.cpp
|
||||||
Then the server is starting
|
Then the server is starting
|
||||||
|
|
|
@ -2,7 +2,7 @@
|
||||||
Feature: llama.cpp server
|
Feature: llama.cpp server
|
||||||
|
|
||||||
Background: Server startup
|
Background: Server startup
|
||||||
Given a server listening on 0.0.0.0:8080
|
Given a server listening on localhost:8080
|
||||||
And a model file stories260K.gguf
|
And a model file stories260K.gguf
|
||||||
And a model alias tinyllama-2
|
And a model alias tinyllama-2
|
||||||
And 42 as server seed
|
And 42 as server seed
|
||||||
|
|
|
@ -6,7 +6,7 @@ Feature: Wrong usage of llama.cpp server
|
||||||
# to cap the number of tokens any completion request can generate
|
# to cap the number of tokens any completion request can generate
|
||||||
# or pass n_predict/max_tokens in the request.
|
# or pass n_predict/max_tokens in the request.
|
||||||
Scenario: Infinite loop
|
Scenario: Infinite loop
|
||||||
Given a server listening on 0.0.0.0:8080
|
Given a server listening on localhost:8080
|
||||||
And a model file stories260K.gguf
|
And a model file stories260K.gguf
|
||||||
# Uncomment below to fix the issue
|
# Uncomment below to fix the issue
|
||||||
#And 64 server max tokens to predict
|
#And 64 server max tokens to predict
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue