Update cancel.feature
This commit is contained in:
parent
231a5e4914
commit
c5a0d57ee5
1 changed files with 0 additions and 2 deletions
|
@ -18,10 +18,8 @@ Feature: Cancellation of llama.cpp server requests
|
||||||
And 64 server max tokens to predict
|
And 64 server max tokens to predict
|
||||||
And prometheus compatible metrics exposed
|
And prometheus compatible metrics exposed
|
||||||
And 300 milliseconds delay in sampler for testing
|
And 300 milliseconds delay in sampler for testing
|
||||||
And no warmup
|
|
||||||
Then the server is starting
|
Then the server is starting
|
||||||
Then the server is healthy
|
Then the server is healthy
|
||||||
# Then the server is healthy with timeout 10 seconds
|
|
||||||
|
|
||||||
|
|
||||||
Scenario Outline: Cancelling an OAI chat completion request frees up slot (streaming <enable_streaming>)
|
Scenario Outline: Cancelling an OAI chat completion request frees up slot (streaming <enable_streaming>)
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue