server: tests: check for infinite loops
This commit is contained in:
parent
c355f76427
commit
367b59a15c
1 changed files with 8 additions and 3 deletions
|
@ -54,6 +54,10 @@ Feature: llama.cpp server
|
|||
It was her greeting to Prince Vassily, a man high in rank
|
||||
and office, who was the first to arrive at her soirée.
|
||||
"""
|
||||
And a prompt:
|
||||
"""
|
||||
Write another very long music lyrics.
|
||||
"""
|
||||
Given concurrent completion requests
|
||||
Then the server is busy
|
||||
And all slots are busy
|
||||
|
@ -65,7 +69,7 @@ Feature: llama.cpp server
|
|||
Scenario: Multi users OAI Compatibility
|
||||
Given a system prompt "You are an AI assistant."
|
||||
And a model tinyllama-2
|
||||
And 1024 max tokens to predict
|
||||
And 512 max tokens to predict
|
||||
And streaming is enabled
|
||||
Given a prompt:
|
||||
"""
|
||||
|
@ -77,11 +81,12 @@ Feature: llama.cpp server
|
|||
"""
|
||||
And a prompt:
|
||||
"""
|
||||
Write yet another very long music lyrics.
|
||||
I believe the meaning of life is
|
||||
"""
|
||||
Given concurrent OAI completions requests
|
||||
Then the server is busy
|
||||
And all slots are busy
|
||||
Then the server is idle
|
||||
And all slots are idle
|
||||
Then all prompts are predicted
|
||||
Then all prompts are predicted
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue