server : initiate tests for later
ggml-ci
This commit is contained in:
parent
84f56f3c45
commit
00b33760aa
2 changed files with 40 additions and 0 deletions
|
@ -3128,6 +3128,7 @@ class JinaBertV2Model(BertModel):
|
|||
|
||||
return super().modify_tensors(data_torch, name, bid)
|
||||
|
||||
|
||||
@Model.register("OpenELMForCausalLM")
|
||||
class OpenELMModel(Model):
|
||||
model_arch = gguf.MODEL_ARCH.OPENELM
|
||||
|
|
39
examples/server/tests/features/rerank.feature
Normal file
39
examples/server/tests/features/rerank.feature
Normal file
|
@ -0,0 +1,39 @@
|
|||
@llama.cpp
|
||||
@rerank
|
||||
Feature: llama.cpp server
|
||||
|
||||
Background: Server startup
|
||||
Given a server listening on localhost:8080
|
||||
And a model url https://huggingface.co/ggml-org/models/resolve/main/jina-reranker-v1-tiny-en/ggml-model-f16.gguf
|
||||
And a model file jina-reranker-v1-tiny-en.gguf
|
||||
And a model alias jina-reranker-v1-tiny-en
|
||||
And 42 as server seed
|
||||
And 2 slots
|
||||
And 128 as batch size
|
||||
And 128 as ubatch size
|
||||
And 512 KV cache size
|
||||
And embeddings extraction
|
||||
Then the server is starting
|
||||
Then the server is healthy
|
||||
|
||||
# TODO: implement some tests
|
||||
# https://github.com/ggerganov/llama.cpp/pull/9510
|
||||
# Scenario: Rerank
|
||||
# Given a prompt:
|
||||
# """
|
||||
# What is panda?
|
||||
# """
|
||||
# And a prompt:
|
||||
# """
|
||||
# Hi.
|
||||
# """
|
||||
# And a prompt:
|
||||
# """
|
||||
# It's a bear.
|
||||
# """
|
||||
# And a prompt:
|
||||
# """
|
||||
# The giant panda (Ailuropoda melanoleuca), sometimes called a panda bear or simply panda, is a bear species endemic to China.
|
||||
# """
|
||||
# When reranking request
|
||||
# Then reranking results are returned
|
Loading…
Add table
Add a link
Reference in a new issue