mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-12-27 06:39:25 +01:00
57bb2c40cd
* server : fix logprobs, make it openai-compatible * update docs * add std::log * return pre-sampling p * sort before apply softmax * add comment * fix test * set p for sampled token * update docs * add --multi-token-probs * update docs * add `post_sampling_probs` option * update docs [no ci] * remove --multi-token-probs * "top_probs" with "post_sampling_probs" * resolve review comments * rename struct token_prob to prob_info * correct comment placement * fix setting prob for sampled token
197 lines
5.9 KiB
Python
197 lines
5.9 KiB
Python
import pytest
|
|
from openai import OpenAI
|
|
from utils import *
|
|
|
|
server = ServerPreset.bert_bge_small()
|
|
|
|
EPSILON = 1e-3
|
|
|
|
@pytest.fixture(scope="module", autouse=True)
|
|
def create_server():
|
|
global server
|
|
server = ServerPreset.bert_bge_small()
|
|
|
|
|
|
def test_embedding_single():
|
|
global server
|
|
server.pooling = 'last'
|
|
server.start()
|
|
res = server.make_request("POST", "/v1/embeddings", data={
|
|
"input": "I believe the meaning of life is",
|
|
})
|
|
assert res.status_code == 200
|
|
assert len(res.body['data']) == 1
|
|
assert 'embedding' in res.body['data'][0]
|
|
assert len(res.body['data'][0]['embedding']) > 1
|
|
|
|
# make sure embedding vector is normalized
|
|
assert abs(sum([x ** 2 for x in res.body['data'][0]['embedding']]) - 1) < EPSILON
|
|
|
|
|
|
def test_embedding_multiple():
|
|
global server
|
|
server.pooling = 'last'
|
|
server.start()
|
|
res = server.make_request("POST", "/v1/embeddings", data={
|
|
"input": [
|
|
"I believe the meaning of life is",
|
|
"Write a joke about AI from a very long prompt which will not be truncated",
|
|
"This is a test",
|
|
"This is another test",
|
|
],
|
|
})
|
|
assert res.status_code == 200
|
|
assert len(res.body['data']) == 4
|
|
for d in res.body['data']:
|
|
assert 'embedding' in d
|
|
assert len(d['embedding']) > 1
|
|
|
|
|
|
@pytest.mark.parametrize(
|
|
"input,is_multi_prompt",
|
|
[
|
|
# do not crash on empty input
|
|
("", False),
|
|
# single prompt
|
|
("string", False),
|
|
([12, 34, 56], False),
|
|
([12, 34, "string", 56, 78], False),
|
|
# multiple prompts
|
|
(["string1", "string2"], True),
|
|
(["string1", [12, 34, 56]], True),
|
|
([[12, 34, 56], [12, 34, 56]], True),
|
|
([[12, 34, 56], [12, "string", 34, 56]], True),
|
|
]
|
|
)
|
|
def test_embedding_mixed_input(input, is_multi_prompt: bool):
|
|
global server
|
|
server.start()
|
|
res = server.make_request("POST", "/v1/embeddings", data={"input": input})
|
|
assert res.status_code == 200
|
|
data = res.body['data']
|
|
if is_multi_prompt:
|
|
assert len(data) == len(input)
|
|
for d in data:
|
|
assert 'embedding' in d
|
|
assert len(d['embedding']) > 1
|
|
else:
|
|
assert 'embedding' in data[0]
|
|
assert len(data[0]['embedding']) > 1
|
|
|
|
|
|
def test_embedding_pooling_none():
|
|
global server
|
|
server.pooling = 'none'
|
|
server.start()
|
|
res = server.make_request("POST", "/embeddings", data={
|
|
"input": "hello hello hello",
|
|
})
|
|
assert res.status_code == 200
|
|
assert 'embedding' in res.body[0]
|
|
assert len(res.body[0]['embedding']) == 5 # 3 text tokens + 2 special
|
|
|
|
# make sure embedding vector is not normalized
|
|
for x in res.body[0]['embedding']:
|
|
assert abs(sum([x ** 2 for x in x]) - 1) > EPSILON
|
|
|
|
|
|
def test_embedding_pooling_none_oai():
|
|
global server
|
|
server.pooling = 'none'
|
|
server.start()
|
|
res = server.make_request("POST", "/v1/embeddings", data={
|
|
"input": "hello hello hello",
|
|
})
|
|
|
|
# /v1/embeddings does not support pooling type 'none'
|
|
assert res.status_code == 400
|
|
assert "error" in res.body
|
|
|
|
|
|
def test_embedding_openai_library_single():
|
|
global server
|
|
server.pooling = 'last'
|
|
server.start()
|
|
client = OpenAI(api_key="dummy", base_url=f"http://{server.server_host}:{server.server_port}/v1")
|
|
res = client.embeddings.create(model="text-embedding-3-small", input="I believe the meaning of life is")
|
|
assert len(res.data) == 1
|
|
assert len(res.data[0].embedding) > 1
|
|
|
|
|
|
def test_embedding_openai_library_multiple():
|
|
global server
|
|
server.pooling = 'last'
|
|
server.start()
|
|
client = OpenAI(api_key="dummy", base_url=f"http://{server.server_host}:{server.server_port}/v1")
|
|
res = client.embeddings.create(model="text-embedding-3-small", input=[
|
|
"I believe the meaning of life is",
|
|
"Write a joke about AI from a very long prompt which will not be truncated",
|
|
"This is a test",
|
|
"This is another test",
|
|
])
|
|
assert len(res.data) == 4
|
|
for d in res.data:
|
|
assert len(d.embedding) > 1
|
|
|
|
|
|
def test_embedding_error_prompt_too_long():
|
|
global server
|
|
server.pooling = 'last'
|
|
server.start()
|
|
res = server.make_request("POST", "/v1/embeddings", data={
|
|
"input": "This is a test " * 512,
|
|
})
|
|
assert res.status_code != 200
|
|
assert "too large" in res.body["error"]["message"]
|
|
|
|
|
|
def test_same_prompt_give_same_result():
|
|
server.pooling = 'last'
|
|
server.start()
|
|
res = server.make_request("POST", "/v1/embeddings", data={
|
|
"input": [
|
|
"I believe the meaning of life is",
|
|
"I believe the meaning of life is",
|
|
"I believe the meaning of life is",
|
|
"I believe the meaning of life is",
|
|
"I believe the meaning of life is",
|
|
],
|
|
})
|
|
assert res.status_code == 200
|
|
assert len(res.body['data']) == 5
|
|
for i in range(1, len(res.body['data'])):
|
|
v0 = res.body['data'][0]['embedding']
|
|
vi = res.body['data'][i]['embedding']
|
|
for x, y in zip(v0, vi):
|
|
assert abs(x - y) < EPSILON
|
|
|
|
|
|
@pytest.mark.parametrize(
|
|
"content,n_tokens",
|
|
[
|
|
("I believe the meaning of life is", 9),
|
|
("This is a test", 6),
|
|
]
|
|
)
|
|
def test_embedding_usage_single(content, n_tokens):
|
|
global server
|
|
server.start()
|
|
res = server.make_request("POST", "/v1/embeddings", data={"input": content})
|
|
assert res.status_code == 200
|
|
assert res.body['usage']['prompt_tokens'] == res.body['usage']['total_tokens']
|
|
assert res.body['usage']['prompt_tokens'] == n_tokens
|
|
|
|
|
|
def test_embedding_usage_multiple():
|
|
global server
|
|
server.start()
|
|
res = server.make_request("POST", "/v1/embeddings", data={
|
|
"input": [
|
|
"I believe the meaning of life is",
|
|
"I believe the meaning of life is",
|
|
],
|
|
})
|
|
assert res.status_code == 200
|
|
assert res.body['usage']['prompt_tokens'] == res.body['usage']['total_tokens']
|
|
assert res.body['usage']['prompt_tokens'] == 2 * 9
|