mirror of
https://github.com/abetlen/llama-cpp-python.git
synced 2023-09-07 17:34:22 +03:00
Fix #16
This commit is contained in:
@@ -262,6 +262,17 @@ class Llama:
|
|||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
|
def embed(self, input: str) -> List[float]:
|
||||||
|
"""Embed a string.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
input: The utf-8 encoded string to embed.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
A list of embeddings
|
||||||
|
"""
|
||||||
|
return list(map(float, self.create_embedding(input)["data"][0]["embedding"]))
|
||||||
|
|
||||||
def _create_completion(
|
def _create_completion(
|
||||||
self,
|
self,
|
||||||
prompt: str,
|
prompt: str,
|
||||||
@@ -341,7 +352,7 @@ class Llama:
|
|||||||
"model": self.model_path,
|
"model": self.model_path,
|
||||||
"choices": [
|
"choices": [
|
||||||
{
|
{
|
||||||
"text": text[start :].decode("utf-8"),
|
"text": text[start:].decode("utf-8"),
|
||||||
"index": 0,
|
"index": 0,
|
||||||
"logprobs": None,
|
"logprobs": None,
|
||||||
"finish_reason": None,
|
"finish_reason": None,
|
||||||
|
|||||||
Reference in New Issue
Block a user