llama_cpp
pixeltable.functions.llama_cpp
create_chat_completion
create_chat_completion(
messages: Json,
*,
model_path: Optional[String] = None,
repo_id: Optional[String] = None,
repo_filename: Optional[String] = None,
args: Optional[Json] = None
) -> Json
Generate a chat completion from a list of messages.
The model can be specified either as a local path, or as a repo_id and repo_filename that reference a pretrained
model on the Hugging Face model hub. Exactly one of model_path
or repo_id
must be provided; if model_path
is provided, then an optional repo_filename
can also be specified.
For additional details, see the llama_cpp create_chat_completions documentation.
Parameters:
-
messages
(Json
) –A list of messages to generate a response for.
-
model_path
(Optional[String]
, default:None
) –Path to the model (if using a local model).
-
repo_id
(Optional[String]
, default:None
) –The Hugging Face model repo id (if using a pretrained model).
-
repo_filename
(Optional[String]
, default:None
) –A filename or glob pattern to match the model file in the repo (optional, if using a pretrained model).
-
args
(Optional[Json]
, default:None
) –Additional arguments to pass to the
create_chat_completions
call, such asmax_tokens
,temperature
,top_p
, andtop_k
. For details, see the llama_cpp create_chat_completions documentation.