Overview
POST /api/generate
Parameters
model
: (required) the model nameprompt
: the prompt to generate a response forsuffix
: the text after the model responseimages
: (optional) a list of base64-encoded images (for multimodal models such as llava
)format
: the format to return a response in. Format can be json
or a JSON schemaoptions
: additional model parameters listed in the documentation for the Modelfile such as temperature
system
: system message to (overrides what is defined in the Modelfile
)template
: the prompt template to use (overrides what is defined in the Modelfile
)stream
: if false
the response will be returned as a single response object, rather than a stream of objectsraw
: if true
no formatting will be applied to the prompt. You may choose to use the raw
parameter if you are specifying a full templated prompt in your request to the APIkeep_alive
: controls how long the model will stay loaded into memory following the request (default: 5m
)context
(deprecated): the context parameter returned from a previous request to /generate
, this can be used to keep a short conversational memoryStructured outputs
format
parameter. The model will generate a response that matches the schema. See the structured outputs example below.JSON mode
format
parameter to json
. This will structure the response as a valid JSON object. See the JSON mode example below.prompt
. Otherwise, the model may generate large amounts whitespace.Modified at 2025-03-14 07:55:41