Resource: Model
Information about a Generative Language Model.
JSON representation |
---|
{ "name": string, "baseModelId": string, "version": string, "displayName": string, "description": string, "inputTokenLimit": integer, "outputTokenLimit": integer, "supportedGenerationMethods": [ string ], "temperature": number, "topP": number, "topK": integer } |
Fields | |
---|---|
name |
Required. The resource name of the Format:
Examples:
|
baseModelId |
Required. The name of the base model, pass this to the generation request. Examples:
|
version |
Required. The version number of the model. This represents the major version |
displayName |
The human-readable name of the model. E.g. "Chat Bison". The name can be up to 128 characters long and can consist of any UTF-8 characters. |
description |
A short description of the model. |
inputTokenLimit |
Maximum number of input tokens allowed for this model. |
outputTokenLimit |
Maximum number of output tokens available for this model. |
supportedGenerationMethods[] |
The model's supported generation methods. The method names are defined as Pascal case strings, such as |
temperature |
Controls the randomness of the output. Values can range over |
topP |
For Nucleus sampling. Nucleus sampling considers the smallest set of tokens whose probability sum is at least |
topK |
For Top-k sampling. Top-k sampling considers the set of |
Methods |
|
---|---|
|
Generates multiple embeddings from the model given input text in a synchronous call. |
|
Runs a model's tokenizer on a string and returns the token count. |
|
Runs a model's tokenizer on a text and returns the token count. |
|
Generates an embedding from the model given an input message. |
|
Generates a response from the model given an input MessagePrompt . |
|
Generates a response from the model given an input message. |
|
Gets information about a specific Model. |
|
Lists models available through the API. |