{
"operation-name" : "add",
"description" : "This operation adds a Google Gemini chat model.",
"request-properties" : {
"allowed-code-execution" : {
"type" : {
"TYPE_MODEL_VALUE" : "BOOLEAN"
},
"description" : "Enables the model to generate and run Python code and learn iteratively from the results until it arrives at a final output.",
"expressions-allowed" : true,
"required" : false,
"nillable" : true,
"stability" : "default"
},
"api-key" : {
"type" : {
"TYPE_MODEL_VALUE" : "STRING"
},
"description" : "API key to authenticate to a Google Gemini chat model.",
"expressions-allowed" : true,
"required" : true,
"nillable" : false,
"min-length" : 1,
"max-length" : 2147483647,
"stability" : "default"
},
"civic-integrity" : {
"type" : {
"TYPE_MODEL_VALUE" : "STRING"
},
"description" : "Defines the threshold to filter out civic integrity content.",
"expressions-allowed" : true,
"required" : false,
"nillable" : true,
"min-length" : 1,
"max-length" : 2147483647,
"allowed" : [
"HARM_BLOCK_THRESHOLD_UNSPECIFIED",
"BLOCK_LOW_AND_ABOVE",
"BLOCK_MEDIUM_AND_ABOVE",
"BLOCK_ONLY_HIGH",
"BLOCK_NONE"
],
"stability" : "default"
},
"connect-timeout" : {
"type" : {
"TYPE_MODEL_VALUE" : "LONG"
},
"description" : "Timeout for the Google Gemini chat model.",
"expressions-allowed" : true,
"required" : false,
"nillable" : true,
"default" : 0,
"unit" : "MILLISECONDS",
"stability" : "default"
},
"dangerous-content" : {
"type" : {
"TYPE_MODEL_VALUE" : "STRING"
},
"description" : "Defines the threshold to filter out dangerous content.",
"expressions-allowed" : true,
"required" : false,
"nillable" : true,
"min-length" : 1,
"max-length" : 2147483647,
"allowed" : [
"HARM_BLOCK_THRESHOLD_UNSPECIFIED",
"BLOCK_LOW_AND_ABOVE",
"BLOCK_MEDIUM_AND_ABOVE",
"BLOCK_ONLY_HIGH",
"BLOCK_NONE"
],
"stability" : "default"
},
"enable-enhanced-civic-answers" : {
"type" : {
"TYPE_MODEL_VALUE" : "BOOLEAN"
},
"description" : "Enables enhanced civic answers. It may not be available for all models.",
"expressions-allowed" : true,
"required" : false,
"nillable" : true,
"stability" : "default"
},
"frequency-penalty" : {
"type" : {
"TYPE_MODEL_VALUE" : "DOUBLE"
},
"description" : "Positive values penalize tokens that repeatedly appear in the generated text, decreasing the probability of repeating content.",
"expressions-allowed" : true,
"required" : false,
"nillable" : true,
"stability" : "default"
},
"harassment" : {
"type" : {
"TYPE_MODEL_VALUE" : "STRING"
},
"description" : "Defines the threshold to filter out harassment content.",
"expressions-allowed" : true,
"required" : false,
"nillable" : true,
"min-length" : 1,
"max-length" : 2147483647,
"allowed" : [
"HARM_BLOCK_THRESHOLD_UNSPECIFIED",
"BLOCK_LOW_AND_ABOVE",
"BLOCK_MEDIUM_AND_ABOVE",
"BLOCK_ONLY_HIGH",
"BLOCK_NONE"
],
"stability" : "default"
},
"hate-speech" : {
"type" : {
"TYPE_MODEL_VALUE" : "STRING"
},
"description" : "Defines the threshold to filter out hate speech content.",
"expressions-allowed" : true,
"required" : false,
"nillable" : true,
"min-length" : 1,
"max-length" : 2147483647,
"allowed" : [
"HARM_BLOCK_THRESHOLD_UNSPECIFIED",
"BLOCK_LOW_AND_ABOVE",
"BLOCK_MEDIUM_AND_ABOVE",
"BLOCK_ONLY_HIGH",
"BLOCK_NONE"
],
"stability" : "default"
},
"include-code-execution-output" : {
"type" : {
"TYPE_MODEL_VALUE" : "BOOLEAN"
},
"description" : "Enables providing the code and the result in the output.",
"expressions-allowed" : true,
"required" : false,
"nillable" : true,
"stability" : "default"
},
"include-thoughts" : {
"type" : {
"TYPE_MODEL_VALUE" : "BOOLEAN"
},
"description" : "Enable the return of the thought summaries of the model.",
"expressions-allowed" : true,
"required" : false,
"nillable" : true,
"stability" : "default"
},
"log-probs" : {
"type" : {
"TYPE_MODEL_VALUE" : "INT"
},
"description" : "Returns the log probabilities of the top candidate tokens at each generation step. The model's chosen token might not be the same as the top candidate token at each step. Specify the number of candidates to return by using an integer value in the range of 1-20. ",
"expressions-allowed" : true,
"required" : false,
"nillable" : true,
"stability" : "default"
},
"log-requests" : {
"type" : {
"TYPE_MODEL_VALUE" : "BOOLEAN"
},
"description" : "Enabling the tracing of requests to Google Gemini",
"expressions-allowed" : true,
"required" : false,
"nillable" : true,
"stability" : "default"
},
"log-responses" : {
"type" : {
"TYPE_MODEL_VALUE" : "BOOLEAN"
},
"description" : "Enabling the tracing of responses to Google Gemini",
"expressions-allowed" : true,
"required" : false,
"nillable" : true,
"stability" : "default"
},
"max-output-token" : {
"type" : {
"TYPE_MODEL_VALUE" : "INT"
},
"description" : "The number of token returned by the Google Gemini chat model.",
"expressions-allowed" : true,
"required" : false,
"nillable" : true,
"stability" : "default"
},
"model-name" : {
"type" : {
"TYPE_MODEL_VALUE" : "STRING"
},
"description" : "Name of the model served by Google Gemini.",
"expressions-allowed" : true,
"required" : true,
"nillable" : false,
"min-length" : 1,
"max-length" : 2147483647,
"stability" : "default"
},
"presence-penalty" : {
"type" : {
"TYPE_MODEL_VALUE" : "DOUBLE"
},
"description" : "Presence penalty applied to the next token's logprobs if the token has already been seen in the response. This penalty is binary on/off and not dependant on the number of times the token is used (after the first). Use frequency-penalty for a penalty that increases with each use.",
"expressions-allowed" : true,
"required" : false,
"nillable" : true,
"stability" : "default"
},
"response-format" : {
"type" : {
"TYPE_MODEL_VALUE" : "STRING"
},
"description" : "The format of the response from Google Gemini.",
"expressions-allowed" : true,
"required" : false,
"nillable" : true,
"allowed" : [
"JSON",
"TEXT"
],
"stability" : "default"
},
"response-log-probs" : {
"type" : {
"TYPE_MODEL_VALUE" : "BOOLEAN"
},
"description" : "If true, returns the log probabilities of the tokens that were chosen by the model at each step.",
"expressions-allowed" : true,
"required" : false,
"nillable" : true,
"stability" : "default"
},
"return-thinking" : {
"type" : {
"TYPE_MODEL_VALUE" : "BOOLEAN"
},
"description" : "Controls whether to return thinking/reasoning text (if available).",
"expressions-allowed" : true,
"required" : false,
"nillable" : true,
"stability" : "default"
},
"seed" : {
"type" : {
"TYPE_MODEL_VALUE" : "INT"
},
"description" : "When seed is fixed to a specific value, the model makes a best effort to provide the same response for repeated requests. Deterministic output isn't guaranteed.",
"expressions-allowed" : true,
"required" : false,
"nillable" : true,
"stability" : "default"
},
"sexually-explicit" : {
"type" : {
"TYPE_MODEL_VALUE" : "STRING"
},
"description" : "Defines the threshold to filter out sexually explicit content.",
"expressions-allowed" : true,
"required" : false,
"nillable" : true,
"min-length" : 1,
"max-length" : 2147483647,
"allowed" : [
"HARM_BLOCK_THRESHOLD_UNSPECIFIED",
"BLOCK_LOW_AND_ABOVE",
"BLOCK_MEDIUM_AND_ABOVE",
"BLOCK_ONLY_HIGH",
"BLOCK_NONE"
],
"stability" : "default"
},
"stop-sequences" : {
"type" : {
"TYPE_MODEL_VALUE" : "LIST"
},
"description" : "List of stop sequences to tell the model to stop generating content.",
"expressions-allowed" : true,
"required" : false,
"nillable" : true,
"min-length" : 0,
"max-length" : 2147483647,
"stability" : "default",
"value-type" : {
"TYPE_MODEL_VALUE" : "STRING"
}
},
"streaming" : {
"type" : {
"TYPE_MODEL_VALUE" : "BOOLEAN"
},
"description" : "Whether to create a token streaming chat language model or not.",
"expressions-allowed" : true,
"required" : false,
"nillable" : true,
"default" : false,
"stability" : "default"
},
"temperature" : {
"type" : {
"TYPE_MODEL_VALUE" : "DOUBLE"
},
"description" : "Temperature of the Google Gemini chat model.",
"expressions-allowed" : true,
"required" : false,
"nillable" : true,
"stability" : "default"
},
"thinking-budget" : {
"type" : {
"TYPE_MODEL_VALUE" : "INT"
},
"description" : "Parameter to guide the model on the number of thinking tokens to use when generating a response.",
"expressions-allowed" : true,
"required" : false,
"nillable" : true,
"stability" : "default"
},
"top-k" : {
"type" : {
"TYPE_MODEL_VALUE" : "INT"
},
"description" : "TopK of the Google Gemini chat model.",
"expressions-allowed" : true,
"required" : false,
"nillable" : true,
"stability" : "default"
},
"top-p" : {
"type" : {
"TYPE_MODEL_VALUE" : "DOUBLE"
},
"description" : "TopP of the Google Gemini chat model.",
"expressions-allowed" : true,
"required" : false,
"nillable" : true,
"stability" : "default"
}
},
"reply-properties" : {},
"stability" : "default",
"read-only" : false,
"runtime-only" : false
}