Skip to content

Instantly share code, notes, and snippets.

@asus4
Last active April 27, 2025 13:52
Show Gist options
  • Save asus4/122847cb7a9b86f959aae293343a15f3 to your computer and use it in GitHub Desktop.
Save asus4/122847cb7a9b86f959aae293343a15f3 to your computer and use it in GitHub Desktop.
{
"models": [
{
"name": "models/chat-bison-001",
"version": "001",
"displayName": "PaLM 2 Chat (Legacy)",
"description": "A legacy text-only model optimized for chat conversations",
"inputTokenLimit": 4096,
"outputTokenLimit": 1024,
"supportedGenerationMethods": [
"generateMessage",
"countMessageTokens"
],
"temperature": 0.25,
"topP": 0.949999988,
"topK": 40
},
{
"name": "models/text-bison-001",
"version": "001",
"displayName": "PaLM 2 (Legacy)",
"description": "A legacy model that understands text and generates text as an output",
"inputTokenLimit": 8196,
"outputTokenLimit": 1024,
"supportedGenerationMethods": [
"generateText",
"countTextTokens",
"createTunedTextModel"
],
"temperature": 0.699999988,
"topP": 0.949999988,
"topK": 40
},
{
"name": "models/embedding-gecko-001",
"version": "001",
"displayName": "Embedding Gecko",
"description": "Obtain a distributed representation of a text.",
"inputTokenLimit": 1024,
"outputTokenLimit": 1,
"supportedGenerationMethods": [
"embedText",
"countTextTokens"
]
},
{
"name": "models/gemini-1.0-pro-vision-latest",
"version": "001",
"displayName": "Gemini 1.0 Pro Vision",
"description": "The original Gemini 1.0 Pro Vision model version which was optimized for image understanding. Gemini 1.0 Pro Vision was deprecated on July 12, 2024. Move to a newer Gemini version.",
"inputTokenLimit": 12288,
"outputTokenLimit": 4096,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 0.400000006,
"topP": 1,
"topK": 32
},
{
"name": "models/gemini-pro-vision",
"version": "001",
"displayName": "Gemini 1.0 Pro Vision",
"description": "The original Gemini 1.0 Pro Vision model version which was optimized for image understanding. Gemini 1.0 Pro Vision was deprecated on July 12, 2024. Move to a newer Gemini version.",
"inputTokenLimit": 12288,
"outputTokenLimit": 4096,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 0.400000006,
"topP": 1,
"topK": 32
},
{
"name": "models/gemini-1.5-pro-latest",
"version": "001",
"displayName": "Gemini 1.5 Pro Latest",
"description": "Alias that points to the most recent production (non-experimental) release of Gemini 1.5 Pro, our mid-size multimodal model that supports up to 2 million tokens.",
"inputTokenLimit": 2000000,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 1,
"maxTemperature": 2,
"topP": 0.949999988,
"topK": 40
},
{
"name": "models/gemini-1.5-pro-001",
"version": "001",
"displayName": "Gemini 1.5 Pro 001",
"description": "Stable version of Gemini 1.5 Pro, our mid-size multimodal model that supports up to 2 million tokens, released in May of 2024.",
"inputTokenLimit": 2000000,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens",
"createCachedContent"
],
"temperature": 1,
"maxTemperature": 2,
"topP": 0.949999988,
"topK": 64
},
{
"name": "models/gemini-1.5-pro-002",
"version": "002",
"displayName": "Gemini 1.5 Pro 002",
"description": "Stable version of Gemini 1.5 Pro, our mid-size multimodal model that supports up to 2 million tokens, released in September of 2024.",
"inputTokenLimit": 2000000,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens",
"createCachedContent"
],
"temperature": 1,
"maxTemperature": 2,
"topP": 0.949999988,
"topK": 40
},
{
"name": "models/gemini-1.5-pro",
"version": "001",
"displayName": "Gemini 1.5 Pro",
"description": "Stable version of Gemini 1.5 Pro, our mid-size multimodal model that supports up to 2 million tokens, released in May of 2024.",
"inputTokenLimit": 2000000,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 1,
"maxTemperature": 2,
"topP": 0.949999988,
"topK": 40
},
{
"name": "models/gemini-1.5-flash-latest",
"version": "001",
"displayName": "Gemini 1.5 Flash Latest",
"description": "Alias that points to the most recent production (non-experimental) release of Gemini 1.5 Flash, our fast and versatile multimodal model for scaling across diverse tasks.",
"inputTokenLimit": 1000000,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 1,
"maxTemperature": 2,
"topP": 0.949999988,
"topK": 40
},
{
"name": "models/gemini-1.5-flash-001",
"version": "001",
"displayName": "Gemini 1.5 Flash 001",
"description": "Stable version of Gemini 1.5 Flash, our fast and versatile multimodal model for scaling across diverse tasks, released in May of 2024.",
"inputTokenLimit": 1000000,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens",
"createCachedContent"
],
"temperature": 1,
"maxTemperature": 2,
"topP": 0.949999988,
"topK": 64
},
{
"name": "models/gemini-1.5-flash-001-tuning",
"version": "001",
"displayName": "Gemini 1.5 Flash 001 Tuning",
"description": "Version of Gemini 1.5 Flash that supports tuning, our fast and versatile multimodal model for scaling across diverse tasks, released in May of 2024.",
"inputTokenLimit": 16384,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens",
"createTunedModel"
],
"temperature": 1,
"maxTemperature": 2,
"topP": 0.949999988,
"topK": 64
},
{
"name": "models/gemini-1.5-flash",
"version": "001",
"displayName": "Gemini 1.5 Flash",
"description": "Alias that points to the most recent stable version of Gemini 1.5 Flash, our fast and versatile multimodal model for scaling across diverse tasks.",
"inputTokenLimit": 1000000,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 1,
"maxTemperature": 2,
"topP": 0.949999988,
"topK": 40
},
{
"name": "models/gemini-1.5-flash-002",
"version": "002",
"displayName": "Gemini 1.5 Flash 002",
"description": "Stable version of Gemini 1.5 Flash, our fast and versatile multimodal model for scaling across diverse tasks, released in September of 2024.",
"inputTokenLimit": 1000000,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens",
"createCachedContent"
],
"temperature": 1,
"maxTemperature": 2,
"topP": 0.949999988,
"topK": 40
},
{
"name": "models/gemini-1.5-flash-8b",
"version": "001",
"displayName": "Gemini 1.5 Flash-8B",
"description": "Stable version of Gemini 1.5 Flash-8B, our smallest and most cost effective Flash model, released in October of 2024.",
"inputTokenLimit": 1000000,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"createCachedContent",
"generateContent",
"countTokens"
],
"temperature": 1,
"maxTemperature": 2,
"topP": 0.949999988,
"topK": 40
},
{
"name": "models/gemini-1.5-flash-8b-001",
"version": "001",
"displayName": "Gemini 1.5 Flash-8B 001",
"description": "Stable version of Gemini 1.5 Flash-8B, our smallest and most cost effective Flash model, released in October of 2024.",
"inputTokenLimit": 1000000,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"createCachedContent",
"generateContent",
"countTokens"
],
"temperature": 1,
"maxTemperature": 2,
"topP": 0.949999988,
"topK": 40
},
{
"name": "models/gemini-1.5-flash-8b-latest",
"version": "001",
"displayName": "Gemini 1.5 Flash-8B Latest",
"description": "Alias that points to the most recent production (non-experimental) release of Gemini 1.5 Flash-8B, our smallest and most cost effective Flash model, released in October of 2024.",
"inputTokenLimit": 1000000,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"createCachedContent",
"generateContent",
"countTokens"
],
"temperature": 1,
"maxTemperature": 2,
"topP": 0.949999988,
"topK": 40
},
{
"name": "models/gemini-1.5-flash-8b-exp-0827",
"version": "001",
"displayName": "Gemini 1.5 Flash 8B Experimental 0827",
"description": "Experimental release (August 27th, 2024) of Gemini 1.5 Flash-8B, our smallest and most cost effective Flash model. Replaced by Gemini-1.5-flash-8b-001 (stable).",
"inputTokenLimit": 1000000,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 1,
"maxTemperature": 2,
"topP": 0.949999988,
"topK": 40
},
{
"name": "models/gemini-1.5-flash-8b-exp-0924",
"version": "001",
"displayName": "Gemini 1.5 Flash 8B Experimental 0924",
"description": "Experimental release (September 24th, 2024) of Gemini 1.5 Flash-8B, our smallest and most cost effective Flash model. Replaced by Gemini-1.5-flash-8b-001 (stable).",
"inputTokenLimit": 1000000,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 1,
"maxTemperature": 2,
"topP": 0.949999988,
"topK": 40
},
{
"name": "models/gemini-2.5-pro-exp-03-25",
"version": "2.5-exp-03-25",
"displayName": "Gemini 2.5 Pro Experimental 03-25",
"description": "Experimental release (March 25th, 2025) of Gemini 2.5 Pro",
"inputTokenLimit": 1048576,
"outputTokenLimit": 65536,
"supportedGenerationMethods": [
"generateContent",
"countTokens",
"createCachedContent"
],
"temperature": 1,
"maxTemperature": 2,
"topP": 0.949999988,
"topK": 64
},
{
"name": "models/gemini-2.5-pro-preview-03-25",
"version": "2.5-preview-03-25",
"displayName": "Gemini 2.5 Pro Preview 03-25",
"description": "Gemini 2.5 Pro Preview 03-25",
"inputTokenLimit": 1048576,
"outputTokenLimit": 65536,
"supportedGenerationMethods": [
"generateContent",
"countTokens",
"createCachedContent"
],
"temperature": 1,
"maxTemperature": 2,
"topP": 0.949999988,
"topK": 64
},
{
"name": "models/gemini-2.5-flash-preview-04-17",
"version": "2.5-preview-04-17",
"displayName": "Gemini 2.5 Flash Preview 04-17",
"description": "Preview release (April 17th, 2025) of Gemini 2.5 Flash",
"inputTokenLimit": 1048576,
"outputTokenLimit": 65536,
"supportedGenerationMethods": [
"generateContent",
"countTokens",
"createCachedContent"
],
"temperature": 1,
"maxTemperature": 2,
"topP": 0.949999988,
"topK": 64
},
{
"name": "models/gemini-2.0-flash-exp",
"version": "2.0",
"displayName": "Gemini 2.0 Flash Experimental",
"description": "Gemini 2.0 Flash Experimental",
"inputTokenLimit": 1048576,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens",
"bidiGenerateContent"
],
"temperature": 1,
"maxTemperature": 2,
"topP": 0.949999988,
"topK": 40
},
{
"name": "models/gemini-2.0-flash",
"version": "2.0",
"displayName": "Gemini 2.0 Flash",
"description": "Gemini 2.0 Flash",
"inputTokenLimit": 1048576,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens",
"createCachedContent"
],
"temperature": 1,
"maxTemperature": 2,
"topP": 0.949999988,
"topK": 40
},
{
"name": "models/gemini-2.0-flash-001",
"version": "2.0",
"displayName": "Gemini 2.0 Flash 001",
"description": "Stable version of Gemini 2.0 Flash, our fast and versatile multimodal model for scaling across diverse tasks, released in January of 2025.",
"inputTokenLimit": 1048576,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens",
"createCachedContent"
],
"temperature": 1,
"maxTemperature": 2,
"topP": 0.949999988,
"topK": 40
},
{
"name": "models/gemini-2.0-flash-lite-001",
"version": "2.0",
"displayName": "Gemini 2.0 Flash-Lite 001",
"description": "Stable version of Gemini 2.0 Flash Lite",
"inputTokenLimit": 1048576,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens",
"createCachedContent"
],
"temperature": 1,
"maxTemperature": 2,
"topP": 0.949999988,
"topK": 40
},
{
"name": "models/gemini-2.0-flash-lite",
"version": "2.0",
"displayName": "Gemini 2.0 Flash-Lite",
"description": "Gemini 2.0 Flash-Lite",
"inputTokenLimit": 1048576,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens",
"createCachedContent"
],
"temperature": 1,
"maxTemperature": 2,
"topP": 0.949999988,
"topK": 40
},
{
"name": "models/gemini-2.0-flash-lite-preview-02-05",
"version": "preview-02-05",
"displayName": "Gemini 2.0 Flash-Lite Preview 02-05",
"description": "Preview release (February 5th, 2025) of Gemini 2.0 Flash Lite",
"inputTokenLimit": 1048576,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens",
"createCachedContent"
],
"temperature": 1,
"maxTemperature": 2,
"topP": 0.949999988,
"topK": 40
},
{
"name": "models/gemini-2.0-flash-lite-preview",
"version": "preview-02-05",
"displayName": "Gemini 2.0 Flash-Lite Preview",
"description": "Preview release (February 5th, 2025) of Gemini 2.0 Flash Lite",
"inputTokenLimit": 1048576,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens",
"createCachedContent"
],
"temperature": 1,
"maxTemperature": 2,
"topP": 0.949999988,
"topK": 40
},
{
"name": "models/gemini-2.0-pro-exp",
"version": "2.5-exp-03-25",
"displayName": "Gemini 2.0 Pro Experimental",
"description": "Experimental release (March 25th, 2025) of Gemini 2.5 Pro",
"inputTokenLimit": 1048576,
"outputTokenLimit": 65536,
"supportedGenerationMethods": [
"generateContent",
"countTokens",
"createCachedContent"
],
"temperature": 1,
"maxTemperature": 2,
"topP": 0.949999988,
"topK": 64
},
{
"name": "models/gemini-2.0-pro-exp-02-05",
"version": "2.5-exp-03-25",
"displayName": "Gemini 2.0 Pro Experimental 02-05",
"description": "Experimental release (March 25th, 2025) of Gemini 2.5 Pro",
"inputTokenLimit": 1048576,
"outputTokenLimit": 65536,
"supportedGenerationMethods": [
"generateContent",
"countTokens",
"createCachedContent"
],
"temperature": 1,
"maxTemperature": 2,
"topP": 0.949999988,
"topK": 64
},
{
"name": "models/gemini-exp-1206",
"version": "2.5-exp-03-25",
"displayName": "Gemini Experimental 1206",
"description": "Experimental release (March 25th, 2025) of Gemini 2.5 Pro",
"inputTokenLimit": 1048576,
"outputTokenLimit": 65536,
"supportedGenerationMethods": [
"generateContent",
"countTokens",
"createCachedContent"
],
"temperature": 1,
"maxTemperature": 2,
"topP": 0.949999988,
"topK": 64
},
{
"name": "models/gemini-2.0-flash-thinking-exp-01-21",
"version": "2.5-preview-04-17",
"displayName": "Gemini 2.5 Flash Preview 04-17",
"description": "Preview release (April 17th, 2025) of Gemini 2.5 Flash",
"inputTokenLimit": 1048576,
"outputTokenLimit": 65536,
"supportedGenerationMethods": [
"generateContent",
"countTokens",
"createCachedContent"
],
"temperature": 1,
"maxTemperature": 2,
"topP": 0.949999988,
"topK": 64
},
{
"name": "models/gemini-2.0-flash-thinking-exp",
"version": "2.5-preview-04-17",
"displayName": "Gemini 2.5 Flash Preview 04-17",
"description": "Preview release (April 17th, 2025) of Gemini 2.5 Flash",
"inputTokenLimit": 1048576,
"outputTokenLimit": 65536,
"supportedGenerationMethods": [
"generateContent",
"countTokens",
"createCachedContent"
],
"temperature": 1,
"maxTemperature": 2,
"topP": 0.949999988,
"topK": 64
},
{
"name": "models/gemini-2.0-flash-thinking-exp-1219",
"version": "2.5-preview-04-17",
"displayName": "Gemini 2.5 Flash Preview 04-17",
"description": "Preview release (April 17th, 2025) of Gemini 2.5 Flash",
"inputTokenLimit": 1048576,
"outputTokenLimit": 65536,
"supportedGenerationMethods": [
"generateContent",
"countTokens",
"createCachedContent"
],
"temperature": 1,
"maxTemperature": 2,
"topP": 0.949999988,
"topK": 64
},
{
"name": "models/learnlm-1.5-pro-experimental",
"version": "001",
"displayName": "LearnLM 1.5 Pro Experimental",
"description": "Alias that points to the most recent stable version of Gemini 1.5 Pro, our mid-size multimodal model that supports up to 2 million tokens.",
"inputTokenLimit": 32767,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 1,
"maxTemperature": 2,
"topP": 0.949999988,
"topK": 64
},
{
"name": "models/learnlm-2.0-flash-experimental",
"version": "2.0",
"displayName": "LearnLM 2.0 Flash Experimental",
"description": "LearnLM 2.0 Flash Experimental",
"inputTokenLimit": 1048576,
"outputTokenLimit": 32768,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 1,
"maxTemperature": 2,
"topP": 0.949999988,
"topK": 64
},
{
"name": "models/gemma-3-1b-it",
"version": "001",
"displayName": "Gemma 3 1B",
"inputTokenLimit": 32768,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 1,
"topP": 0.949999988,
"topK": 64
},
{
"name": "models/gemma-3-4b-it",
"version": "001",
"displayName": "Gemma 3 4B",
"inputTokenLimit": 32768,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 1,
"topP": 0.949999988,
"topK": 64
},
{
"name": "models/gemma-3-12b-it",
"version": "001",
"displayName": "Gemma 3 12B",
"inputTokenLimit": 32768,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 1,
"topP": 0.949999988,
"topK": 64
},
{
"name": "models/gemma-3-27b-it",
"version": "001",
"displayName": "Gemma 3 27B",
"inputTokenLimit": 131072,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 1,
"topP": 0.949999988,
"topK": 64
},
{
"name": "models/embedding-001",
"version": "001",
"displayName": "Embedding 001",
"description": "Obtain a distributed representation of a text.",
"inputTokenLimit": 2048,
"outputTokenLimit": 1,
"supportedGenerationMethods": [
"embedContent"
]
},
{
"name": "models/text-embedding-004",
"version": "004",
"displayName": "Text Embedding 004",
"description": "Obtain a distributed representation of a text.",
"inputTokenLimit": 2048,
"outputTokenLimit": 1,
"supportedGenerationMethods": [
"embedContent"
]
},
{
"name": "models/gemini-embedding-exp-03-07",
"version": "exp-03-07",
"displayName": "Gemini Embedding Experimental 03-07",
"description": "Obtain a distributed representation of a text.",
"inputTokenLimit": 8192,
"outputTokenLimit": 1,
"supportedGenerationMethods": [
"embedContent",
"countTextTokens"
]
},
{
"name": "models/gemini-embedding-exp",
"version": "exp-03-07",
"displayName": "Gemini Embedding Experimental",
"description": "Obtain a distributed representation of a text.",
"inputTokenLimit": 8192,
"outputTokenLimit": 1,
"supportedGenerationMethods": [
"embedContent",
"countTextTokens"
]
},
{
"name": "models/aqa",
"version": "001",
"displayName": "Model that performs Attributed Question Answering.",
"description": "Model trained to return answers to questions that are grounded in provided sources, along with estimating answerable probability.",
"inputTokenLimit": 7168,
"outputTokenLimit": 1024,
"supportedGenerationMethods": [
"generateAnswer"
],
"temperature": 0.200000003,
"topP": 1,
"topK": 40
},
{
"name": "models/imagen-3.0-generate-002",
"version": "002",
"displayName": "Imagen 3.0 002 model",
"description": "Vertex served Imagen 3.0 002 model",
"inputTokenLimit": 480,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"predict"
]
},
{
"name": "models/veo-2.0-generate-001",
"version": "2.0",
"displayName": "Veo 2",
"description": "Vertex served Veo 2 model.",
"inputTokenLimit": 480,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"predictLongRunning"
]
},
{
"name": "models/gemini-2.0-flash-live-001",
"version": "001",
"displayName": "Gemini 2.0 Flash 001",
"description": "Gemini 2.0 Flash 001",
"inputTokenLimit": 131072,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"bidiGenerateContent",
"countTokens"
],
"temperature": 1,
"maxTemperature": 2,
"topP": 0.949999988,
"topK": 64
}
]
}
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment