A RetroSearch Logo

Home - News ( United States | United Kingdom | Italy | Germany ) - Football scores

Search Query:

Showing content from http://cloud.google.com/vertex-ai/generative-ai/docs/models/tune-text-models-supervised below:

About supervised fine-tuning for Gemini models | Generative AI on Vertex AI

About supervised fine-tuning for Gemini models

Stay organized with collections Save and categorize content based on your preferences.

Supervised fine-tuning is a good option when you have a well-defined task with available labeled data. It's particularly effective for domain-specific applications where the language or content significantly differs from the data the large model was originally trained on. You can tune text, image, audio, and document data types.

Supervised fine-tuning adapts model behavior with a labeled dataset. This process adjusts the model's weights to minimize the difference between its predictions and the actual labels. For example, it can improve model performance for the following types of tasks:

For a discussion of the top tuning use cases, check out the blog post Hundreds of organizations are fine-tuning Gemini models. Here's their favorite use cases.

To learn more, see When to use supervised fine-tuning for Gemini.

Supported models

The following Gemini models support supervised fine-tuning:

For models that support thinking, we suggest setting the thinking budget to off or its lowest value. This can improve performance and reduce costs for tuned tasks. During supervised fine-tuning, the model learns from the training data and omits the thinking process. Therefore, the resulting tuned model can perform tuned tasks effectively without a thinking budget.

Limitations Gemini 2.5 Flash
Gemini 2.5 Flash-Lite Specification Value Maximum input and output training tokens 131,072 Maximum input and output serving tokens Same as base Gemini model Maximum validation dataset size 5000 examples Maximum training dataset file size 1GB for JSONL Maximum training dataset size 1M text-only examples or 300K multimodal examples Adapter size Supported values are 1, 2, 4, 8, and 16 Gemini 2.5 Pro Specification Value Maximum input and output training tokens 131,072 Maximum input and output serving tokens Same as base Gemini model Maximum validation dataset size 5000 examples Maximum training dataset file size 1GB for JSONL Maximum training dataset size 1M text-only examples or 300K multimodal examples Adapter size Supported values are 1, 2, 4, and 8 Gemini 2.0 Flash
Gemini 2.0 Flash-Lite Specification Value Maximum input and output training tokens 131,072 Maximum input and output serving tokens Same as base Gemini model Maximum validation dataset size 5000 examples Maximum training dataset file size 1GB for JSONL Maximum training dataset size 1M text-only examples or 300K multimodal examples Adapter size Supported values are 1, 2, 4, and 8 Known issues Use cases for using supervised fine-tuning

Foundation models work well when the expected output or task can be clearly and concisely defined in a prompt and the prompt consistently produces the expected output. If you want a model to learn something niche or specific that deviates from general patterns, then you might want to consider tuning that model. For example, you can use model tuning to teach the model the following:

The following examples are use cases that are difficult to capture with only prompt instructions:

You can also tune a model in the following situations:

Configure a tuning job region

User data, such as the transformed dataset and the tuned model, is stored in the tuning job region. During tuning, computation could be offloaded to other US or EU regions for available accelerators. The offloading is transparent to users.

Quota

Quota is enforced on the number of concurrent tuning jobs. Every project comes with a default quota to run at least one tuning job. This is a global quota, shared across all available regions and supported models. If you want to run more jobs concurrently, you need to request additional quota for Global concurrent tuning jobs.

Pricing

Pricing for Gemini supervised fine-tuning can be found here: Vertex AI pricing.

The number of training tokens is calculated by multiplying the number of tokens in your training dataset by the number of epochs. After tuning, inference (prediction request) costs for the tuned model still apply. Inference pricing is the same for each stable version of Gemini. For more information, see Available Gemini stable model versions.

What's next

Except as otherwise noted, the content of this page is licensed under the Creative Commons Attribution 4.0 License, and code samples are licensed under the Apache 2.0 License. For details, see the Google Developers Site Policies. Java is a registered trademark of Oracle and/or its affiliates.

Last updated 2025-08-14 UTC.

[[["Easy to understand","easyToUnderstand","thumb-up"],["Solved my problem","solvedMyProblem","thumb-up"],["Other","otherUp","thumb-up"]],[["Hard to understand","hardToUnderstand","thumb-down"],["Incorrect information or sample code","incorrectInformationOrSampleCode","thumb-down"],["Missing the information/samples I need","missingTheInformationSamplesINeed","thumb-down"],["Other","otherDown","thumb-down"]],["Last updated 2025-08-14 UTC."],[],[]]


RetroSearch is an open source project built by @garambo | Open a GitHub Issue

Search and Browse the WWW like it's 1997 | Search results from DuckDuckGo

HTML: 3.2 | Encoding: UTF-8 | Version: 0.7.4