Creating a New Fine-Tuning Model
The Create New Fine-Tuning Model page allows users to fine-tune a base model with custom datasets and hyperparameters to suit their specific needs. The process involves selecting a base model, uploading training and validation datasets, and configuring hyperparameters.
Steps to Create a New Fine-Tuning Model:
-
Base Model Selection: Choose a base model from the provided list of pre-trained models. Popular options like
llama-2-70b
,gpt-4-mini
, and other state-of-the-art models are available for selection. The already fine-tuned models are displayed at the top for easy access.- Fine-tuned models show their completion status, making it easy to manage or re-use them.
- The modal gallery shows a variety of models that users can filter, search, and select based on their needs.
Select Base Models:
The following are most popular models available for fine-tuning:
Model Name | Type | Updated At | Size |
---|---|---|---|
llama-2-70b-chat | Text Generation | Aug 22, 2024 | 325.7 GB |
llama-2-70b | Text Generation | Aug 16, 2024 | 320.4 GB |
llama-2-13b-chat | Text Generation | Aug 20, 2024 | 78.1 GB |
code-llama-34b | Code Generation | Aug 07, 2024 | 94.3 GB |
gpt-4-mini | Chat Completion | Jul 18, 2024 | 120.5 GB |
gpt-35-turbo-16k | Chat Completion | Jul 28, 2024 | 14.7 GB |
gpt-35-turbo-instruct | Chat Completion | Aug 02, 2024 | 8.9 GB |
gpt-4o-2024-08-06 | Chat Completion | Aug 06, 2024 | 9.1 GB |
text-embedding-ada | Embeddings | May 20, 2024 | 2.7 GB |
You can search, filter, and select the base models depending on your task.
-
Name the Model: Enter a custom name for your fine-tuned model in the Name field. Naming helps identify and distinguish fine-tuned models, especially when managing multiple versions.
-
Upload Training Data: Upload the dataset that will be used to train the model. You can either upload a new dataset or choose from an existing dataset already available in your dataset library. This flexibility allows quick adaptation for different projects.
-
Upload Validation Data: In the next step, upload a validation dataset. This data is essential to evaluate how well the model generalizes to unseen data during the training process.
- You can upload new data or select an existing validation set.
- Alternatively, you can choose None if you don't want to use validation at this stage (although this is not recommended).
-
Configure Hyperparameters: Customize key hyperparameters to fine-tune the model's training process:
- Batch Size: The number of samples processed before the model is updated.
- Learning Rate Multiplier: Adjusts how much to change the model in response to the error each time the model weights are updated.
- Number of Epochs: Set the number of complete passes through the training dataset.
- These settings control how fast the model learns and how precise the tuning process becomes.
-
Optional Settings: Customize additional settings:
- Suffix: Add a custom suffix to experiment names for easier identification.
- Seed: Set a seed for reproducibility or leave it as random. Seeds are used to ensure consistent training results when needed.
Key Benefits:
- Model Customization: Fine-tune models based on specific data, significantly improving the accuracy and relevance of generated responses.
- Flexible Dataset Management: Easily upload new datasets or reuse existing ones from the dataset library.
- Full Hyperparameter Control: Adjust critical settings like batch size, learning rate, and epochs to optimize performance.
Adjusting hyperparameters carefully can lead to faster and more efficient fine-tuning, depending on the size of your dataset and the complexity of the base model.
Example Process:
- Base Model Selection: Choose a pre-trained model like
llama-2-70b-chat
, which is optimized for text generation tasks such as conversational agents. - Data Upload: Upload the training data, for example, a custom dataset containing chat interactions, to fine-tune the model's conversational capabilities.
- Configure Hyperparameters: Set a learning rate suitable for gradual tuning and a batch size that matches the memory available on the hardware.
- Final Steps: Once all settings are configured, click Save to begin the fine-tuning process.
The screenshot above illustrates how fine-tuned models (if any) appear at the top for quick reference. New fine-tuning models can be created by selecting a base model from the gallery below.