Back to Developer Roadmap

Fine Tuning Vs More Data

src/data/question-groups/data-science/content/fine-tuning-vs-more-data.md

4.01.1 KB
Original Source

Fine-tuning a model is the process of adapting a pre-trained model for specific tasks or use cases. The reasoning behind fine-tuning is that it is easier and cheaper to improve the capabilities of a pre-trained base model that has already learned road knowledge about the task than it is to train a new model from scratch.

Generalization is a measure of how your model performs in predicting unseen data. Generalizing with more data is improving the model's ability to make predictions on new data rather than the data it was trained on.

Choosing whether to generalize with more data or fine-tune to achieve your goal depends on the specific situation.

For specialization with fine-tuning:

  • It is better when high performance is needed on a very specific task or domain.
  • It is more efficient to use when you have limited resources, but good data for a specific task.
  • It can achieve strong results with smaller models.

For generalization with more data:

  • It is better for models that need to handle a wide range of tasks.
  • It is great for situations where overfitting will be a problem.