src/data/question-groups/data-science/content/fine-tuning-vs-more-data.md
Fine-tuning a model is the process of adapting a pre-trained model for specific tasks or use cases. The reasoning behind fine-tuning is that it is easier and cheaper to improve the capabilities of a pre-trained base model that has already learned road knowledge about the task than it is to train a new model from scratch.
Generalization is a measure of how your model performs in predicting unseen data. Generalizing with more data is improving the model's ability to make predictions on new data rather than the data it was trained on.
Choosing whether to generalize with more data or fine-tune to achieve your goal depends on the specific situation.
For specialization with fine-tuning:
For generalization with more data: