Instruction: Describe what overfitting is and how it can affect a model's performance.
Context: This question tests the candidate's understanding of a common problem in machine learning models and their knowledge of how to identify and address it.
The way I'd explain it in an interview is this: Overfitting is when a model learns patterns that are too specific to the training data, including noise, quirks, or accidental correlations that do not hold on new data. As a result, the model can look excellent during training and still fail when it sees unseen examples.
A useful way to recognize it is the gap between training performance and validation performance. If the training score keeps improving while validation performance stalls or gets worse, the model is usually learning detail that does not generalize. That is why overfitting is less about memorization in a narrow sense and more about poor generalization.
What matters in an interview is not only knowing the definition, but being able to connect it back to how it changes modeling, evaluation, or deployment decisions in practice.
A weak answer says overfitting means the model memorizes the training set, without connecting that to the practical symptom: poor performance on unseen data.
easy
easy
easy
easy
easy
easy