
Research on Overfitting Problem and Correction in Machine Learning
Author(s) -
Chao Bu,
Zengping Zhang
Publication year - 2020
Publication title -
journal of physics. conference series
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.21
H-Index - 85
eISSN - 1742-6596
pISSN - 1742-6588
DOI - 10.1088/1742-6596/1693/1/012100
Subject(s) - overfitting , generalization , computer science , artificial intelligence , machine learning , regularization (linguistics) , process (computing) , polynomial , structural risk minimization , key (lock) , artificial neural network , mathematics , mathematical analysis , operating system , computer security
Machine learning is the key technology of artificial intelligence, which uses learning and training data model to find the problem to achieve the law. In practical applications, there is always a difference between the input data of the model and the training data. Based on the biased training data, overfitting will occur, which will cause the machine learning to fail to achieve the expected goals. Generalization is the process of ensuring that the model can fully reflect the characteristics of the actual input data. Therefore, the ability of the machine learning model depends to a large extent on the effectiveness of generalization. In order to fully understand the learning and generalization model, we conducted a study by the polynomial curve fitting. First, based on the training data, we analyzed using matrix theory and analytical solutions derived polynomial fit. Then, numerical and maximum likelihood theoretical analysis was carried out for the overfitting problem. Finally, combined with the objective function, a typical regularization method to overcome overfitting and improve generalization ability is elaborated.